mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 06:12:50 +08:00
tiny fix
Former-commit-id: ce43386080fa1535672f4d879ffe6c4360a1ef7d
This commit is contained in:
parent
ce08b4a7ec
commit
93f2e35035
@ -101,7 +101,7 @@ def _init_adapter(
|
|||||||
logger.info("Fine-tuning method: LoRA")
|
logger.info("Fine-tuning method: LoRA")
|
||||||
lastest_checkpoint = None
|
lastest_checkpoint = None
|
||||||
|
|
||||||
assert os.path.exists(model_args.checkpoint_dir[0], CONFIG_NAME), \
|
assert os.path.exists(os.path.join(model_args.checkpoint_dir[0], CONFIG_NAME)), \
|
||||||
"The given checkpoint is not a LoRA checkpoint, please specify `--finetuning_type full/freeze` instead."
|
"The given checkpoint is not a LoRA checkpoint, please specify `--finetuning_type full/freeze` instead."
|
||||||
|
|
||||||
if model_args.checkpoint_dir is not None:
|
if model_args.checkpoint_dir is not None:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user