mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 20:52:59 +08:00
fix bug
Former-commit-id: 714d936dfbe022c4f2cfa6ff643e3482a3f96012
This commit is contained in:
parent
8b8671817f
commit
9ff7c99eb1
@ -294,6 +294,7 @@ def _create_loraplus_optimizer(
|
|||||||
dict(params=param_dict["embedding"], lr=finetuning_args.loraplus_lr_embedding, **decay_args),
|
dict(params=param_dict["embedding"], lr=finetuning_args.loraplus_lr_embedding, **decay_args),
|
||||||
]
|
]
|
||||||
optimizer = optim_class(param_groups, **optim_kwargs)
|
optimizer = optim_class(param_groups, **optim_kwargs)
|
||||||
|
logger.info("Using LoRA+ optimizer with loraplus lr ratio {:.2f}.".format(finetuning_args.loraplus_lr_ratio))
|
||||||
return optimizer
|
return optimizer
|
||||||
|
|
||||||
|
|
||||||
@ -303,7 +304,7 @@ def create_custom_optimzer(
|
|||||||
training_args: "Seq2SeqTrainingArguments",
|
training_args: "Seq2SeqTrainingArguments",
|
||||||
finetuning_args: "FinetuningArguments",
|
finetuning_args: "FinetuningArguments",
|
||||||
) -> Optional["torch.optim.Optimizer"]:
|
) -> Optional["torch.optim.Optimizer"]:
|
||||||
if not finetuning_args.use_galore:
|
if finetuning_args.use_galore:
|
||||||
return _create_galore_optimizer(model, dataset, training_args, finetuning_args)
|
return _create_galore_optimizer(model, dataset, training_args, finetuning_args)
|
||||||
|
|
||||||
if finetuning_args.loraplus_lr_ratio is not None:
|
if finetuning_args.loraplus_lr_ratio is not None:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user