mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-11-06 02:42:15 +08:00
tiny fix
Former-commit-id: d24ea58c1a44b94227f4cb60f13fc1dd79997d01
This commit is contained in:
parent
d8ab75ee44
commit
dd6e9b3cc1
@ -39,7 +39,7 @@ def run_sft(
|
|||||||
# Override the decoding parameters of Seq2SeqTrainer
|
# Override the decoding parameters of Seq2SeqTrainer
|
||||||
training_args_dict = training_args.to_dict()
|
training_args_dict = training_args.to_dict()
|
||||||
training_args_dict.update(dict(
|
training_args_dict.update(dict(
|
||||||
generation_max_length=training_args.generation_max_length or data_args.max_target_length,
|
generation_max_length=training_args.generation_max_length or data_args.cutoff_len,
|
||||||
generation_num_beams=data_args.eval_num_beams or training_args.generation_num_beams
|
generation_num_beams=data_args.eval_num_beams or training_args.generation_num_beams
|
||||||
))
|
))
|
||||||
training_args = Seq2SeqTrainingArguments(**training_args_dict)
|
training_args = Seq2SeqTrainingArguments(**training_args_dict)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user