mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 08:38:09 +08:00
tiny fix
Former-commit-id: d24ea58c1a44b94227f4cb60f13fc1dd79997d01
This commit is contained in:
parent
dc68c313ee
commit
e145a2ce0c
@ -39,7 +39,7 @@ def run_sft(
|
|||||||
# Override the decoding parameters of Seq2SeqTrainer
|
# Override the decoding parameters of Seq2SeqTrainer
|
||||||
training_args_dict = training_args.to_dict()
|
training_args_dict = training_args.to_dict()
|
||||||
training_args_dict.update(dict(
|
training_args_dict.update(dict(
|
||||||
generation_max_length=training_args.generation_max_length or data_args.max_target_length,
|
generation_max_length=training_args.generation_max_length or data_args.cutoff_len,
|
||||||
generation_num_beams=data_args.eval_num_beams or training_args.generation_num_beams
|
generation_num_beams=data_args.eval_num_beams or training_args.generation_num_beams
|
||||||
))
|
))
|
||||||
training_args = Seq2SeqTrainingArguments(**training_args_dict)
|
training_args = Seq2SeqTrainingArguments(**training_args_dict)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user