diff --git a/src/llmtuner/tuner/dpo/workflow.py b/src/llmtuner/tuner/dpo/workflow.py index 63968604..c4acb331 100644 --- a/src/llmtuner/tuner/dpo/workflow.py +++ b/src/llmtuner/tuner/dpo/workflow.py @@ -60,9 +60,9 @@ def run_dpo( plot_loss(training_args.output_dir, keys=["loss", "eval_loss"]) if training_args.push_to_hub: - trainer.push_to_hub(**generate_model_card()) + trainer.push_to_hub(**generate_model_card(model_args, data_args, finetuning_args)) else: - trainer.create_model_card(**generate_model_card()) + trainer.create_model_card(**generate_model_card(model_args, data_args, finetuning_args)) # Evaluation if training_args.do_eval: diff --git a/src/llmtuner/tuner/pt/workflow.py b/src/llmtuner/tuner/pt/workflow.py index 002d2dd1..c7edff21 100644 --- a/src/llmtuner/tuner/pt/workflow.py +++ b/src/llmtuner/tuner/pt/workflow.py @@ -46,9 +46,9 @@ def run_pt( plot_loss(training_args.output_dir, keys=["loss", "eval_loss"]) if training_args.push_to_hub: - trainer.push_to_hub(**generate_model_card()) + trainer.push_to_hub(**generate_model_card(model_args, data_args, finetuning_args)) else: - trainer.create_model_card(**generate_model_card()) + trainer.create_model_card(**generate_model_card(model_args, data_args, finetuning_args)) # Evaluation if training_args.do_eval: diff --git a/src/llmtuner/tuner/rm/workflow.py b/src/llmtuner/tuner/rm/workflow.py index c95f1cb6..eedec5e7 100644 --- a/src/llmtuner/tuner/rm/workflow.py +++ b/src/llmtuner/tuner/rm/workflow.py @@ -54,9 +54,9 @@ def run_rm( plot_loss(training_args.output_dir, keys=["loss", "eval_loss"]) if training_args.push_to_hub: - trainer.push_to_hub(**generate_model_card()) + trainer.push_to_hub(**generate_model_card(model_args, data_args, finetuning_args)) else: - trainer.create_model_card(**generate_model_card()) + trainer.create_model_card(**generate_model_card(model_args, data_args, finetuning_args)) # Evaluation if training_args.do_eval: diff --git a/src/llmtuner/tuner/sft/workflow.py b/src/llmtuner/tuner/sft/workflow.py index dc22904b..04b37ac7 100644 --- a/src/llmtuner/tuner/sft/workflow.py +++ b/src/llmtuner/tuner/sft/workflow.py @@ -73,9 +73,9 @@ def run_sft( plot_loss(training_args.output_dir, keys=["loss", "eval_loss"]) if training_args.push_to_hub: - trainer.push_to_hub(**generate_model_card()) + trainer.push_to_hub(**generate_model_card(model_args, data_args, finetuning_args)) else: - trainer.create_model_card(**generate_model_card()) + trainer.create_model_card(**generate_model_card(model_args, data_args, finetuning_args)) # Evaluation if training_args.do_eval: