fix dpo metrics

This commit is contained in:
hiyouga
2024-11-02 19:22:11 +08:00
parent 07e5088851
commit 4270f7dfb9
7 changed files with 143 additions and 58 deletions

View File

@@ -81,7 +81,7 @@ def run_kto(
trainer.save_metrics("train", train_result.metrics)
trainer.save_state()
if trainer.is_world_process_zero() and finetuning_args.plot_loss:
plot_loss(training_args.output_dir, keys=["loss", "eval_loss", "train/rewards/chosen"])
plot_loss(training_args.output_dir, keys=["loss", "eval_loss", "rewards/chosen"])
# Evaluation
if training_args.do_eval: