From 574a7d175c5a9285e23cbeed6338ee782b3fdd52 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Fri, 3 Nov 2023 00:44:53 +0800 Subject: [PATCH] fix #1290 Former-commit-id: eb9d9e104ad5cdfc23f9cc48f20068c674762c47 --- src/llmtuner/tuner/core/parser.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/llmtuner/tuner/core/parser.py b/src/llmtuner/tuner/core/parser.py index 71b2c810..603fc1bc 100644 --- a/src/llmtuner/tuner/core/parser.py +++ b/src/llmtuner/tuner/core/parser.py @@ -106,8 +106,8 @@ def get_train_args( if finetuning_args.stage in ["rm", "ppo"] and training_args.resume_from_checkpoint is not None: raise ValueError("RM and PPO stages do not support `resume_from_checkpoint`.") - if finetuning_args.stage in ["ppo", "dpo"] and not training_args.do_train: - raise ValueError("PPO and DPO stages can only be performed at training.") + if finetuning_args.stage == "ppo" and not training_args.do_train: + raise ValueError("PPO training does not support evaluation.") if finetuning_args.stage in ["rm", "dpo"]: for dataset_attr in data_args.dataset_list: