From 6af2f1f21775fbf0fed2d7e4f7ec86c50d0edbd2 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Thu, 18 Jan 2024 14:45:37 +0800 Subject: [PATCH] fix rm dataset Former-commit-id: 5edf7cce0eb4cd569438bb406fe6f148eb05ee46 --- src/llmtuner/hparams/parser.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/src/llmtuner/hparams/parser.py b/src/llmtuner/hparams/parser.py index cba9c690..a050b89a 100644 --- a/src/llmtuner/hparams/parser.py +++ b/src/llmtuner/hparams/parser.py @@ -120,9 +120,6 @@ def get_train_args(args: Optional[Dict[str, Any]] = None) -> _TRAIN_CLS: if finetuning_args.stage == "ppo" and not training_args.do_train: raise ValueError("PPO training does not support evaluation, use the SFT stage to evaluate models.") - if finetuning_args.stage in ["rm", "dpo"] and (not all([data_attr.ranking for data_attr in data_args.dataset_list])): - raise ValueError("Please use ranked datasets for reward modeling or DPO training.") - if finetuning_args.stage == "ppo" and model_args.shift_attn: raise ValueError("PPO training is incompatible with S^2-Attn.")