mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 12:42:51 +08:00
add deepspeed check in PPO training
Former-commit-id: ed1c2c5557bb2714c3341294f0ea86f6496d4b0c
This commit is contained in:
parent
e6fa0229f4
commit
5030f05126
@ -119,6 +119,9 @@ def get_train_args(
|
|||||||
if general_args.stage == "ppo" and model_args.reward_model is None:
|
if general_args.stage == "ppo" and model_args.reward_model is None:
|
||||||
raise ValueError("Reward model is necessary for PPO training.")
|
raise ValueError("Reward model is necessary for PPO training.")
|
||||||
|
|
||||||
|
if general_args.stage == "ppo" and training_args.deepspeed is not None:
|
||||||
|
raise ValueError("PPO training is incompatible with DeepSpeed, use Accelerate instead.")
|
||||||
|
|
||||||
if general_args.stage == "ppo" and data_args.streaming:
|
if general_args.stage == "ppo" and data_args.streaming:
|
||||||
raise ValueError("Streaming mode does not suppport PPO training currently.")
|
raise ValueError("Streaming mode does not suppport PPO training currently.")
|
||||||
|
|
||||||
|
@ -39,6 +39,9 @@ class PPOPeftTrainer(PPOTrainer, PeftTrainer):
|
|||||||
**kwargs
|
**kwargs
|
||||||
):
|
):
|
||||||
PPOTrainer.__init__(self, **kwargs)
|
PPOTrainer.__init__(self, **kwargs)
|
||||||
|
if getattr(self.accelerator.state, "deepspeed_plugin", None) is not None:
|
||||||
|
raise ValueError("PPOTrainer is incompatible with DeepSpeed.")
|
||||||
|
|
||||||
self.args = training_args
|
self.args = training_args
|
||||||
self.finetuning_args = finetuning_args
|
self.finetuning_args = finetuning_args
|
||||||
self.generating_args = generating_args
|
self.generating_args = generating_args
|
||||||
|
Loading…
x
Reference in New Issue
Block a user