mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 12:42:51 +08:00
fix bug
Former-commit-id: 8b681ee273c28813c599d9d55b2a3540c8ac257d
This commit is contained in:
parent
64eead3fb1
commit
6493558c3b
@ -66,7 +66,7 @@ class CustomPPOTrainer(PPOTrainer, Trainer):
|
|||||||
if self.args.max_steps > 0:
|
if self.args.max_steps > 0:
|
||||||
logger.info("max_steps is given, it will override any value given in num_train_epochs")
|
logger.info("max_steps is given, it will override any value given in num_train_epochs")
|
||||||
|
|
||||||
if reward_model is not None:
|
if finetuning_args.reward_model_type == "full":
|
||||||
if self.is_deepspeed_enabled:
|
if self.is_deepspeed_enabled:
|
||||||
if not (
|
if not (
|
||||||
getattr(reward_model.pretrained_model, "is_loaded_in_8bit", False)
|
getattr(reward_model.pretrained_model, "is_loaded_in_8bit", False)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user