diff --git a/src/llamafactory/train/ppo/trainer.py b/src/llamafactory/train/ppo/trainer.py index 6a05b704..31d461e3 100644 --- a/src/llamafactory/train/ppo/trainer.py +++ b/src/llamafactory/train/ppo/trainer.py @@ -106,7 +106,8 @@ class CustomPPOTrainer(PPOTrainer, Trainer): DistributedDataParallelKwargs(find_unused_parameters=training_args.ddp_find_unused_parameters) ] ppo_config.accelerator_kwargs["deepspeed_plugin"] = training_args.deepspeed_plugin - if ppo_config.log_with == "tensorboard": # tensorboard raises error about accelerator_kwargs + if ppo_config.log_with is not None: + logger.warning("PPOTrainer cannot use external logger when DeepSpeed is enabled.") ppo_config.log_with = None # Create optimizer and scheduler