mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-14 19:06:26 +08:00
Update finetuning_args.py
This commit is contained in:
@@ -74,6 +74,10 @@ class RLHFArguments:
|
|||||||
default=None,
|
default=None,
|
||||||
metadata={"help": "Log with either 'wandb' or 'tensorboard' in PPO training."}
|
metadata={"help": "Log with either 'wandb' or 'tensorboard' in PPO training."}
|
||||||
)
|
)
|
||||||
|
ppo_epochs: Optional[int] = field(
|
||||||
|
default=4,
|
||||||
|
metadata={"help": "Number of optimisation epochs per batch of samples"},
|
||||||
|
)
|
||||||
ppo_score_norm: Optional[bool] = field(
|
ppo_score_norm: Optional[bool] = field(
|
||||||
default=False,
|
default=False,
|
||||||
metadata={"help": "Use score normalization in PPO training."}
|
metadata={"help": "Use score normalization in PPO training."}
|
||||||
|
|||||||
Reference in New Issue
Block a user