fix reserved label len

This commit is contained in:
hiyouga
2024-02-04 17:54:26 +08:00
parent 19d33ede13
commit db0ab4d601
3 changed files with 33 additions and 13 deletions

View File

@@ -21,10 +21,10 @@ class DataArguments:
default="train", metadata={"help": "Which dataset split to use for training and evaluation."}
)
cutoff_len: Optional[int] = field(
default=1024, metadata={"help": "The maximum length of the model inputs after tokenization."}
default=1024, metadata={"help": "The cutoff length of the model inputs after tokenization."}
)
reserved_label_len: Optional[int] = field(
default=1, metadata={"help": "The maximum length reserved for label after tokenization."}
default=1, metadata={"help": "The minimum cutoff length reserved for label after tokenization."}
)
train_on_prompt: Optional[bool] = field(
default=False, metadata={"help": "Whether to disable the mask on the prompt or not."}
@@ -57,7 +57,7 @@ class DataArguments:
ignore_pad_token_for_loss: Optional[bool] = field(
default=True,
metadata={
"help": "Whether to ignore the tokens corresponding to padded labels in the loss computation or not."
"help": "Whether or not to ignore the tokens corresponding to padded labels in the loss computation."
},
)
val_size: Optional[float] = field(