mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-17 12:20:37 +08:00
implement efficient packing without cross-contamination attention
Former-commit-id: b2c367bc61
This commit is contained in:
@@ -84,6 +84,10 @@ class DataArguments:
|
||||
"help": "Whether or not to pack the sequences in training. Will automatically enable in pre-training."
|
||||
},
|
||||
)
|
||||
efficient_packing: Optional[bool] = field(
|
||||
default=None,
|
||||
metadata={"help": "Whether or not to pack the sequences without cross-contamination attention for efficient training."},
|
||||
)
|
||||
tokenized_path: Optional[str] = field(
|
||||
default=None,
|
||||
metadata={"help": "Path to save or load the tokenized datasets."},
|
||||
|
||||
Reference in New Issue
Block a user