mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-17 20:30:36 +08:00
support BLOOM models
This commit is contained in:
@@ -9,9 +9,9 @@ from transformers.tokenization_utils import PreTrainedTokenizer
|
||||
from .other import IGNORE_INDEX
|
||||
|
||||
|
||||
class DataCollatorForLLaMA(DataCollatorWithPadding):
|
||||
class DynamicDataCollatorWithPadding(DataCollatorWithPadding):
|
||||
r"""
|
||||
Data collator for LLaMA. It is capable of dynamically padding for batched data.
|
||||
Inherits DataCollatorWithPadding. It is capable of dynamically padding for batched data.
|
||||
"""
|
||||
def __init__(
|
||||
self,
|
||||
|
||||
Reference in New Issue
Block a user