mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-22 13:42:51 +08:00
tiny fix
Former-commit-id: 44747cebd28d0b800196f032e18d2f4ff51ee5b3
This commit is contained in:
parent
4987aa32ba
commit
cbb93a2b47
@ -15,8 +15,6 @@
|
||||
from functools import partial
|
||||
from typing import TYPE_CHECKING, Callable, Literal, Optional, Tuple
|
||||
|
||||
from datasets.arrow_writer import OptimizedTypedSequence, TypedSequence
|
||||
|
||||
from .processors.feedback import preprocess_feedback_dataset
|
||||
from .processors.pairwise import preprocess_pairwise_dataset, print_pairwise_dataset_example
|
||||
from .processors.pretrain import preprocess_pretrain_dataset
|
||||
@ -53,6 +51,7 @@ def get_preprocess_and_print_func(
|
||||
elif stage == "sft" and not training_args.predict_with_generate:
|
||||
if data_args.packing:
|
||||
if data_args.neat_packing:
|
||||
from datasets.arrow_writer import OptimizedTypedSequence, TypedSequence
|
||||
|
||||
def __init__(self, data, **kwargs):
|
||||
return TypedSequence.__init__(
|
||||
|
Loading…
x
Reference in New Issue
Block a user