mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-11-06 02:42:15 +08:00
Fix tokenizer max length (#6632)
Former-commit-id: 1807c7ba033985490aa7c8c39d880da6af983b92
This commit is contained in:
parent
068d44b509
commit
1e2b1cedec
@ -85,6 +85,9 @@ def load_tokenizer(model_args: "ModelArguments") -> "TokenizerModule":
|
|||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise OSError("Failed to load tokenizer.") from e
|
raise OSError("Failed to load tokenizer.") from e
|
||||||
|
|
||||||
|
if model_args.model_max_length is not None and tokenizer.model_max_length != model_args.model_max_length:
|
||||||
|
tokenizer.model_max_length = model_args.model_max_length
|
||||||
|
|
||||||
if model_args.new_special_tokens is not None:
|
if model_args.new_special_tokens is not None:
|
||||||
num_added_tokens = tokenizer.add_special_tokens(
|
num_added_tokens = tokenizer.add_special_tokens(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user