mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 04:32:50 +08:00
tiny fix
Former-commit-id: 0239d29fa02a88b50f27caa706834f3c3ce0262d
This commit is contained in:
parent
9ead5a2d21
commit
2e42e38ff2
@ -189,7 +189,7 @@ def resize_embedding_layer(model: "PreTrainedModel", tokenizer: "PreTrainedToken
|
||||
Resize token embeddings.
|
||||
"""
|
||||
old_vocab_size = model.get_input_embeddings().weight.size(0)
|
||||
new_vocab_size = len(tokenizer)
|
||||
if new_vocab_size != old_vocab_size:
|
||||
model.resize_token_embeddings(new_vocab_size, pad_to_multiple_of=64)
|
||||
logger.info("Resized embedding tokens from {} to {}.".format(old_vocab_size, new_vocab_size))
|
||||
if len(tokenizer) != old_vocab_size:
|
||||
model.resize_token_embeddings(len(tokenizer), pad_to_multiple_of=64)
|
||||
new_vocab_size = model.get_input_embeddings().weight.size(0)
|
||||
logger.info("Resized token embeddings from {} to {}.".format(old_vocab_size, new_vocab_size))
|
||||
|
Loading…
x
Reference in New Issue
Block a user