mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 00:28:10 +08:00
parent
2542b62d77
commit
6f9d55b8eb
@ -182,7 +182,7 @@ def resize_embedding_layer(model: "PreTrainedModel", tokenizer: "PreTrainedToken
|
|||||||
return
|
return
|
||||||
|
|
||||||
old_vocab_size = model.get_input_embeddings().weight.size(0)
|
old_vocab_size = model.get_input_embeddings().weight.size(0)
|
||||||
if len(tokenizer) != old_vocab_size:
|
if len(tokenizer) > old_vocab_size:
|
||||||
model.resize_token_embeddings(len(tokenizer), pad_to_multiple_of=64)
|
model.resize_token_embeddings(len(tokenizer), pad_to_multiple_of=64)
|
||||||
new_vocab_size = model.get_input_embeddings().weight.size(0)
|
new_vocab_size = model.get_input_embeddings().weight.size(0)
|
||||||
logger.info("Resized token embeddings from {} to {}.".format(old_vocab_size, new_vocab_size))
|
logger.info("Resized token embeddings from {} to {}.".format(old_vocab_size, new_vocab_size))
|
||||||
|
Loading…
x
Reference in New Issue
Block a user