[train] fix adjusting logits size after adding special tokens (#8823)

This commit is contained in:
davidlightmysterion 2025-08-05 20:35:07 +08:00 committed by GitHub
parent b3445a75f0
commit 48615ddb07
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -67,4 +67,5 @@ def resize_embedding_layer(model: "PreTrainedModel", tokenizer: "PreTrainedToken
_noisy_mean_initialization(model.get_input_embeddings().weight.data, num_new_tokens)
_noisy_mean_initialization(model.get_output_embeddings().weight.data, num_new_tokens)
model.config.vocab_size = new_embedding_size
logger.info_rank0(f"Resized token embeddings from {current_embedding_size} to {new_embedding_size}.")