From b2c3001f8e642ddbeeef47341272c37f9e46c0e1 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Sun, 5 Nov 2023 12:21:07 +0800 Subject: [PATCH] fix #1365 Former-commit-id: 0277d120e62164bb7fa1d6043b8fcc52c881fe96 --- src/llmtuner/tuner/core/loader.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/llmtuner/tuner/core/loader.py b/src/llmtuner/tuner/core/loader.py index 6389c705..507d862e 100644 --- a/src/llmtuner/tuner/core/loader.py +++ b/src/llmtuner/tuner/core/loader.py @@ -196,6 +196,7 @@ def load_model_and_tokenizer( # Fix LM head (for ChatGLM2 and ChatGLM3) if getattr(config, "model_type", None) == "chatglm": setattr(model, "lm_head", model.transformer.output_layer) + setattr(model, "_keys_to_ignore_on_save", ["lm_head.weight"]) # Register auto class to save the custom code files. if isinstance(config, PretrainedConfig) and "AutoConfig" in getattr(config, "auto_map", {}):