From 0d488085b8dbe71dccc9b87818f6309406aa9eb2 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Sun, 5 Nov 2023 12:21:07 +0800 Subject: [PATCH] fix #1365 Former-commit-id: 63ff909310eec1db64bd4bcc1ce013dbd96f0f99 --- src/llmtuner/tuner/core/loader.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/llmtuner/tuner/core/loader.py b/src/llmtuner/tuner/core/loader.py index 6389c705..507d862e 100644 --- a/src/llmtuner/tuner/core/loader.py +++ b/src/llmtuner/tuner/core/loader.py @@ -196,6 +196,7 @@ def load_model_and_tokenizer( # Fix LM head (for ChatGLM2 and ChatGLM3) if getattr(config, "model_type", None) == "chatglm": setattr(model, "lm_head", model.transformer.output_layer) + setattr(model, "_keys_to_ignore_on_save", ["lm_head.weight"]) # Register auto class to save the custom code files. if isinstance(config, PretrainedConfig) and "AutoConfig" in getattr(config, "auto_map", {}):