From f00ad6b4f8c9efc7ea553417838fe8ecb046bbc0 Mon Sep 17 00:00:00 2001 From: A-Cepheus <60658915+A-Cepheus@users.noreply.github.com> Date: Mon, 22 Jan 2024 16:04:39 +0800 Subject: [PATCH] =?UTF-8?q?=F0=9F=90=9E=20fix:=20typo?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Former-commit-id: b06a31e76ac623ea89ecc567dd9e9ec88dfdb585 --- src/llmtuner/model/patcher.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llmtuner/model/patcher.py b/src/llmtuner/model/patcher.py index 52690e68..892a9165 100644 --- a/src/llmtuner/model/patcher.py +++ b/src/llmtuner/model/patcher.py @@ -284,7 +284,7 @@ def patch_model( if is_trainable: _prepare_model_for_training(model, model_args) - if getattr(config, "model_type", None) == "mixtral" and is_deepspeed_zero3_enabled(): + if getattr(model.config, "model_type", None) == "mixtral" and is_deepspeed_zero3_enabled(): from deepspeed.utils import set_z3_leaf_modules from transformers.models.mixtral.modeling_mixtral import MixtralSparseMoeBlock set_z3_leaf_modules(model, [MixtralSparseMoeBlock])