From e09d68985f55ecd1ed826ddba976c28fca9c37be Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Wed, 15 May 2024 15:37:07 +0800 Subject: [PATCH] Update patcher.py Former-commit-id: 5a0c8a8d343adb15b510f65286ee08f33b1b2751 --- src/llmtuner/model/patcher.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/src/llmtuner/model/patcher.py b/src/llmtuner/model/patcher.py index bddea594..70aed709 100644 --- a/src/llmtuner/model/patcher.py +++ b/src/llmtuner/model/patcher.py @@ -17,7 +17,7 @@ from .utils.moe import add_z3_leaf_module, configure_moe from .utils.quantization import configure_quantization from .utils.rope import configure_rope from .utils.valuehead import prepare_valuehead_model -from .utils.visual import autocast_projector_dtype, configure_hidden_size, configure_visual +from .utils.visual import autocast_projector_dtype, configure_visual_model if TYPE_CHECKING: @@ -54,8 +54,7 @@ def patch_config( configure_longlora(config, model_args, is_trainable) configure_quantization(config, tokenizer, model_args, init_kwargs) configure_moe(config, model_args, is_trainable) - configure_hidden_size(config) - configure_visual(config, model_args) + configure_visual_model(config) if model_args.use_cache and not is_trainable: setattr(config, "use_cache", True)