From 035e98035cbce8e8e960b3b3594d6fdb5257f771 Mon Sep 17 00:00:00 2001 From: Kingsley <82590017+Kuangdd01@users.noreply.github.com> Date: Sun, 27 Apr 2025 23:12:28 +0800 Subject: [PATCH] fix attn patch for kimivl (#7867) --- src/llamafactory/model/model_utils/attention.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/src/llamafactory/model/model_utils/attention.py b/src/llamafactory/model/model_utils/attention.py index 0f95d3f2..860bf891 100644 --- a/src/llamafactory/model/model_utils/attention.py +++ b/src/llamafactory/model/model_utils/attention.py @@ -69,6 +69,9 @@ def configure_attn_implementation( if getattr(config, "model_type", None) == "internlm2": # special case for custom models setattr(config, "attn_implementation", requested_attn_implementation) + elif getattr(config, "model_type", None) == "kimi_vl": + setattr(config.vision_config, "_attn_implementation", requested_attn_implementation) + setattr(config.text_config, "_attn_implementation", requested_attn_implementation) else: setattr(config, "_attn_implementation", requested_attn_implementation)