From 2d421c57bf85c043f290b757c1f4d0ffe633f23f Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Tue, 1 Apr 2025 00:15:12 +0800 Subject: [PATCH] [data] fix qwen2.5 omni collator (#7553) --- src/llamafactory/data/collator.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llamafactory/data/collator.py b/src/llamafactory/data/collator.py index 93c7349c..93958892 100644 --- a/src/llamafactory/data/collator.py +++ b/src/llamafactory/data/collator.py @@ -193,7 +193,7 @@ class MultiModalDataCollatorForSeq2Seq(DataCollatorForSeq2Seq): if "second_per_grid_ts" in mm_inputs: # for qwen2vl rope_index_kwargs["second_per_grid_ts"] = mm_inputs.get("second_per_grid_ts") - if getattr(self.model.config, "model_type", None) == "qwen2_5_omni": # for qwen2omni + if getattr(self.model.config, "model_type", None) == "qwen2_5_omni_thinker": # for qwen2omni feature_attention_mask = mm_inputs.get("feature_attention_mask", None) if feature_attention_mask is not None: audio_feature_lengths = torch.sum(