From 3501257780b1cedac84c86393559ee89b0984cf0 Mon Sep 17 00:00:00 2001 From: BUAADreamer <1428195643@qq.com> Date: Sat, 28 Sep 2024 00:59:14 +0800 Subject: [PATCH] add tests Former-commit-id: f3be3d21e7efcb2f596ee96f5a97bb53438f2d89 --- tests/data/test_mm_plugin.py | 17 ----------------- 1 file changed, 17 deletions(-) diff --git a/tests/data/test_mm_plugin.py b/tests/data/test_mm_plugin.py index 722d093e..a48ad795 100644 --- a/tests/data/test_mm_plugin.py +++ b/tests/data/test_mm_plugin.py @@ -136,23 +136,6 @@ def test_llava_plugin(): _check_plugin(**check_inputs) -def test_idefics2_plugin(): - tokenizer, processor = _load_tokenizer_module(model_name_or_path="HuggingFaceM4/idefics2-8b") - idefics2_plugin = get_mm_plugin(name="idefics2", image_token="") - check_inputs = {"plugin": idefics2_plugin, "tokenizer": tokenizer, "processor": processor} - mm_messages = copy.deepcopy(MM_MESSAGES) - fake_image_token = processor.fake_image_token.content - image_str = f"{fake_image_token}{'' * processor.image_seq_len}{fake_image_token}" - image_str = image_str * 5 - for message in mm_messages: - content = message["content"] - content = content.replace("", image_str) - content = content.replace(f"{fake_image_token}{fake_image_token}", f"{fake_image_token}") - message["content"] = content - check_inputs["expected_mm_inputs"] = _get_mm_inputs(processor) - _check_plugin(**check_inputs) - - def test_llava_next_plugin(): tokenizer, processor = _load_tokenizer_module(model_name_or_path="llava-hf/llava-v1.6-vicuna-7b-hf") llava_next_plugin = get_mm_plugin(name="llava_next", image_token="")