mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-23 22:32:54 +08:00
Merge pull request #4706 from T-Atlas/main
chore: Update vllm_engine.py to support vllm version >= 0.5.1 Former-commit-id: 563a27dab7e66d9454c6a09404c354d9fca06908
This commit is contained in:
commit
553e517f0f
@ -18,7 +18,7 @@ from typing import TYPE_CHECKING, Any, AsyncGenerator, AsyncIterator, Dict, List
|
|||||||
from ..data import get_template_and_fix_tokenizer
|
from ..data import get_template_and_fix_tokenizer
|
||||||
from ..extras.logging import get_logger
|
from ..extras.logging import get_logger
|
||||||
from ..extras.misc import get_device_count
|
from ..extras.misc import get_device_count
|
||||||
from ..extras.packages import is_vllm_available, is_vllm_version_greater_than_0_5
|
from ..extras.packages import is_vllm_available, is_vllm_version_greater_than_0_5, is_vllm_version_greater_than_0_5_1
|
||||||
from ..model import load_config, load_tokenizer
|
from ..model import load_config, load_tokenizer
|
||||||
from ..model.model_utils.quantization import QuantizationMethod
|
from ..model.model_utils.quantization import QuantizationMethod
|
||||||
from ..model.model_utils.visual import LlavaMultiModalProjectorForYiVLForVLLM
|
from ..model.model_utils.visual import LlavaMultiModalProjectorForYiVLForVLLM
|
||||||
@ -29,7 +29,9 @@ if is_vllm_available():
|
|||||||
from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams
|
from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams
|
||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
|
|
||||||
if is_vllm_version_greater_than_0_5():
|
if is_vllm_version_greater_than_0_5_1():
|
||||||
|
pass
|
||||||
|
elif is_vllm_version_greater_than_0_5():
|
||||||
from vllm.multimodal.image import ImagePixelData
|
from vllm.multimodal.image import ImagePixelData
|
||||||
else:
|
else:
|
||||||
from vllm.sequence import MultiModalData
|
from vllm.sequence import MultiModalData
|
||||||
@ -130,7 +132,9 @@ class VllmEngine(BaseEngine):
|
|||||||
if self.processor is not None and image is not None: # add image features
|
if self.processor is not None and image is not None: # add image features
|
||||||
image_processor: "BaseImageProcessor" = getattr(self.processor, "image_processor")
|
image_processor: "BaseImageProcessor" = getattr(self.processor, "image_processor")
|
||||||
pixel_values = image_processor(image, return_tensors="pt")["pixel_values"]
|
pixel_values = image_processor(image, return_tensors="pt")["pixel_values"]
|
||||||
if is_vllm_version_greater_than_0_5():
|
if is_vllm_version_greater_than_0_5_1():
|
||||||
|
multi_modal_data = {"image": pixel_values}
|
||||||
|
elif is_vllm_version_greater_than_0_5():
|
||||||
multi_modal_data = ImagePixelData(image=pixel_values)
|
multi_modal_data = ImagePixelData(image=pixel_values)
|
||||||
else: # TODO: remove vllm 0.4.3 support
|
else: # TODO: remove vllm 0.4.3 support
|
||||||
multi_modal_data = MultiModalData(type=MultiModalData.Type.IMAGE, data=pixel_values)
|
multi_modal_data = MultiModalData(type=MultiModalData.Type.IMAGE, data=pixel_values)
|
||||||
|
@ -81,3 +81,8 @@ def is_vllm_available():
|
|||||||
@lru_cache
|
@lru_cache
|
||||||
def is_vllm_version_greater_than_0_5():
|
def is_vllm_version_greater_than_0_5():
|
||||||
return _get_package_version("vllm") >= version.parse("0.5.0")
|
return _get_package_version("vllm") >= version.parse("0.5.0")
|
||||||
|
|
||||||
|
|
||||||
|
@lru_cache
|
||||||
|
def is_vllm_version_greater_than_0_5_1():
|
||||||
|
return _get_package_version("vllm") >= version.parse("0.5.1")
|
||||||
|
Loading…
x
Reference in New Issue
Block a user