mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-22 22:02:51 +08:00
parent
bc5e97295e
commit
f8497921fe
@ -32,7 +32,7 @@ RUN EXTRA_PACKAGES="metrics"; \
|
|||||||
EXTRA_PACKAGES="${EXTRA_PACKAGES},deepspeed"; \
|
EXTRA_PACKAGES="${EXTRA_PACKAGES},deepspeed"; \
|
||||||
fi; \
|
fi; \
|
||||||
pip install -e .[$EXTRA_PACKAGES] && \
|
pip install -e .[$EXTRA_PACKAGES] && \
|
||||||
pip uninstall -y transformer-engine
|
pip uninstall -y transformer-engine flash-attn
|
||||||
|
|
||||||
# Set up volumes
|
# Set up volumes
|
||||||
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
|
VOLUME [ "/root/.cache/huggingface/", "/app/data", "/app/output" ]
|
||||||
|
@ -13,10 +13,11 @@ from .base_engine import BaseEngine, Response
|
|||||||
if is_vllm_available():
|
if is_vllm_available():
|
||||||
from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams
|
from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams
|
||||||
from vllm.lora.request import LoRARequest
|
from vllm.lora.request import LoRARequest
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from vllm.multimodal import MultiModalData # vllm==0.5.0
|
from vllm.multimodal import MultiModalData # type: ignore (for vllm>=0.5.0)
|
||||||
except ImportError:
|
except ImportError:
|
||||||
from vllm.sequence import MultiModalData # vllm<0.5.0
|
from vllm.sequence import MultiModalData # for vllm<0.5.0
|
||||||
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user