From e7556b591ebc1d04079034631eaa6c7dd0b9f38d Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Thu, 6 Mar 2025 15:25:08 +0800 Subject: [PATCH] [deps] upgrade vllm (#7183) Former-commit-id: d739fddb10f51f422c12b1e7d5d7208309bf6c91 --- README.md | 4 ++-- README_zh.md | 4 ++-- setup.py | 2 +- src/llamafactory/hparams/parser.py | 2 +- 4 files changed, 6 insertions(+), 6 deletions(-) diff --git a/README.md b/README.md index 7b699b3b..5d7d0f10 100644 --- a/README.md +++ b/README.md @@ -393,7 +393,7 @@ huggingface-cli login | Mandatory | Minimum | Recommend | | ------------ | ------- | --------- | | python | 3.9 | 3.10 | -| torch | 1.13.1 | 2.4.0 | +| torch | 1.13.1 | 2.5.1 | | transformers | 4.41.2 | 4.49.0 | | datasets | 2.16.0 | 3.2.0 | | accelerate | 0.34.0 | 1.2.1 | @@ -405,7 +405,7 @@ huggingface-cli login | CUDA | 11.6 | 12.2 | | deepspeed | 0.10.0 | 0.16.2 | | bitsandbytes | 0.39.0 | 0.43.1 | -| vllm | 0.4.3 | 0.7.2 | +| vllm | 0.4.3 | 0.7.3 | | flash-attn | 2.3.0 | 2.7.2 | ### Hardware Requirement diff --git a/README_zh.md b/README_zh.md index 9ac5418b..0417bccf 100644 --- a/README_zh.md +++ b/README_zh.md @@ -395,7 +395,7 @@ huggingface-cli login | 必需项 | 至少 | 推荐 | | ------------ | ------- | --------- | | python | 3.9 | 3.10 | -| torch | 1.13.1 | 2.4.0 | +| torch | 1.13.1 | 2.5.1 | | transformers | 4.41.2 | 4.49.0 | | datasets | 2.16.0 | 3.2.0 | | accelerate | 0.34.0 | 1.2.1 | @@ -407,7 +407,7 @@ huggingface-cli login | CUDA | 11.6 | 12.2 | | deepspeed | 0.10.0 | 0.16.2 | | bitsandbytes | 0.39.0 | 0.43.1 | -| vllm | 0.4.3 | 0.7.2 | +| vllm | 0.4.3 | 0.7.3 | | flash-attn | 2.3.0 | 2.7.2 | ### 硬件依赖 diff --git a/setup.py b/setup.py index 8e55af97..5bf8d0f3 100644 --- a/setup.py +++ b/setup.py @@ -54,7 +54,7 @@ extra_require = { "gptq": ["optimum>=1.17.0", "auto-gptq>=0.5.0"], "awq": ["autoawq"], "aqlm": ["aqlm[gpu]>=1.1.0"], - "vllm": ["vllm>=0.4.3,<=0.7.2"], + "vllm": ["vllm>=0.4.3,<=0.7.3"], "galore": ["galore-torch"], "apollo": ["apollo-torch"], "badam": ["badam>=1.2.1"], diff --git a/src/llamafactory/hparams/parser.py b/src/llamafactory/hparams/parser.py index 911b78ec..b40148fd 100644 --- a/src/llamafactory/hparams/parser.py +++ b/src/llamafactory/hparams/parser.py @@ -137,7 +137,7 @@ def _check_extra_dependencies( check_version("mixture-of-depth>=1.1.6", mandatory=True) if model_args.infer_backend == "vllm": - check_version("vllm>=0.4.3,<=0.7.2") + check_version("vllm>=0.4.3,<=0.7.3") check_version("vllm", mandatory=True) if finetuning_args.use_galore: