mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-14 10:56:56 +08:00
[model] fix kv cache (#7564)
This commit is contained in:
@@ -56,7 +56,7 @@ def vllm_infer(
|
||||
|
||||
Usage: python vllm_infer.py --model_name_or_path meta-llama/Llama-2-7b-hf --template llama --dataset alpaca_en_demo
|
||||
"""
|
||||
check_version("vllm>=0.4.3,<=0.7.3")
|
||||
check_version("vllm>=0.4.3,<=0.8.2")
|
||||
if pipeline_parallel_size > get_device_count():
|
||||
raise ValueError("Pipeline parallel size should be smaller than the number of gpus.")
|
||||
|
||||
|
||||
Reference in New Issue
Block a user