mirror of
				https://github.com/hiyouga/LLaMA-Factory.git
				synced 2025-11-04 18:02:19 +08:00 
			
		
		
		
	[script] fix vllm version (#7193)
Former-commit-id: ababdde597b2b9bf0ab3f30f036bc8d97de07f03
This commit is contained in:
		
							parent
							
								
									2635794727
								
							
						
					
					
						commit
						f4ec4fa6ad
					
				@ -55,7 +55,7 @@ def vllm_infer(
 | 
			
		||||
    Performs batch generation using vLLM engine, which supports tensor parallelism.
 | 
			
		||||
    Usage: python vllm_infer.py --model_name_or_path meta-llama/Llama-2-7b-hf --template llama --dataset alpaca_en_demo
 | 
			
		||||
    """
 | 
			
		||||
    check_version("vllm>=0.4.3,<=0.7.2")
 | 
			
		||||
    check_version("vllm>=0.4.3,<=0.7.3")
 | 
			
		||||
    if pipeline_parallel_size > get_device_count():
 | 
			
		||||
        raise ValueError("Pipeline parallel size should be smaller than the number of gpus.")
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user