mirror of
				https://github.com/hiyouga/LLaMA-Factory.git
				synced 2025-11-04 18:02:19 +08:00 
			
		
		
		
	fix docker flashattn
Former-commit-id: 0966f5d4616a3877a6b921976dc39e8799831d36
This commit is contained in:
		
							parent
							
								
									42e7489713
								
							
						
					
					
						commit
						a1437c15f7
					
				@ -4,6 +4,7 @@ FROM nvcr.io/nvidia/pytorch:24.02-py3
 | 
			
		||||
 | 
			
		||||
# Define environments
 | 
			
		||||
ENV MAX_JOBS=4
 | 
			
		||||
ENV FLASH_ATTENTION_FORCE_BUILD=TRUE
 | 
			
		||||
 | 
			
		||||
# Define installation arguments
 | 
			
		||||
ARG INSTALL_BNB=false
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user