mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-14 23:58:11 +08:00
fix docker flashattn
Former-commit-id: 0966f5d4616a3877a6b921976dc39e8799831d36
This commit is contained in:
parent
42e7489713
commit
a1437c15f7
@ -4,6 +4,7 @@ FROM nvcr.io/nvidia/pytorch:24.02-py3
|
|||||||
|
|
||||||
# Define environments
|
# Define environments
|
||||||
ENV MAX_JOBS=4
|
ENV MAX_JOBS=4
|
||||||
|
ENV FLASH_ATTENTION_FORCE_BUILD=TRUE
|
||||||
|
|
||||||
# Define installation arguments
|
# Define installation arguments
|
||||||
ARG INSTALL_BNB=false
|
ARG INSTALL_BNB=false
|
||||||
|
Loading…
x
Reference in New Issue
Block a user