mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-02 03:32:50 +08:00
fix docker flashattn
Former-commit-id: e3141f5f1b435d12c71d8b1fc6ade6e69deead71
This commit is contained in:
parent
d3b7c489f2
commit
cba994b9dc
@ -4,6 +4,7 @@ FROM nvcr.io/nvidia/pytorch:24.02-py3
|
||||
|
||||
# Define environments
|
||||
ENV MAX_JOBS=4
|
||||
ENV FLASH_ATTENTION_FORCE_BUILD=TRUE
|
||||
|
||||
# Define installation arguments
|
||||
ARG INSTALL_BNB=false
|
||||
|
Loading…
x
Reference in New Issue
Block a user