mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-07-31 10:42:50 +08:00
59 lines
1.5 KiB
Docker
59 lines
1.5 KiB
Docker
# https://hub.docker.com/r/ascendai/cann/tags
|
|
ARG BASE_IMAGE=ascendai/cann:8.0.0-910b-ubuntu22.04-py3.11
|
|
FROM ${BASE_IMAGE}
|
|
|
|
# Installation arguments
|
|
ARG PIP_INDEX=https://pypi.org/simple
|
|
ARG EXTRAS=torch-npu,metrics
|
|
ARG HTTP_PROXY=""
|
|
|
|
# Define environments
|
|
ENV MAX_JOBS=16
|
|
ENV FLASH_ATTENTION_FORCE_BUILD=TRUE
|
|
ENV VLLM_WORKER_MULTIPROC_METHOD=spawn
|
|
ENV DEBIAN_FRONTEND=noninteractive
|
|
ENV NODE_OPTIONS=""
|
|
ENV PIP_ROOT_USER_ACTION=ignore
|
|
ENV http_proxy="${HTTP_PROXY}"
|
|
ENV https_proxy="${HTTP_PROXY}"
|
|
|
|
# Use Bash instead of default /bin/sh
|
|
SHELL ["/bin/bash", "-c"]
|
|
|
|
# Set the working directory
|
|
WORKDIR /app
|
|
|
|
# Change pip source
|
|
RUN pip config set global.index-url "${PIP_INDEX}" && \
|
|
pip config set global.extra-index-url "${PIP_INDEX}" && \
|
|
python -m pip install --upgrade pip
|
|
|
|
# Install the requirements
|
|
COPY requirements.txt /app
|
|
RUN pip install --no-cache-dir -r requirements.txt
|
|
|
|
# Copy the rest of the application into the image
|
|
COPY . /app
|
|
|
|
# Install LLaMA Factory
|
|
RUN pip install --no-cache-dir -e ".[${EXTRAS}]" --no-build-isolation
|
|
|
|
# Set up volumes
|
|
VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/root/.cache/openmind", "/app/shared_data", "/app/output" ]
|
|
|
|
# Expose port 7860 for LLaMA Board
|
|
ENV GRADIO_SERVER_PORT=7860
|
|
EXPOSE 7860
|
|
|
|
# Expose port 8000 for API service
|
|
ENV API_PORT=8000
|
|
EXPOSE 8000
|
|
|
|
# unset proxy
|
|
ENV http_proxy=
|
|
ENV https_proxy=
|
|
|
|
# Reset pip config
|
|
RUN pip config unset global.index-url && \
|
|
pip config unset global.extra-index-url
|