# https://hub.docker.com/r/ascendai/cann/tags ARG BASE_IMAGE=ascendai/cann:8.0.0-910b-ubuntu22.04-py3.11 FROM ${BASE_IMAGE} # Installation arguments ARG PIP_INDEX=https://pypi.org/simple ARG EXTRAS=torch-npu,metrics ARG HTTP_PROXY="" # Define environments ENV MAX_JOBS=16 ENV FLASH_ATTENTION_FORCE_BUILD=TRUE ENV VLLM_WORKER_MULTIPROC_METHOD=spawn ENV DEBIAN_FRONTEND=noninteractive ENV NODE_OPTIONS="" ENV PIP_ROOT_USER_ACTION=ignore ENV http_proxy="${HTTP_PROXY}" ENV https_proxy="${HTTP_PROXY}" # Use Bash instead of default /bin/sh SHELL ["/bin/bash", "-c"] # Set the working directory WORKDIR /app # Change pip source RUN pip config set global.index-url "${PIP_INDEX}" && \ pip config set global.extra-index-url "${PIP_INDEX}" && \ python -m pip install --upgrade pip # Install the requirements COPY requirements.txt /app RUN pip install --no-cache-dir -r requirements.txt # Copy the rest of the application into the image COPY . /app # Install LLaMA Factory RUN pip install --no-cache-dir -e ".[${EXTRAS}]" --no-build-isolation # Set up volumes VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/root/.cache/openmind", "/app/shared_data", "/app/output" ] # Expose port 7860 for LLaMA Board ENV GRADIO_SERVER_PORT=7860 EXPOSE 7860 # Expose port 8000 for API service ENV API_PORT=8000 EXPOSE 8000 # unset proxy ENV http_proxy= ENV https_proxy= # Reset pip config RUN pip config unset global.index-url && \ pip config unset global.extra-index-url