Spaces:
Runtime error
Runtime error
| # Default use the NVIDIA official image with PyTorch 2.6.0 | |
| # https://docs.nvidia.com/deeplearning/frameworks/pytorch-release-notes/index.html | |
| ARG BASE_IMAGE=nvcr.io/nvidia/pytorch:24.12-py3 | |
| FROM ${BASE_IMAGE} | |
| # Define environments | |
| ENV MAX_JOBS=4 | |
| ENV FLASH_ATTENTION_FORCE_BUILD=TRUE | |
| ENV VLLM_WORKER_MULTIPROC_METHOD=spawn | |
| # Define installation arguments | |
| ARG INSTALL_BNB=false | |
| ARG INSTALL_VLLM=false | |
| ARG INSTALL_DEEPSPEED=false | |
| ARG INSTALL_FLASHATTN=false | |
| ARG INSTALL_LIGER_KERNEL=false | |
| ARG INSTALL_HQQ=false | |
| ARG INSTALL_EETQ=false | |
| ARG PIP_INDEX=https://pypi.org/simple | |
| ARG HTTP_PROXY= | |
| # Set the working directory | |
| WORKDIR /app | |
| # Set http proxy | |
| RUN if [ -n "$HTTP_PROXY" ]; then \ | |
| echo "Configuring proxy..."; \ | |
| export http_proxy=$HTTP_PROXY; \ | |
| export https_proxy=$HTTP_PROXY; \ | |
| fi | |
| # Install the requirements | |
| COPY requirements.txt /app | |
| RUN pip config set global.index-url "$PIP_INDEX" && \ | |
| pip config set global.extra-index-url "$PIP_INDEX" && \ | |
| python -m pip install --upgrade pip && \ | |
| if [ -n "$HTTP_PROXY" ]; then \ | |
| python -m pip install --proxy=$HTTP_PROXY -r requirements.txt; \ | |
| else \ | |
| python -m pip install -r requirements.txt; \ | |
| fi | |
| # Copy the rest of the application into the image | |
| COPY . /app | |
| # Install the LLaMA Factory | |
| RUN EXTRA_PACKAGES="metrics"; \ | |
| if [ "$INSTALL_BNB" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},bitsandbytes"; \ | |
| fi; \ | |
| if [ "$INSTALL_VLLM" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},vllm"; \ | |
| fi; \ | |
| if [ "$INSTALL_DEEPSPEED" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},deepspeed"; \ | |
| fi; \ | |
| if [ "$INSTALL_LIGER_KERNEL" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},liger-kernel"; \ | |
| fi; \ | |
| if [ "$INSTALL_HQQ" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},hqq"; \ | |
| fi; \ | |
| if [ "$INSTALL_EETQ" == "true" ]; then \ | |
| EXTRA_PACKAGES="${EXTRA_PACKAGES},eetq"; \ | |
| fi; \ | |
| if [ -n "$HTTP_PROXY" ]; then \ | |
| pip install --proxy=$HTTP_PROXY -e ".[$EXTRA_PACKAGES]"; \ | |
| else \ | |
| pip install -e ".[$EXTRA_PACKAGES]"; \ | |
| fi | |
| # Rebuild flash attention | |
| RUN pip uninstall -y transformer-engine flash-attn && \ | |
| if [ "$INSTALL_FLASHATTN" == "true" ]; then \ | |
| pip uninstall -y ninja && \ | |
| if [ -n "$HTTP_PROXY" ]; then \ | |
| pip install --proxy=$HTTP_PROXY ninja && \ | |
| pip install --proxy=$HTTP_PROXY --no-cache-dir flash-attn --no-build-isolation; \ | |
| else \ | |
| pip install ninja && \ | |
| pip install --no-cache-dir flash-attn --no-build-isolation; \ | |
| fi; \ | |
| fi | |
| # Unset http proxy | |
| RUN if [ -n "$HTTP_PROXY" ]; then \ | |
| unset http_proxy; \ | |
| unset https_proxy; \ | |
| fi | |
| # Set up volumes | |
| VOLUME [ "/root/.cache/huggingface", "/root/.cache/modelscope", "/app/data", "/app/output" ] | |
| # Expose port 7860 for the LLaMA Board | |
| ENV GRADIO_SERVER_PORT 7860 | |
| EXPOSE 7860 | |
| # Expose port 8000 for the API service | |
| ENV API_PORT 8000 | |
| EXPOSE 8000 | |