gguf-test-2 / Dockerfile
broadfield-dev's picture
Update Dockerfile
c7a088c verified
# Use a Python 3.10 base image with Debian Bookworm
FROM python:3.10-bookworm
# Set environment variables to avoid Python buffering and ensure non-interactive installs
ENV PYTHONUNBUFFERED=1 \
DEBIAN_FRONTEND=noninteractive
# Install system dependencies, including OpenBLAS, Git, and build tools
RUN apt-get update && apt-get install -y \
git \
build-essential \
cmake \
libopenblas-dev \
&& rm -rf /var/lib/apt/lists/*
# Create a non-root user with configurable UID/GID
ARG USER_ID=1000
ARG GROUP_ID=1000
RUN groupadd -g ${GROUP_ID} appuser && \
useradd -m -u ${USER_ID} -g ${GROUP_ID} -s /bin/bash appuser
# Set up Hugging Face cache directory and /app with proper permissions
RUN mkdir -p /home/appuser/.cache/huggingface /app && \
chown -R appuser:appuser /home/appuser/.cache /app
# Set working directory
WORKDIR /app
# Clone llama-cpp-python repository with submodules as appuser
USER appuser
RUN git clone --recursive https://github.com/abetlen/llama-cpp-python.git /app/llama-cpp-python
# Set working directory to llama-cpp-python
WORKDIR /app/llama-cpp-python
# Update llama.cpp submodule to the latest version as appuser
RUN git submodule update --remote vendor/llama.cpp
# Set environment variables for building with OpenBLAS
ENV FORCE_CMAKE=1
ENV CMAKE_ARGS="-DGGML_BLAS=ON -DGGML_BLAS_VENDOR=OpenBLAS -DLLAMA_CURL=OFF"
# Install llama-cpp-python from source as appuser
RUN pip install . --user --upgrade --force-reinstall --no-cache-dir
# Switch to root to copy application code (to ensure permissions)
USER root
COPY --chown=appuser:appuser app.py /app/
# Install additional Python dependencies for your app as appuser
USER appuser
RUN pip install --user gradio huggingface_hub
# Set working directory for the application
WORKDIR /app
# Expose port for Gradio
EXPOSE 7860
# Run the application as appuser
CMD ["python", "app.py"]