Spaces:
Sleeping
Sleeping
# Use NVIDIA CUDA base image with Python support (ONLY for GPU Spaces) | |
FROM nvidia/cuda:12.1.1-runtime-ubuntu22.04 | |
# Set working directory | |
WORKDIR /app | |
# Install Python and system dependencies | |
RUN apt-get update && apt-get install -y \ | |
python3.10 \ | |
python3-pip \ | |
python3.10-venv \ | |
python3.10-dev \ | |
build-essential \ | |
cmake \ | |
git \ | |
curl \ | |
wget \ | |
&& rm -rf /var/lib/apt/lists/* | |
# Make python3.10 the default | |
RUN update-alternatives --install /usr/bin/python python /usr/bin/python3.10 1 | |
RUN update-alternatives --install /usr/bin/pip pip /usr/bin/pip3 1 | |
# Upgrade pip | |
RUN pip install --upgrade pip | |
# Install prebuilt llama-cpp-python CUDA wheel | |
RUN CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install --no-cache-dir llama-cpp-python==0.2.90 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu121 | |
# Copy requirements.txt and install remaining dependencies | |
COPY requirements.txt . | |
RUN pip install --no-cache-dir -r requirements.txt | |
# Create a non-root user | |
RUN useradd -m -u 1000 user | |
# Create cache directory with proper permissions | |
RUN mkdir -p /app/.cache/huggingface | |
RUN chown -R user:user /app | |
# Copy app code | |
COPY . . | |
RUN chown -R user:user /app | |
# Create models directory | |
RUN mkdir -p models && chown -R user:user models | |
# Switch to non-root user | |
USER user | |
# Environment variables | |
ENV GRADIO_SERVER_NAME="0.0.0.0" | |
ENV GRADIO_SERVER_PORT=7860 | |
ENV CUDA_VISIBLE_DEVICES=0 | |
ENV HF_HOME=/app/.cache/huggingface | |
ENV TRANSFORMERS_CACHE=/app/.cache/huggingface | |
# Expose Gradio port | |
EXPOSE 7860 | |
# Start app | |
CMD ["python", "app.py"] |