# Base image: Hugging Face TGI FROM ghcr.io/huggingface/text-generation-inference:3.0.2 # Set working directory WORKDIR /app # Minimal storage directories RUN mkdir -p /data && chmod 777 /data RUN mkdir -p /.cache && chmod 777 /.cache RUN mkdir -p /.triton && chmod 777 /.triton # Expose the API port EXPOSE 8080 # Optional: Set Hugging Face token for private model access ARG HF_TOKEN ENV HF_TOKEN=${HF_TOKEN} # Model: Replace with any small or quantized TGI-compatible model # Example: TheBloke/Mistral-7B-Instruct-v0.1-GPTQ or TinyLlama/TinyLlama-1.1B-Chat CMD ["--model-id", "TinyLlama/TinyLlama-1.1B-Chat-v1.0", "--port", "8080"]