Issue Deploying Phi-4-mini-instruct on SageMaker (TGI): Container Health Check Fails

#30
by aamirfaaiz - opened

Hi there,

I’ve been trying to deploy 'microsoft/Phi-4-mini-instruct' on Amazon SageMaker using the Hugging Face LLM Inference container (TGI backend), but the endpoint consistently fails with a 'ping health check' error.

Here’s a summary of what I’m doing:

  • Using: get_huggingface_llm_image_uri(backend="huggingface", version="1.2.0")
  • Instance type: ml.g5.2xlarge

The endpoint consistently fails with:
The primary container for production variant AllTraffic did not pass the ping health check.

Sign up or log in to comment