Issue Deploying Phi-4-mini-instruct on SageMaker (TGI): Container Health Check Fails
#30
by
aamirfaaiz
- opened
Hi there,
I’ve been trying to deploy 'microsoft/Phi-4-mini-instruct' on Amazon SageMaker using the Hugging Face LLM Inference container (TGI backend), but the endpoint consistently fails with a 'ping health check' error.
Here’s a summary of what I’m doing:
- Using:
get_huggingface_llm_image_uri(backend="huggingface", version="1.2.0")
- Instance type:
ml.g5.2xlarge
The endpoint consistently fails with:
The primary container for production variant AllTraffic did not pass the ping health check.