Precompiled TensorRT Engines by Qlip
Precompiled Qlip engines for accelerated diffusion model inference. Base model weights are not included.
For supported models, benchmarks, and usage instructions see ComfyUI-Qlip.
Usage
ComfyUI (automatic download)
In the Qlip Engines Loader node, set the hf_repo input. Engines are downloaded once and cached locally.
Manual download
pip install huggingface-hub
huggingface-cli download TheStageAI/<repo-name> \
--local-dir ./engines \
--include "models/H100/<variant>/*"
Installation
# Qlip core
pip install 'qlip.core[nvidia]' \
--extra-index-url https://thestage.jfrog.io/artifactory/api/pypi/pypi-thestage-ai-production/simple
# elastic_models (LoRA runtime support)
pip install 'thestage-elastic-models[nvidia]' \
--extra-index-url https://thestage.jfrog.io/artifactory/api/pypi/pypi-thestage-ai-production/simple
Requirements
- NVIDIA GPU with CUDA 12.x
- TensorRT 10.13.3.9 (must match compilation version)
- Engines are GPU-architecture specific — recompile after changing hardware
License
Proprietary. Powered by TheStage AI.
- Downloads last month
- 7
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support