Mixtral-8x22B-Instruct-NVFP4

NVFP4 quantized version of mistralai/Mixtral-8x22B-Instruct-v0.1 optimized for NVIDIA DGX/Hopper+ architectures (H100, H200, GB10, etc.).

Quantization Details

  • Format: NVFP4 (4-bit floating point)
  • Quantized using: NVIDIA TensorRT Model Optimizer 0.35.0
  • Hardware: 2× NVIDIA H200 SXM (188GB VRAM each)
  • Original precision: BF16/FP16
  • Compatible with: vLLM 0.10+, NVIDIA NGC containers

Usage with vLLM

vllm serve tbhot3ww/Mixtral-8x22B-Instruct-NVFP4 \
  --quantization modelopt_fp4 \
  --max-model-len 8192 \
  --gpu-memory-utilization 0.95

Requirements

  • NVIDIA GPU with Hopper architecture or newer (compute capability ≥ 9.0)
  • CUDA 12.0+
  • vLLM 0.10 or newer with ModelOpt support

Original Model

For architecture details, training data, intended use, and limitations, see the original model card.

Quantization Notes

This model uses NVIDIA's NVFP4 format which provides 4-bit quantization with minimal quality degradation. Best performance is achieved on NVIDIA Hopper+ GPUs with native FP4 tensor core support.

Downloads last month
12
Safetensors
Model size
71B params
Tensor type
BF16
·
F8_E4M3
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tbhot3ww/Mixtral-8x22B-Instruct-NVFP4

Quantized
(6)
this model