This is a MXFP4_MOE quantization of the model Qwen3-VL-235B-A22B-Thinking

Original model: https://huggingface.co/unsloth/Qwen3-VL-235B-A22B-Thinking-1M

This is the version from unloth that has expanded the context size from 256k to 1M.

Download the latest llama.cpp to use them.

Downloads last month
1,162
GGUF
Model size
235B params
Architecture
qwen3vlmoe
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for noctrex/Qwen3-VL-235B-A22B-Thinking-1M-MXFP4_MOE-GGUF

Quantized
(13)
this model

Collection including noctrex/Qwen3-VL-235B-A22B-Thinking-1M-MXFP4_MOE-GGUF