GGUF
conversational

This is basically a test to see if the conversion and inference in llama.cpp works fine It seems to work though i wont add more quant sizes for now

Since this is merely a quantization of the original model the license of the original model still applies!

Downloads last month
71
GGUF
Model size
752M params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for QuantStack/InternVL3_5-1B-gguf

Collection including QuantStack/InternVL3_5-1B-gguf