Post
3341
Introducing AWQ and GPTQ quantized versions of SmolVLM from Hugging Face!
These models only had their text models quantized, and had a 50% model size reduction (4GB~2GB) while keeping model degradation under 1% on the DocVQA benchmark.
#huggingface #smolvlm #smollm
ronantakizawa/SmolVLM-Instruct-awq
ronantakizawa/SmolVLM-Instruct-gptq
These models only had their text models quantized, and had a 50% model size reduction (4GB~2GB) while keeping model degradation under 1% on the DocVQA benchmark.
#huggingface #smolvlm #smollm
ronantakizawa/SmolVLM-Instruct-awq
ronantakizawa/SmolVLM-Instruct-gptq