Granite 4.0 Micro (GGUF)

This repository contains models that have been converted to the GGUF format with various quantizations from an IBM Granite base model.

Please reference the base model's full model card here: https://huggingface.co/ibm-granite/granite-4.0-micro

Downloads last month
5,798
GGUF
Model size
3B params
Architecture
granite
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ibm-granite/granite-4.0-micro-GGUF

Quantized
(20)
this model

Collection including ibm-granite/granite-4.0-micro-GGUF