 
				google/gemma-3-4b-it-qat-q4_0-gguf
			Image-Text-to-Text
			• 
		
				4B
			• 
	
				Updated
					
				
				• 
					
					3.61k
				
	
				• 
					
					204
				
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory
 
				 
				 
				 
				 
				 
				 
				 
				Note ^ GGUFs to be used in llama.cpp and Ollama. We strongly recommend using the IT models.
 
				 
				 
				 
				 
				 
				 
				Note ^ unquantized QAT-based checkpoints that allow quantizing while retaining similar quality to half precision