
google/gemma-3-4b-it-qat-q4_0-gguf
Image-Text-to-Text
•
Updated
•
16.9k
•
113
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory
Note ^ GGUFs to be used in llama.cpp and Ollama. We strongly recommend using the IT models.
Note ^ unquantized QAT-based checkpoints that allow quantizing while retaining similar quality to half precision