Gemma 3
Collection
Collection of quants for Google's Gemma 3
โข
3 items
โข
Updated
This repository contains GGUF quantized versions of Google's Gemma 3 27B pretrained model, optimized for efficient deployment across various hardware configurations.
Model | Size (GB) | Compression Ratio | Size Reduction |
---|---|---|---|
Q8_0 | 26.7 GB | 53% | 47% |
Q6_K | 20.6 GB | 41% | 59% |
Q5_K | 17.9 GB | 36% | 64% |
Q4_K | 15.4 GB | 31% | 69% |
Q3_K | 12.5 GB | 25% | 75% |
Q2_K | 9.8 GB | 19% | 81% |
These models can be used with llama.cpp and its various interfaces. Example:
# Running with llama-gemma3-cli.exe (adjust paths as needed)
./llama-gemma3-cli --model Google.Gemma-3-27b-pt.q5_k.gguf --ctx-size 4096 --temp 0.7 --prompt "Write a short story about a robot who discovers it has feelings."
This model is released under the same Gemma license as the original model.
This quantized set is derived from Google's Gemma 3 27B pretrained model.
The pre-trained Gemma 3 27B model achieves the following benchmark results:
@article{gemma_2025,
title={Gemma 3},
url={https://goo.gle/Gemma3Report},
publisher={Kaggle},
author={Gemma Team},
year={2025}
}
@misc{gemma3_quantization_2025,
title={Quantized Versions of Google's Gemma 3 27B Model},
author={Lex-au},
year={2025},
month={March},
note={Quantized models (Q8_0, Q6_K, Q5_K, Q4_K, Q3_K, Q2_K) derived from Google's Gemma 3 27B},
url={https://huggingface.co/lex-au}
}
Base model
google/gemma-3-27b-pt