gemma-2-9b GGUF

Llama.cpp version b3259 was used for hf to gguf conversion.

Original model: https://huggingface.co/google/gemma-2-9b

Available Precisions:

  • f16
  • q8_0

License

Gemma Terms of Use applies the same as the original model.

Downloads last month
22
GGUF
Model size
9.24B params
Architecture
gemma2
Hardware compatibility
Log In to view the estimation

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support