ArtusDev/allura-org_Gemma-3-Glitter-4B-GGUF
โจ Overview :3
This repository contains GGUF format model files converted from allura-org/Gemma-3-Glitter-4B.
The conversion was performed by ArtusDev using llama.cpp, specifically utilizing the imatrix quantization option for potentially improved performance.
๐ Original Model Details ^_^
For more information about the model please refer to the original model card. It's pretty neat (empty)!
๐ฌ Instruct Format >.<
This model uses a custom Gemma 2/3 instruct format. It has been trained to recognize an optional system role.
<start_of_turn>system
{optional system prompt here}<end_of_turn>
<start_of_turn>user
{User messages. You can also place the system prompt here.}<end_of_turn>
<start_of_turn>model
{Model's response}<end_of_turn>
Note: Always ensure the format strictly adheres to the required tokens and structure for optimal model performance. Don't mess it up :3!
GGUF Quantizations (imatrix) by ArtusDev >:3
- Downloads last month
- 18
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Model tree for ArtusDev/allura-org_Gemma-3-Glitter-4B-GGUF
Base model
google/gemma-3-4b-pt
Finetuned
google/gemma-3-4b-it
Finetuned
allura-org/Gemma-3-Glitter-4B