Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,25 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: google/gemma-2-9b-it
|
3 |
+
inference: false
|
4 |
+
license: apache-2.0
|
5 |
+
model_name: Gemma-2-9B-Instruct-4Bit-GPTQ
|
6 |
+
pipeline_tag: text-generation
|
7 |
+
quantized_by: Granther
|
8 |
+
tags:
|
9 |
+
- gptq
|
10 |
+
---
|
11 |
+
|
12 |
+
# Gemma-2-9B-Instruct-4Bit-GPTQ
|
13 |
+
- Original Model: [gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it)
|
14 |
+
- Model Creator: [google](https://huggingface.co/google)
|
15 |
+
|
16 |
+
## Quantization
|
17 |
+
- This model was quantized with the Auto-GPTQ library
|
18 |
+
|
19 |
+
|
20 |
+
## Metrics
|
21 |
+
|
22 |
+
| Benchmark | Metric | Gemma 2 GPTQ | Gemma 2 9B it |
|
23 |
+
| ------------------| ---------- | ----------- | -------------- |
|
24 |
+
| [PIQA](piqa) | 0-shot | 80.52 | 80.79 |
|
25 |
+
| [MMLU](mmlu) | 5-shot | 52.0 | 50.00 |
|