Update README.md
Browse files
README.md
CHANGED
@@ -1,13 +1,38 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
inference: false
|
|
|
3 |
---
|
4 |
# jeiku/Average_Normie_v2_l3_8B AWQ
|
5 |
|
6 |
-
** PROCESSING .... ETA 30mins **
|
7 |
-
|
8 |
- Model creator: [jeiku](https://huggingface.co/jeiku)
|
9 |
- Original model: [Average_Normie_v2_l3_8B](https://huggingface.co/jeiku/Average_Normie_v2_l3_8B)
|
10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
### About AWQ
|
12 |
|
13 |
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B
|
4 |
+
- vicgalle/Roleplay-Llama-3-8B
|
5 |
+
- cgato/L3-TheSpice-8b-v0.1.3
|
6 |
+
- ResplendentAI/Kei_Llama3_8B
|
7 |
+
library_name: transformers
|
8 |
+
tags:
|
9 |
+
- mergekit
|
10 |
+
- merge
|
11 |
+
- 4-bit
|
12 |
+
- AWQ
|
13 |
+
- text-generation
|
14 |
+
- autotrain_compatible
|
15 |
+
- endpoints_compatible
|
16 |
+
pipeline_tag: text-generation
|
17 |
inference: false
|
18 |
+
quantized_by: Suparious
|
19 |
---
|
20 |
# jeiku/Average_Normie_v2_l3_8B AWQ
|
21 |
|
|
|
|
|
22 |
- Model creator: [jeiku](https://huggingface.co/jeiku)
|
23 |
- Original model: [Average_Normie_v2_l3_8B](https://huggingface.co/jeiku/Average_Normie_v2_l3_8B)
|
24 |
|
25 |
+
## Model Summary
|
26 |
+
|
27 |
+
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
28 |
+
|
29 |
+
This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [ResplendentAI/Kei_Llama3_8B](https://huggingface.co/ResplendentAI/Kei_Llama3_8B) as a base.
|
30 |
+
|
31 |
+
The following models were included in the merge:
|
32 |
+
* [ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-v0.7-L3-8B)
|
33 |
+
* [vicgalle/Roleplay-Llama-3-8B](https://huggingface.co/vicgalle/Roleplay-Llama-3-8B)
|
34 |
+
* [cgato/L3-TheSpice-8b-v0.1.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.1.3)
|
35 |
+
|
36 |
### About AWQ
|
37 |
|
38 |
AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
|