RichardErkhov commited on
Commit
e5acfe9
β€’
1 Parent(s): 479985b

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +295 -0
README.md ADDED
@@ -0,0 +1,295 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Daredevil-8B - GGUF
11
+ - Model creator: https://huggingface.co/mlabonne/
12
+ - Original model: https://huggingface.co/mlabonne/Daredevil-8B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [Daredevil-8B.Q2_K.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q2_K.gguf) | Q2_K | 2.96GB |
18
+ | [Daredevil-8B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
19
+ | [Daredevil-8B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.IQ3_S.gguf) | IQ3_S | 3.43GB |
20
+ | [Daredevil-8B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
21
+ | [Daredevil-8B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.IQ3_M.gguf) | IQ3_M | 3.52GB |
22
+ | [Daredevil-8B.Q3_K.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q3_K.gguf) | Q3_K | 3.74GB |
23
+ | [Daredevil-8B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
24
+ | [Daredevil-8B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
25
+ | [Daredevil-8B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
26
+ | [Daredevil-8B.Q4_0.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q4_0.gguf) | Q4_0 | 4.34GB |
27
+ | [Daredevil-8B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
28
+ | [Daredevil-8B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
29
+ | [Daredevil-8B.Q4_K.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q4_K.gguf) | Q4_K | 4.58GB |
30
+ | [Daredevil-8B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
31
+ | [Daredevil-8B.Q4_1.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q4_1.gguf) | Q4_1 | 4.78GB |
32
+ | [Daredevil-8B.Q5_0.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q5_0.gguf) | Q5_0 | 5.21GB |
33
+ | [Daredevil-8B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
34
+ | [Daredevil-8B.Q5_K.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q5_K.gguf) | Q5_K | 5.34GB |
35
+ | [Daredevil-8B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
36
+ | [Daredevil-8B.Q5_1.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q5_1.gguf) | Q5_1 | 5.65GB |
37
+ | [Daredevil-8B.Q6_K.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q6_K.gguf) | Q6_K | 6.14GB |
38
+ | [Daredevil-8B.Q8_0.gguf](https://huggingface.co/RichardErkhov/mlabonne_-_Daredevil-8B-gguf/blob/main/Daredevil-8B.Q8_0.gguf) | Q8_0 | 7.95GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: other
46
+ tags:
47
+ - merge
48
+ - mergekit
49
+ - lazymergekit
50
+ base_model:
51
+ - nbeerbower/llama-3-stella-8B
52
+ - Hastagaras/llama-3-8b-okay
53
+ - nbeerbower/llama-3-gutenberg-8B
54
+ - openchat/openchat-3.6-8b-20240522
55
+ - Kukedlc/NeuralLLaMa-3-8b-DT-v0.1
56
+ - cstr/llama3-8b-spaetzle-v20
57
+ - mlabonne/ChimeraLlama-3-8B-v3
58
+ - flammenai/Mahou-1.1-llama3-8B
59
+ - KingNish/KingNish-Llama3-8b
60
+ model-index:
61
+ - name: Daredevil-8B
62
+ results:
63
+ - task:
64
+ type: text-generation
65
+ name: Text Generation
66
+ dataset:
67
+ name: AI2 Reasoning Challenge (25-Shot)
68
+ type: ai2_arc
69
+ config: ARC-Challenge
70
+ split: test
71
+ args:
72
+ num_few_shot: 25
73
+ metrics:
74
+ - type: acc_norm
75
+ value: 68.86
76
+ name: normalized accuracy
77
+ source:
78
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
79
+ name: Open LLM Leaderboard
80
+ - task:
81
+ type: text-generation
82
+ name: Text Generation
83
+ dataset:
84
+ name: HellaSwag (10-Shot)
85
+ type: hellaswag
86
+ split: validation
87
+ args:
88
+ num_few_shot: 10
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 84.5
92
+ name: normalized accuracy
93
+ source:
94
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: MMLU (5-Shot)
101
+ type: cais/mmlu
102
+ config: all
103
+ split: test
104
+ args:
105
+ num_few_shot: 5
106
+ metrics:
107
+ - type: acc
108
+ value: 69.24
109
+ name: accuracy
110
+ source:
111
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
112
+ name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: TruthfulQA (0-shot)
118
+ type: truthful_qa
119
+ config: multiple_choice
120
+ split: validation
121
+ args:
122
+ num_few_shot: 0
123
+ metrics:
124
+ - type: mc2
125
+ value: 59.89
126
+ source:
127
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
128
+ name: Open LLM Leaderboard
129
+ - task:
130
+ type: text-generation
131
+ name: Text Generation
132
+ dataset:
133
+ name: Winogrande (5-shot)
134
+ type: winogrande
135
+ config: winogrande_xl
136
+ split: validation
137
+ args:
138
+ num_few_shot: 5
139
+ metrics:
140
+ - type: acc
141
+ value: 78.45
142
+ name: accuracy
143
+ source:
144
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
145
+ name: Open LLM Leaderboard
146
+ - task:
147
+ type: text-generation
148
+ name: Text Generation
149
+ dataset:
150
+ name: GSM8k (5-shot)
151
+ type: gsm8k
152
+ config: main
153
+ split: test
154
+ args:
155
+ num_few_shot: 5
156
+ metrics:
157
+ - type: acc
158
+ value: 73.54
159
+ name: accuracy
160
+ source:
161
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/Daredevil-8B
162
+ name: Open LLM Leaderboard
163
+ ---
164
+
165
+ # Daredevil-8B
166
+
167
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/gFEhcIDSKa3AWpkNfH91q.jpeg)
168
+
169
+ Daredevil-8B is a mega-merge designed to maximize MMLU. On 27 May 24, it is the Llama 3 8B model with the **highest MMLU score**.
170
+ From my experience, a high MMLU score is all you need with Llama 3 models.
171
+
172
+ It is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
173
+ * [nbeerbower/llama-3-stella-8B](https://huggingface.co/nbeerbower/llama-3-stella-8B)
174
+ * [Hastagaras/llama-3-8b-okay](https://huggingface.co/Hastagaras/llama-3-8b-okay)
175
+ * [nbeerbower/llama-3-gutenberg-8B](https://huggingface.co/nbeerbower/llama-3-gutenberg-8B)
176
+ * [openchat/openchat-3.6-8b-20240522](https://huggingface.co/openchat/openchat-3.6-8b-20240522)
177
+ * [Kukedlc/NeuralLLaMa-3-8b-DT-v0.1](https://huggingface.co/Kukedlc/NeuralLLaMa-3-8b-DT-v0.1)
178
+ * [cstr/llama3-8b-spaetzle-v20](https://huggingface.co/cstr/llama3-8b-spaetzle-v20)
179
+ * [mlabonne/ChimeraLlama-3-8B-v3](https://huggingface.co/mlabonne/ChimeraLlama-3-8B-v3)
180
+ * [flammenai/Mahou-1.1-llama3-8B](https://huggingface.co/flammenai/Mahou-1.1-llama3-8B)
181
+ * [KingNish/KingNish-Llama3-8b](https://huggingface.co/KingNish/KingNish-Llama3-8b)
182
+
183
+ Thanks to nbeerbower, Hastagaras, openchat, Kukedlc, cstr, flammenai, and KingNish for their merges. Special thanks to Charles Goddard and Arcee.ai for MergeKit.
184
+
185
+ ## πŸ”Ž Applications
186
+
187
+ You can use it as an improved version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct).
188
+
189
+ This is a censored model. For an uncensored version, see [mlabonne/Daredevil-8B-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated).
190
+
191
+ Tested on LM Studio using the "Llama 3" preset.
192
+
193
+ ## ⚑ Quantization
194
+
195
+ * **GGUF**: https://huggingface.co/mlabonne/Daredevil-8B-GGUF
196
+
197
+ ## πŸ† Evaluation
198
+
199
+ ### Open LLM Leaderboard
200
+
201
+ Daredevil-8B is the best-performing 8B model on the Open LLM Leaderboard in terms of MMLU score (27 May 24).
202
+
203
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/xFKhGdSaIxL9_tcJPhM5w.png)
204
+
205
+ ### Nous
206
+
207
+ Daredevil-8B is the best-performing 8B model on Nous' benchmark suite (evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval), 27 May 24). See the entire leaderboard [here](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard).
208
+
209
+ | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
210
+ |---|---:|---:|---:|---:|---:|
211
+ | [**mlabonne/Daredevil-8B**](https://huggingface.co/mlabonne/Daredevil-8B) [πŸ“„](https://gist.github.com/mlabonne/080f9c5f153ea57a7ab7d932cf896f21) | **55.87** | **44.13** | **73.52** | **59.05** | **46.77** |
212
+ | [mlabonne/Daredevil-8B-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated) [πŸ“„](https://gist.github.com/mlabonne/32cdd8460804662c856bcb2a20acd49e) | 55.06 | 43.29 | 73.33 | 57.47 | 46.17 |
213
+ | [mlabonne/Llama-3-8B-Instruct-abliterated-dpomix](https://huggingface.co/mlabonne/Llama-3-8B-Instruct-abliterated-dpomix) [πŸ“„](https://gist.github.com/mlabonne/d711548df70e2c04771cc68ab33fe2b9) | 52.26 | 41.6 | 69.95 | 54.22 | 43.26 |
214
+ | [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) [πŸ“„](https://gist.github.com/mlabonne/8329284d86035e6019edb11eb0933628) | 51.34 | 41.22 | 69.86 | 51.65 | 42.64 |
215
+ | [failspy/Meta-Llama-3-8B-Instruct-abliterated-v3](https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3) [πŸ“„](https://gist.github.com/mlabonne/f46cce0262443365e4cce2b6fa7507fc) | 51.21 | 40.23 | 69.5 | 52.44 | 42.69 |
216
+ | [mlabonne/OrpoLlama-3-8B](https://huggingface.co/mlabonne/OrpoLlama-3-8B) [πŸ“„](https://gist.github.com/mlabonne/22896a1ae164859931cc8f4858c97f6f) | 48.63 | 34.17 | 70.59 | 52.39 | 37.36 |
217
+ | [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) [πŸ“„](https://gist.github.com/mlabonne/616b6245137a9cfc4ea80e4c6e55d847) | 45.42 | 31.1 | 69.95 | 43.91 | 36.7 |
218
+
219
+ ## 🌳 Model family tree
220
+
221
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/ekwRGgnjzEOyprT8sEBFt.png)
222
+
223
+ ## 🧩 Configuration
224
+
225
+ ```yaml
226
+ models:
227
+ - model: NousResearch/Meta-Llama-3-8B
228
+ # No parameters necessary for base model
229
+ - model: nbeerbower/llama-3-stella-8B
230
+ parameters:
231
+ density: 0.6
232
+ weight: 0.16
233
+ - model: Hastagaras/llama-3-8b-okay
234
+ parameters:
235
+ density: 0.56
236
+ weight: 0.1
237
+ - model: nbeerbower/llama-3-gutenberg-8B
238
+ parameters:
239
+ density: 0.6
240
+ weight: 0.18
241
+ - model: openchat/openchat-3.6-8b-20240522
242
+ parameters:
243
+ density: 0.56
244
+ weight: 0.12
245
+ - model: Kukedlc/NeuralLLaMa-3-8b-DT-v0.1
246
+ parameters:
247
+ density: 0.58
248
+ weight: 0.18
249
+ - model: cstr/llama3-8b-spaetzle-v20
250
+ parameters:
251
+ density: 0.56
252
+ weight: 0.08
253
+ - model: mlabonne/ChimeraLlama-3-8B-v3
254
+ parameters:
255
+ density: 0.56
256
+ weight: 0.08
257
+ - model: flammenai/Mahou-1.1-llama3-8B
258
+ parameters:
259
+ density: 0.55
260
+ weight: 0.05
261
+ - model: KingNish/KingNish-Llama3-8b
262
+ parameters:
263
+ density: 0.55
264
+ weight: 0.05
265
+ merge_method: dare_ties
266
+ base_model: NousResearch/Meta-Llama-3-8B
267
+ dtype: bfloat16
268
+ ```
269
+
270
+ ## πŸ’» Usage
271
+
272
+ ```python
273
+ !pip install -qU transformers accelerate
274
+
275
+ from transformers import AutoTokenizer
276
+ import transformers
277
+ import torch
278
+
279
+ model = "mlabonne/Daredevil-8B"
280
+ messages = [{"role": "user", "content": "What is a large language model?"}]
281
+
282
+ tokenizer = AutoTokenizer.from_pretrained(model)
283
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
284
+ pipeline = transformers.pipeline(
285
+ "text-generation",
286
+ model=model,
287
+ torch_dtype=torch.bfloat16,
288
+ device_map="auto",
289
+ )
290
+
291
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
292
+ print(outputs[0]["generated_text"])
293
+ ```
294
+
295
+