RichardErkhov commited on
Commit
fee0a4f
1 Parent(s): 63a5d93

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +274 -0
README.md ADDED
@@ -0,0 +1,274 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ OpenDolphinHermes_Llama2_7B - GGUF
11
+ - Model creator: https://huggingface.co/sethuiyer/
12
+ - Original model: https://huggingface.co/sethuiyer/OpenDolphinHermes_Llama2_7B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [OpenDolphinHermes_Llama2_7B.Q2_K.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q2_K.gguf) | Q2_K | 2.36GB |
18
+ | [OpenDolphinHermes_Llama2_7B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.IQ3_XS.gguf) | IQ3_XS | 2.6GB |
19
+ | [OpenDolphinHermes_Llama2_7B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.IQ3_S.gguf) | IQ3_S | 2.75GB |
20
+ | [OpenDolphinHermes_Llama2_7B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q3_K_S.gguf) | Q3_K_S | 2.75GB |
21
+ | [OpenDolphinHermes_Llama2_7B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.IQ3_M.gguf) | IQ3_M | 2.9GB |
22
+ | [OpenDolphinHermes_Llama2_7B.Q3_K.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q3_K.gguf) | Q3_K | 3.07GB |
23
+ | [OpenDolphinHermes_Llama2_7B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q3_K_M.gguf) | Q3_K_M | 3.07GB |
24
+ | [OpenDolphinHermes_Llama2_7B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q3_K_L.gguf) | Q3_K_L | 3.35GB |
25
+ | [OpenDolphinHermes_Llama2_7B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.IQ4_XS.gguf) | IQ4_XS | 3.4GB |
26
+ | [OpenDolphinHermes_Llama2_7B.Q4_0.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q4_0.gguf) | Q4_0 | 3.56GB |
27
+ | [OpenDolphinHermes_Llama2_7B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.IQ4_NL.gguf) | IQ4_NL | 3.58GB |
28
+ | [OpenDolphinHermes_Llama2_7B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q4_K_S.gguf) | Q4_K_S | 3.59GB |
29
+ | [OpenDolphinHermes_Llama2_7B.Q4_K.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q4_K.gguf) | Q4_K | 3.8GB |
30
+ | [OpenDolphinHermes_Llama2_7B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q4_K_M.gguf) | Q4_K_M | 3.8GB |
31
+ | [OpenDolphinHermes_Llama2_7B.Q4_1.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q4_1.gguf) | Q4_1 | 3.95GB |
32
+ | [OpenDolphinHermes_Llama2_7B.Q5_0.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q5_0.gguf) | Q5_0 | 4.33GB |
33
+ | [OpenDolphinHermes_Llama2_7B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q5_K_S.gguf) | Q5_K_S | 4.33GB |
34
+ | [OpenDolphinHermes_Llama2_7B.Q5_K.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q5_K.gguf) | Q5_K | 4.45GB |
35
+ | [OpenDolphinHermes_Llama2_7B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q5_K_M.gguf) | Q5_K_M | 4.45GB |
36
+ | [OpenDolphinHermes_Llama2_7B.Q5_1.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q5_1.gguf) | Q5_1 | 4.72GB |
37
+ | [OpenDolphinHermes_Llama2_7B.Q6_K.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q6_K.gguf) | Q6_K | 5.15GB |
38
+ | [OpenDolphinHermes_Llama2_7B.Q8_0.gguf](https://huggingface.co/RichardErkhov/sethuiyer_-_OpenDolphinHermes_Llama2_7B-gguf/blob/main/OpenDolphinHermes_Llama2_7B.Q8_0.gguf) | Q8_0 | 6.67GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - en
47
+ license: llama2
48
+ library_name: transformers
49
+ tags:
50
+ - merge
51
+ - mergekit
52
+ - lazymergekit
53
+ datasets:
54
+ - teknium/openhermes
55
+ - cognitivecomputations/dolphin
56
+ base_model:
57
+ - cognitivecomputations/dolphin-llama2-7b
58
+ - Tensoic/Llama-2-openhermes
59
+ pipeline_tag: text-generation
60
+ model-index:
61
+ - name: OpenDolphinHermes_Llama2_7B
62
+ results:
63
+ - task:
64
+ type: text-generation
65
+ name: Text Generation
66
+ dataset:
67
+ name: AI2 Reasoning Challenge (25-Shot)
68
+ type: ai2_arc
69
+ config: ARC-Challenge
70
+ split: test
71
+ args:
72
+ num_few_shot: 25
73
+ metrics:
74
+ - type: acc_norm
75
+ value: 55.03
76
+ name: normalized accuracy
77
+ source:
78
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
79
+ name: Open LLM Leaderboard
80
+ - task:
81
+ type: text-generation
82
+ name: Text Generation
83
+ dataset:
84
+ name: HellaSwag (10-Shot)
85
+ type: hellaswag
86
+ split: validation
87
+ args:
88
+ num_few_shot: 10
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 78.74
92
+ name: normalized accuracy
93
+ source:
94
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: MMLU (5-Shot)
101
+ type: cais/mmlu
102
+ config: all
103
+ split: test
104
+ args:
105
+ num_few_shot: 5
106
+ metrics:
107
+ - type: acc
108
+ value: 52.25
109
+ name: accuracy
110
+ source:
111
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
112
+ name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: TruthfulQA (0-shot)
118
+ type: truthful_qa
119
+ config: multiple_choice
120
+ split: validation
121
+ args:
122
+ num_few_shot: 0
123
+ metrics:
124
+ - type: mc2
125
+ value: 46.1
126
+ source:
127
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
128
+ name: Open LLM Leaderboard
129
+ - task:
130
+ type: text-generation
131
+ name: Text Generation
132
+ dataset:
133
+ name: Winogrande (5-shot)
134
+ type: winogrande
135
+ config: winogrande_xl
136
+ split: validation
137
+ args:
138
+ num_few_shot: 5
139
+ metrics:
140
+ - type: acc
141
+ value: 73.16
142
+ name: accuracy
143
+ source:
144
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
145
+ name: Open LLM Leaderboard
146
+ - task:
147
+ type: text-generation
148
+ name: Text Generation
149
+ dataset:
150
+ name: GSM8k (5-shot)
151
+ type: gsm8k
152
+ config: main
153
+ split: test
154
+ args:
155
+ num_few_shot: 5
156
+ metrics:
157
+ - type: acc
158
+ value: 20.17
159
+ name: accuracy
160
+ source:
161
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/OpenDolphinHermes_Llama2_7B
162
+ name: Open LLM Leaderboard
163
+ ---
164
+
165
+ # OpenDolphinHermes_Llama2_7B
166
+
167
+
168
+ <p align="center">
169
+ <img src="https://huggingface.co/sethuiyer/OpenDolphinHermes_Llama2_7B/resolve/main/dolphin_hermes.webp" height="256px" alt="SynthIQ">
170
+ </p>
171
+
172
+ mergekit SLERP of these two models
173
+ * [cognitivecomputations/dolphin-llama2-7b](https://huggingface.co/cognitivecomputations/dolphin-llama2-7b)
174
+ * [Tensoic/Llama-2-openhermes](https://huggingface.co/Tensoic/Llama-2-openhermes)
175
+
176
+ ## 🧩 Configuration
177
+
178
+ ```yaml
179
+ slices:
180
+ - sources:
181
+ - model: cognitivecomputations/dolphin-llama2-7b
182
+ layer_range: [0, 32]
183
+ - model: Tensoic/Llama-2-openhermes
184
+ layer_range: [0, 32]
185
+ merge_method: slerp
186
+ base_model: Tensoic/Llama-2-openhermes
187
+ parameters:
188
+ t:
189
+ - filter: self_attn
190
+ value: [0, 0.5, 0.3, 0.7, 1]
191
+ - filter: mlp
192
+ value: [1, 0.5, 0.7, 0.3, 0]
193
+ - value: 0.5
194
+ dtype: bfloat16
195
+ ```
196
+
197
+ # Prompt Template (ChatML)
198
+ ```text
199
+ <|im_start|>system
200
+ You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe.
201
+ Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content.
202
+ Please ensure that your responses are socially unbiased and positive in nature.
203
+
204
+ If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct.
205
+ If you don't know the answer to a question, please don't share false information.
206
+ <|im_end|>
207
+ <|im_start|>user
208
+ { .Prompt}
209
+ <|im_end|>
210
+ <|im_start|>assistant
211
+ ```
212
+
213
+ # OpenLLM Leaderboard
214
+
215
+ | T | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
216
+ |---|--------------------------------------------|---------|------|-----------|-------|------------|------------|-------|
217
+ | 0 | meta-llama/llama-2-13b-hf | 55.69 | 59.39 | 82.13 | 55.77 | 37.38 | 76.64 | 22.82 |
218
+ | 1 | sethuiyer/OpenDolphinHermes_Llama2_7B | 54.24 | 55.03| 78.74 | 52.25 | 46.1 | 73.16 | 20.17 |
219
+ | 2 | togethercomputer/Llama-2-7B-32K-Instruct | 50.02 | 51.11| 78.51 | 46.11 | 44.86 | 73.88 | 5.69 |
220
+ | 3 | togethercomputer/LLaMa-2-7B-32K | 47.07 | 47.53| 76.14 | 43.33 | 39.23 | 71.9 | 4.32 |
221
+
222
+ ## Why?
223
+
224
+ I wanted a LLaMa2-7B model which is as good as base LLaMa2-13B model.
225
+
226
+ ## 💻 Usage
227
+
228
+ ```python
229
+ !pip install -qU transformers accelerate
230
+
231
+ from transformers import AutoTokenizer
232
+ import transformers
233
+ import torch
234
+
235
+ model = "sethuiyer/OpenDolphinHermes_Llama2_7B"
236
+ messages = [{"role": "user", "content": "What is a large language model?"}]
237
+
238
+ tokenizer = AutoTokenizer.from_pretrained(model)
239
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
240
+ pipeline = transformers.pipeline(
241
+ "text-generation",
242
+ model=model,
243
+ torch_dtype=torch.float16,
244
+ device_map="auto",
245
+ )
246
+
247
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
248
+ print(outputs[0]["generated_text"])
249
+ ```
250
+
251
+ Output:
252
+ ```text
253
+ A large language model is a type of artificial intelligence system that has been trained on a massive amount of data, often millions or even billions of words, to learn the patterns and relationships between words and phrases.
254
+ These models can then be used to generate new text, understand and translate languages, and perform various natural language processing tasks.
255
+ They have become increasingly popular in recent years due to advances in machine learning technology and their ability to achieve high levels of accuracy and performance on natural language processing tasks.
256
+ Examples of large language models include GPT-2, BERT, and T5.
257
+ ```
258
+ ## Thanks
259
+ Thanks to Google Colab for the compute.
260
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
261
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_sethuiyer__OpenDolphinHermes_Llama2_7B)
262
+
263
+ | Metric |Value|
264
+ |---------------------------------|----:|
265
+ |Avg. |54.24|
266
+ |AI2 Reasoning Challenge (25-Shot)|55.03|
267
+ |HellaSwag (10-Shot) |78.74|
268
+ |MMLU (5-Shot) |52.25|
269
+ |TruthfulQA (0-shot) |46.10|
270
+ |Winogrande (5-shot) |73.16|
271
+ |GSM8k (5-shot) |20.17|
272
+
273
+
274
+