Smokeweaver commited on
Commit
c3c133a
1 Parent(s): 0924197

add model card

Browse files
Files changed (1) hide show
  1. README.md +235 -1
README.md CHANGED
@@ -1,3 +1,237 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ license: cc-by-nc-4.0
5
+ tags:
6
+ - merge
7
+ - lazymergekit
8
+ - dpo
9
+ - rlhf
10
+ - quantized
11
+ - 4-bit
12
+ - AWQ
13
+ - text-generation
14
+ - autotrain_compatible
15
+ - endpoints_compatible
16
+ - chatml
17
+ dataset:
18
+ - mlabonne/truthy-dpo-v0.1
19
+ - mlabonne/distilabel-intel-orca-dpo-pairs
20
+ - mlabonne/chatml-OpenHermes2.5-dpo-binarized-alpha
21
+ base_model:
22
+ - mlabonne/NeuralMonarch-7B
23
+ model-index:
24
+ - name: AlphaMonarch-7B
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 73.04
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 89.18
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 64.4
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 77.91
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 84.69
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 66.72
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/AlphaMonarch-7B
125
+ name: Open LLM Leaderboard
126
+ library_name: transformers
127
+ model_creator: mlabonne
128
+ model_name: AlphaMonarch-7B
129
+ model_type: mistral
130
+ pipeline_tag: text-generation
131
+ inference: false
132
+ prompt_template: '<|im_start|>system
133
+
134
+ {system_message}<|im_end|>
135
+
136
+ <|im_start|>user
137
+
138
+ {prompt}<|im_end|>
139
+
140
+ <|im_start|>assistant
141
+
142
+ '
143
+ quantized_by: Suparious
144
  ---
145
+ # mlabonne/AlphaMonarch-7B AWQ
146
+
147
+ - Model creator: [mlabonne](https://huggingface.co/mlabonne)
148
+ - Original model: [AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
149
+
150
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/TI7C8F2gk43gmI9U2L0uk.jpeg)
151
+
152
+ ## Model Summary
153
+
154
+ **tl;dr: AlphaMonarch-7B is a new DPO merge that retains all the reasoning abilities of the very best merges and significantly improves its conversational abilities. Kind of the best of both worlds in a 7B model. 🎉**
155
+
156
+ AlphaMonarch-7B is a DPO fine-tuned of [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B/) using the [argilla/OpenHermes2.5-dpo-binarized-alpha](https://huggingface.co/datasets/argilla/OpenHermes2.5-dpo-binarized-alpha) preference dataset.
157
+
158
+ It is based on a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
159
+ * [mlabonne/OmniTruthyBeagle-7B-v0](https://huggingface.co/mlabonne/OmniTruthyBeagle-7B-v0)
160
+ * [mlabonne/NeuBeagle-7B](https://huggingface.co/mlabonne/NeuBeagle-7B)
161
+ * [mlabonne/NeuralOmniBeagle-7B](https://huggingface.co/mlabonne/NeuralOmniBeagle-7B)
162
+
163
+ Special thanks to [Jon Durbin](https://huggingface.co/jondurbin), [Intel](https://huggingface.co/Intel), [Argilla](https://huggingface.co/argilla), and [Teknium](https://huggingface.co/teknium) for the preference datasets.
164
+
165
+ **Try the demo**: https://huggingface.co/spaces/mlabonne/AlphaMonarch-7B-GGUF-Chat
166
+
167
+ ## How to use
168
+
169
+ ### Install the necessary packages
170
+
171
+ ```bash
172
+ pip install --upgrade autoawq autoawq-kernels
173
+ ```
174
+
175
+ ### Example Python code
176
+
177
+ ```python
178
+ from awq import AutoAWQForCausalLM
179
+ from transformers import AutoTokenizer, TextStreamer
180
+
181
+ model_path = "solidrust/AlphaMonarch-7B-AWQ"
182
+ system_message = "You are Alpha, incarnated as a powerful AI."
183
+
184
+ # Load model
185
+ model = AutoAWQForCausalLM.from_quantized(model_path,
186
+ fuse_layers=True)
187
+ tokenizer = AutoTokenizer.from_pretrained(model_path,
188
+ trust_remote_code=True)
189
+ streamer = TextStreamer(tokenizer,
190
+ skip_prompt=True,
191
+ skip_special_tokens=True)
192
+
193
+ # Convert prompt to tokens
194
+ prompt_template = """\
195
+ <|im_start|>system
196
+ {system_message}<|im_end|>
197
+ <|im_start|>user
198
+ {prompt}<|im_end|>
199
+ <|im_start|>assistant"""
200
+
201
+ prompt = "You're standing on the surface of the Earth. "\
202
+ "You walk one mile south, one mile west and one mile north. "\
203
+ "You end up exactly where you started. Where are you?"
204
+
205
+ tokens = tokenizer(prompt_template.format(system_message=system_message,prompt=prompt),
206
+ return_tensors='pt').input_ids.cuda()
207
+
208
+ # Generate output
209
+ generation_output = model.generate(tokens,
210
+ streamer=streamer,
211
+ max_new_tokens=512)
212
+
213
+ ```
214
+
215
+ ### About AWQ
216
+
217
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
218
+
219
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
220
+
221
+ It is supported by:
222
+
223
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
224
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
225
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
226
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
227
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
228
+
229
+ ## Prompt template: ChatML
230
+
231
+ ```plaintext
232
+ <|im_start|>system
233
+ {system_message}<|im_end|>
234
+ <|im_start|>user
235
+ {prompt}<|im_end|>
236
+ <|im_start|>assistant
237
+ ```