Suparious commited on
Commit
9e023aa
1 Parent(s): ab74d22

Add model card

Browse files
Files changed (1) hide show
  1. README.md +265 -1
README.md CHANGED
@@ -1,3 +1,267 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ - Mistral
9
+ - instruct
10
+ - finetune
11
+ - chatml
12
+ - gpt4
13
+ - synthetic data
14
+ - science
15
+ - physics
16
+ - chemistry
17
+ - biology
18
+ - math
19
+ - quantized
20
+ - 4-bit
21
+ - AWQ
22
+ - autotrain_compatible
23
+ - endpoints_compatible
24
+ - text-generation-inference
25
+ base_model: alpindale/Mistral-7B-v0.2-hf
26
+ datasets:
27
+ - allenai/ai2_arc
28
+ - camel-ai/physics
29
+ - camel-ai/chemistry
30
+ - camel-ai/biology
31
+ - camel-ai/math
32
+ - metaeval/reclor
33
+ - openbookqa
34
+ - mandyyyyii/scibench
35
+ - derek-thomas/ScienceQA
36
+ - TIGER-Lab/ScienceEval
37
+ - jondurbin/airoboros-3.2
38
+ - LDJnr/Capybara
39
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
40
+ - STEM-AI-mtl/Electrical-engineering
41
+ - knowrohit07/saraswati-stem
42
+ - sablo/oasst2_curated
43
+ - lmsys/lmsys-chat-1m
44
+ - TIGER-Lab/MathInstruct
45
+ - bigbio/med_qa
46
+ - meta-math/MetaMathQA-40K
47
+ - openbookqa
48
+ - piqa
49
+ - metaeval/reclor
50
+ - derek-thomas/ScienceQA
51
+ - scibench
52
+ - sciq
53
+ - Open-Orca/SlimOrca
54
+ - migtissera/Synthia-v1.3
55
+ - TIGER-Lab/ScienceEval
56
+ - allenai/WildChat
57
+ - microsoft/orca-math-word-problems-200k
58
+ - openchat/openchat_sharegpt4_dataset
59
+ - teknium/GPTeacher-General-Instruct
60
+ - m-a-p/CodeFeedback-Filtered-Instruction
61
+ - totally-not-an-llm/EverythingLM-data-V3
62
+ - HuggingFaceH4/no_robots
63
+ - OpenAssistant/oasst_top1_2023-08-25
64
+ - WizardLM/WizardLM_evol_instruct_70k
65
+ model-index:
66
+ - name: Einstein-v6-7B
67
+ results:
68
+ - task:
69
+ type: text-generation
70
+ name: Text Generation
71
+ dataset:
72
+ name: AI2 Reasoning Challenge (25-Shot)
73
+ type: ai2_arc
74
+ config: ARC-Challenge
75
+ split: test
76
+ args:
77
+ num_few_shot: 25
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 63.57
81
+ name: normalized accuracy
82
+ source:
83
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: HellaSwag (10-Shot)
90
+ type: hellaswag
91
+ split: validation
92
+ args:
93
+ num_few_shot: 10
94
+ metrics:
95
+ - type: acc_norm
96
+ value: 82.76
97
+ name: normalized accuracy
98
+ source:
99
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
100
+ name: Open LLM Leaderboard
101
+ - task:
102
+ type: text-generation
103
+ name: Text Generation
104
+ dataset:
105
+ name: MMLU (5-Shot)
106
+ type: cais/mmlu
107
+ config: all
108
+ split: test
109
+ args:
110
+ num_few_shot: 5
111
+ metrics:
112
+ - type: acc
113
+ value: 62.23
114
+ name: accuracy
115
+ source:
116
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
117
+ name: Open LLM Leaderboard
118
+ - task:
119
+ type: text-generation
120
+ name: Text Generation
121
+ dataset:
122
+ name: TruthfulQA (0-shot)
123
+ type: truthful_qa
124
+ config: multiple_choice
125
+ split: validation
126
+ args:
127
+ num_few_shot: 0
128
+ metrics:
129
+ - type: mc2
130
+ value: 52.02
131
+ source:
132
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
133
+ name: Open LLM Leaderboard
134
+ - task:
135
+ type: text-generation
136
+ name: Text Generation
137
+ dataset:
138
+ name: Winogrande (5-shot)
139
+ type: winogrande
140
+ config: winogrande_xl
141
+ split: validation
142
+ args:
143
+ num_few_shot: 5
144
+ metrics:
145
+ - type: acc
146
+ value: 78.61
147
+ name: accuracy
148
+ source:
149
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
150
+ name: Open LLM Leaderboard
151
+ - task:
152
+ type: text-generation
153
+ name: Text Generation
154
+ dataset:
155
+ name: GSM8k (5-shot)
156
+ type: gsm8k
157
+ config: main
158
+ split: test
159
+ args:
160
+ num_few_shot: 5
161
+ metrics:
162
+ - type: acc
163
+ value: 63.53
164
+ name: accuracy
165
+ source:
166
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6-7B
167
+ name: Open LLM Leaderboard
168
+ quantized_by: Suparious
169
+ pipeline_tag: text-generation
170
+ model_creator: Weyaxi
171
+ model_name: Einstein-v6-7B
172
+ inference: false
173
+ prompt_template: |
174
+ <|im_start|>system
175
+ {system_message}<|im_end|>
176
+ <|im_start|>user
177
+ {prompt}<|im_end|>
178
+ <|im_start|>assistant
179
  ---
180
+ # Weyaxi/Einstein-v6-7B AWQ
181
+
182
+ **UPLOAD IN PROGRESS**
183
+
184
+ - Model creator: [Weyaxi](https://huggingface.co/Weyaxi)
185
+ - Original model: [Einstein-v6-7B](https://huggingface.co/Weyaxi/Einstein-v6-7B)
186
+
187
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/CxDk4KKhQqL-Pg0AMn1gb.png)
188
+
189
+ ## Model Summary
190
+
191
+ This model is a full fine-tuned version of [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf) on diverse datasets.
192
+
193
+ This model is finetuned using `8xRTX3090` + `1xRTXA6000` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
194
+
195
+ This model's training was sponsored by [sablo.ai](https://sablo.ai).
196
+
197
+ ## How to use
198
+
199
+ ### Install the necessary packages
200
+
201
+ ```bash
202
+ pip install --upgrade autoawq autoawq-kernels
203
+ ```
204
+
205
+ ### Example Python code
206
+
207
+ ```python
208
+ from awq import AutoAWQForCausalLM
209
+ from transformers import AutoTokenizer, TextStreamer
210
+
211
+ model_path = "solidrust/Einstein-v5-v0.2-7B-AWQ"
212
+ system_message = "You are Alpert Einstein, incarnated a powerful AI."
213
+
214
+ # Load model
215
+ model = AutoAWQForCausalLM.from_quantized(model_path,
216
+ fuse_layers=True)
217
+ tokenizer = AutoTokenizer.from_pretrained(model_path,
218
+ trust_remote_code=True)
219
+ streamer = TextStreamer(tokenizer,
220
+ skip_prompt=True,
221
+ skip_special_tokens=True)
222
+
223
+ # Convert prompt to tokens
224
+ prompt_template = """\
225
+ <|im_start|>system
226
+ {system_message}<|im_end|>
227
+ <|im_start|>user
228
+ {prompt}<|im_end|>
229
+ <|im_start|>assistant"""
230
+
231
+ prompt = "You're standing on the surface of the Earth. "\
232
+ "You walk one mile south, one mile west and one mile north. "\
233
+ "You end up exactly where you started. Where are you?"
234
+
235
+ tokens = tokenizer(prompt_template.format(system_message=system_message,prompt=prompt),
236
+ return_tensors='pt').input_ids.cuda()
237
+
238
+ # Generate output
239
+ generation_output = model.generate(tokens,
240
+ streamer=streamer,
241
+ max_new_tokens=512)
242
+
243
+ ```
244
+
245
+ ### About AWQ
246
+
247
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
248
+
249
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
250
+
251
+ It is supported by:
252
+
253
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
254
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
255
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
256
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
257
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
258
+
259
+ ## Prompt template: ChatML
260
+
261
+ ```plaintext
262
+ <|im_start|>system
263
+ {system_message}<|im_end|>
264
+ <|im_start|>user
265
+ {prompt}<|im_end|>
266
+ <|im_start|>assistant
267
+ ```