RichardErkhov commited on
Commit
d141626
·
verified ·
1 Parent(s): 197342c

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +236 -0
README.md ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ calme-2.3-llama3.1-70b - GGUF
11
+ - Model creator: https://huggingface.co/MaziyarPanahi/
12
+ - Original model: https://huggingface.co/MaziyarPanahi/calme-2.3-llama3.1-70b/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [calme-2.3-llama3.1-70b.Q2_K.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q2_K.gguf) | Q2_K | 24.56GB |
18
+ | [calme-2.3-llama3.1-70b.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.IQ3_XS.gguf) | IQ3_XS | 27.29GB |
19
+ | [calme-2.3-llama3.1-70b.IQ3_S.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.IQ3_S.gguf) | IQ3_S | 28.79GB |
20
+ | [calme-2.3-llama3.1-70b.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q3_K_S.gguf) | Q3_K_S | 28.79GB |
21
+ | [calme-2.3-llama3.1-70b.IQ3_M.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.IQ3_M.gguf) | IQ3_M | 29.74GB |
22
+ | [calme-2.3-llama3.1-70b.Q3_K.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q3_K.gguf) | Q3_K | 31.91GB |
23
+ | [calme-2.3-llama3.1-70b.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q3_K_M.gguf) | Q3_K_M | 31.91GB |
24
+ | [calme-2.3-llama3.1-70b.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q3_K_L.gguf) | Q3_K_L | 34.59GB |
25
+ | [calme-2.3-llama3.1-70b.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.IQ4_XS.gguf) | IQ4_XS | 35.64GB |
26
+ | [calme-2.3-llama3.1-70b.Q4_0.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/blob/main/calme-2.3-llama3.1-70b.Q4_0.gguf) | Q4_0 | 37.22GB |
27
+ | [calme-2.3-llama3.1-70b.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | IQ4_NL | 37.58GB |
28
+ | [calme-2.3-llama3.1-70b.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q4_K_S | 37.58GB |
29
+ | [calme-2.3-llama3.1-70b.Q4_K.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q4_K | 39.6GB |
30
+ | [calme-2.3-llama3.1-70b.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q4_K_M | 39.6GB |
31
+ | [calme-2.3-llama3.1-70b.Q4_1.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q4_1 | 41.27GB |
32
+ | [calme-2.3-llama3.1-70b.Q5_0.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q5_0 | 45.32GB |
33
+ | [calme-2.3-llama3.1-70b.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q5_K_S | 45.32GB |
34
+ | [calme-2.3-llama3.1-70b.Q5_K.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q5_K | 46.52GB |
35
+ | [calme-2.3-llama3.1-70b.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q5_K_M | 46.52GB |
36
+ | [calme-2.3-llama3.1-70b.Q5_1.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q5_1 | 49.36GB |
37
+ | [calme-2.3-llama3.1-70b.Q6_K.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q6_K | 53.91GB |
38
+ | [calme-2.3-llama3.1-70b.Q8_0.gguf](https://huggingface.co/RichardErkhov/MaziyarPanahi_-_calme-2.3-llama3.1-70b-gguf/tree/main/) | Q8_0 | 69.83GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - en
47
+ library_name: transformers
48
+ tags:
49
+ - chat
50
+ - llama
51
+ - facebook
52
+ - llaam3
53
+ - finetune
54
+ - chatml
55
+ base_model: meta-llama/Meta-Llama-3.1-70B-Instruct
56
+ datasets:
57
+ - MaziyarPanahi/truthy-dpo-v0.1-axolotl
58
+ model_name: calme-2.3-llama3.1-70b
59
+ pipeline_tag: text-generation
60
+ inference: false
61
+ model_creator: MaziyarPanahi
62
+ quantized_by: MaziyarPanahi
63
+ model-index:
64
+ - name: calme-2.3-llama3.1-70b
65
+ results:
66
+ - task:
67
+ type: text-generation
68
+ name: Text Generation
69
+ dataset:
70
+ name: IFEval (0-Shot)
71
+ type: HuggingFaceH4/ifeval
72
+ args:
73
+ num_few_shot: 0
74
+ metrics:
75
+ - type: inst_level_strict_acc and prompt_level_strict_acc
76
+ value: 86.05
77
+ name: strict accuracy
78
+ source:
79
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
80
+ name: Open LLM Leaderboard
81
+ - task:
82
+ type: text-generation
83
+ name: Text Generation
84
+ dataset:
85
+ name: BBH (3-Shot)
86
+ type: BBH
87
+ args:
88
+ num_few_shot: 3
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 55.59
92
+ name: normalized accuracy
93
+ source:
94
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: MATH Lvl 5 (4-Shot)
101
+ type: hendrycks/competition_math
102
+ args:
103
+ num_few_shot: 4
104
+ metrics:
105
+ - type: exact_match
106
+ value: 21.45
107
+ name: exact match
108
+ source:
109
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
110
+ name: Open LLM Leaderboard
111
+ - task:
112
+ type: text-generation
113
+ name: Text Generation
114
+ dataset:
115
+ name: GPQA (0-shot)
116
+ type: Idavidrein/gpqa
117
+ args:
118
+ num_few_shot: 0
119
+ metrics:
120
+ - type: acc_norm
121
+ value: 12.53
122
+ name: acc_norm
123
+ source:
124
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
125
+ name: Open LLM Leaderboard
126
+ - task:
127
+ type: text-generation
128
+ name: Text Generation
129
+ dataset:
130
+ name: MuSR (0-shot)
131
+ type: TAUR-Lab/MuSR
132
+ args:
133
+ num_few_shot: 0
134
+ metrics:
135
+ - type: acc_norm
136
+ value: 17.74
137
+ name: acc_norm
138
+ source:
139
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
140
+ name: Open LLM Leaderboard
141
+ - task:
142
+ type: text-generation
143
+ name: Text Generation
144
+ dataset:
145
+ name: MMLU-PRO (5-shot)
146
+ type: TIGER-Lab/MMLU-Pro
147
+ config: main
148
+ split: test
149
+ args:
150
+ num_few_shot: 5
151
+ metrics:
152
+ - type: acc
153
+ value: 48.48
154
+ name: accuracy
155
+ source:
156
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=MaziyarPanahi/calme-2.3-llama3.1-70b
157
+ name: Open LLM Leaderboard
158
+ ---
159
+
160
+ <img src="./calme-2.webp" alt="Calme-2 Models" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
161
+
162
+ # MaziyarPanahi/calme-2.3-llama3.1-70b
163
+
164
+ This model is a fine-tuned version of the powerful `meta-llama/Meta-Llama-3.1-70B-Instruct`, pushing the boundaries of natural language understanding and generation even further. My goal was to create a versatile and robust model that excels across a wide range of benchmarks and real-world applications.
165
+
166
+
167
+ ## Use Cases
168
+
169
+ This model is suitable for a wide range of applications, including but not limited to:
170
+
171
+ - Advanced question-answering systems
172
+ - Intelligent chatbots and virtual assistants
173
+ - Content generation and summarization
174
+ - Code generation and analysis
175
+ - Complex problem-solving and decision support
176
+
177
+ # ⚡ Quantized GGUF
178
+
179
+ coming soon!
180
+
181
+ # 🏆 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
182
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_MaziyarPanahi__calme-2.3-llama3.1-70b)
183
+
184
+ | Metric |Value|
185
+ |-------------------|----:|
186
+ |Avg. |40.30|
187
+ |IFEval (0-Shot) |86.05|
188
+ |BBH (3-Shot) |55.59|
189
+ |MATH Lvl 5 (4-Shot)|21.45|
190
+ |GPQA (0-shot) |12.53|
191
+ |MuSR (0-shot) |17.74|
192
+ |MMLU-PRO (5-shot) |48.48|
193
+
194
+
195
+
196
+ This model uses `ChatML` prompt template:
197
+
198
+ ```
199
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
200
+
201
+ {system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
202
+
203
+ {prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
204
+
205
+ ```
206
+
207
+ # How to use
208
+
209
+
210
+ ```python
211
+
212
+ # Use a pipeline as a high-level helper
213
+
214
+ from transformers import pipeline
215
+
216
+ messages = [
217
+ {"role": "user", "content": "Who are you?"},
218
+ ]
219
+ pipe = pipeline("text-generation", model="MaziyarPanahi/calme-2.3-llama3.1-70b")
220
+ pipe(messages)
221
+
222
+
223
+ # Load model directly
224
+
225
+ from transformers import AutoTokenizer, AutoModelForCausalLM
226
+
227
+ tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-2.3-llama3.1-70b")
228
+ model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-2.3-llama3.1-70b")
229
+ ```
230
+
231
+
232
+ # Ethical Considerations
233
+
234
+ As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.
235
+
236
+