Text Generation
Transformers
Safetensors
English
mistral
axolotl
Generated from Trainer
Mistral
instruct
finetune
chatml
gpt4
synthetic data
science
physics
chemistry
biology
math
conversational
Eval Results
Inference Endpoints
text-generation-inference
Weyaxi commited on
Commit
7eecd98
1 Parent(s): d8ee461

update model card, nice :)

Browse files
Files changed (1) hide show
  1. README.md +31 -18
README.md CHANGED
@@ -64,7 +64,8 @@ model-index:
64
  value: 64.68
65
  name: normalized accuracy
66
  source:
67
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
68
  name: Open LLM Leaderboard
69
  - task:
70
  type: text-generation
@@ -80,7 +81,8 @@ model-index:
80
  value: 83.75
81
  name: normalized accuracy
82
  source:
83
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
84
  name: Open LLM Leaderboard
85
  - task:
86
  type: text-generation
@@ -97,7 +99,8 @@ model-index:
97
  value: 62.31
98
  name: accuracy
99
  source:
100
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
101
  name: Open LLM Leaderboard
102
  - task:
103
  type: text-generation
@@ -113,7 +116,8 @@ model-index:
113
  - type: mc2
114
  value: 55.15
115
  source:
116
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
117
  name: Open LLM Leaderboard
118
  - task:
119
  type: text-generation
@@ -130,7 +134,8 @@ model-index:
130
  value: 76.24
131
  name: accuracy
132
  source:
133
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
134
  name: Open LLM Leaderboard
135
  - task:
136
  type: text-generation
@@ -147,8 +152,11 @@ model-index:
147
  value: 57.62
148
  name: accuracy
149
  source:
150
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
 
151
  name: Open LLM Leaderboard
 
 
152
  ---
153
 
154
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/U0zyXVGj-O8a7KP3BvPue.png)
@@ -311,17 +319,6 @@ Quantizationed versions of this model is available.
311
 
312
  - https://huggingface.co/bartowski/Einstein-v4-7B-exl2
313
 
314
- You can switch up branches in the repo to use the one you want
315
-
316
- | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
317
- | ----- | ---- | ------- | ------ | ------ | ------ | ------------ |
318
- | [8_0](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/8_0) | 8.0 | 8.0 | 8.4 GB | 9.8 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. |
319
- | [6_5](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/6_5) | 6.5 | 8.0 | 7.2 GB | 8.6 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. |
320
- | [5_0](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/5_0) | 5.0 | 6.0 | 6.0 GB | 7.4 GB | 9.4 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. |
321
- | [4_25](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/4_25) | 4.25 | 6.0 | 5.3 GB | 6.7 GB | 8.7 GB | GPTQ equivalent bits per weight, slightly higher quality. |
322
- | [3_5](https://huggingface.co/bartowski/Einstein-v4-7B-exl2/tree/3_5) | 3.5 | 6.0 | 4.7 GB | 6.1 GB | 8.1 GB | Lower quality, only use if you have to. |
323
-
324
-
325
  # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
326
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
327
 
@@ -335,6 +332,22 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
335
  |Winogrande (5-shot) |76.24|
336
  |GSM8k (5-shot) |57.62|
337
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
338
  # 🤖 Additional information about training
339
 
340
  This model is full fine-tuned for 1.5 epoch.
@@ -360,4 +373,4 @@ Thanks to all open source AI community.
360
 
361
  If you would like to support me:
362
 
363
- [☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
 
64
  value: 64.68
65
  name: normalized accuracy
66
  source:
67
+ url: >-
68
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
69
  name: Open LLM Leaderboard
70
  - task:
71
  type: text-generation
 
81
  value: 83.75
82
  name: normalized accuracy
83
  source:
84
+ url: >-
85
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
86
  name: Open LLM Leaderboard
87
  - task:
88
  type: text-generation
 
99
  value: 62.31
100
  name: accuracy
101
  source:
102
+ url: >-
103
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
104
  name: Open LLM Leaderboard
105
  - task:
106
  type: text-generation
 
116
  - type: mc2
117
  value: 55.15
118
  source:
119
+ url: >-
120
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
121
  name: Open LLM Leaderboard
122
  - task:
123
  type: text-generation
 
134
  value: 76.24
135
  name: accuracy
136
  source:
137
+ url: >-
138
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
139
  name: Open LLM Leaderboard
140
  - task:
141
  type: text-generation
 
152
  value: 57.62
153
  name: accuracy
154
  source:
155
+ url: >-
156
+ https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v4-7B
157
  name: Open LLM Leaderboard
158
+ language:
159
+ - en
160
  ---
161
 
162
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/U0zyXVGj-O8a7KP3BvPue.png)
 
319
 
320
  - https://huggingface.co/bartowski/Einstein-v4-7B-exl2
321
 
 
 
 
 
 
 
 
 
 
 
 
322
  # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
323
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-7B)
324
 
 
332
  |Winogrande (5-shot) |76.24|
333
  |GSM8k (5-shot) |57.62|
334
 
335
+ # 📚 Some resources, discussions and reviews aboout this model
336
+
337
+ #### 🐦 Announcement tweet:
338
+
339
+ https://twitter.com/Weyaxi/status/1765851433448944125
340
+
341
+ #### 🔍 Reddit post in r/LocalLLaMA:
342
+
343
+ - https://www.reddit.com/r/LocalLLaMA/comments/1b9gmvl/meet_einsteinv47b_mistralbased_sft_model_using/
344
+
345
+ #### ▶️ Youtube Videos
346
+
347
+ - https://www.youtube.com/watch?v=-3YWgHJIORE&t=18s
348
+
349
+ - https://www.youtube.com/watch?v=Xo2ySU8gja0
350
+
351
  # 🤖 Additional information about training
352
 
353
  This model is full fine-tuned for 1.5 epoch.
 
373
 
374
  If you would like to support me:
375
 
376
+ [☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)