Update README.md
Browse files
README.md
CHANGED
@@ -86,53 +86,6 @@ Currently, we are working on more suitable benchmarks for Spanish, French, Germa
|
|
86 |
<details>
|
87 |
<summary>Evaluation results</summary>
|
88 |
|
89 |
-
### All languages
|
90 |
-
|
91 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
92 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
93 |
-
| Mistral-7B-v0.1 | 0.5277 | 0.6825 | 0.7687 | 0.6287 | 0.6519 |
|
94 |
-
| leo-mistral-hessianai-7b | 0.4614 | 0.6423 | 0.6524 | 0.5440 | 0.5750 |
|
95 |
-
| Occiglot-7B-EU5 | 0.5083 | 0.7191 | 0.6758 | 0.5432 | 0.6116 |
|
96 |
-
|
97 |
-
### English
|
98 |
-
|
99 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
100 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
101 |
-
| Mistral-7B-v0.1 | 0.6143 | 0.8344 | 0.8444 | 0.6351 | 0.7321 |
|
102 |
-
| leo-mistral-hessianai-7b | 0.5213 | 0.7779 | 0.7356 | 0.5508 | 0.6464 |
|
103 |
-
| Occiglot-7B-EU5 | 0.5307 | 0.7900 | 0.7267 | 0.5467 | 0.6485 |
|
104 |
-
|
105 |
-
### German
|
106 |
-
|
107 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
108 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
109 |
-
| Mistral-7B-v0.1 | 0.4765 | 0.6101 | 0.7411 | 0.5274 | 0.5888 |
|
110 |
-
| leo-mistral-hessianai-7b | 0.4739 | 0.6818 | 0.6900 | 0.4887 | 0.5836 |
|
111 |
-
| Occiglot-7B-EU5 | 0.4944 | 0.6667 | 0.6467 | 0.4833 | 0.5728 |
|
112 |
-
|
113 |
-
### Spanish
|
114 |
-
|
115 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
116 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
117 |
-
| Mistral-7B-v0.1 | 0.5256 | 0.6728 | 0.7478 | 0.5432 | 0.6224 |
|
118 |
-
| leo-mistral-hessianai-7b | 0.4436 | 0.5970 | 0.6178 | 0.4359 | 0.5236 |
|
119 |
-
| Occiglot-7B-EU5 | 0.5085 | 0.7255 | 0.6778 | 0.4997 | 0.6029 |
|
120 |
-
|
121 |
-
### French
|
122 |
-
|
123 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
124 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
125 |
-
| Mistral-7B-v0.1 | 0.5244 | 0.6651 | 0.7744 | 0.5413 | 0.6263 |
|
126 |
-
| leo-mistral-hessianai-7b | 0.4354 | 0.5967 | 0.6222 | 0.4326 | 0.5217 |
|
127 |
-
| Occiglot-7B-EU5 | 0.5064 | 0.7125 | 0.6756 | 0.4959 | 0.5976 |
|
128 |
-
|
129 |
-
### Italian
|
130 |
-
|
131 |
-
| **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
|
132 |
-
|--------------------------|-------------------|---------------|--------------|----------|---------|
|
133 |
-
| Mistral-7B-v0.1 | 0.4979 | 0.6303 | 0.7356 | 0.5372 | 0.6002 |
|
134 |
-
| leo-mistral-hessianai-7b | 0.4328 | 0.5580 | 0.5967 | 0.4311 | 0.5047 |
|
135 |
-
| Occiglot-7B-EU5 | 0.5013 | 0.7008 | 0.6522 | 0.4949 | 0.5873 |
|
136 |
|
137 |
</details>
|
138 |
|
|
|
86 |
<details>
|
87 |
<summary>Evaluation results</summary>
|
88 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
89 |
|
90 |
</details>
|
91 |
|