mbrack commited on
Commit
084e588
1 Parent(s): 24a8ca9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -47
README.md CHANGED
@@ -86,53 +86,6 @@ Currently, we are working on more suitable benchmarks for Spanish, French, Germa
86
  <details>
87
  <summary>Evaluation results</summary>
88
 
89
- ### All languages
90
-
91
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
92
- |--------------------------|-------------------|---------------|--------------|----------|---------|
93
- | Mistral-7B-v0.1 | 0.5277 | 0.6825 | 0.7687 | 0.6287 | 0.6519 |
94
- | leo-mistral-hessianai-7b | 0.4614 | 0.6423 | 0.6524 | 0.5440 | 0.5750 |
95
- | Occiglot-7B-EU5 | 0.5083 | 0.7191 | 0.6758 | 0.5432 | 0.6116 |
96
-
97
- ### English
98
-
99
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
100
- |--------------------------|-------------------|---------------|--------------|----------|---------|
101
- | Mistral-7B-v0.1 | 0.6143 | 0.8344 | 0.8444 | 0.6351 | 0.7321 |
102
- | leo-mistral-hessianai-7b | 0.5213 | 0.7779 | 0.7356 | 0.5508 | 0.6464 |
103
- | Occiglot-7B-EU5 | 0.5307 | 0.7900 | 0.7267 | 0.5467 | 0.6485 |
104
-
105
- ### German
106
-
107
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
108
- |--------------------------|-------------------|---------------|--------------|----------|---------|
109
- | Mistral-7B-v0.1 | 0.4765 | 0.6101 | 0.7411 | 0.5274 | 0.5888 |
110
- | leo-mistral-hessianai-7b | 0.4739 | 0.6818 | 0.6900 | 0.4887 | 0.5836 |
111
- | Occiglot-7B-EU5 | 0.4944 | 0.6667 | 0.6467 | 0.4833 | 0.5728 |
112
-
113
- ### Spanish
114
-
115
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
116
- |--------------------------|-------------------|---------------|--------------|----------|---------|
117
- | Mistral-7B-v0.1 | 0.5256 | 0.6728 | 0.7478 | 0.5432 | 0.6224 |
118
- | leo-mistral-hessianai-7b | 0.4436 | 0.5970 | 0.6178 | 0.4359 | 0.5236 |
119
- | Occiglot-7B-EU5 | 0.5085 | 0.7255 | 0.6778 | 0.4997 | 0.6029 |
120
-
121
- ### French
122
-
123
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
124
- |--------------------------|-------------------|---------------|--------------|----------|---------|
125
- | Mistral-7B-v0.1 | 0.5244 | 0.6651 | 0.7744 | 0.5413 | 0.6263 |
126
- | leo-mistral-hessianai-7b | 0.4354 | 0.5967 | 0.6222 | 0.4326 | 0.5217 |
127
- | Occiglot-7B-EU5 | 0.5064 | 0.7125 | 0.6756 | 0.4959 | 0.5976 |
128
-
129
- ### Italian
130
-
131
- | **model_name** | **arc_challenge** | **hellaswag** | **belebele** | **mmlu** | **avg** |
132
- |--------------------------|-------------------|---------------|--------------|----------|---------|
133
- | Mistral-7B-v0.1 | 0.4979 | 0.6303 | 0.7356 | 0.5372 | 0.6002 |
134
- | leo-mistral-hessianai-7b | 0.4328 | 0.5580 | 0.5967 | 0.4311 | 0.5047 |
135
- | Occiglot-7B-EU5 | 0.5013 | 0.7008 | 0.6522 | 0.4949 | 0.5873 |
136
 
137
  </details>
138
 
 
86
  <details>
87
  <summary>Evaluation results</summary>
88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
 
90
  </details>
91