Update README.md
Browse files
README.md
CHANGED
@@ -102,6 +102,12 @@ Compared to Aleph Alpha Luminous Models
|
|
102 |
### BBH:
|
103 |
![BBH](https://vago-solutions.de/wp-content/uploads/2023/11/bbh.png "SauerkrautLM-7b-HerO BBH")
|
104 |
*performed with newest Language Model Evaluation Harness
|
|
|
|
|
|
|
|
|
|
|
|
|
105 |
### MT-Bench (German):
|
106 |
![MT-Bench German Diagram](https://vago-solutions.de/wp-content/uploads/2023/11/MT-Bench-German.png "SauerkrautLM-7b-HerO MT-Bench German Diagram")
|
107 |
```
|
@@ -162,11 +168,6 @@ SauerkrautLM-3b-v1 2.581250
|
|
162 |
open_llama_3b_v2 1.456250
|
163 |
Llama-2-7b 1.181250
|
164 |
```
|
165 |
-
### MMLU:
|
166 |
-
![MMLU](https://vago-solutions.de/wp-content/uploads/2023/11/MMLU-Benchmark.png "SauerkrautLM-7b-HerO MMLU")
|
167 |
-
### TruthfulQA:
|
168 |
-
![TruthfulQA](https://vago-solutions.de/wp-content/uploads/2023/11/Truthfulqa-Benchmark.png "SauerkrautLM-7b-HerO TruthfulQA")
|
169 |
-
|
170 |
### MT-Bench (English):
|
171 |
![MT-Bench English Diagram](https://vago-solutions.de/wp-content/uploads/2023/11/MT-Bench-Englisch.png "SauerkrautLM-7b-HerO MT-Bench English Diagram")
|
172 |
```
|
@@ -194,6 +195,8 @@ SauerkrautLM-7b-HerO <--- 7.409375
|
|
194 |
Mistral-7B-OpenOrca 6.915625
|
195 |
neural-chat-7b-v3-1 6.812500
|
196 |
```
|
|
|
|
|
197 |
### Additional German Benchmark results:
|
198 |
![GermanBenchmarks](https://vago-solutions.de/wp-content/uploads/2023/11/German-benchmarks.png "SauerkrautLM-7b-HerO German Benchmarks")
|
199 |
*performed with newest Language Model Evaluation Harness
|
|
|
102 |
### BBH:
|
103 |
![BBH](https://vago-solutions.de/wp-content/uploads/2023/11/bbh.png "SauerkrautLM-7b-HerO BBH")
|
104 |
*performed with newest Language Model Evaluation Harness
|
105 |
+
|
106 |
+
### MMLU:
|
107 |
+
![MMLU](https://vago-solutions.de/wp-content/uploads/2023/11/MMLU-Benchmark.png "SauerkrautLM-7b-HerO MMLU")
|
108 |
+
### TruthfulQA:
|
109 |
+
![TruthfulQA](https://vago-solutions.de/wp-content/uploads/2023/11/Truthfulqa-Benchmark.png "SauerkrautLM-7b-HerO TruthfulQA")
|
110 |
+
|
111 |
### MT-Bench (German):
|
112 |
![MT-Bench German Diagram](https://vago-solutions.de/wp-content/uploads/2023/11/MT-Bench-German.png "SauerkrautLM-7b-HerO MT-Bench German Diagram")
|
113 |
```
|
|
|
168 |
open_llama_3b_v2 1.456250
|
169 |
Llama-2-7b 1.181250
|
170 |
```
|
|
|
|
|
|
|
|
|
|
|
171 |
### MT-Bench (English):
|
172 |
![MT-Bench English Diagram](https://vago-solutions.de/wp-content/uploads/2023/11/MT-Bench-Englisch.png "SauerkrautLM-7b-HerO MT-Bench English Diagram")
|
173 |
```
|
|
|
195 |
Mistral-7B-OpenOrca 6.915625
|
196 |
neural-chat-7b-v3-1 6.812500
|
197 |
```
|
198 |
+
|
199 |
+
|
200 |
### Additional German Benchmark results:
|
201 |
![GermanBenchmarks](https://vago-solutions.de/wp-content/uploads/2023/11/German-benchmarks.png "SauerkrautLM-7b-HerO German Benchmarks")
|
202 |
*performed with newest Language Model Evaluation Harness
|