ChuckMcSneed commited on
Commit
abf42d5
1 Parent(s): 136d8ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -1
README.md CHANGED
@@ -18,4 +18,14 @@ This is a 32k version of Sao10K/WinterGoddess-1.4x-70B-L2, extended using method
18
  | D | 3 | 0 |
19
  | S | 2.75 | 1.5 |
20
  | P | 5.5 | 2.25 |
21
- | Total | 14.75 | 8.25 |
 
 
 
 
 
 
 
 
 
 
 
18
  | D | 3 | 0 |
19
  | S | 2.75 | 1.5 |
20
  | P | 5.5 | 2.25 |
21
+ | Total | 14.75 | 8.25 |
22
+
23
+ ### Open LLM leaderboard
24
+ [Leaderboard on Huggingface](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
25
+ |Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
26
+ |---------------------------------------|-------|-----|---------|-----|----------|----------|-----|
27
+ |Sao10K/WinterGoddess-1.4x-70B-L2 |73.23 |72.78|90.11 |71.12|65.76 |85 |54.59|
28
+ |ChuckMcSneed/WinterGoddess-1.4x-70b-32k|69.4 |71.16|89.12 |66.42|63.87 |82.56 |43.29|
29
+ |Difference |3.83 |1.62 |0.99 |4.7 |1.89 |2.44 |11.3 |
30
+
31
+ Here the losses seem far less brutal than on my bench. It seems that extending with LongLORA kills MMLU and GSM8K performance.