Commit
•
bcc175a
1
Parent(s):
7eb7a89
Update README.md
Browse files
README.md
CHANGED
@@ -35,7 +35,7 @@ According to the leaderboard description, here are the benchmarks used for the e
|
|
35 |
## Leaderboard Highlights (as of August 17, 2023)
|
36 |
- Godzilla 2 70B ranks 4th, worldwide, in the Open LLM Leaderboard.
|
37 |
- Godzilla 2 70B ranks #3 in the ARC challenge.
|
38 |
-
- Godzilla 2 70B ranks #
|
39 |
- *Godzilla 2 70B beats GPT-3.5 (ChatGPT) in terms of average performance and the HellaSwag benchmark (87.53 > 85.5).
|
40 |
- *Godzilla 2 70B outperforms GPT-3.5 (ChatGPT) and GPT-4 on the TruthfulQA benchmark (61.54 for G2-70B, 47 for GPT-3.5, 59 for GPT-4).
|
41 |
- *Godzilla 2 70B is on par with GPT-3.5 (ChatGPT) on the MMLU benchmark (<0.12%).
|
|
|
35 |
## Leaderboard Highlights (as of August 17, 2023)
|
36 |
- Godzilla 2 70B ranks 4th, worldwide, in the Open LLM Leaderboard.
|
37 |
- Godzilla 2 70B ranks #3 in the ARC challenge.
|
38 |
+
- Godzilla 2 70B ranks #5 in the TruthfulQA benchmark.
|
39 |
- *Godzilla 2 70B beats GPT-3.5 (ChatGPT) in terms of average performance and the HellaSwag benchmark (87.53 > 85.5).
|
40 |
- *Godzilla 2 70B outperforms GPT-3.5 (ChatGPT) and GPT-4 on the TruthfulQA benchmark (61.54 for G2-70B, 47 for GPT-3.5, 59 for GPT-4).
|
41 |
- *Godzilla 2 70B is on par with GPT-3.5 (ChatGPT) on the MMLU benchmark (<0.12%).
|