jaspercatapang
commited on
Commit
•
7eb7a89
1
Parent(s):
39b533d
Update README.md
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ datasets:
|
|
14 |
Released August 11, 2023
|
15 |
|
16 |
## Model Description
|
17 |
-
GodziLLa 2 70B is an experimental combination of various proprietary LoRAs from Maya Philippines and [Guanaco LLaMA 2 1K dataset](https://huggingface.co/datasets/mlabonne/guanaco-llama2-1k), with LLaMA 2 70B. This model's primary purpose is to stress test the limitations of composite, instruction-following LLMs and observe its performance with respect to other LLMs available on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). This model debuted in the leaderboard at rank #
|
18 |
![Godzilla Happy GIF](https://i.pinimg.com/originals/81/3a/e0/813ae09a30f0bc44130cd2c834fe2eba.gif)
|
19 |
|
20 |
## Open LLM Leaderboard Metrics
|
@@ -33,7 +33,7 @@ According to the leaderboard description, here are the benchmarks used for the e
|
|
33 |
- [TruthfulQA](https://arxiv.org/abs/2109.07958) (0-shot) - a test to measure a model’s propensity to reproduce falsehoods commonly found online.
|
34 |
|
35 |
## Leaderboard Highlights (as of August 17, 2023)
|
36 |
-
- Godzilla 2 70B ranks
|
37 |
- Godzilla 2 70B ranks #3 in the ARC challenge.
|
38 |
- Godzilla 2 70B ranks #6 in the TruthfulQA benchmark.
|
39 |
- *Godzilla 2 70B beats GPT-3.5 (ChatGPT) in terms of average performance and the HellaSwag benchmark (87.53 > 85.5).
|
|
|
14 |
Released August 11, 2023
|
15 |
|
16 |
## Model Description
|
17 |
+
GodziLLa 2 70B is an experimental combination of various proprietary LoRAs from Maya Philippines and [Guanaco LLaMA 2 1K dataset](https://huggingface.co/datasets/mlabonne/guanaco-llama2-1k), with LLaMA 2 70B. This model's primary purpose is to stress test the limitations of composite, instruction-following LLMs and observe its performance with respect to other LLMs available on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). This model debuted in the leaderboard at rank #4 (August 17, 2023).
|
18 |
![Godzilla Happy GIF](https://i.pinimg.com/originals/81/3a/e0/813ae09a30f0bc44130cd2c834fe2eba.gif)
|
19 |
|
20 |
## Open LLM Leaderboard Metrics
|
|
|
33 |
- [TruthfulQA](https://arxiv.org/abs/2109.07958) (0-shot) - a test to measure a model’s propensity to reproduce falsehoods commonly found online.
|
34 |
|
35 |
## Leaderboard Highlights (as of August 17, 2023)
|
36 |
+
- Godzilla 2 70B ranks 4th, worldwide, in the Open LLM Leaderboard.
|
37 |
- Godzilla 2 70B ranks #3 in the ARC challenge.
|
38 |
- Godzilla 2 70B ranks #6 in the TruthfulQA benchmark.
|
39 |
- *Godzilla 2 70B beats GPT-3.5 (ChatGPT) in terms of average performance and the HellaSwag benchmark (87.53 > 85.5).
|