Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
# LLaMa-30b-instruct model card
|
2 |
|
3 |
**Model Developers**
|
@@ -41,7 +56,7 @@
|
|
41 |
## Evaluation Results
|
42 |
|
43 |
**Overview**
|
44 |
-
- We conducted a performance evaluation based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). We evaluated our model on four benchmark datasets, which include ARC-Challenge, HellaSwag, MMLU, and TruthfulQA. We used the lm-evaluation-harness repository, specifically commit b281b0921b636bc36ad05c0b0b0763bd6dd43463
|
45 |
|
46 |
- **Main Results**
|
47 |
|
@@ -64,21 +79,4 @@
|
|
64 |
**Why Upstage LLM?**
|
65 |
- [Upstage](https://en.upstage.ai)'s LLM research has yielded remarkable results. Our 30B model size outperforms all models worldwide with less than 65B, establishing itself as the leading performer. Recognizing the immense potential for private LLM adoption within companies, we invite you to effortlessly implement a private LLM and fine-tune it with your own data. For a seamless and tailored solution, please don't hesitate to reach out to us [(click here to mail)].
|
66 |
|
67 |
-
[
|
68 |
-
[2]: https://github.com/facebookresearch/llama/tree/llama_v1
|
69 |
-
[3]: https://huggingface.co/upstage/llama-30b-instruct
|
70 |
-
[4]: https://huggingface.co/upstage/llama-30b-instruct-2048
|
71 |
-
[5]: https://huggingface.co/upstage/llama-65b-instruct
|
72 |
-
[6]: https://docs.google.com/forms/d/e/1FAIpQLSfqNECQnMkycAp2jP4Z9TFX0cGR4uf7b_fBxjY_OjhJILlKGA/viewform
|
73 |
-
[7]: https://huggingface.co/upstage/llama-30b-instruct-2048/discussions
|
74 |
-
[8]: https://huggingface.co/datasets/openbookqa
|
75 |
-
[9]: https://huggingface.co/datasets/sciq
|
76 |
-
[10]: https://huggingface.co/datasets/Open-Orca/OpenOrca
|
77 |
-
[11]: https://huggingface.co/datasets/metaeval/ScienceQA_text_only
|
78 |
-
[12]: https://huggingface.co/datasets/GAIR/lima
|
79 |
-
[13]: https://github.com/microsoft/DeepSpeed
|
80 |
-
[14]: https://huggingface.co/docs/transformers/main_classes/trainer
|
81 |
-
[15]: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
|
82 |
-
[(click here to mail)]: mailto:contact@upstage.ai
|
83 |
-
|
84 |
-
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- sciq
|
4 |
+
- metaeval/ScienceQA_text_only
|
5 |
+
- openbookqa
|
6 |
+
- GAIR/lima
|
7 |
+
- Open-Orca/OpenOrca
|
8 |
+
language:
|
9 |
+
- en
|
10 |
+
tags:
|
11 |
+
- upstage
|
12 |
+
- llama
|
13 |
+
- instruct
|
14 |
+
- instruction
|
15 |
+
---
|
16 |
# LLaMa-30b-instruct model card
|
17 |
|
18 |
**Model Developers**
|
|
|
56 |
## Evaluation Results
|
57 |
|
58 |
**Overview**
|
59 |
+
- We conducted a performance evaluation based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). We evaluated our model on four benchmark datasets, which include ARC-Challenge, HellaSwag, MMLU, and TruthfulQA. We used the lm-evaluation-harness repository, specifically commit `b281b0921b636bc36ad05c0b0b0763bd6dd43463`. We can reproduce the evaluation environments using the command below:
|
60 |
|
61 |
- **Main Results**
|
62 |
|
|
|
79 |
**Why Upstage LLM?**
|
80 |
- [Upstage](https://en.upstage.ai)'s LLM research has yielded remarkable results. Our 30B model size outperforms all models worldwide with less than 65B, establishing itself as the leading performer. Recognizing the immense potential for private LLM adoption within companies, we invite you to effortlessly implement a private LLM and fine-tune it with your own data. For a seamless and tailored solution, please don't hesitate to reach out to us [(click here to mail)].
|
81 |
|
82 |
+
[(click here to mail)]: mailto:contact@upstage.ai
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|