wonhosong commited on
Commit
a8e7703
·
1 Parent(s): 4db6307

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -19
README.md CHANGED
@@ -1,3 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # LLaMa-30b-instruct model card
2
 
3
  **Model Developers**
@@ -41,7 +56,7 @@
41
  ## Evaluation Results
42
 
43
  **Overview**
44
- - We conducted a performance evaluation based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). We evaluated our model on four benchmark datasets, which include ARC-Challenge, HellaSwag, MMLU, and TruthfulQA. We used the lm-evaluation-harness repository, specifically commit b281b0921b636bc36ad05c0b0b0763bd6dd43463. We can reproduce the evaluation environments using the command below:
45
 
46
  - **Main Results**
47
 
@@ -64,21 +79,4 @@
64
  **Why Upstage LLM?**
65
  - [Upstage](https://en.upstage.ai)'s LLM research has yielded remarkable results. Our 30B model size outperforms all models worldwide with less than 65B, establishing itself as the leading performer. Recognizing the immense potential for private LLM adoption within companies, we invite you to effortlessly implement a private LLM and fine-tune it with your own data. For a seamless and tailored solution, please don't hesitate to reach out to us [(click here to mail)].
66
 
67
- [1]: https://en.upstage.ai
68
- [2]: https://github.com/facebookresearch/llama/tree/llama_v1
69
- [3]: https://huggingface.co/upstage/llama-30b-instruct
70
- [4]: https://huggingface.co/upstage/llama-30b-instruct-2048
71
- [5]: https://huggingface.co/upstage/llama-65b-instruct
72
- [6]: https://docs.google.com/forms/d/e/1FAIpQLSfqNECQnMkycAp2jP4Z9TFX0cGR4uf7b_fBxjY_OjhJILlKGA/viewform
73
- [7]: https://huggingface.co/upstage/llama-30b-instruct-2048/discussions
74
- [8]: https://huggingface.co/datasets/openbookqa
75
- [9]: https://huggingface.co/datasets/sciq
76
- [10]: https://huggingface.co/datasets/Open-Orca/OpenOrca
77
- [11]: https://huggingface.co/datasets/metaeval/ScienceQA_text_only
78
- [12]: https://huggingface.co/datasets/GAIR/lima
79
- [13]: https://github.com/microsoft/DeepSpeed
80
- [14]: https://huggingface.co/docs/transformers/main_classes/trainer
81
- [15]: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
82
- [(click here to mail)]: mailto:contact@upstage.ai
83
-
84
-
 
1
+ ---
2
+ datasets:
3
+ - sciq
4
+ - metaeval/ScienceQA_text_only
5
+ - openbookqa
6
+ - GAIR/lima
7
+ - Open-Orca/OpenOrca
8
+ language:
9
+ - en
10
+ tags:
11
+ - upstage
12
+ - llama
13
+ - instruct
14
+ - instruction
15
+ ---
16
  # LLaMa-30b-instruct model card
17
 
18
  **Model Developers**
 
56
  ## Evaluation Results
57
 
58
  **Overview**
59
+ - We conducted a performance evaluation based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). We evaluated our model on four benchmark datasets, which include ARC-Challenge, HellaSwag, MMLU, and TruthfulQA. We used the lm-evaluation-harness repository, specifically commit `b281b0921b636bc36ad05c0b0b0763bd6dd43463`. We can reproduce the evaluation environments using the command below:
60
 
61
  - **Main Results**
62
 
 
79
  **Why Upstage LLM?**
80
  - [Upstage](https://en.upstage.ai)'s LLM research has yielded remarkable results. Our 30B model size outperforms all models worldwide with less than 65B, establishing itself as the leading performer. Recognizing the immense potential for private LLM adoption within companies, we invite you to effortlessly implement a private LLM and fine-tune it with your own data. For a seamless and tailored solution, please don't hesitate to reach out to us [(click here to mail)].
81
 
82
+ [(click here to mail)]: mailto:contact@upstage.ai