wonhosong commited on
Commit
788a7f3
1 Parent(s): be44a37

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +81 -2
README.md CHANGED
@@ -8,6 +8,85 @@ tags:
8
  - instruction
9
  pipeline_tag: text-generation
10
  ---
11
- Please refer to the model card of [upstage/llama-30b-instruct](https://huggingface.co/upstage/llama-30b-instruct) as this one is almost the same.
12
 
13
- **License** This model is under a Non-commercial Bespoke License and governed by the Meta license. You should only use this repository if you have been granted access to the model by filling out [this form](https://docs.google.com/forms/d/e/1FAIpQLSfqNECQnMkycAp2jP4Z9TFX0cGR4uf7b_fBxjY_OjhJILlKGA/viewform), but have either lost your copy of the weights or encountered issues converting them to the Transformers format.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  - instruction
9
  pipeline_tag: text-generation
10
  ---
11
+ # LLaMa-30b-instruct-2048 model card
12
 
13
+ ## Model Details
14
+
15
+ ### Model Developers
16
+ - [Upstage](https://en.upstage.ai)
17
+
18
+ ### Backbone Model
19
+ - [LLaMA](https://github.com/facebookresearch/llama/tree/llama_v1)
20
+
21
+ ### Variations
22
+ - It has different model parameter sizes and sequence lengths: [30B/1024](https://huggingface.co/upstage/llama-30b-instruct), [30B/2048](https://huggingface.co/upstage/llama-30b-instruct-2048), [65B/1024](https://huggingface.co/upstage/llama-65b-instruct).
23
+
24
+ ### Input
25
+ - Models solely process textual input.
26
+
27
+ ### Output
28
+ - Models solely generate textual output.
29
+
30
+ ### License
31
+ - This model is under a **Non-commercial** Bespoke License and governed by the Meta license. You should only use this repository if you have been granted access to the model by filling out [this form](https://docs.google.com/forms/d/e/1FAIpQLSfqNECQnMkycAp2jP4Z9TFX0cGR4uf7b_fBxjY_OjhJILlKGA/viewform), but have either lost your copy of the weights or encountered issues converting them to the Transformers format.
32
+
33
+ ### Where to send comments
34
+ - Instructions on how to provide feedback or comments on a model can be found by opening an issue in the [Hugging Face community's model repository](https://huggingface.co/upstage/llama-30b-instruct-2048/discussions).
35
+
36
+ ## Dataset Details
37
+
38
+ ### Used Datasets
39
+ - [openbookqa](https://huggingface.co/datasets/openbookqa)
40
+ - [sciq](https://huggingface.co/datasets/sciq)
41
+ - [Open-Orca/OpenOrca](https://huggingface.co/datasets/Open-Orca/OpenOrca)
42
+ - [metaeval/ScienceQA_text_only](https://huggingface.co/datasets/metaeval/ScienceQA_text_only)
43
+ - [GAIR/lima](https://huggingface.co/datasets/GAIR/lima)
44
+
45
+ ## Hardware and Software
46
+
47
+ ### Hardware
48
+ - We utilized an A100 for training our model.
49
+
50
+ ### Training Factors
51
+ - We fine-tuned this model using a combination of the [DeepSpeed library](https://github.com/microsoft/DeepSpeed) and the [HuggingFace trainer](https://huggingface.co/docs/transformers/main_classes/trainer).
52
+
53
+ ## Evaluation Results
54
+
55
+ ### Overview
56
+ - We conducted a performance evaluation based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
57
+ We evaluated our model on four benchmark datasets, which include `ARC-Challenge`, `HellaSwag`, `MMLU`, and `TruthfulQA`.
58
+ We used the [lm-evaluation-harness repository](https://github.com/EleutherAI/lm-evaluation-harness), specifically commit [b281b0921b636bc36ad05c0b0b0763bd6dd43463](https://github.com/EleutherAI/lm-evaluation-harness/tree/b281b0921b636bc36ad05c0b0b0763bd6dd43463).
59
+
60
+ ### Main Results
61
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA |
62
+ |-----------------------------------------------|---------|-------|-----------|-------|------------|
63
+ | llama-65b-instruct (***Ours***, ***Local Reproduction***) | **69.4** | **67.6** | **86.5** | **64.9** | **58.8** |
64
+ | llama-30b-instruct-2048 (***Ours***, ***Open LLM Leaderboard***) | 67.0 | 64.9 | 84.9 | 61.9 | 56.3 |
65
+ | falcon-40b-instruct | 63.4 | 61.6 | 84.3 | 55.4 | 52.5 |
66
+ | llama-30b-instruct (***Ours***, ***Open LLM Leaderboard***) | 63.2 | 56.7 | 84.0 | 59.0 | 53.1 |
67
+ | llama-65b | 62.1 | 57.6 | 84.3 | 63.4 | 43.0 |
68
+
69
+ ### Scripts
70
+ - Prepare evaluation environments:
71
+ ```
72
+ # clone the repository
73
+ git clone https://github.com/EleutherAI/lm-evaluation-harness.git
74
+
75
+ # check out the specific commit
76
+ git checkout b281b0921b636bc36ad05c0b0b0763bd6dd43463
77
+
78
+ # change to the repository directory
79
+ cd lm-evaluation-harness
80
+ ```
81
+
82
+ ## Ethical Issues
83
+
84
+ ### Ethical Considerations
85
+ - There were no ethical issues involved, as we did not include the benchmark test set or the training set in the model's training process.
86
+
87
+ ## Contact Us
88
+
89
+ ### Why Upstage LLM?
90
+ - [Upstage](https://en.upstage.ai)'s LLM research has yielded remarkable results. Our 30B model size outperforms all models worldwide with less than 65B, establishing itself as the leading performer. Recognizing the immense potential for private LLM adoption within companies, we invite you to effortlessly implement a private LLM and fine-tune it with your own data. For a seamless and tailored solution, please don't hesitate to reach out to us [(click here to mail)].
91
+
92
+ [(click here to mail)]: mailto:contact@upstage.ai