Files changed (1) hide show
  1. README.md +129 -11
README.md CHANGED
@@ -1,23 +1,128 @@
1
  ---
 
 
 
2
  license: other
 
 
 
 
 
 
3
  license_name: yi-license
4
  license_link: LICENSE.md
5
  extra_gated_heading: Access beomi/Yi-Ko-6B on Hugging Face
6
  extra_gated_button_content: Submit
7
  extra_gated_fields:
8
  I agree to share my name, email address and username: checkbox
9
- I confirm that I understand this project is for research purposes only, and confirm that I agree to follow the LICENSE of this model: checkbox
10
- language:
11
- - en
12
- - ko
13
  pipeline_tag: text-generation
14
  inference: false
15
- tags:
16
- - pytorch
17
- - Yi-Ko
18
- - 01-ai
19
- - Yi
20
- library_name: transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  ---
22
 
23
  > Update @ 2024.01.29 New Model [beomi/Yi-Ko-DUS-9B](https://huggingface.co/beomi/Yi-Ko-DUS-9B) Released! 🎉
@@ -120,4 +225,17 @@ Please use this bibtex below:
120
 
121
  ## Acknowledgement
122
 
123
- The training is supported by [TPU Research Cloud](https://sites.research.google/trc/) program.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ - ko
5
  license: other
6
+ library_name: transformers
7
+ tags:
8
+ - pytorch
9
+ - Yi-Ko
10
+ - 01-ai
11
+ - Yi
12
  license_name: yi-license
13
  license_link: LICENSE.md
14
  extra_gated_heading: Access beomi/Yi-Ko-6B on Hugging Face
15
  extra_gated_button_content: Submit
16
  extra_gated_fields:
17
  I agree to share my name, email address and username: checkbox
18
+ ? I confirm that I understand this project is for research purposes only, and confirm
19
+ that I agree to follow the LICENSE of this model
20
+ : checkbox
 
21
  pipeline_tag: text-generation
22
  inference: false
23
+ model-index:
24
+ - name: Yi-Ko-6B
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 48.89
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 74.48
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 55.72
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 37.09
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 72.93
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 12.51
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=beomi/Yi-Ko-6B
125
+ name: Open LLM Leaderboard
126
  ---
127
 
128
  > Update @ 2024.01.29 New Model [beomi/Yi-Ko-DUS-9B](https://huggingface.co/beomi/Yi-Ko-DUS-9B) Released! 🎉
 
225
 
226
  ## Acknowledgement
227
 
228
+ The training is supported by [TPU Research Cloud](https://sites.research.google/trc/) program.
229
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
230
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_beomi__Yi-Ko-6B)
231
+
232
+ | Metric |Value|
233
+ |---------------------------------|----:|
234
+ |Avg. |50.27|
235
+ |AI2 Reasoning Challenge (25-Shot)|48.89|
236
+ |HellaSwag (10-Shot) |74.48|
237
+ |MMLU (5-Shot) |55.72|
238
+ |TruthfulQA (0-shot) |37.09|
239
+ |Winogrande (5-shot) |72.93|
240
+ |GSM8k (5-shot) |12.51|
241
+