leaderboard-pr-bot commited on
Commit
4c3d3c6
1 Parent(s): 0d5e341

Adding Evaluation Results

Browse files

This is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr

The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.

If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions

Files changed (1) hide show
  1. README.md +128 -11
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
 
 
2
  license: apache-2.0
3
  datasets:
4
  - HuggingFaceTB/cosmopedia
5
- language:
6
- - en
7
  inference:
8
  parameters:
9
  temperature: 0.6
@@ -11,15 +11,118 @@ inference:
11
  top_k: 50
12
  repetition_penalty: 1.2
13
  widget:
14
- - text: 'Photosynthesis is'
15
- example_title: Textbook
16
- group: Completion
17
- - text: '<s> [INST] How to take care of plants? [/INST] '
18
- example_title: Wikihow
19
- group: Completion
20
- - text: '<s> [INST] Generate a story about a flying cat [/INST] '
21
- example_title: Story
22
- group: Completion
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  ---
24
 
25
  # Model Summary
@@ -107,3 +210,17 @@ This is a small 1.8B model trained on synthetic data, so it might hallucinate, g
107
  The training loss:
108
 
109
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/rJobY7F6tqTAvIox1ZGKR.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: apache-2.0
5
  datasets:
6
  - HuggingFaceTB/cosmopedia
 
 
7
  inference:
8
  parameters:
9
  temperature: 0.6
 
11
  top_k: 50
12
  repetition_penalty: 1.2
13
  widget:
14
+ - text: Photosynthesis is
15
+ example_title: Textbook
16
+ group: Completion
17
+ - text: '<s> [INST] How to take care of plants? [/INST] '
18
+ example_title: Wikihow
19
+ group: Completion
20
+ - text: '<s> [INST] Generate a story about a flying cat [/INST] '
21
+ example_title: Story
22
+ group: Completion
23
+ model-index:
24
+ - name: cosmo-1b
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 38.57
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 55.13
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 26.69
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 38.15
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 55.49
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 5.53
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceTB/cosmo-1b
125
+ name: Open LLM Leaderboard
126
  ---
127
 
128
  # Model Summary
 
210
  The training loss:
211
 
212
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61c141342aac764ce1654e43/rJobY7F6tqTAvIox1ZGKR.png)
213
+
214
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
215
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_HuggingFaceTB__cosmo-1b)
216
+
217
+ | Metric |Value|
218
+ |---------------------------------|----:|
219
+ |Avg. |36.59|
220
+ |AI2 Reasoning Challenge (25-Shot)|38.57|
221
+ |HellaSwag (10-Shot) |55.13|
222
+ |MMLU (5-Shot) |26.69|
223
+ |TruthfulQA (0-shot) |38.15|
224
+ |Winogrande (5-shot) |55.49|
225
+ |GSM8k (5-shot) | 5.53|
226
+