Severian leaderboard-pr-bot commited on
Commit
54769e3
1 Parent(s): 0943653

Adding Evaluation Results (#3)

Browse files

- Adding Evaluation Results (757f85dc72bd5a4ff03cbb150d2634928d4a225e)


Co-authored-by: Open LLM Leaderboard PR Bot <leaderboard-pr-bot@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +124 -8
README.md CHANGED
@@ -1,11 +1,5 @@
1
  ---
2
  license: artistic-2.0
3
- datasets:
4
- - Severian/Biomimicry
5
- - emrgnt-cmplxty/sciphi-textbooks-are-all-you-need
6
- - fmars/wiki_stem
7
- - fblgit/tree-of-knowledge
8
- - Severian/Bio-Design-Process
9
  tags:
10
  - chemistry
11
  - biology
@@ -17,9 +11,118 @@ tags:
17
  - biomimicry
18
  - fauna
19
  - flora
20
- pipeline_tag: text-generation
 
 
 
 
 
21
  metrics:
22
  - accuracy
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  ---
24
  # ANIMA-Phi-Neptune-Mistral-7B: Biomimicry Enhanced LLM
25
 
@@ -186,4 +289,17 @@ that help protect our environment from contamination.
186
 
187
  This project is licensed under Artistic-2.0
188
 
189
- **This model is for research purposes only and restricted from any commercial use**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: artistic-2.0
 
 
 
 
 
 
3
  tags:
4
  - chemistry
5
  - biology
 
11
  - biomimicry
12
  - fauna
13
  - flora
14
+ datasets:
15
+ - Severian/Biomimicry
16
+ - emrgnt-cmplxty/sciphi-textbooks-are-all-you-need
17
+ - fmars/wiki_stem
18
+ - fblgit/tree-of-knowledge
19
+ - Severian/Bio-Design-Process
20
  metrics:
21
  - accuracy
22
+ pipeline_tag: text-generation
23
+ model-index:
24
+ - name: ANIMA-Phi-Neptune-Mistral-7B-v4
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 55.46
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 77.63
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 53.12
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 59.01
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 73.48
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 14.94
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
125
+ name: Open LLM Leaderboard
126
  ---
127
  # ANIMA-Phi-Neptune-Mistral-7B: Biomimicry Enhanced LLM
128
 
 
289
 
290
  This project is licensed under Artistic-2.0
291
 
292
+ **This model is for research purposes only and restricted from any commercial use**
293
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
294
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Severian__ANIMA-Phi-Neptune-Mistral-7B-v4)
295
+
296
+ | Metric |Value|
297
+ |---------------------------------|----:|
298
+ |Avg. |55.61|
299
+ |AI2 Reasoning Challenge (25-Shot)|55.46|
300
+ |HellaSwag (10-Shot) |77.63|
301
+ |MMLU (5-Shot) |53.12|
302
+ |TruthfulQA (0-shot) |59.01|
303
+ |Winogrande (5-shot) |73.48|
304
+ |GSM8k (5-shot) |14.94|
305
+