Adding Evaluation Results

#1
by leaderboard-pr-bot - opened
Files changed (1) hide show
  1. README.md +114 -5
README.md CHANGED
@@ -1,4 +1,9 @@
1
  ---
 
 
 
 
 
2
  base_model:
3
  - failspy/Meta-Llama-3-8B-Instruct-abliterated-v3
4
  - VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
@@ -6,11 +11,101 @@ base_model:
6
  - nbeerbower/llama-3-gutenberg-8B
7
  - jpacifico/French-Alpaca-Llama3-8B-Instruct-v1.0
8
  - meta-llama/Meta-Llama-3-8B-Instruct
9
- library_name: transformers
10
- license: apache-2.0
11
- tags:
12
- - mergekit
13
- - merge
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ---
15
  # Model Merge Parameters
16
  Base model: meta-llama/Meta-Llama-3-8B-Instruct
@@ -27,3 +122,17 @@ gamma: 0.1
27
  normalize: true
28
  weight: 1.0
29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
  base_model:
8
  - failspy/Meta-Llama-3-8B-Instruct-abliterated-v3
9
  - VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
 
11
  - nbeerbower/llama-3-gutenberg-8B
12
  - jpacifico/French-Alpaca-Llama3-8B-Instruct-v1.0
13
  - meta-llama/Meta-Llama-3-8B-Instruct
14
+ model-index:
15
+ - name: Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
16
+ results:
17
+ - task:
18
+ type: text-generation
19
+ name: Text Generation
20
+ dataset:
21
+ name: IFEval (0-Shot)
22
+ type: HuggingFaceH4/ifeval
23
+ args:
24
+ num_few_shot: 0
25
+ metrics:
26
+ - type: inst_level_strict_acc and prompt_level_strict_acc
27
+ value: 42.74
28
+ name: strict accuracy
29
+ source:
30
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
31
+ name: Open LLM Leaderboard
32
+ - task:
33
+ type: text-generation
34
+ name: Text Generation
35
+ dataset:
36
+ name: BBH (3-Shot)
37
+ type: BBH
38
+ args:
39
+ num_few_shot: 3
40
+ metrics:
41
+ - type: acc_norm
42
+ value: 30.51
43
+ name: normalized accuracy
44
+ source:
45
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
46
+ name: Open LLM Leaderboard
47
+ - task:
48
+ type: text-generation
49
+ name: Text Generation
50
+ dataset:
51
+ name: MATH Lvl 5 (4-Shot)
52
+ type: hendrycks/competition_math
53
+ args:
54
+ num_few_shot: 4
55
+ metrics:
56
+ - type: exact_match
57
+ value: 7.02
58
+ name: exact match
59
+ source:
60
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
61
+ name: Open LLM Leaderboard
62
+ - task:
63
+ type: text-generation
64
+ name: Text Generation
65
+ dataset:
66
+ name: GPQA (0-shot)
67
+ type: Idavidrein/gpqa
68
+ args:
69
+ num_few_shot: 0
70
+ metrics:
71
+ - type: acc_norm
72
+ value: 7.83
73
+ name: acc_norm
74
+ source:
75
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
76
+ name: Open LLM Leaderboard
77
+ - task:
78
+ type: text-generation
79
+ name: Text Generation
80
+ dataset:
81
+ name: MuSR (0-shot)
82
+ type: TAUR-Lab/MuSR
83
+ args:
84
+ num_few_shot: 0
85
+ metrics:
86
+ - type: acc_norm
87
+ value: 11.4
88
+ name: acc_norm
89
+ source:
90
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: MMLU-PRO (5-shot)
97
+ type: TIGER-Lab/MMLU-Pro
98
+ config: main
99
+ split: test
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 30.44
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=johnsutor/Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1
108
+ name: Open LLM Leaderboard
109
  ---
110
  # Model Merge Parameters
111
  Base model: meta-llama/Meta-Llama-3-8B-Instruct
 
122
  normalize: true
123
  weight: 1.0
124
 
125
+
126
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
127
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_johnsutor__Llama-3-8B-Instruct_breadcrumbs-density-0.3-gamma-0.1)
128
+
129
+ | Metric |Value|
130
+ |-------------------|----:|
131
+ |Avg. |21.66|
132
+ |IFEval (0-Shot) |42.74|
133
+ |BBH (3-Shot) |30.51|
134
+ |MATH Lvl 5 (4-Shot)| 7.02|
135
+ |GPQA (0-shot) | 7.83|
136
+ |MuSR (0-shot) |11.40|
137
+ |MMLU-PRO (5-shot) |30.44|
138
+