Files changed (1) hide show
  1. README.md +113 -5
README.md CHANGED
@@ -1,15 +1,110 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - en
5
- base_model:
6
- - mistralai/Mistral-7B-v0.3
7
- pipeline_tag: text-classification
8
  library_name: transformers
9
  tags:
10
  - moe
11
  - mergekit
12
  - MoErges
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
  Model Name: Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial - Mixture of Experts (MoE)
15
 
@@ -47,4 +142,17 @@ This model can be used for a wide range of applications:
47
  Limitations
48
 
49
  • The model may occasionally generate responses that are not entirely contextually appropriate, especially in cases requiring highly specialized domain knowledge.
50
- • Despite its 24-bit precision, it may not perform well with extremely large datasets or tasks that require higher precision levels.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  language:
3
  - en
4
+ license: apache-2.0
 
 
5
  library_name: transformers
6
  tags:
7
  - moe
8
  - mergekit
9
  - MoErges
10
+ base_model:
11
+ - mistralai/Mistral-7B-v0.3
12
+ pipeline_tag: text-classification
13
+ model-index:
14
+ - name: MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
15
+ results:
16
+ - task:
17
+ type: text-generation
18
+ name: Text Generation
19
+ dataset:
20
+ name: IFEval (0-Shot)
21
+ type: HuggingFaceH4/ifeval
22
+ args:
23
+ num_few_shot: 0
24
+ metrics:
25
+ - type: inst_level_strict_acc and prompt_level_strict_acc
26
+ value: 16.97
27
+ name: strict accuracy
28
+ source:
29
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
30
+ name: Open LLM Leaderboard
31
+ - task:
32
+ type: text-generation
33
+ name: Text Generation
34
+ dataset:
35
+ name: BBH (3-Shot)
36
+ type: BBH
37
+ args:
38
+ num_few_shot: 3
39
+ metrics:
40
+ - type: acc_norm
41
+ value: 8.87
42
+ name: normalized accuracy
43
+ source:
44
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
45
+ name: Open LLM Leaderboard
46
+ - task:
47
+ type: text-generation
48
+ name: Text Generation
49
+ dataset:
50
+ name: MATH Lvl 5 (4-Shot)
51
+ type: hendrycks/competition_math
52
+ args:
53
+ num_few_shot: 4
54
+ metrics:
55
+ - type: exact_match
56
+ value: 0.3
57
+ name: exact match
58
+ source:
59
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
60
+ name: Open LLM Leaderboard
61
+ - task:
62
+ type: text-generation
63
+ name: Text Generation
64
+ dataset:
65
+ name: GPQA (0-shot)
66
+ type: Idavidrein/gpqa
67
+ args:
68
+ num_few_shot: 0
69
+ metrics:
70
+ - type: acc_norm
71
+ value: 1.23
72
+ name: acc_norm
73
+ source:
74
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: MuSR (0-shot)
81
+ type: TAUR-Lab/MuSR
82
+ args:
83
+ num_few_shot: 0
84
+ metrics:
85
+ - type: acc_norm
86
+ value: 7.85
87
+ name: acc_norm
88
+ source:
89
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
90
+ name: Open LLM Leaderboard
91
+ - task:
92
+ type: text-generation
93
+ name: Text Generation
94
+ dataset:
95
+ name: MMLU-PRO (5-shot)
96
+ type: TIGER-Lab/MMLU-Pro
97
+ config: main
98
+ split: test
99
+ args:
100
+ num_few_shot: 5
101
+ metrics:
102
+ - type: acc
103
+ value: 4.21
104
+ name: accuracy
105
+ source:
106
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial
107
+ name: Open LLM Leaderboard
108
  ---
109
  Model Name: Marsouuu/MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial - Mixture of Experts (MoE)
110
 
 
142
  Limitations
143
 
144
  • The model may occasionally generate responses that are not entirely contextually appropriate, especially in cases requiring highly specialized domain knowledge.
145
+ • Despite its 24-bit precision, it may not perform well with extremely large datasets or tasks that require higher precision levels.
146
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
147
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Marsouuu__MistralBase-4x7B-MoE-ECE-PRYMMAL-Martial)
148
+
149
+ | Metric |Value|
150
+ |-------------------|----:|
151
+ |Avg. | 6.57|
152
+ |IFEval (0-Shot) |16.97|
153
+ |BBH (3-Shot) | 8.87|
154
+ |MATH Lvl 5 (4-Shot)| 0.30|
155
+ |GPQA (0-shot) | 1.23|
156
+ |MuSR (0-shot) | 7.85|
157
+ |MMLU-PRO (5-shot) | 4.21|
158
+