munish0838 commited on
Commit
d111282
·
verified ·
1 Parent(s): 04099a2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +194 -0
README.md ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ base_model:
9
+ - huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
10
+ - EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
11
+ - v000000/Qwen2.5-Lumen-14B
12
+ - qwen/Qwen2.5-14b
13
+ - arcee-ai/SuperNova-Medius
14
+ - allura-org/TQ2.5-14B-Aletheia-v1
15
+ model-index:
16
+ - name: Q2.5-Veltha-14B
17
+ results:
18
+ - task:
19
+ type: text-generation
20
+ name: Text Generation
21
+ dataset:
22
+ name: IFEval (0-Shot)
23
+ type: HuggingFaceH4/ifeval
24
+ args:
25
+ num_few_shot: 0
26
+ metrics:
27
+ - type: inst_level_strict_acc and prompt_level_strict_acc
28
+ value: 82.92
29
+ name: strict accuracy
30
+ source:
31
+ url: >-
32
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
33
+ name: Open LLM Leaderboard
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: BBH (3-Shot)
39
+ type: BBH
40
+ args:
41
+ num_few_shot: 3
42
+ metrics:
43
+ - type: acc_norm
44
+ value: 49.75
45
+ name: normalized accuracy
46
+ source:
47
+ url: >-
48
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
49
+ name: Open LLM Leaderboard
50
+ - task:
51
+ type: text-generation
52
+ name: Text Generation
53
+ dataset:
54
+ name: MATH Lvl 5 (4-Shot)
55
+ type: hendrycks/competition_math
56
+ args:
57
+ num_few_shot: 4
58
+ metrics:
59
+ - type: exact_match
60
+ value: 28.02
61
+ name: exact match
62
+ source:
63
+ url: >-
64
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
65
+ name: Open LLM Leaderboard
66
+ - task:
67
+ type: text-generation
68
+ name: Text Generation
69
+ dataset:
70
+ name: GPQA (0-shot)
71
+ type: Idavidrein/gpqa
72
+ args:
73
+ num_few_shot: 0
74
+ metrics:
75
+ - type: acc_norm
76
+ value: 14.54
77
+ name: acc_norm
78
+ source:
79
+ url: >-
80
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
81
+ name: Open LLM Leaderboard
82
+ - task:
83
+ type: text-generation
84
+ name: Text Generation
85
+ dataset:
86
+ name: MuSR (0-shot)
87
+ type: TAUR-Lab/MuSR
88
+ args:
89
+ num_few_shot: 0
90
+ metrics:
91
+ - type: acc_norm
92
+ value: 12.26
93
+ name: acc_norm
94
+ source:
95
+ url: >-
96
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
97
+ name: Open LLM Leaderboard
98
+ - task:
99
+ type: text-generation
100
+ name: Text Generation
101
+ dataset:
102
+ name: MMLU-PRO (5-shot)
103
+ type: TIGER-Lab/MMLU-Pro
104
+ config: main
105
+ split: test
106
+ args:
107
+ num_few_shot: 5
108
+ metrics:
109
+ - type: acc
110
+ value: 47.76
111
+ name: accuracy
112
+ source:
113
+ url: >-
114
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B
115
+ name: Open LLM Leaderboard
116
+ new_version: djuna/Q2.5-Veltha-14B-0.5
117
+
118
+ ---
119
+
120
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
121
+
122
+
123
+ # QuantFactory/Q2.5-Veltha-14B-GGUF
124
+ This is quantized version of [djuna/Q2.5-Veltha-14B](https://huggingface.co/djuna/Q2.5-Veltha-14B) created using llama.cpp
125
+
126
+ # Original Model Card
127
+
128
+ # merge
129
+
130
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
131
+
132
+ ## Merge Details
133
+ ### Merge Method
134
+
135
+ This model was merged using the della_linear merge method using [qwen/Qwen2.5-14b](https://huggingface.co/qwen/Qwen2.5-14b) as a base.
136
+
137
+ ### Models Merged
138
+
139
+ The following models were included in the merge:
140
+ * [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2)
141
+ * [EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2)
142
+ * [v000000/Qwen2.5-Lumen-14B](https://huggingface.co/v000000/Qwen2.5-Lumen-14B)
143
+ * [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius)
144
+ * [allura-org/TQ2.5-14B-Aletheia-v1](https://huggingface.co/allura-org/TQ2.5-14B-Aletheia-v1)
145
+
146
+ ### Configuration
147
+
148
+ The following YAML configuration was used to produce this model:
149
+
150
+ ```yaml
151
+ merge_method: della_linear
152
+ dtype: float32
153
+ out_dtype: bfloat16
154
+ parameters:
155
+ epsilon: 0.04
156
+ lambda: 1.05
157
+ normalize: true
158
+ base_model: qwen/Qwen2.5-14b
159
+ tokenizer_source: arcee-ai/SuperNova-Medius
160
+ models:
161
+ - model: arcee-ai/SuperNova-Medius
162
+ parameters:
163
+ weight: 10
164
+ density: 1
165
+ - model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
166
+ parameters:
167
+ weight: 7
168
+ density: 0.5
169
+ - model: v000000/Qwen2.5-Lumen-14B
170
+ parameters:
171
+ weight: 7
172
+ density: 0.4
173
+ - model: allura-org/TQ2.5-14B-Aletheia-v1
174
+ parameters:
175
+ weight: 8
176
+ density: 0.4
177
+ - model: huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
178
+ parameters:
179
+ weight: 8
180
+ density: 0.45
181
+ ```
182
+
183
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
184
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/djuna__Q2.5-Veltha-14B-details)
185
+
186
+ | Metric |Value|
187
+ |-------------------|----:|
188
+ |Avg. |39.21|
189
+ |IFEval (0-Shot) |82.92|
190
+ |BBH (3-Shot) |49.75|
191
+ |MATH Lvl 5 (4-Shot)|28.02|
192
+ |GPQA (0-shot) |14.54|
193
+ |MuSR (0-shot) |12.26|
194
+ |MMLU-PRO (5-shot) |47.76|