Files changed (1) hide show
  1. README.md +112 -4
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
2
  license: other
3
- license_name: tongyi-qianwen
4
- license_link: >-
5
- https://huggingface.co/Qwen/Qwen1.5-110B/blob/main/LICENSE
6
- base_model: Qwen/Qwen2-72B
7
  tags:
8
  - generated_from_trainer
9
  - axolotl
 
10
  datasets:
11
  - cognitivecomputations/Dolphin-2.9
12
  - teknium/OpenHermes-2.5
@@ -16,6 +13,103 @@ datasets:
16
  - microsoft/orca-math-word-problems-200k
17
  - Locutusque/function-calling-chatml
18
  - internlm/Agent-FLAN
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ---
20
 
21
  # Dolphin 2.9.2 Qwen2 72B 🐬
@@ -441,3 +535,17 @@ special_tokens:
441
 
442
  ```
443
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: other
 
 
 
 
3
  tags:
4
  - generated_from_trainer
5
  - axolotl
6
+ base_model: Qwen/Qwen2-72B
7
  datasets:
8
  - cognitivecomputations/Dolphin-2.9
9
  - teknium/OpenHermes-2.5
 
13
  - microsoft/orca-math-word-problems-200k
14
  - Locutusque/function-calling-chatml
15
  - internlm/Agent-FLAN
16
+ license_name: tongyi-qianwen
17
+ license_link: https://huggingface.co/Qwen/Qwen1.5-110B/blob/main/LICENSE
18
+ model-index:
19
+ - name: dolphin-2.9.2-qwen2-72b
20
+ results:
21
+ - task:
22
+ type: text-generation
23
+ name: Text Generation
24
+ dataset:
25
+ name: IFEval (0-Shot)
26
+ type: HuggingFaceH4/ifeval
27
+ args:
28
+ num_few_shot: 0
29
+ metrics:
30
+ - type: inst_level_strict_acc and prompt_level_strict_acc
31
+ value: 40.38
32
+ name: strict accuracy
33
+ source:
34
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
35
+ name: Open LLM Leaderboard
36
+ - task:
37
+ type: text-generation
38
+ name: Text Generation
39
+ dataset:
40
+ name: BBH (3-Shot)
41
+ type: BBH
42
+ args:
43
+ num_few_shot: 3
44
+ metrics:
45
+ - type: acc_norm
46
+ value: 47.7
47
+ name: normalized accuracy
48
+ source:
49
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
50
+ name: Open LLM Leaderboard
51
+ - task:
52
+ type: text-generation
53
+ name: Text Generation
54
+ dataset:
55
+ name: MATH Lvl 5 (4-Shot)
56
+ type: hendrycks/competition_math
57
+ args:
58
+ num_few_shot: 4
59
+ metrics:
60
+ - type: exact_match
61
+ value: 21.37
62
+ name: exact match
63
+ source:
64
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
65
+ name: Open LLM Leaderboard
66
+ - task:
67
+ type: text-generation
68
+ name: Text Generation
69
+ dataset:
70
+ name: GPQA (0-shot)
71
+ type: Idavidrein/gpqa
72
+ args:
73
+ num_few_shot: 0
74
+ metrics:
75
+ - type: acc_norm
76
+ value: 16.0
77
+ name: acc_norm
78
+ source:
79
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
80
+ name: Open LLM Leaderboard
81
+ - task:
82
+ type: text-generation
83
+ name: Text Generation
84
+ dataset:
85
+ name: MuSR (0-shot)
86
+ type: TAUR-Lab/MuSR
87
+ args:
88
+ num_few_shot: 0
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 17.04
92
+ name: acc_norm
93
+ source:
94
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: MMLU-PRO (5-shot)
101
+ type: TIGER-Lab/MMLU-Pro
102
+ config: main
103
+ split: test
104
+ args:
105
+ num_few_shot: 5
106
+ metrics:
107
+ - type: acc
108
+ value: 49.52
109
+ name: accuracy
110
+ source:
111
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=cognitivecomputations/dolphin-2.9.2-qwen2-72b
112
+ name: Open LLM Leaderboard
113
  ---
114
 
115
  # Dolphin 2.9.2 Qwen2 72B 🐬
 
535
 
536
  ```
537
 
538
+
539
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
540
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_cognitivecomputations__dolphin-2.9.2-qwen2-72b)
541
+
542
+ | Metric |Value|
543
+ |-------------------|----:|
544
+ |Avg. |32.00|
545
+ |IFEval (0-Shot) |40.38|
546
+ |BBH (3-Shot) |47.70|
547
+ |MATH Lvl 5 (4-Shot)|21.37|
548
+ |GPQA (0-shot) |16.00|
549
+ |MuSR (0-shot) |17.04|
550
+ |MMLU-PRO (5-shot) |49.52|
551
+