mkurman commited on
Commit
053f09e
1 Parent(s): 27a77f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +166 -3
README.md CHANGED
@@ -1,3 +1,166 @@
1
- ---
2
- license: llama3.2
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.2
3
+ base_model:
4
+ - meta-llama/Llama-3.2-1B-Instruct
5
+ model-index:
6
+ - name: Llama-3.2-SUN-2.4B-v1.0.0
7
+ results:
8
+ - task:
9
+ type: text-generation
10
+ name: Text Generation
11
+ dataset:
12
+ name: IFEval (0-Shot)
13
+ type: HuggingFaceH4/ifeval
14
+ args:
15
+ num_few_shot: 0
16
+ metrics:
17
+ - type: inst_level_strict_acc and prompt_level_strict_acc
18
+ value: 56.37
19
+ name: strict accuracy
20
+ source:
21
+ url: >-
22
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
23
+ name: Open LLM Leaderboard
24
+ - task:
25
+ type: text-generation
26
+ name: Text Generation
27
+ dataset:
28
+ name: BBH (3-Shot)
29
+ type: BBH
30
+ args:
31
+ num_few_shot: 3
32
+ metrics:
33
+ - type: acc_norm
34
+ value: 7.21
35
+ name: normalized accuracy
36
+ source:
37
+ url: >-
38
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: MATH Lvl 5 (4-Shot)
45
+ type: hendrycks/competition_math
46
+ args:
47
+ num_few_shot: 4
48
+ metrics:
49
+ - type: exact_match
50
+ value: 4.83
51
+ name: exact match
52
+ source:
53
+ url: >-
54
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
55
+ name: Open LLM Leaderboard
56
+ - task:
57
+ type: text-generation
58
+ name: Text Generation
59
+ dataset:
60
+ name: GPQA (0-shot)
61
+ type: Idavidrein/gpqa
62
+ args:
63
+ num_few_shot: 0
64
+ metrics:
65
+ - type: acc_norm
66
+ value: 1.01
67
+ name: acc_norm
68
+ source:
69
+ url: >-
70
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
71
+ name: Open LLM Leaderboard
72
+ - task:
73
+ type: text-generation
74
+ name: Text Generation
75
+ dataset:
76
+ name: MuSR (0-shot)
77
+ type: TAUR-Lab/MuSR
78
+ args:
79
+ num_few_shot: 0
80
+ metrics:
81
+ - type: acc_norm
82
+ value: 3.02
83
+ name: acc_norm
84
+ source:
85
+ url: >-
86
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MMLU-PRO (5-shot)
93
+ type: TIGER-Lab/MMLU-Pro
94
+ config: main
95
+ split: test
96
+ args:
97
+ num_few_shot: 5
98
+ metrics:
99
+ - type: acc
100
+ value: 6.03
101
+ name: accuracy
102
+ source:
103
+ url: >-
104
+ https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=meditsolutions/Llama-3.2-SUN-2.4B-v1.0.0
105
+ name: Open LLM Leaderboard
106
+ datasets:
107
+ - argilla/OpenHermesPreferences
108
+ - argilla/magpie-ultra-v0.1
109
+ - argilla/Capybara-Preferences-Filtered
110
+ - mlabonne/open-perfectblend
111
+ - HuggingFaceTB/everyday-conversations-llama3.1-2k
112
+ - WizardLMTeam/WizardLM_evol_instruct_V2_196k
113
+ - ProlificAI/social-reasoning-rlhf
114
+ language:
115
+ - en
116
+ ---
117
+
118
+ # MedIT SUN 2.5B
119
+
120
+ <div align="center">
121
+ <img src="https://i.ibb.co/3Y4fz6m/imagine-image-c9f7de3a-753f-4a65-9e8f-002bf03a30cd.png" alt="Llama-3.2-MedIT-SUN-2.5B" style="border-radius: 10px; box-shadow: 0 4px 8px 0 rgba(0, 0, 0, 0.2), 0 6px 20px 0 rgba(0, 0, 0, 0.19); max-width: 100%; height: auto;">
122
+ </div>
123
+
124
+ **Base Model**
125
+ - Llama 3.2 1B
126
+
127
+ **Extended Size**
128
+ - 1B to 2.5B parameters
129
+
130
+ **Extension Method**
131
+ - Proprietary technique developed by MedIT Solutions
132
+
133
+ **Fine-tuning**
134
+ - Open (or open subsets allowing for commercial use) open datasets from HF
135
+ - Open (or open subsets allowing for commercial use) SFT datasets from HF
136
+
137
+ **Training Status**
138
+ - Current version: chat-1.0.0
139
+
140
+ **Key Features**
141
+ - Built on Llama 3.2 architecture
142
+ - Expanded from 1B to 2.47B parameters
143
+ - Optimized for open-ended conversations
144
+ - Incorporates supervised fine-tuning for improved performance
145
+
146
+ **Use Case**
147
+ - General conversation and task-oriented interactions
148
+
149
+ **Limitations**
150
+ As the model is still in training, performance and capabilities may vary. Users should be aware that the model is not in its final form and may exhibit inconsistencies or limitations typical of in-progress AI models.
151
+
152
+ **Disclaimer and Safety Considerations**
153
+ The Model is designed to be used as a smart assistant but not as a knowledge source within your applications, systems, or environments. It is not intended to provide 100% accurate answers, especially in scenarios where high precision and accuracy are crucial.
154
+
155
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
156
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/meditsolutions__Llama-3.2-SUN-2.4B-v1.0.0-details)
157
+
158
+ | Metric |Value|
159
+ |-------------------|----:|
160
+ |Avg. |13.08|
161
+ |IFEval (0-Shot) |56.37|
162
+ |BBH (3-Shot) | 7.21|
163
+ |MATH Lvl 5 (4-Shot)| 4.83|
164
+ |GPQA (0-shot) | 1.01|
165
+ |MuSR (0-shot) | 3.02|
166
+ |MMLU-PRO (5-shot) | 6.03|