Upload folder using huggingface_hub
Browse files- .mdl +0 -0
- .msc +0 -0
- .mv +1 -0
- README.md +4 -0
- XinYuan-Qwen2-1_5B_summary.csv +254 -0
- added_tokens.json +5 -0
- all_results.json +12 -0
- config.json +28 -0
- configuration.json +0 -0
- eval_results.json +7 -0
- generation_config.json +6 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- special_tokens_map.json +20 -0
- tokenizer.json +0 -0
- tokenizer_config.json +44 -0
- train_results.json +8 -0
- trainer_log.jsonl +46 -0
- trainer_state.json +357 -0
- training_args.bin +3 -0
- training_loss.png +0 -0
- vocab.json +0 -0
.mdl
ADDED
Binary file (48 Bytes). View file
|
|
.msc
ADDED
Binary file (1.39 kB). View file
|
|
.mv
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
Revision:master,CreatedAt:1724494558
|
README.md
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
base_model: Qwen2-1.5B
|
4 |
+
---
|
XinYuan-Qwen2-1_5B_summary.csv
ADDED
@@ -0,0 +1,254 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
dataset,version,metric,mode,XinYuan-Qwen2-1_5B-0822
|
2 |
+
lukaemon_mmlu_college_biology,caec7d,accuracy,gen,53.47
|
3 |
+
lukaemon_mmlu_college_chemistry,520aa6,accuracy,gen,40.00
|
4 |
+
lukaemon_mmlu_college_computer_science,99c216,accuracy,gen,45.00
|
5 |
+
lukaemon_mmlu_college_mathematics,678751,accuracy,gen,34.00
|
6 |
+
lukaemon_mmlu_college_physics,4f382c,accuracy,gen,29.41
|
7 |
+
lukaemon_mmlu_electrical_engineering,770ce3,accuracy,gen,55.17
|
8 |
+
lukaemon_mmlu_astronomy,d3ee01,accuracy,gen,50.66
|
9 |
+
lukaemon_mmlu_anatomy,72183b,accuracy,gen,54.07
|
10 |
+
lukaemon_mmlu_abstract_algebra,2db373,accuracy,gen,38.00
|
11 |
+
lukaemon_mmlu_machine_learning,0283bb,accuracy,gen,39.29
|
12 |
+
lukaemon_mmlu_clinical_knowledge,cb3218,accuracy,gen,57.74
|
13 |
+
lukaemon_mmlu_global_facts,ab07b6,accuracy,gen,33.00
|
14 |
+
lukaemon_mmlu_management,80876d,accuracy,gen,80.58
|
15 |
+
lukaemon_mmlu_nutrition,4543bd,accuracy,gen,65.03
|
16 |
+
lukaemon_mmlu_marketing,7394e3,accuracy,gen,82.05
|
17 |
+
lukaemon_mmlu_professional_accounting,444b7f,accuracy,gen,43.97
|
18 |
+
lukaemon_mmlu_high_school_geography,0780e6,accuracy,gen,69.19
|
19 |
+
lukaemon_mmlu_international_law,cf3179,accuracy,gen,72.73
|
20 |
+
lukaemon_mmlu_moral_scenarios,f6dbe2,accuracy,gen,26.37
|
21 |
+
lukaemon_mmlu_computer_security,ce7550,accuracy,gen,67.00
|
22 |
+
lukaemon_mmlu_high_school_microeconomics,04d21a,accuracy,gen,56.30
|
23 |
+
lukaemon_mmlu_professional_law,5f7e6c,accuracy,gen,42.37
|
24 |
+
lukaemon_mmlu_medical_genetics,881ef5,accuracy,gen,63.00
|
25 |
+
lukaemon_mmlu_professional_psychology,221a16,accuracy,gen,53.43
|
26 |
+
lukaemon_mmlu_jurisprudence,001f24,accuracy,gen,69.44
|
27 |
+
lukaemon_mmlu_world_religions,232c09,accuracy,gen,69.01
|
28 |
+
lukaemon_mmlu_philosophy,08042b,accuracy,gen,60.45
|
29 |
+
lukaemon_mmlu_virology,12e270,accuracy,gen,43.98
|
30 |
+
lukaemon_mmlu_high_school_chemistry,ae8820,accuracy,gen,49.75
|
31 |
+
lukaemon_mmlu_public_relations,e7d39b,accuracy,gen,58.18
|
32 |
+
lukaemon_mmlu_high_school_macroeconomics,a01685,accuracy,gen,56.15
|
33 |
+
lukaemon_mmlu_human_sexuality,42407c,accuracy,gen,67.18
|
34 |
+
lukaemon_mmlu_elementary_mathematics,269926,accuracy,gen,38.10
|
35 |
+
lukaemon_mmlu_high_school_physics,93278f,accuracy,gen,33.77
|
36 |
+
lukaemon_mmlu_high_school_computer_science,9965a5,accuracy,gen,56.00
|
37 |
+
lukaemon_mmlu_high_school_european_history,eefc90,accuracy,gen,65.45
|
38 |
+
lukaemon_mmlu_business_ethics,1dec08,accuracy,gen,66.00
|
39 |
+
lukaemon_mmlu_moral_disputes,a2173e,accuracy,gen,58.09
|
40 |
+
lukaemon_mmlu_high_school_statistics,8f3f3a,accuracy,gen,42.59
|
41 |
+
lukaemon_mmlu_miscellaneous,935647,accuracy,gen,70.50
|
42 |
+
lukaemon_mmlu_formal_logic,cfcb0c,accuracy,gen,37.30
|
43 |
+
lukaemon_mmlu_high_school_government_and_politics,3c52f9,accuracy,gen,75.65
|
44 |
+
lukaemon_mmlu_prehistory,bbb197,accuracy,gen,57.72
|
45 |
+
lukaemon_mmlu_security_studies,9b1743,accuracy,gen,70.61
|
46 |
+
lukaemon_mmlu_high_school_biology,37b125,accuracy,gen,63.55
|
47 |
+
lukaemon_mmlu_logical_fallacies,9cebb0,accuracy,gen,68.10
|
48 |
+
lukaemon_mmlu_high_school_world_history,048e7e,accuracy,gen,71.31
|
49 |
+
lukaemon_mmlu_professional_medicine,857144,accuracy,gen,49.26
|
50 |
+
lukaemon_mmlu_high_school_mathematics,ed4dc0,accuracy,gen,33.33
|
51 |
+
lukaemon_mmlu_college_medicine,38709e,accuracy,gen,49.71
|
52 |
+
lukaemon_mmlu_high_school_us_history,8932df,accuracy,gen,69.12
|
53 |
+
lukaemon_mmlu_sociology,c266a2,accuracy,gen,71.14
|
54 |
+
lukaemon_mmlu_econometrics,d1134d,accuracy,gen,35.09
|
55 |
+
lukaemon_mmlu_high_school_psychology,7db114,accuracy,gen,73.39
|
56 |
+
lukaemon_mmlu_human_aging,82a410,accuracy,gen,60.09
|
57 |
+
lukaemon_mmlu_us_foreign_policy,528cfe,accuracy,gen,78.00
|
58 |
+
lukaemon_mmlu_conceptual_physics,63588e,accuracy,gen,45.11
|
59 |
+
mmlu_pro_math,736233,accuracy,gen,17.25
|
60 |
+
mmlu_pro_physics,736233,accuracy,gen,13.16
|
61 |
+
mmlu_pro_chemistry,736233,accuracy,gen,9.54
|
62 |
+
mmlu_pro_law,736233,accuracy,gen,17.44
|
63 |
+
mmlu_pro_engineering,736233,accuracy,gen,11.35
|
64 |
+
mmlu_pro_other,736233,accuracy,gen,25.11
|
65 |
+
mmlu_pro_economics,736233,accuracy,gen,29.50
|
66 |
+
mmlu_pro_health,736233,accuracy,gen,21.64
|
67 |
+
mmlu_pro_psychology,736233,accuracy,gen,30.83
|
68 |
+
mmlu_pro_business,736233,accuracy,gen,17.11
|
69 |
+
mmlu_pro_biology,736233,accuracy,gen,31.52
|
70 |
+
mmlu_pro_philosophy,736233,accuracy,gen,20.84
|
71 |
+
mmlu_pro_computer_science,736233,accuracy,gen,23.90
|
72 |
+
mmlu_pro_history,736233,accuracy,gen,22.31
|
73 |
+
gsm8k,1d7fe4,accuracy,gen,57.62
|
74 |
+
ceval-computer_network,db9ce2,accuracy,gen,52.63
|
75 |
+
ceval-operating_system,1c2571,accuracy,gen,52.63
|
76 |
+
ceval-computer_architecture,a74dad,accuracy,gen,61.90
|
77 |
+
ceval-college_programming,4ca32a,accuracy,gen,67.57
|
78 |
+
ceval-college_physics,963fa8,accuracy,gen,36.84
|
79 |
+
ceval-college_chemistry,e78857,accuracy,gen,50.00
|
80 |
+
ceval-advanced_mathematics,ce03e2,accuracy,gen,57.89
|
81 |
+
ceval-probability_and_statistics,65e812,accuracy,gen,27.78
|
82 |
+
ceval-discrete_mathematics,e894ae,accuracy,gen,31.25
|
83 |
+
ceval-electrical_engineer,ae42b9,accuracy,gen,48.65
|
84 |
+
ceval-metrology_engineer,ee34ea,accuracy,gen,83.33
|
85 |
+
ceval-high_school_mathematics,1dc5bf,accuracy,gen,16.67
|
86 |
+
ceval-high_school_physics,adf25f,accuracy,gen,63.16
|
87 |
+
ceval-high_school_chemistry,2ed27f,accuracy,gen,68.42
|
88 |
+
ceval-high_school_biology,8e2b9a,accuracy,gen,84.21
|
89 |
+
ceval-middle_school_mathematics,bee8d5,accuracy,gen,63.16
|
90 |
+
ceval-middle_school_biology,86817c,accuracy,gen,85.71
|
91 |
+
ceval-middle_school_physics,8accf6,accuracy,gen,78.95
|
92 |
+
ceval-middle_school_chemistry,167a15,accuracy,gen,85.00
|
93 |
+
ceval-veterinary_medicine,b4e08d,accuracy,gen,78.26
|
94 |
+
ceval-college_economics,f3f4e6,accuracy,gen,72.73
|
95 |
+
ceval-business_administration,c1614e,accuracy,gen,72.73
|
96 |
+
ceval-marxism,cf874c,accuracy,gen,89.47
|
97 |
+
ceval-mao_zedong_thought,51c7a4,accuracy,gen,87.50
|
98 |
+
ceval-education_science,591fee,accuracy,gen,82.76
|
99 |
+
ceval-teacher_qualification,4e4ced,accuracy,gen,86.36
|
100 |
+
ceval-high_school_politics,5c0de2,accuracy,gen,78.95
|
101 |
+
ceval-high_school_geography,865461,accuracy,gen,73.68
|
102 |
+
ceval-middle_school_politics,5be3e7,accuracy,gen,85.71
|
103 |
+
ceval-middle_school_geography,8a63be,accuracy,gen,91.67
|
104 |
+
ceval-modern_chinese_history,fc01af,accuracy,gen,86.96
|
105 |
+
ceval-ideological_and_moral_cultivation,a2aa4a,accuracy,gen,100.00
|
106 |
+
ceval-logic,f5b022,accuracy,gen,63.64
|
107 |
+
ceval-law,a110a1,accuracy,gen,70.83
|
108 |
+
ceval-chinese_language_and_literature,0f8b68,accuracy,gen,52.17
|
109 |
+
ceval-art_studies,2a1300,accuracy,gen,69.70
|
110 |
+
ceval-professional_tour_guide,4e673e,accuracy,gen,82.76
|
111 |
+
ceval-legal_professional,ce8787,accuracy,gen,60.87
|
112 |
+
ceval-high_school_chinese,315705,accuracy,gen,68.42
|
113 |
+
ceval-high_school_history,7eb30a,accuracy,gen,80.00
|
114 |
+
ceval-middle_school_history,48ab4a,accuracy,gen,95.45
|
115 |
+
ceval-civil_servant,87d061,accuracy,gen,59.57
|
116 |
+
ceval-sports_science,70f27b,accuracy,gen,73.68
|
117 |
+
ceval-plant_protection,8941f9,accuracy,gen,63.64
|
118 |
+
ceval-basic_medicine,c409d6,accuracy,gen,84.21
|
119 |
+
ceval-clinical_medicine,49e82d,accuracy,gen,72.73
|
120 |
+
ceval-urban_and_rural_planner,95b885,accuracy,gen,67.39
|
121 |
+
ceval-accountant,002837,accuracy,gen,67.35
|
122 |
+
ceval-fire_engineer,bc23f5,accuracy,gen,61.29
|
123 |
+
ceval-environmental_impact_assessment_engineer,c64e2d,accuracy,gen,67.74
|
124 |
+
ceval-tax_accountant,3a5e3c,accuracy,gen,71.43
|
125 |
+
ceval-physician,6e277d,accuracy,gen,75.51
|
126 |
+
cmmlu-agronomy,4c7f2c,accuracy,gen,62.13
|
127 |
+
cmmlu-anatomy,ea09bf,accuracy,gen,75.00
|
128 |
+
cmmlu-ancient_chinese,f7c97f,accuracy,gen,32.93
|
129 |
+
cmmlu-arts,dd77b8,accuracy,gen,86.25
|
130 |
+
cmmlu-astronomy,1e49db,accuracy,gen,34.55
|
131 |
+
cmmlu-business_ethics,dc78cb,accuracy,gen,62.68
|
132 |
+
cmmlu-chinese_civil_service_exam,1de82c,accuracy,gen,66.25
|
133 |
+
cmmlu-chinese_driving_rule,b8a42b,accuracy,gen,96.18
|
134 |
+
cmmlu-chinese_food_culture,2d568a,accuracy,gen,59.56
|
135 |
+
cmmlu-chinese_foreign_policy,dc2427,accuracy,gen,68.22
|
136 |
+
cmmlu-chinese_history,4cc7ed,accuracy,gen,82.97
|
137 |
+
cmmlu-chinese_literature,af3c41,accuracy,gen,54.90
|
138 |
+
cmmlu-chinese_teacher_qualification,87de11,accuracy,gen,86.03
|
139 |
+
cmmlu-clinical_knowledge,c55b1d,accuracy,gen,62.03
|
140 |
+
cmmlu-college_actuarial_science,d3c360,accuracy,gen,37.74
|
141 |
+
cmmlu-college_education,df8790,accuracy,gen,80.37
|
142 |
+
cmmlu-college_engineering_hydrology,673f23,accuracy,gen,65.09
|
143 |
+
cmmlu-college_law,524c3a,accuracy,gen,62.04
|
144 |
+
cmmlu-college_mathematics,e4ebad,accuracy,gen,25.71
|
145 |
+
cmmlu-college_medical_statistics,55af35,accuracy,gen,58.49
|
146 |
+
cmmlu-college_medicine,702f48,accuracy,gen,70.70
|
147 |
+
cmmlu-computer_science,637007,accuracy,gen,69.12
|
148 |
+
cmmlu-computer_security,932b6b,accuracy,gen,87.72
|
149 |
+
cmmlu-conceptual_physics,cfc077,accuracy,gen,78.23
|
150 |
+
cmmlu-construction_project_management,968a4a,accuracy,gen,55.40
|
151 |
+
cmmlu-economics,ddaf7c,accuracy,gen,74.21
|
152 |
+
cmmlu-education,c35963,accuracy,gen,74.85
|
153 |
+
cmmlu-electrical_engineering,70e98a,accuracy,gen,77.91
|
154 |
+
cmmlu-elementary_chinese,cbcd6a,accuracy,gen,69.05
|
155 |
+
cmmlu-elementary_commonsense,a67f37,accuracy,gen,74.75
|
156 |
+
cmmlu-elementary_information_and_technology,d34d2a,accuracy,gen,86.97
|
157 |
+
cmmlu-elementary_mathematics,a9d403,accuracy,gen,43.91
|
158 |
+
cmmlu-ethnology,31955f,accuracy,gen,65.19
|
159 |
+
cmmlu-food_science,741d8e,accuracy,gen,64.34
|
160 |
+
cmmlu-genetics,c326f7,accuracy,gen,51.14
|
161 |
+
cmmlu-global_facts,0a1236,accuracy,gen,63.09
|
162 |
+
cmmlu-high_school_biology,2be811,accuracy,gen,75.15
|
163 |
+
cmmlu-high_school_chemistry,d63c05,accuracy,gen,54.55
|
164 |
+
cmmlu-high_school_geography,5cd489,accuracy,gen,74.58
|
165 |
+
cmmlu-high_school_mathematics,6b2087,accuracy,gen,37.80
|
166 |
+
cmmlu-high_school_physics,3df353,accuracy,gen,52.73
|
167 |
+
cmmlu-high_school_politics,7a88d8,accuracy,gen,67.83
|
168 |
+
cmmlu-human_sexuality,54ac98,accuracy,gen,63.49
|
169 |
+
cmmlu-international_law,0f5d40,accuracy,gen,54.59
|
170 |
+
cmmlu-journalism,a4f6a0,accuracy,gen,65.70
|
171 |
+
cmmlu-jurisprudence,7843da,accuracy,gen,76.40
|
172 |
+
cmmlu-legal_and_moral_basis,f906b0,accuracy,gen,96.26
|
173 |
+
cmmlu-logical,15a71b,accuracy,gen,62.60
|
174 |
+
cmmlu-machine_learning,bc6ad4,accuracy,gen,61.48
|
175 |
+
cmmlu-management,e5e8db,accuracy,gen,78.57
|
176 |
+
cmmlu-marketing,8b4c18,accuracy,gen,78.89
|
177 |
+
cmmlu-marxist_theory,75eb79,accuracy,gen,95.24
|
178 |
+
cmmlu-modern_chinese,83a9b7,accuracy,gen,47.41
|
179 |
+
cmmlu-nutrition,adfff7,accuracy,gen,71.03
|
180 |
+
cmmlu-philosophy,75e22d,accuracy,gen,74.29
|
181 |
+
cmmlu-professional_accounting,0edc91,accuracy,gen,84.57
|
182 |
+
cmmlu-professional_law,d24af5,accuracy,gen,63.51
|
183 |
+
cmmlu-professional_medicine,134139,accuracy,gen,63.30
|
184 |
+
cmmlu-professional_psychology,ec920e,accuracy,gen,79.31
|
185 |
+
cmmlu-public_relations,70ee06,accuracy,gen,67.82
|
186 |
+
cmmlu-security_study,45f96f,accuracy,gen,82.96
|
187 |
+
cmmlu-sociology,485285,accuracy,gen,66.81
|
188 |
+
cmmlu-sports_science,838cfe,accuracy,gen,64.85
|
189 |
+
cmmlu-traditional_chinese_medicine,3bbf64,accuracy,gen,76.76
|
190 |
+
cmmlu-virology,8925bf,accuracy,gen,72.19
|
191 |
+
cmmlu-world_history,57c97c,accuracy,gen,76.40
|
192 |
+
cmmlu-world_religions,1d0f4b,accuracy,gen,65.00
|
193 |
+
math,265cce,accuracy,gen,22.70
|
194 |
+
mbpp,830460,score,gen,33.80
|
195 |
+
mbpp,830460,pass,gen,169.00
|
196 |
+
mbpp,830460,timeout,gen,0.00
|
197 |
+
mbpp,830460,failed,gen,70.00
|
198 |
+
mbpp,830460,wrong_answer,gen,261.00
|
199 |
+
GPQA_extended,4baadb,accuracy,gen,29.30
|
200 |
+
GPQA_main,4baadb,accuracy,gen,28.35
|
201 |
+
GPQA_diamond,4baadb,accuracy,gen,26.26
|
202 |
+
bbh-temporal_sequences,e43931,score,gen,16.40
|
203 |
+
bbh-disambiguation_qa,d52c61,score,gen,43.20
|
204 |
+
bbh-date_understanding,a8000b,score,gen,46.80
|
205 |
+
bbh-tracking_shuffled_objects_three_objects,7964c0,score,gen,32.40
|
206 |
+
bbh-penguins_in_a_table,fceb27,score,gen,36.99
|
207 |
+
bbh-geometric_shapes,503c8f,score,gen,32.00
|
208 |
+
bbh-snarks,42d6ca,score,gen,55.06
|
209 |
+
bbh-ruin_names,408de8,score,gen,26.80
|
210 |
+
bbh-tracking_shuffled_objects_seven_objects,7964c0,score,gen,11.20
|
211 |
+
bbh-tracking_shuffled_objects_five_objects,7964c0,score,gen,16.40
|
212 |
+
bbh-logical_deduction_three_objects,45ebc5,score,gen,43.60
|
213 |
+
bbh-hyperbaton,5e5016,score,gen,64.80
|
214 |
+
bbh-logical_deduction_five_objects,45ebc5,score,gen,26.00
|
215 |
+
bbh-logical_deduction_seven_objects,45ebc5,score,gen,22.80
|
216 |
+
bbh-movie_recommendation,cc2fde,score,gen,47.60
|
217 |
+
bbh-salient_translation_error_detection,5b5f35,score,gen,30.00
|
218 |
+
bbh-reasoning_about_colored_objects,1cb761,score,gen,39.60
|
219 |
+
bbh-multistep_arithmetic_two,30f91e,score,gen,19.60
|
220 |
+
bbh-navigate,1576d9,score,gen,64.80
|
221 |
+
bbh-dyck_languages,805bea,score,gen,0.80
|
222 |
+
bbh-word_sorting,9a3f78,score,gen,3.60
|
223 |
+
bbh-sports_understanding,d3fa77,score,gen,57.20
|
224 |
+
bbh-boolean_expressions,612c92,score,gen,73.20
|
225 |
+
bbh-object_counting,781e5c,score,gen,56.40
|
226 |
+
bbh-formal_fallacies,eada96,score,gen,51.60
|
227 |
+
bbh-causal_judgement,89eaa4,score,gen,51.87
|
228 |
+
bbh-web_of_lies,0c0441,score,gen,65.20
|
229 |
+
IFEval,3321a3,Prompt-level-strict-accuracy,gen,24.03
|
230 |
+
IFEval,3321a3,Inst-level-strict-accuracy,gen,37.05
|
231 |
+
IFEval,3321a3,Prompt-level-loose-accuracy,gen,25.14
|
232 |
+
IFEval,3321a3,Inst-level-loose-accuracy,gen,38.49
|
233 |
+
ARC-c,1e0de5,accuracy,gen,71.19
|
234 |
+
hellaswag,6faab5,accuracy,gen,48.75
|
235 |
+
openai_humaneval,8e312c,humaneval_pass@1,gen,41.46
|
236 |
+
mmlu-humanities,-,naive_average,gen,59.04
|
237 |
+
mmlu-stem,-,naive_average,gen,45.70
|
238 |
+
mmlu-social-science,-,naive_average,gen,63.69
|
239 |
+
mmlu-other,-,naive_average,gen,58.84
|
240 |
+
mmlu,-,naive_average,gen,55.53
|
241 |
+
mmlu-weighted,-,weighted_average,gen,54.14
|
242 |
+
cmmlu-humanities,-,naive_average,gen,70.56
|
243 |
+
cmmlu-stem,-,naive_average,gen,57.10
|
244 |
+
cmmlu-social-science,-,naive_average,gen,69.72
|
245 |
+
cmmlu-other,-,naive_average,gen,73.06
|
246 |
+
cmmlu-china-specific,-,naive_average,gen,66.90
|
247 |
+
cmmlu,-,naive_average,gen,67.43
|
248 |
+
ceval-stem,-,naive_average,gen,59.70
|
249 |
+
ceval-social-science,-,naive_average,gen,82.16
|
250 |
+
ceval-humanities,-,naive_average,gen,75.53
|
251 |
+
ceval-other,-,naive_average,gen,69.50
|
252 |
+
ceval-hard,-,naive_average,gen,44.00
|
253 |
+
ceval,-,naive_average,gen,69.44
|
254 |
+
bbh,-,naive_average,gen,38.37
|
added_tokens.json
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|endoftext|>": 151643,
|
3 |
+
"<|im_end|>": 151645,
|
4 |
+
"<|im_start|>": 151644
|
5 |
+
}
|
all_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.9999310202110782,
|
3 |
+
"eval_loss": 0.8731282353401184,
|
4 |
+
"eval_runtime": 6594.4281,
|
5 |
+
"eval_samples_per_second": 3.908,
|
6 |
+
"eval_steps_per_second": 0.489,
|
7 |
+
"total_flos": 1.7598525261526073e+19,
|
8 |
+
"train_loss": 0.8787032284220849,
|
9 |
+
"train_runtime": 87969.9071,
|
10 |
+
"train_samples_per_second": 2.637,
|
11 |
+
"train_steps_per_second": 0.005
|
12 |
+
}
|
config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/nas/data/yanxin/qwen/Qwen2-1.5B",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 1536,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 8960,
|
13 |
+
"max_position_embeddings": 131072,
|
14 |
+
"max_window_layers": 21,
|
15 |
+
"model_type": "qwen2",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 28,
|
18 |
+
"num_key_value_heads": 2,
|
19 |
+
"rms_norm_eps": 1e-06,
|
20 |
+
"rope_theta": 1000000.0,
|
21 |
+
"sliding_window": null,
|
22 |
+
"tie_word_embeddings": false,
|
23 |
+
"torch_dtype": "bfloat16",
|
24 |
+
"transformers_version": "4.43.4",
|
25 |
+
"use_cache": false,
|
26 |
+
"use_sliding_window": false,
|
27 |
+
"vocab_size": 151936
|
28 |
+
}
|
configuration.json
ADDED
File without changes
|
eval_results.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.9999310202110782,
|
3 |
+
"eval_loss": 0.8731282353401184,
|
4 |
+
"eval_runtime": 6594.4281,
|
5 |
+
"eval_samples_per_second": 3.908,
|
6 |
+
"eval_steps_per_second": 0.489
|
7 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 151643,
|
3 |
+
"eos_token_id": 151643,
|
4 |
+
"max_new_tokens": 2048,
|
5 |
+
"transformers_version": "4.43.4"
|
6 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af82783ecff5472cddbab0ec2ab26cd348687565307e68eb339172e2b1cd1ab1
|
3 |
+
size 3554214752
|
special_tokens_map.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>"
|
5 |
+
],
|
6 |
+
"eos_token": {
|
7 |
+
"content": "<|im_end|>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false
|
12 |
+
},
|
13 |
+
"pad_token": {
|
14 |
+
"content": "<|endoftext|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false
|
19 |
+
}
|
20 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
}
|
28 |
+
},
|
29 |
+
"additional_special_tokens": [
|
30 |
+
"<|im_start|>",
|
31 |
+
"<|im_end|>"
|
32 |
+
],
|
33 |
+
"bos_token": null,
|
34 |
+
"chat_template": "{% set system_message = 'You are a helpful assistant.' %}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ '<|im_start|>system\n' + system_message + '<|im_end|>\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\n' + content + '<|im_end|>\n<|im_start|>assistant\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
|
35 |
+
"clean_up_tokenization_spaces": false,
|
36 |
+
"eos_token": "<|im_end|>",
|
37 |
+
"errors": "replace",
|
38 |
+
"model_max_length": 32768,
|
39 |
+
"pad_token": "<|endoftext|>",
|
40 |
+
"padding_side": "right",
|
41 |
+
"split_special_tokens": false,
|
42 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
43 |
+
"unk_token": null
|
44 |
+
}
|
train_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 0.9999310202110782,
|
3 |
+
"total_flos": 1.7598525261526073e+19,
|
4 |
+
"train_loss": 0.8787032284220849,
|
5 |
+
"train_runtime": 87969.9071,
|
6 |
+
"train_samples_per_second": 2.637,
|
7 |
+
"train_steps_per_second": 0.005
|
8 |
+
}
|
trainer_log.jsonl
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{"current_steps": 10, "total_steps": 453, "loss": 1.0047, "learning_rate": 1.0869565217391305e-05, "epoch": 0.022073532454990687, "percentage": 2.21, "elapsed_time": "0:36:23", "remaining_time": "1 day, 2:51:49", "throughput": "0.00", "total_tokens": 0}
|
2 |
+
{"current_steps": 20, "total_steps": 453, "loss": 0.9551, "learning_rate": 2.173913043478261e-05, "epoch": 0.044147064909981375, "percentage": 4.42, "elapsed_time": "1:08:30", "remaining_time": "1 day, 0:43:03", "throughput": "0.00", "total_tokens": 0}
|
3 |
+
{"current_steps": 30, "total_steps": 453, "loss": 0.9291, "learning_rate": 3.260869565217392e-05, "epoch": 0.06622059736497206, "percentage": 6.62, "elapsed_time": "1:40:40", "remaining_time": "23:39:34", "throughput": "0.00", "total_tokens": 0}
|
4 |
+
{"current_steps": 40, "total_steps": 453, "loss": 0.914, "learning_rate": 4.347826086956522e-05, "epoch": 0.08829412981996275, "percentage": 8.83, "elapsed_time": "2:12:48", "remaining_time": "22:51:17", "throughput": "0.00", "total_tokens": 0}
|
5 |
+
{"current_steps": 50, "total_steps": 453, "loss": 0.9187, "learning_rate": 4.9988084660498037e-05, "epoch": 0.11036766227495344, "percentage": 11.04, "elapsed_time": "2:44:54", "remaining_time": "22:09:11", "throughput": "0.00", "total_tokens": 0}
|
6 |
+
{"current_steps": 60, "total_steps": 453, "loss": 0.9183, "learning_rate": 4.985416749673074e-05, "epoch": 0.13244119472994412, "percentage": 13.25, "elapsed_time": "3:17:01", "remaining_time": "21:30:30", "throughput": "0.00", "total_tokens": 0}
|
7 |
+
{"current_steps": 70, "total_steps": 453, "loss": 0.9048, "learning_rate": 4.957223915853709e-05, "epoch": 0.1545147271849348, "percentage": 15.45, "elapsed_time": "3:49:07", "remaining_time": "20:53:36", "throughput": "0.00", "total_tokens": 0}
|
8 |
+
{"current_steps": 80, "total_steps": 453, "loss": 0.8957, "learning_rate": 4.9143978581429445e-05, "epoch": 0.1765882596399255, "percentage": 17.66, "elapsed_time": "4:21:14", "remaining_time": "20:18:01", "throughput": "0.00", "total_tokens": 0}
|
9 |
+
{"current_steps": 90, "total_steps": 453, "loss": 0.8952, "learning_rate": 4.857193613652711e-05, "epoch": 0.1986617920949162, "percentage": 19.87, "elapsed_time": "4:53:20", "remaining_time": "19:43:06", "throughput": "0.00", "total_tokens": 0}
|
10 |
+
{"current_steps": 100, "total_steps": 453, "loss": 0.8866, "learning_rate": 4.78595184426236e-05, "epoch": 0.22073532454990688, "percentage": 22.08, "elapsed_time": "5:25:26", "remaining_time": "19:08:48", "throughput": "0.00", "total_tokens": 0}
|
11 |
+
{"current_steps": 110, "total_steps": 453, "loss": 0.8895, "learning_rate": 4.7010968079140294e-05, "epoch": 0.24280885700489757, "percentage": 24.28, "elapsed_time": "5:57:33", "remaining_time": "18:34:54", "throughput": "0.00", "total_tokens": 0}
|
12 |
+
{"current_steps": 120, "total_steps": 453, "loss": 0.8815, "learning_rate": 4.6031338320779534e-05, "epoch": 0.26488238945988823, "percentage": 26.49, "elapsed_time": "6:29:39", "remaining_time": "18:01:17", "throughput": "0.00", "total_tokens": 0}
|
13 |
+
{"current_steps": 130, "total_steps": 453, "loss": 0.8808, "learning_rate": 4.492646304433711e-05, "epoch": 0.28695592191487895, "percentage": 28.7, "elapsed_time": "7:01:47", "remaining_time": "17:27:58", "throughput": "0.00", "total_tokens": 0}
|
14 |
+
{"current_steps": 140, "total_steps": 453, "loss": 0.8716, "learning_rate": 4.3702921986884574e-05, "epoch": 0.3090294543698696, "percentage": 30.91, "elapsed_time": "7:33:53", "remaining_time": "16:54:46", "throughput": "0.00", "total_tokens": 0}
|
15 |
+
{"current_steps": 150, "total_steps": 453, "loss": 0.8823, "learning_rate": 4.236800156221536e-05, "epoch": 0.33110298682486033, "percentage": 33.11, "elapsed_time": "8:06:09", "remaining_time": "16:22:02", "throughput": "0.00", "total_tokens": 0}
|
16 |
+
{"current_steps": 160, "total_steps": 453, "loss": 0.8725, "learning_rate": 4.092965146890002e-05, "epoch": 0.353176519279851, "percentage": 35.32, "elapsed_time": "8:38:26", "remaining_time": "15:49:23", "throughput": "0.00", "total_tokens": 0}
|
17 |
+
{"current_steps": 170, "total_steps": 453, "loss": 0.8684, "learning_rate": 3.9396437348357684e-05, "epoch": 0.3752500517348417, "percentage": 37.53, "elapsed_time": "9:10:44", "remaining_time": "15:16:49", "throughput": "0.00", "total_tokens": 0}
|
18 |
+
{"current_steps": 180, "total_steps": 453, "loss": 0.89, "learning_rate": 3.777748977487366e-05, "epoch": 0.3973235841898324, "percentage": 39.74, "elapsed_time": "9:43:00", "remaining_time": "14:44:13", "throughput": "0.00", "total_tokens": 0}
|
19 |
+
{"current_steps": 190, "total_steps": 453, "loss": 0.8706, "learning_rate": 3.608244988133713e-05, "epoch": 0.41939711664482304, "percentage": 41.94, "elapsed_time": "10:15:16", "remaining_time": "14:11:40", "throughput": "0.00", "total_tokens": 0}
|
20 |
+
{"current_steps": 200, "total_steps": 453, "loss": 0.872, "learning_rate": 3.432141194450772e-05, "epoch": 0.44147064909981376, "percentage": 44.15, "elapsed_time": "10:47:32", "remaining_time": "13:39:08", "throughput": "0.00", "total_tokens": 0}
|
21 |
+
{"current_steps": 210, "total_steps": 453, "loss": 0.8682, "learning_rate": 3.2504863271726286e-05, "epoch": 0.4635441815548044, "percentage": 46.36, "elapsed_time": "11:19:47", "remaining_time": "13:06:37", "throughput": "0.00", "total_tokens": 0}
|
22 |
+
{"current_steps": 220, "total_steps": 453, "loss": 0.8794, "learning_rate": 3.064362174705578e-05, "epoch": 0.48561771400979514, "percentage": 48.57, "elapsed_time": "11:51:52", "remaining_time": "12:33:56", "throughput": "0.00", "total_tokens": 0}
|
23 |
+
{"current_steps": 230, "total_steps": 453, "loss": 0.8716, "learning_rate": 2.8748771408776466e-05, "epoch": 0.5076912464647858, "percentage": 50.77, "elapsed_time": "12:23:42", "remaining_time": "12:01:04", "throughput": "0.00", "total_tokens": 0}
|
24 |
+
{"current_steps": 240, "total_steps": 453, "loss": 0.8807, "learning_rate": 2.683159644188339e-05, "epoch": 0.5297647789197765, "percentage": 52.98, "elapsed_time": "12:56:15", "remaining_time": "11:28:56", "throughput": "0.00", "total_tokens": 0}
|
25 |
+
{"current_steps": 250, "total_steps": 453, "loss": 0.8701, "learning_rate": 2.4903513978673077e-05, "epoch": 0.5518383113747672, "percentage": 55.19, "elapsed_time": "13:31:28", "remaining_time": "10:58:55", "throughput": "0.00", "total_tokens": 0}
|
26 |
+
{"current_steps": 260, "total_steps": 453, "loss": 0.8632, "learning_rate": 2.2976006107604482e-05, "epoch": 0.5739118438297579, "percentage": 57.4, "elapsed_time": "14:05:07", "remaining_time": "10:27:20", "throughput": "0.00", "total_tokens": 0}
|
27 |
+
{"current_steps": 270, "total_steps": 453, "loss": 0.8761, "learning_rate": 2.1060551495333818e-05, "epoch": 0.5959853762847486, "percentage": 59.6, "elapsed_time": "14:38:13", "remaining_time": "9:55:14", "throughput": "0.00", "total_tokens": 0}
|
28 |
+
{"current_steps": 280, "total_steps": 453, "loss": 0.8611, "learning_rate": 1.9168557029126963e-05, "epoch": 0.6180589087397392, "percentage": 61.81, "elapsed_time": "15:11:17", "remaining_time": "9:23:02", "throughput": "0.00", "total_tokens": 0}
|
29 |
+
{"current_steps": 290, "total_steps": 453, "loss": 0.8616, "learning_rate": 1.7311289886731408e-05, "epoch": 0.6401324411947299, "percentage": 64.02, "elapsed_time": "15:44:24", "remaining_time": "8:50:49", "throughput": "0.00", "total_tokens": 0}
|
30 |
+
{"current_steps": 300, "total_steps": 453, "loss": 0.8603, "learning_rate": 1.549981043824425e-05, "epoch": 0.6622059736497207, "percentage": 66.23, "elapsed_time": "16:17:12", "remaining_time": "8:18:22", "throughput": "0.00", "total_tokens": 0}
|
31 |
+
{"current_steps": 310, "total_steps": 453, "loss": 0.8608, "learning_rate": 1.3744906379558165e-05, "epoch": 0.6842795061047113, "percentage": 68.43, "elapsed_time": "16:49:23", "remaining_time": "7:45:37", "throughput": "0.00", "total_tokens": 0}
|
32 |
+
{"current_steps": 320, "total_steps": 453, "loss": 0.8703, "learning_rate": 1.2057028489632682e-05, "epoch": 0.706353038559702, "percentage": 70.64, "elapsed_time": "17:21:22", "remaining_time": "7:12:49", "throughput": "0.00", "total_tokens": 0}
|
33 |
+
{"current_steps": 330, "total_steps": 453, "loss": 0.8518, "learning_rate": 1.0446228394168356e-05, "epoch": 0.7284265710146927, "percentage": 72.85, "elapsed_time": "17:53:16", "remaining_time": "6:40:02", "throughput": "0.00", "total_tokens": 0}
|
34 |
+
{"current_steps": 340, "total_steps": 453, "loss": 0.8609, "learning_rate": 8.922098706312548e-06, "epoch": 0.7505001034696834, "percentage": 75.06, "elapsed_time": "18:25:12", "remaining_time": "6:07:19", "throughput": "0.00", "total_tokens": 0}
|
35 |
+
{"current_steps": 350, "total_steps": 453, "loss": 0.8504, "learning_rate": 7.493715900870027e-06, "epoch": 0.7725736359246741, "percentage": 77.26, "elapsed_time": "18:57:05", "remaining_time": "5:34:37", "throughput": "0.00", "total_tokens": 0}
|
36 |
+
{"current_steps": 360, "total_steps": 453, "loss": 0.8614, "learning_rate": 6.169586262213081e-06, "epoch": 0.7946471683796648, "percentage": 79.47, "elapsed_time": "19:28:49", "remaining_time": "5:01:56", "throughput": "0.00", "total_tokens": 0}
|
37 |
+
{"current_steps": 370, "total_steps": 453, "loss": 0.8574, "learning_rate": 4.957595227781395e-06, "epoch": 0.8167207008346554, "percentage": 81.68, "elapsed_time": "20:00:39", "remaining_time": "4:29:20", "throughput": "0.00", "total_tokens": 0}
|
38 |
+
{"current_steps": 380, "total_steps": 453, "loss": 0.8501, "learning_rate": 3.864960428840375e-06, "epoch": 0.8387942332896461, "percentage": 83.89, "elapsed_time": "20:32:39", "remaining_time": "3:56:48", "throughput": "0.00", "total_tokens": 0}
|
39 |
+
{"current_steps": 390, "total_steps": 453, "loss": 0.85, "learning_rate": 2.8981887081491576e-06, "epoch": 0.8608677657446369, "percentage": 86.09, "elapsed_time": "21:04:29", "remaining_time": "3:24:15", "throughput": "0.00", "total_tokens": 0}
|
40 |
+
{"current_steps": 400, "total_steps": 453, "loss": 0.8599, "learning_rate": 2.0630373705058407e-06, "epoch": 0.8829412981996275, "percentage": 88.3, "elapsed_time": "21:36:17", "remaining_time": "2:51:45", "throughput": "0.00", "total_tokens": 0}
|
41 |
+
{"current_steps": 410, "total_steps": 453, "loss": 0.8627, "learning_rate": 1.3644798969302403e-06, "epoch": 0.9050148306546182, "percentage": 90.51, "elapsed_time": "22:08:05", "remaining_time": "2:19:17", "throughput": "0.00", "total_tokens": 0}
|
42 |
+
{"current_steps": 420, "total_steps": 453, "loss": 0.8567, "learning_rate": 8.066763266625282e-07, "epoch": 0.9270883631096088, "percentage": 92.72, "elapsed_time": "22:39:54", "remaining_time": "1:46:51", "throughput": "0.00", "total_tokens": 0}
|
43 |
+
{"current_steps": 430, "total_steps": 453, "loss": 0.8484, "learning_rate": 3.929484833584546e-07, "epoch": 0.9491618955645996, "percentage": 94.92, "elapsed_time": "23:11:46", "remaining_time": "1:14:26", "throughput": "0.00", "total_tokens": 0}
|
44 |
+
{"current_steps": 440, "total_steps": 453, "loss": 0.8591, "learning_rate": 1.2576019301373532e-07, "epoch": 0.9712354280195903, "percentage": 97.13, "elapsed_time": "23:43:34", "remaining_time": "0:42:03", "throughput": "0.00", "total_tokens": 0}
|
45 |
+
{"current_steps": 450, "total_steps": 453, "loss": 0.8493, "learning_rate": 6.702611423550775e-09, "epoch": 0.993308960474581, "percentage": 99.34, "elapsed_time": "1 day, 0:15:27", "remaining_time": "0:09:42", "throughput": "0.00", "total_tokens": 0}
|
46 |
+
{"current_steps": 453, "total_steps": 453, "epoch": 0.9999310202110782, "percentage": 100.0, "elapsed_time": "1 day, 0:26:09", "remaining_time": "0:00:00", "throughput": "0.00", "total_tokens": 0}
|
trainer_state.json
ADDED
@@ -0,0 +1,357 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": null,
|
3 |
+
"best_model_checkpoint": null,
|
4 |
+
"epoch": 0.9999310202110782,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 453,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.022073532454990687,
|
13 |
+
"grad_norm": 1.5836448669433594,
|
14 |
+
"learning_rate": 1.0869565217391305e-05,
|
15 |
+
"loss": 1.0047,
|
16 |
+
"step": 10
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 0.044147064909981375,
|
20 |
+
"grad_norm": 0.4156560003757477,
|
21 |
+
"learning_rate": 2.173913043478261e-05,
|
22 |
+
"loss": 0.9551,
|
23 |
+
"step": 20
|
24 |
+
},
|
25 |
+
{
|
26 |
+
"epoch": 0.06622059736497206,
|
27 |
+
"grad_norm": 0.33475786447525024,
|
28 |
+
"learning_rate": 3.260869565217392e-05,
|
29 |
+
"loss": 0.9291,
|
30 |
+
"step": 30
|
31 |
+
},
|
32 |
+
{
|
33 |
+
"epoch": 0.08829412981996275,
|
34 |
+
"grad_norm": 0.721441924571991,
|
35 |
+
"learning_rate": 4.347826086956522e-05,
|
36 |
+
"loss": 0.914,
|
37 |
+
"step": 40
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 0.11036766227495344,
|
41 |
+
"grad_norm": 0.2514866292476654,
|
42 |
+
"learning_rate": 4.9988084660498037e-05,
|
43 |
+
"loss": 0.9187,
|
44 |
+
"step": 50
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 0.13244119472994412,
|
48 |
+
"grad_norm": 0.21546441316604614,
|
49 |
+
"learning_rate": 4.985416749673074e-05,
|
50 |
+
"loss": 0.9183,
|
51 |
+
"step": 60
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"epoch": 0.1545147271849348,
|
55 |
+
"grad_norm": 0.2580913007259369,
|
56 |
+
"learning_rate": 4.957223915853709e-05,
|
57 |
+
"loss": 0.9048,
|
58 |
+
"step": 70
|
59 |
+
},
|
60 |
+
{
|
61 |
+
"epoch": 0.1765882596399255,
|
62 |
+
"grad_norm": 0.24421757459640503,
|
63 |
+
"learning_rate": 4.9143978581429445e-05,
|
64 |
+
"loss": 0.8957,
|
65 |
+
"step": 80
|
66 |
+
},
|
67 |
+
{
|
68 |
+
"epoch": 0.1986617920949162,
|
69 |
+
"grad_norm": 0.23899255692958832,
|
70 |
+
"learning_rate": 4.857193613652711e-05,
|
71 |
+
"loss": 0.8952,
|
72 |
+
"step": 90
|
73 |
+
},
|
74 |
+
{
|
75 |
+
"epoch": 0.22073532454990688,
|
76 |
+
"grad_norm": 0.25667503476142883,
|
77 |
+
"learning_rate": 4.78595184426236e-05,
|
78 |
+
"loss": 0.8866,
|
79 |
+
"step": 100
|
80 |
+
},
|
81 |
+
{
|
82 |
+
"epoch": 0.24280885700489757,
|
83 |
+
"grad_norm": 0.3306167423725128,
|
84 |
+
"learning_rate": 4.7010968079140294e-05,
|
85 |
+
"loss": 0.8895,
|
86 |
+
"step": 110
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 0.26488238945988823,
|
90 |
+
"grad_norm": 0.2200348675251007,
|
91 |
+
"learning_rate": 4.6031338320779534e-05,
|
92 |
+
"loss": 0.8815,
|
93 |
+
"step": 120
|
94 |
+
},
|
95 |
+
{
|
96 |
+
"epoch": 0.28695592191487895,
|
97 |
+
"grad_norm": 0.28366169333457947,
|
98 |
+
"learning_rate": 4.492646304433711e-05,
|
99 |
+
"loss": 0.8808,
|
100 |
+
"step": 130
|
101 |
+
},
|
102 |
+
{
|
103 |
+
"epoch": 0.3090294543698696,
|
104 |
+
"grad_norm": 0.22618895769119263,
|
105 |
+
"learning_rate": 4.3702921986884574e-05,
|
106 |
+
"loss": 0.8716,
|
107 |
+
"step": 140
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"epoch": 0.33110298682486033,
|
111 |
+
"grad_norm": 0.27280193567276,
|
112 |
+
"learning_rate": 4.236800156221536e-05,
|
113 |
+
"loss": 0.8823,
|
114 |
+
"step": 150
|
115 |
+
},
|
116 |
+
{
|
117 |
+
"epoch": 0.353176519279851,
|
118 |
+
"grad_norm": 0.19905485212802887,
|
119 |
+
"learning_rate": 4.092965146890002e-05,
|
120 |
+
"loss": 0.8725,
|
121 |
+
"step": 160
|
122 |
+
},
|
123 |
+
{
|
124 |
+
"epoch": 0.3752500517348417,
|
125 |
+
"grad_norm": 0.3659760355949402,
|
126 |
+
"learning_rate": 3.9396437348357684e-05,
|
127 |
+
"loss": 0.8684,
|
128 |
+
"step": 170
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"epoch": 0.3973235841898324,
|
132 |
+
"grad_norm": 0.2508715093135834,
|
133 |
+
"learning_rate": 3.777748977487366e-05,
|
134 |
+
"loss": 0.89,
|
135 |
+
"step": 180
|
136 |
+
},
|
137 |
+
{
|
138 |
+
"epoch": 0.41939711664482304,
|
139 |
+
"grad_norm": 0.24242979288101196,
|
140 |
+
"learning_rate": 3.608244988133713e-05,
|
141 |
+
"loss": 0.8706,
|
142 |
+
"step": 190
|
143 |
+
},
|
144 |
+
{
|
145 |
+
"epoch": 0.44147064909981376,
|
146 |
+
"grad_norm": 0.191656693816185,
|
147 |
+
"learning_rate": 3.432141194450772e-05,
|
148 |
+
"loss": 0.872,
|
149 |
+
"step": 200
|
150 |
+
},
|
151 |
+
{
|
152 |
+
"epoch": 0.4635441815548044,
|
153 |
+
"grad_norm": 0.2087477147579193,
|
154 |
+
"learning_rate": 3.2504863271726286e-05,
|
155 |
+
"loss": 0.8682,
|
156 |
+
"step": 210
|
157 |
+
},
|
158 |
+
{
|
159 |
+
"epoch": 0.48561771400979514,
|
160 |
+
"grad_norm": 0.2208724468946457,
|
161 |
+
"learning_rate": 3.064362174705578e-05,
|
162 |
+
"loss": 0.8794,
|
163 |
+
"step": 220
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"epoch": 0.5076912464647858,
|
167 |
+
"grad_norm": 0.19049623608589172,
|
168 |
+
"learning_rate": 2.8748771408776466e-05,
|
169 |
+
"loss": 0.8716,
|
170 |
+
"step": 230
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"epoch": 0.5297647789197765,
|
174 |
+
"grad_norm": 0.18646596372127533,
|
175 |
+
"learning_rate": 2.683159644188339e-05,
|
176 |
+
"loss": 0.8807,
|
177 |
+
"step": 240
|
178 |
+
},
|
179 |
+
{
|
180 |
+
"epoch": 0.5518383113747672,
|
181 |
+
"grad_norm": 0.17940568923950195,
|
182 |
+
"learning_rate": 2.4903513978673077e-05,
|
183 |
+
"loss": 0.8701,
|
184 |
+
"step": 250
|
185 |
+
},
|
186 |
+
{
|
187 |
+
"epoch": 0.5739118438297579,
|
188 |
+
"grad_norm": 0.17862474918365479,
|
189 |
+
"learning_rate": 2.2976006107604482e-05,
|
190 |
+
"loss": 0.8632,
|
191 |
+
"step": 260
|
192 |
+
},
|
193 |
+
{
|
194 |
+
"epoch": 0.5959853762847486,
|
195 |
+
"grad_norm": 0.16970194876194,
|
196 |
+
"learning_rate": 2.1060551495333818e-05,
|
197 |
+
"loss": 0.8761,
|
198 |
+
"step": 270
|
199 |
+
},
|
200 |
+
{
|
201 |
+
"epoch": 0.6180589087397392,
|
202 |
+
"grad_norm": 0.1763421595096588,
|
203 |
+
"learning_rate": 1.9168557029126963e-05,
|
204 |
+
"loss": 0.8611,
|
205 |
+
"step": 280
|
206 |
+
},
|
207 |
+
{
|
208 |
+
"epoch": 0.6401324411947299,
|
209 |
+
"grad_norm": 0.16400456428527832,
|
210 |
+
"learning_rate": 1.7311289886731408e-05,
|
211 |
+
"loss": 0.8616,
|
212 |
+
"step": 290
|
213 |
+
},
|
214 |
+
{
|
215 |
+
"epoch": 0.6622059736497207,
|
216 |
+
"grad_norm": 0.17227979004383087,
|
217 |
+
"learning_rate": 1.549981043824425e-05,
|
218 |
+
"loss": 0.8603,
|
219 |
+
"step": 300
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"epoch": 0.6842795061047113,
|
223 |
+
"grad_norm": 0.15701688826084137,
|
224 |
+
"learning_rate": 1.3744906379558165e-05,
|
225 |
+
"loss": 0.8608,
|
226 |
+
"step": 310
|
227 |
+
},
|
228 |
+
{
|
229 |
+
"epoch": 0.706353038559702,
|
230 |
+
"grad_norm": 0.15198074281215668,
|
231 |
+
"learning_rate": 1.2057028489632682e-05,
|
232 |
+
"loss": 0.8703,
|
233 |
+
"step": 320
|
234 |
+
},
|
235 |
+
{
|
236 |
+
"epoch": 0.7284265710146927,
|
237 |
+
"grad_norm": 0.14645570516586304,
|
238 |
+
"learning_rate": 1.0446228394168356e-05,
|
239 |
+
"loss": 0.8518,
|
240 |
+
"step": 330
|
241 |
+
},
|
242 |
+
{
|
243 |
+
"epoch": 0.7505001034696834,
|
244 |
+
"grad_norm": 0.14851906895637512,
|
245 |
+
"learning_rate": 8.922098706312548e-06,
|
246 |
+
"loss": 0.8609,
|
247 |
+
"step": 340
|
248 |
+
},
|
249 |
+
{
|
250 |
+
"epoch": 0.7725736359246741,
|
251 |
+
"grad_norm": 0.14792883396148682,
|
252 |
+
"learning_rate": 7.493715900870027e-06,
|
253 |
+
"loss": 0.8504,
|
254 |
+
"step": 350
|
255 |
+
},
|
256 |
+
{
|
257 |
+
"epoch": 0.7946471683796648,
|
258 |
+
"grad_norm": 0.14906203746795654,
|
259 |
+
"learning_rate": 6.169586262213081e-06,
|
260 |
+
"loss": 0.8614,
|
261 |
+
"step": 360
|
262 |
+
},
|
263 |
+
{
|
264 |
+
"epoch": 0.8167207008346554,
|
265 |
+
"grad_norm": 0.13984321057796478,
|
266 |
+
"learning_rate": 4.957595227781395e-06,
|
267 |
+
"loss": 0.8574,
|
268 |
+
"step": 370
|
269 |
+
},
|
270 |
+
{
|
271 |
+
"epoch": 0.8387942332896461,
|
272 |
+
"grad_norm": 0.13922333717346191,
|
273 |
+
"learning_rate": 3.864960428840375e-06,
|
274 |
+
"loss": 0.8501,
|
275 |
+
"step": 380
|
276 |
+
},
|
277 |
+
{
|
278 |
+
"epoch": 0.8608677657446369,
|
279 |
+
"grad_norm": 0.14093540608882904,
|
280 |
+
"learning_rate": 2.8981887081491576e-06,
|
281 |
+
"loss": 0.85,
|
282 |
+
"step": 390
|
283 |
+
},
|
284 |
+
{
|
285 |
+
"epoch": 0.8829412981996275,
|
286 |
+
"grad_norm": 0.13516093790531158,
|
287 |
+
"learning_rate": 2.0630373705058407e-06,
|
288 |
+
"loss": 0.8599,
|
289 |
+
"step": 400
|
290 |
+
},
|
291 |
+
{
|
292 |
+
"epoch": 0.9050148306546182,
|
293 |
+
"grad_norm": 0.1319512128829956,
|
294 |
+
"learning_rate": 1.3644798969302403e-06,
|
295 |
+
"loss": 0.8627,
|
296 |
+
"step": 410
|
297 |
+
},
|
298 |
+
{
|
299 |
+
"epoch": 0.9270883631096088,
|
300 |
+
"grad_norm": 0.13131946325302124,
|
301 |
+
"learning_rate": 8.066763266625282e-07,
|
302 |
+
"loss": 0.8567,
|
303 |
+
"step": 420
|
304 |
+
},
|
305 |
+
{
|
306 |
+
"epoch": 0.9491618955645996,
|
307 |
+
"grad_norm": 0.13254371285438538,
|
308 |
+
"learning_rate": 3.929484833584546e-07,
|
309 |
+
"loss": 0.8484,
|
310 |
+
"step": 430
|
311 |
+
},
|
312 |
+
{
|
313 |
+
"epoch": 0.9712354280195903,
|
314 |
+
"grad_norm": 0.1311715841293335,
|
315 |
+
"learning_rate": 1.2576019301373532e-07,
|
316 |
+
"loss": 0.8591,
|
317 |
+
"step": 440
|
318 |
+
},
|
319 |
+
{
|
320 |
+
"epoch": 0.993308960474581,
|
321 |
+
"grad_norm": 0.13352562487125397,
|
322 |
+
"learning_rate": 6.702611423550775e-09,
|
323 |
+
"loss": 0.8493,
|
324 |
+
"step": 450
|
325 |
+
},
|
326 |
+
{
|
327 |
+
"epoch": 0.9999310202110782,
|
328 |
+
"step": 453,
|
329 |
+
"total_flos": 1.7598525261526073e+19,
|
330 |
+
"train_loss": 0.8787032284220849,
|
331 |
+
"train_runtime": 87969.9071,
|
332 |
+
"train_samples_per_second": 2.637,
|
333 |
+
"train_steps_per_second": 0.005
|
334 |
+
}
|
335 |
+
],
|
336 |
+
"logging_steps": 10,
|
337 |
+
"max_steps": 453,
|
338 |
+
"num_input_tokens_seen": 0,
|
339 |
+
"num_train_epochs": 1,
|
340 |
+
"save_steps": 500,
|
341 |
+
"stateful_callbacks": {
|
342 |
+
"TrainerControl": {
|
343 |
+
"args": {
|
344 |
+
"should_epoch_stop": false,
|
345 |
+
"should_evaluate": false,
|
346 |
+
"should_log": false,
|
347 |
+
"should_save": true,
|
348 |
+
"should_training_stop": true
|
349 |
+
},
|
350 |
+
"attributes": {}
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"total_flos": 1.7598525261526073e+19,
|
354 |
+
"train_batch_size": 2,
|
355 |
+
"trial_name": null,
|
356 |
+
"trial_params": null
|
357 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7f781cb5d1232a49b441084c9a80a267386ad0a75c168898f8838a171a1cd61
|
3 |
+
size 6584
|
training_loss.png
ADDED
![]() |
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|