RikkiXu commited on
Commit
e87862d
1 Parent(s): 8f503d9

Model save

Browse files
README.md CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # zephyr-7b-dpo-full
15
 
16
- This model was trained from scratch on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -32,7 +32,7 @@ More information needed
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - learning_rate: 5e-09
36
  - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
@@ -44,7 +44,7 @@ The following hyperparameters were used during training:
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
47
- - num_epochs: 1
48
 
49
  ### Training results
50
 
@@ -54,5 +54,5 @@ The following hyperparameters were used during training:
54
 
55
  - Transformers 4.39.3
56
  - Pytorch 2.1.2+cu118
57
- - Datasets 2.19.1
58
  - Tokenizers 0.15.2
 
13
 
14
  # zephyr-7b-dpo-full
15
 
16
+ This model was trained from scratch on the None dataset.
17
 
18
  ## Model description
19
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 1e-08
36
  - train_batch_size: 8
37
  - eval_batch_size: 8
38
  - seed: 42
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
47
+ - num_epochs: 2
48
 
49
  ### Training results
50
 
 
54
 
55
  - Transformers 4.39.3
56
  - Pytorch 2.1.2+cu118
57
+ - Datasets 2.16.1
58
  - Tokenizers 0.15.2
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.0,
3
- "train_loss": 0.7162073644002279,
4
- "train_runtime": 4512.4874,
5
  "train_samples": 38445,
6
- "train_samples_per_second": 8.52,
7
- "train_steps_per_second": 0.033
8
  }
 
1
  {
2
+ "epoch": 2.0,
3
+ "train_loss": 0.6411590957641602,
4
+ "train_runtime": 8942.072,
5
  "train_samples": 38445,
6
+ "train_samples_per_second": 8.599,
7
+ "train_steps_per_second": 0.034
8
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c198eb9df08ec1da45189177d9165731677bcf8abcb6d6938bc58bf75d03977d
3
  size 4943178720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e95be6f8c0aa9a00e2dec7409b07bb3971685d185f076a462b8f78d03a1941fd
3
  size 4943178720
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57087d3701236ddd033204862662fca1d328a1d40449da21fe52584fb50fefff
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65d756c6023aac10ba06a637406ed4d2e2f4752c360d7d38f9d57ab38052466e
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d22d542aa9e2e706d1339613cd7b76b3452024d5fd48f29b7543bbc9e61fc003
3
  size 4540532728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4dee0a340ae6ecce1b95adc398b7196eb2db45a4ad1ef0dc042944445a81bb6
3
  size 4540532728
runs/Jun17_01-59-50_n136-129-074/events.out.tfevents.1718560924.n136-129-074.1300229.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa590a072804e4e12986ff8579b5512a4d087ed195a8a7fdab8596761c3285b4
3
- size 26092
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d38c6a12ddad6fd5a37a8c85d17efcf257f35b1cf09dbb5272269fa21ba4220
3
+ size 26446
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.0,
3
- "train_loss": 0.7162073644002279,
4
- "train_runtime": 4512.4874,
5
  "train_samples": 38445,
6
- "train_samples_per_second": 8.52,
7
- "train_steps_per_second": 0.033
8
  }
 
1
  {
2
+ "epoch": 2.0,
3
+ "train_loss": 0.6411590957641602,
4
+ "train_runtime": 8942.072,
5
  "train_samples": 38445,
6
+ "train_samples_per_second": 8.599,
7
+ "train_steps_per_second": 0.034
8
  }
trainer_state.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9983361064891847,
5
  "eval_steps": 500,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
- "grad_norm": 1151.6129333090275,
14
  "learning_rate": 3.333333333333333e-10,
15
  "logits/chosen": -4.106247425079346,
16
  "logits/rejected": -4.200438499450684,
@@ -25,243 +25,468 @@
25
  },
26
  {
27
  "epoch": 0.07,
28
- "grad_norm": 1175.0279383615512,
29
  "learning_rate": 3.3333333333333334e-09,
30
- "logits/chosen": -4.217203617095947,
31
- "logits/rejected": -4.32081413269043,
32
- "logps/chosen": -334.6714172363281,
33
- "logps/rejected": -313.4322509765625,
34
- "loss": 0.7317,
35
- "rewards/accuracies": 0.3923611044883728,
36
- "rewards/chosen": -0.044815655797719955,
37
- "rewards/margins": -0.03612741455435753,
38
- "rewards/rejected": -0.008688241243362427,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.13,
43
- "grad_norm": 1253.9538230101048,
44
- "learning_rate": 4.983095894354858e-09,
45
- "logits/chosen": -4.266427040100098,
46
- "logits/rejected": -4.4187798500061035,
47
- "logps/chosen": -313.9143371582031,
48
- "logps/rejected": -288.782470703125,
49
- "loss": 0.732,
50
- "rewards/accuracies": 0.4781250059604645,
51
- "rewards/chosen": -0.000932177877984941,
52
- "rewards/margins": -0.012739461846649647,
53
- "rewards/rejected": 0.011807283386588097,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.2,
58
- "grad_norm": 1147.308654200848,
59
- "learning_rate": 4.849231551964771e-09,
60
- "logits/chosen": -4.301184177398682,
61
- "logits/rejected": -4.36545991897583,
62
- "logps/chosen": -308.3140563964844,
63
- "logps/rejected": -285.6829833984375,
64
- "loss": 0.7352,
65
- "rewards/accuracies": 0.46562498807907104,
66
- "rewards/chosen": -0.018803134560585022,
67
- "rewards/margins": 0.008298242464661598,
68
- "rewards/rejected": -0.02710137702524662,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.27,
73
- "grad_norm": 1152.01480268075,
74
- "learning_rate": 4.588719528532341e-09,
75
- "logits/chosen": -4.186487674713135,
76
- "logits/rejected": -4.270204544067383,
77
- "logps/chosen": -332.467041015625,
78
- "logps/rejected": -307.530517578125,
79
- "loss": 0.722,
80
- "rewards/accuracies": 0.5531250238418579,
81
- "rewards/chosen": 0.012754167430102825,
82
- "rewards/margins": 0.04999501258134842,
83
- "rewards/rejected": -0.03724084421992302,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.33,
88
- "grad_norm": 1161.340877574017,
89
- "learning_rate": 4.215604094671834e-09,
90
- "logits/chosen": -4.197475910186768,
91
- "logits/rejected": -4.366654396057129,
92
- "logps/chosen": -333.69940185546875,
93
- "logps/rejected": -309.05511474609375,
94
- "loss": 0.7263,
95
- "rewards/accuracies": 0.512499988079071,
96
- "rewards/chosen": 0.022939234972000122,
97
- "rewards/margins": -0.01653190515935421,
98
- "rewards/rejected": 0.03947114199399948,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.4,
103
- "grad_norm": 1135.5326964714952,
104
- "learning_rate": 3.7500000000000005e-09,
105
- "logits/chosen": -4.186155796051025,
106
- "logits/rejected": -4.269167900085449,
107
- "logps/chosen": -323.9843444824219,
108
- "logps/rejected": -308.364990234375,
109
- "loss": 0.7133,
110
- "rewards/accuracies": 0.484375,
111
- "rewards/chosen": 0.009113344363868237,
112
- "rewards/margins": -0.005971288774162531,
113
- "rewards/rejected": 0.015084633603692055,
114
  "step": 60
115
  },
116
  {
117
  "epoch": 0.47,
118
- "grad_norm": 1170.598014061488,
119
- "learning_rate": 3.2170080817777257e-09,
120
- "logits/chosen": -4.14601469039917,
121
- "logits/rejected": -4.301178932189941,
122
- "logps/chosen": -344.8753967285156,
123
- "logps/rejected": -314.6096496582031,
124
- "loss": 0.7182,
125
- "rewards/accuracies": 0.581250011920929,
126
- "rewards/chosen": 0.08767497539520264,
127
- "rewards/margins": 0.08200599253177643,
128
- "rewards/rejected": 0.005668987520039082,
129
  "step": 70
130
  },
131
  {
132
  "epoch": 0.53,
133
- "grad_norm": 1088.822875661876,
134
- "learning_rate": 2.6453620722761897e-09,
135
- "logits/chosen": -4.220850944519043,
136
- "logits/rejected": -4.3867106437683105,
137
- "logps/chosen": -334.7141418457031,
138
- "logps/rejected": -305.6958923339844,
139
- "loss": 0.7175,
140
- "rewards/accuracies": 0.515625,
141
- "rewards/chosen": 0.06255482137203217,
142
- "rewards/margins": 0.04721928387880325,
143
- "rewards/rejected": 0.015335534699261189,
144
  "step": 80
145
  },
146
  {
147
  "epoch": 0.6,
148
- "grad_norm": 1120.1472438016,
149
- "learning_rate": 2.0658795558326744e-09,
150
- "logits/chosen": -4.226521968841553,
151
- "logits/rejected": -4.347161293029785,
152
- "logps/chosen": -327.5687561035156,
153
- "logps/rejected": -308.19744873046875,
154
- "loss": 0.7128,
155
- "rewards/accuracies": 0.518750011920929,
156
- "rewards/chosen": 0.03935312479734421,
157
- "rewards/margins": 0.012340927496552467,
158
- "rewards/rejected": 0.02701219543814659,
159
  "step": 90
160
  },
161
  {
162
  "epoch": 0.67,
163
- "grad_norm": 1147.49194388717,
164
- "learning_rate": 1.5098005849021078e-09,
165
- "logits/chosen": -4.208071708679199,
166
- "logits/rejected": -4.417771339416504,
167
- "logps/chosen": -320.9470520019531,
168
- "logps/rejected": -290.018310546875,
169
- "loss": 0.712,
170
- "rewards/accuracies": 0.5218750238418579,
171
- "rewards/chosen": 0.060337893664836884,
172
- "rewards/margins": 0.03545590117573738,
173
- "rewards/rejected": 0.0248819962143898,
174
  "step": 100
175
  },
176
  {
177
  "epoch": 0.73,
178
- "grad_norm": 1131.801124941311,
179
- "learning_rate": 1.0071035207430352e-09,
180
- "logits/chosen": -4.222798824310303,
181
- "logits/rejected": -4.275721549987793,
182
- "logps/chosen": -315.00030517578125,
183
- "logps/rejected": -303.7385559082031,
184
- "loss": 0.7022,
185
- "rewards/accuracies": 0.543749988079071,
186
- "rewards/chosen": 0.08098876476287842,
187
- "rewards/margins": 0.037886131554841995,
188
- "rewards/rejected": 0.04310264065861702,
189
  "step": 110
190
  },
191
  {
192
  "epoch": 0.8,
193
- "grad_norm": 1201.8232728734756,
194
- "learning_rate": 5.848888922025553e-10,
195
- "logits/chosen": -4.147335052490234,
196
- "logits/rejected": -4.3030242919921875,
197
- "logps/chosen": -339.2781677246094,
198
- "logps/rejected": -313.5749206542969,
199
- "loss": 0.7129,
200
- "rewards/accuracies": 0.543749988079071,
201
- "rewards/chosen": 0.08869560062885284,
202
- "rewards/margins": 0.048744406551122665,
203
- "rewards/rejected": 0.03995119035243988,
204
  "step": 120
205
  },
206
  {
207
  "epoch": 0.87,
208
- "grad_norm": 1182.4206440855232,
209
- "learning_rate": 2.659183991914696e-10,
210
- "logits/chosen": -4.173482418060303,
211
- "logits/rejected": -4.367284297943115,
212
- "logps/chosen": -324.4493713378906,
213
- "logps/rejected": -305.2012634277344,
214
- "loss": 0.7,
215
- "rewards/accuracies": 0.5062500238418579,
216
- "rewards/chosen": 0.08546491712331772,
217
- "rewards/margins": 0.01908993348479271,
218
- "rewards/rejected": 0.06637498736381531,
219
  "step": 130
220
  },
221
  {
222
  "epoch": 0.93,
223
- "grad_norm": 1158.9214495930962,
224
- "learning_rate": 6.738782355044049e-11,
225
- "logits/chosen": -4.290203094482422,
226
- "logits/rejected": -4.35637092590332,
227
- "logps/chosen": -308.0101318359375,
228
- "logps/rejected": -297.37701416015625,
229
- "loss": 0.7082,
230
- "rewards/accuracies": 0.5406249761581421,
231
- "rewards/chosen": 0.12010886520147324,
232
- "rewards/margins": 0.05074785277247429,
233
- "rewards/rejected": 0.06936100870370865,
234
  "step": 140
235
  },
236
  {
237
  "epoch": 1.0,
238
- "grad_norm": 1208.2615961890754,
239
- "learning_rate": 0.0,
240
- "logits/chosen": -4.257304668426514,
241
- "logits/rejected": -4.3310017585754395,
242
- "logps/chosen": -312.7613830566406,
243
- "logps/rejected": -300.3546447753906,
244
- "loss": 0.7025,
245
- "rewards/accuracies": 0.5562499761581421,
246
- "rewards/chosen": 0.10821112245321274,
247
- "rewards/margins": 0.047078292816877365,
248
- "rewards/rejected": 0.06113281846046448,
249
  "step": 150
250
  },
251
  {
252
- "epoch": 1.0,
253
- "step": 150,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
254
  "total_flos": 0.0,
255
- "train_loss": 0.7162073644002279,
256
- "train_runtime": 4512.4874,
257
- "train_samples_per_second": 8.52,
258
- "train_steps_per_second": 0.033
259
  }
260
  ],
261
  "logging_steps": 10,
262
- "max_steps": 150,
263
  "num_input_tokens_seen": 0,
264
- "num_train_epochs": 1,
265
  "save_steps": 100,
266
  "total_flos": 0.0,
267
  "train_batch_size": 8,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.9966722129783694,
5
  "eval_steps": 500,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.01,
13
+ "grad_norm": 1151.6184415193795,
14
  "learning_rate": 3.333333333333333e-10,
15
  "logits/chosen": -4.106247425079346,
16
  "logits/rejected": -4.200438499450684,
 
25
  },
26
  {
27
  "epoch": 0.07,
28
+ "grad_norm": 1153.8530059959064,
29
  "learning_rate": 3.3333333333333334e-09,
30
+ "logits/chosen": -4.217168807983398,
31
+ "logits/rejected": -4.321505069732666,
32
+ "logps/chosen": -334.6739501953125,
33
+ "logps/rejected": -313.41986083984375,
34
+ "loss": 0.7234,
35
+ "rewards/accuracies": 0.3784722089767456,
36
+ "rewards/chosen": -0.04607396200299263,
37
+ "rewards/margins": -0.04357295483350754,
38
+ "rewards/rejected": -0.0025010076351463795,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.13,
43
+ "grad_norm": 1270.1091311718026,
44
+ "learning_rate": 6.666666666666667e-09,
45
+ "logits/chosen": -4.26615571975708,
46
+ "logits/rejected": -4.41886043548584,
47
+ "logps/chosen": -313.93829345703125,
48
+ "logps/rejected": -288.78863525390625,
49
+ "loss": 0.7299,
50
+ "rewards/accuracies": 0.5062500238418579,
51
+ "rewards/chosen": -0.012899210676550865,
52
+ "rewards/margins": -0.021626513451337814,
53
+ "rewards/rejected": 0.00872730277478695,
54
  "step": 20
55
  },
56
  {
57
  "epoch": 0.2,
58
+ "grad_norm": 1195.4358872902142,
59
+ "learning_rate": 1e-08,
60
+ "logits/chosen": -4.3016462326049805,
61
+ "logits/rejected": -4.365716457366943,
62
+ "logps/chosen": -308.2979431152344,
63
+ "logps/rejected": -285.63018798828125,
64
+ "loss": 0.7311,
65
+ "rewards/accuracies": 0.49687498807907104,
66
+ "rewards/chosen": -0.010719490237534046,
67
+ "rewards/margins": -0.010031750425696373,
68
+ "rewards/rejected": -0.000687739229761064,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.27,
73
+ "grad_norm": 1144.0076561795222,
74
+ "learning_rate": 9.966191788709716e-09,
75
+ "logits/chosen": -4.187338829040527,
76
+ "logits/rejected": -4.271176338195801,
77
+ "logps/chosen": -332.39453125,
78
+ "logps/rejected": -307.4621276855469,
79
+ "loss": 0.7199,
80
+ "rewards/accuracies": 0.5375000238418579,
81
+ "rewards/chosen": 0.04900529980659485,
82
+ "rewards/margins": 0.05206792429089546,
83
+ "rewards/rejected": -0.0030626237858086824,
84
  "step": 40
85
  },
86
  {
87
  "epoch": 0.33,
88
+ "grad_norm": 1135.1845096891816,
89
+ "learning_rate": 9.86522435289912e-09,
90
+ "logits/chosen": -4.197329044342041,
91
+ "logits/rejected": -4.366620063781738,
92
+ "logps/chosen": -333.64678955078125,
93
+ "logps/rejected": -309.0525817871094,
94
+ "loss": 0.7189,
95
+ "rewards/accuracies": 0.515625,
96
+ "rewards/chosen": 0.04922889173030853,
97
+ "rewards/margins": 0.008506924845278263,
98
+ "rewards/rejected": 0.040721967816352844,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.4,
103
+ "grad_norm": 1124.4015217482226,
104
+ "learning_rate": 9.698463103929542e-09,
105
+ "logits/chosen": -4.186924457550049,
106
+ "logits/rejected": -4.269418239593506,
107
+ "logps/chosen": -323.7723693847656,
108
+ "logps/rejected": -308.26971435546875,
109
+ "loss": 0.7082,
110
+ "rewards/accuracies": 0.5249999761581421,
111
+ "rewards/chosen": 0.11511299759149551,
112
+ "rewards/margins": 0.052407026290893555,
113
+ "rewards/rejected": 0.06270597130060196,
114
  "step": 60
115
  },
116
  {
117
  "epoch": 0.47,
118
+ "grad_norm": 1133.1247984062034,
119
+ "learning_rate": 9.468163201617062e-09,
120
+ "logits/chosen": -4.147147178649902,
121
+ "logits/rejected": -4.3019304275512695,
122
+ "logps/chosen": -344.58563232421875,
123
+ "logps/rejected": -314.4212951660156,
124
+ "loss": 0.7005,
125
+ "rewards/accuracies": 0.590624988079071,
126
+ "rewards/chosen": 0.23254117369651794,
127
+ "rewards/margins": 0.13268980383872986,
128
+ "rewards/rejected": 0.09985135495662689,
129
  "step": 70
130
  },
131
  {
132
  "epoch": 0.53,
133
+ "grad_norm": 1057.1503651106625,
134
+ "learning_rate": 9.177439057064682e-09,
135
+ "logits/chosen": -4.223555564880371,
136
+ "logits/rejected": -4.389444828033447,
137
+ "logps/chosen": -334.3599853515625,
138
+ "logps/rejected": -305.4422607421875,
139
+ "loss": 0.6844,
140
+ "rewards/accuracies": 0.59375,
141
+ "rewards/chosen": 0.23962649703025818,
142
+ "rewards/margins": 0.09747296571731567,
143
+ "rewards/rejected": 0.1421535313129425,
144
  "step": 80
145
  },
146
  {
147
  "epoch": 0.6,
148
+ "grad_norm": 1055.57995971272,
149
+ "learning_rate": 8.830222215594889e-09,
150
+ "logits/chosen": -4.2292633056640625,
151
+ "logits/rejected": -4.349828243255615,
152
+ "logps/chosen": -327.07330322265625,
153
+ "logps/rejected": -307.81707763671875,
154
+ "loss": 0.6772,
155
+ "rewards/accuracies": 0.5562499761581421,
156
+ "rewards/chosen": 0.2870681881904602,
157
+ "rewards/margins": 0.06986425817012787,
158
+ "rewards/rejected": 0.21720390021800995,
159
  "step": 90
160
  },
161
  {
162
  "epoch": 0.67,
163
+ "grad_norm": 1124.231150209405,
164
+ "learning_rate": 8.431208189343668e-09,
165
+ "logits/chosen": -4.214221000671387,
166
+ "logits/rejected": -4.424824237823486,
167
+ "logps/chosen": -320.45928955078125,
168
+ "logps/rejected": -289.71466064453125,
169
+ "loss": 0.6738,
170
+ "rewards/accuracies": 0.581250011920929,
171
+ "rewards/chosen": 0.3042110502719879,
172
+ "rewards/margins": 0.12749677896499634,
173
+ "rewards/rejected": 0.1767142415046692,
174
  "step": 100
175
  },
176
  {
177
  "epoch": 0.73,
178
+ "grad_norm": 1045.9935050945592,
179
+ "learning_rate": 7.98579295851393e-09,
180
+ "logits/chosen": -4.233684062957764,
181
+ "logits/rejected": -4.286379814147949,
182
+ "logps/chosen": -314.39251708984375,
183
+ "logps/rejected": -303.2556457519531,
184
+ "loss": 0.6623,
185
+ "rewards/accuracies": 0.5687500238418579,
186
+ "rewards/chosen": 0.3848657011985779,
187
+ "rewards/margins": 0.10032075643539429,
188
+ "rewards/rejected": 0.2845449447631836,
189
  "step": 110
190
  },
191
  {
192
  "epoch": 0.8,
193
+ "grad_norm": 1107.1293654591734,
194
+ "learning_rate": 7.500000000000001e-09,
195
+ "logits/chosen": -4.151052951812744,
196
+ "logits/rejected": -4.307383060455322,
197
+ "logps/chosen": -338.46759033203125,
198
+ "logps/rejected": -313.0832214355469,
199
+ "loss": 0.6634,
200
+ "rewards/accuracies": 0.625,
201
+ "rewards/chosen": 0.49397316575050354,
202
+ "rewards/margins": 0.20816774666309357,
203
+ "rewards/rejected": 0.2858053743839264,
204
  "step": 120
205
  },
206
  {
207
  "epoch": 0.87,
208
+ "grad_norm": 1113.1665283582252,
209
+ "learning_rate": 6.980398830195784e-09,
210
+ "logits/chosen": -4.174288272857666,
211
+ "logits/rejected": -4.36756706237793,
212
+ "logps/chosen": -323.51800537109375,
213
+ "logps/rejected": -304.7021484375,
214
+ "loss": 0.6392,
215
+ "rewards/accuracies": 0.637499988079071,
216
+ "rewards/chosen": 0.5511430501937866,
217
+ "rewards/margins": 0.23519937694072723,
218
+ "rewards/rejected": 0.31594371795654297,
219
  "step": 130
220
  },
221
  {
222
  "epoch": 0.93,
223
+ "grad_norm": 1022.5201031232197,
224
+ "learning_rate": 6.434016163555451e-09,
225
+ "logits/chosen": -4.292388916015625,
226
+ "logits/rejected": -4.358359336853027,
227
+ "logps/chosen": -307.1917419433594,
228
+ "logps/rejected": -296.78192138671875,
229
+ "loss": 0.6402,
230
+ "rewards/accuracies": 0.6000000238418579,
231
+ "rewards/chosen": 0.5292906761169434,
232
+ "rewards/margins": 0.16237936913967133,
233
+ "rewards/rejected": 0.36691129207611084,
234
  "step": 140
235
  },
236
  {
237
  "epoch": 1.0,
238
+ "grad_norm": 1051.3521300860295,
239
+ "learning_rate": 5.868240888334653e-09,
240
+ "logits/chosen": -4.260380268096924,
241
+ "logits/rejected": -4.333888053894043,
242
+ "logps/chosen": -311.844482421875,
243
+ "logps/rejected": -299.74261474609375,
244
+ "loss": 0.6387,
245
+ "rewards/accuracies": 0.6187499761581421,
246
+ "rewards/chosen": 0.5666605830192566,
247
+ "rewards/margins": 0.1995116025209427,
248
+ "rewards/rejected": 0.3671489655971527,
249
  "step": 150
250
  },
251
  {
252
+ "epoch": 1.06,
253
+ "grad_norm": 1029.1548288848146,
254
+ "learning_rate": 5.290724144552379e-09,
255
+ "logits/chosen": -4.191515922546387,
256
+ "logits/rejected": -4.348960876464844,
257
+ "logps/chosen": -326.90582275390625,
258
+ "logps/rejected": -305.37689208984375,
259
+ "loss": 0.6032,
260
+ "rewards/accuracies": 0.6937500238418579,
261
+ "rewards/chosen": 0.6744669079780579,
262
+ "rewards/margins": 0.37094420194625854,
263
+ "rewards/rejected": 0.3035227358341217,
264
+ "step": 160
265
+ },
266
+ {
267
+ "epoch": 1.13,
268
+ "grad_norm": 991.2604083787021,
269
+ "learning_rate": 4.709275855447621e-09,
270
+ "logits/chosen": -4.241927623748779,
271
+ "logits/rejected": -4.339847564697266,
272
+ "logps/chosen": -314.4874572753906,
273
+ "logps/rejected": -295.49151611328125,
274
+ "loss": 0.6151,
275
+ "rewards/accuracies": 0.6937500238418579,
276
+ "rewards/chosen": 0.6504887938499451,
277
+ "rewards/margins": 0.3018999397754669,
278
+ "rewards/rejected": 0.3485889136791229,
279
+ "step": 170
280
+ },
281
+ {
282
+ "epoch": 1.2,
283
+ "grad_norm": 976.0494957890444,
284
+ "learning_rate": 4.131759111665349e-09,
285
+ "logits/chosen": -4.174123287200928,
286
+ "logits/rejected": -4.313396453857422,
287
+ "logps/chosen": -325.78045654296875,
288
+ "logps/rejected": -304.82989501953125,
289
+ "loss": 0.6147,
290
+ "rewards/accuracies": 0.6656249761581421,
291
+ "rewards/chosen": 0.6252425909042358,
292
+ "rewards/margins": 0.27976280450820923,
293
+ "rewards/rejected": 0.3454797863960266,
294
+ "step": 180
295
+ },
296
+ {
297
+ "epoch": 1.26,
298
+ "grad_norm": 949.77837430784,
299
+ "learning_rate": 3.56598383644455e-09,
300
+ "logits/chosen": -4.231461524963379,
301
+ "logits/rejected": -4.367356777191162,
302
+ "logps/chosen": -326.8230895996094,
303
+ "logps/rejected": -299.7027587890625,
304
+ "loss": 0.5969,
305
+ "rewards/accuracies": 0.6968749761581421,
306
+ "rewards/chosen": 0.7501948475837708,
307
+ "rewards/margins": 0.35921385884284973,
308
+ "rewards/rejected": 0.39098095893859863,
309
+ "step": 190
310
+ },
311
+ {
312
+ "epoch": 1.33,
313
+ "grad_norm": 963.9694294192556,
314
+ "learning_rate": 3.0196011698042157e-09,
315
+ "logits/chosen": -4.249307155609131,
316
+ "logits/rejected": -4.317980766296387,
317
+ "logps/chosen": -313.9615783691406,
318
+ "logps/rejected": -293.1615905761719,
319
+ "loss": 0.5922,
320
+ "rewards/accuracies": 0.6781250238418579,
321
+ "rewards/chosen": 0.7425140142440796,
322
+ "rewards/margins": 0.3114904761314392,
323
+ "rewards/rejected": 0.4310235381126404,
324
+ "step": 200
325
+ },
326
+ {
327
+ "epoch": 1.4,
328
+ "grad_norm": 1038.4899227109647,
329
+ "learning_rate": 2.5000000000000013e-09,
330
+ "logits/chosen": -4.174517631530762,
331
+ "logits/rejected": -4.288783073425293,
332
+ "logps/chosen": -327.461181640625,
333
+ "logps/rejected": -304.9690246582031,
334
+ "loss": 0.5952,
335
+ "rewards/accuracies": 0.699999988079071,
336
+ "rewards/chosen": 0.7787834405899048,
337
+ "rewards/margins": 0.3926675319671631,
338
+ "rewards/rejected": 0.3861159086227417,
339
+ "step": 210
340
+ },
341
+ {
342
+ "epoch": 1.46,
343
+ "grad_norm": 992.6867793357102,
344
+ "learning_rate": 2.0142070414860704e-09,
345
+ "logits/chosen": -4.165514945983887,
346
+ "logits/rejected": -4.316833019256592,
347
+ "logps/chosen": -358.1416931152344,
348
+ "logps/rejected": -323.22998046875,
349
+ "loss": 0.5994,
350
+ "rewards/accuracies": 0.6781250238418579,
351
+ "rewards/chosen": 0.8851108551025391,
352
+ "rewards/margins": 0.35754603147506714,
353
+ "rewards/rejected": 0.5275647640228271,
354
+ "step": 220
355
+ },
356
+ {
357
+ "epoch": 1.53,
358
+ "grad_norm": 918.0076197376386,
359
+ "learning_rate": 1.5687918106563326e-09,
360
+ "logits/chosen": -4.234222412109375,
361
+ "logits/rejected": -4.369565486907959,
362
+ "logps/chosen": -328.3675537109375,
363
+ "logps/rejected": -303.726806640625,
364
+ "loss": 0.5919,
365
+ "rewards/accuracies": 0.65625,
366
+ "rewards/chosen": 0.8034540414810181,
367
+ "rewards/margins": 0.2974655032157898,
368
+ "rewards/rejected": 0.5059884786605835,
369
+ "step": 230
370
+ },
371
+ {
372
+ "epoch": 1.6,
373
+ "grad_norm": 944.3107996293463,
374
+ "learning_rate": 1.1697777844051105e-09,
375
+ "logits/chosen": -4.202380180358887,
376
+ "logits/rejected": -4.353602409362793,
377
+ "logps/chosen": -326.75604248046875,
378
+ "logps/rejected": -297.03619384765625,
379
+ "loss": 0.588,
380
+ "rewards/accuracies": 0.659375011920929,
381
+ "rewards/chosen": 0.8500032424926758,
382
+ "rewards/margins": 0.3546527922153473,
383
+ "rewards/rejected": 0.4953504502773285,
384
+ "step": 240
385
+ },
386
+ {
387
+ "epoch": 1.66,
388
+ "grad_norm": 944.2690060176395,
389
+ "learning_rate": 8.225609429353187e-10,
390
+ "logits/chosen": -4.173297882080078,
391
+ "logits/rejected": -4.368635654449463,
392
+ "logps/chosen": -321.2724914550781,
393
+ "logps/rejected": -293.50311279296875,
394
+ "loss": 0.5856,
395
+ "rewards/accuracies": 0.731249988079071,
396
+ "rewards/chosen": 0.8207529783248901,
397
+ "rewards/margins": 0.40889453887939453,
398
+ "rewards/rejected": 0.4118584990501404,
399
+ "step": 250
400
+ },
401
+ {
402
+ "epoch": 1.73,
403
+ "grad_norm": 957.8767493880096,
404
+ "learning_rate": 5.318367983829391e-10,
405
+ "logits/chosen": -4.18589448928833,
406
+ "logits/rejected": -4.284361839294434,
407
+ "logps/chosen": -348.6470031738281,
408
+ "logps/rejected": -326.4360046386719,
409
+ "loss": 0.5895,
410
+ "rewards/accuracies": 0.690625011920929,
411
+ "rewards/chosen": 0.8914071917533875,
412
+ "rewards/margins": 0.37524908781051636,
413
+ "rewards/rejected": 0.5161582231521606,
414
+ "step": 260
415
+ },
416
+ {
417
+ "epoch": 1.8,
418
+ "grad_norm": 949.249423709826,
419
+ "learning_rate": 3.015368960704584e-10,
420
+ "logits/chosen": -4.271857738494873,
421
+ "logits/rejected": -4.4089555740356445,
422
+ "logps/chosen": -325.19049072265625,
423
+ "logps/rejected": -302.7436828613281,
424
+ "loss": 0.5938,
425
+ "rewards/accuracies": 0.6781250238418579,
426
+ "rewards/chosen": 0.9044697880744934,
427
+ "rewards/margins": 0.38229116797447205,
428
+ "rewards/rejected": 0.522178590297699,
429
+ "step": 270
430
+ },
431
+ {
432
+ "epoch": 1.86,
433
+ "grad_norm": 932.1311904270524,
434
+ "learning_rate": 1.3477564710088098e-10,
435
+ "logits/chosen": -4.163296222686768,
436
+ "logits/rejected": -4.306557655334473,
437
+ "logps/chosen": -327.57257080078125,
438
+ "logps/rejected": -306.2264099121094,
439
+ "loss": 0.5865,
440
+ "rewards/accuracies": 0.715624988079071,
441
+ "rewards/chosen": 0.8713304400444031,
442
+ "rewards/margins": 0.4181024134159088,
443
+ "rewards/rejected": 0.45322805643081665,
444
+ "step": 280
445
+ },
446
+ {
447
+ "epoch": 1.93,
448
+ "grad_norm": 884.8124763448137,
449
+ "learning_rate": 3.380821129028488e-11,
450
+ "logits/chosen": -4.1040120124816895,
451
+ "logits/rejected": -4.260913372039795,
452
+ "logps/chosen": -334.97686767578125,
453
+ "logps/rejected": -312.01617431640625,
454
+ "loss": 0.5924,
455
+ "rewards/accuracies": 0.6812499761581421,
456
+ "rewards/chosen": 0.9016163945198059,
457
+ "rewards/margins": 0.4117993414402008,
458
+ "rewards/rejected": 0.48981720209121704,
459
+ "step": 290
460
+ },
461
+ {
462
+ "epoch": 2.0,
463
+ "grad_norm": 959.7419745329076,
464
+ "learning_rate": 0.0,
465
+ "logits/chosen": -4.31270170211792,
466
+ "logits/rejected": -4.468858242034912,
467
+ "logps/chosen": -314.4400939941406,
468
+ "logps/rejected": -284.8276672363281,
469
+ "loss": 0.5825,
470
+ "rewards/accuracies": 0.6875,
471
+ "rewards/chosen": 0.82757169008255,
472
+ "rewards/margins": 0.36657077074050903,
473
+ "rewards/rejected": 0.4610009789466858,
474
+ "step": 300
475
+ },
476
+ {
477
+ "epoch": 2.0,
478
+ "step": 300,
479
  "total_flos": 0.0,
480
+ "train_loss": 0.6411590957641602,
481
+ "train_runtime": 8942.072,
482
+ "train_samples_per_second": 8.599,
483
+ "train_steps_per_second": 0.034
484
  }
485
  ],
486
  "logging_steps": 10,
487
+ "max_steps": 300,
488
  "num_input_tokens_seen": 0,
489
+ "num_train_epochs": 2,
490
  "save_steps": 100,
491
  "total_flos": 0.0,
492
  "train_batch_size": 8,