RikkiXu commited on
Commit
2c388ad
1 Parent(s): c5bf38c

Model save

Browse files
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- base_model: princeton-nlp/Mistral-7B-Base-SFT-DPO
3
  tags:
4
  - trl
5
  - dpo
@@ -14,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # zephyr-7b-dpo-full
16
 
17
- This model is a fine-tuned version of [princeton-nlp/Mistral-7B-Base-SFT-DPO](https://huggingface.co/princeton-nlp/Mistral-7B-Base-SFT-DPO) on an unknown dataset.
18
 
19
  ## Model description
20
 
@@ -34,14 +33,14 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 5e-07
37
- - train_batch_size: 8
38
- - eval_batch_size: 8
39
  - seed: 42
40
  - distributed_type: multi-GPU
41
  - num_devices: 8
42
  - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 256
44
- - total_eval_batch_size: 64
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: cosine
47
  - lr_scheduler_warmup_ratio: 0.1
@@ -55,5 +54,5 @@ The following hyperparameters were used during training:
55
 
56
  - Transformers 4.39.3
57
  - Pytorch 2.1.2+cu118
58
- - Datasets 2.19.1
59
  - Tokenizers 0.15.2
 
1
  ---
 
2
  tags:
3
  - trl
4
  - dpo
 
13
 
14
  # zephyr-7b-dpo-full
15
 
16
+ This model was trained from scratch on the None dataset.
17
 
18
  ## Model description
19
 
 
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 5e-07
36
+ - train_batch_size: 4
37
+ - eval_batch_size: 4
38
  - seed: 42
39
  - distributed_type: multi-GPU
40
  - num_devices: 8
41
  - gradient_accumulation_steps: 4
42
+ - total_train_batch_size: 128
43
+ - total_eval_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
 
54
 
55
  - Transformers 4.39.3
56
  - Pytorch 2.1.2+cu118
57
+ - Datasets 2.16.1
58
  - Tokenizers 0.15.2
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.2785977178812027,
4
- "train_runtime": 11929.9898,
5
- "train_samples": 102360,
6
- "train_samples_per_second": 8.58,
7
- "train_steps_per_second": 0.034
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.36373490798167696,
4
+ "train_runtime": 5835.088,
5
+ "train_samples": 50000,
6
+ "train_samples_per_second": 8.569,
7
+ "train_steps_per_second": 0.067
8
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41b405c6b39fce54d81023410da5b0175bc34f1b707551cc87bb57315a19139d
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e5deb8630443e9400286b2122958aab3235892baddf289d4dbde4dc184fd0e5
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4fcfc77bc0cce12435a691bee318c376a963ab3c60e50f0201871ef7f9f1899
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fb2f3e09768a1370109befb94e79cce23f327b3f41a617eeb25535926889cff
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec4cc12a5a582b9be0861f8da52397528a0b5094e4e53a00c5c10ad9fdc740da
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36444111aaaa0874ed9be86eac12cf8dca245ebd1c686a45563cfc788c09cba5
3
  size 4540516344
runs/Jun22_01-21-37_n136-129-074/events.out.tfevents.1718990517.n136-129-074.300436.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db3c793612bcb36a08b9f60a4bdb334073f7e777c6a4b1c21d99e49e19f625ab
3
- size 26099
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a652924d436028cbb10a75eca64eb43de02d697386ecb827f353baa2e09f3c9
3
+ size 32645
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.2785977178812027,
4
- "train_runtime": 11929.9898,
5
- "train_samples": 102360,
6
- "train_samples_per_second": 8.58,
7
- "train_steps_per_second": 0.034
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.36373490798167696,
4
+ "train_runtime": 5835.088,
5
+ "train_samples": 50000,
6
+ "train_samples_per_second": 8.569,
7
+ "train_steps_per_second": 0.067
8
  }
trainer_state.json CHANGED
@@ -1,21 +1,21 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.0,
5
- "eval_steps": 500,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
- "grad_norm": 24.862652137264853,
14
- "learning_rate": 1.25e-08,
15
- "logits/chosen": -0.5811702013015747,
16
- "logits/rejected": -0.11655431985855103,
17
- "logps/chosen": -351.5902099609375,
18
- "logps/rejected": -240.969970703125,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
@@ -25,621 +25,606 @@
25
  },
26
  {
27
  "epoch": 0.03,
28
- "grad_norm": 23.69292682023629,
29
- "learning_rate": 1.25e-07,
30
- "logits/chosen": 0.26120826601982117,
31
- "logits/rejected": 0.23706814646720886,
32
- "logps/chosen": -333.1805419921875,
33
- "logps/rejected": -244.67898559570312,
34
- "loss": 0.6922,
35
- "rewards/accuracies": 0.5173611044883728,
36
- "rewards/chosen": 0.0021614907309412956,
37
- "rewards/margins": 0.0021554920822381973,
38
- "rewards/rejected": 5.998538654239383e-06,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.05,
43
- "grad_norm": 18.203526649945516,
44
- "learning_rate": 2.5e-07,
45
- "logits/chosen": -0.017204787582159042,
46
- "logits/rejected": 0.1991611272096634,
47
- "logps/chosen": -320.430908203125,
48
- "logps/rejected": -234.376220703125,
49
- "loss": 0.669,
50
  "rewards/accuracies": 0.737500011920929,
51
- "rewards/chosen": 0.033605434000492096,
52
- "rewards/margins": 0.04716432839632034,
53
- "rewards/rejected": -0.01355889905244112,
54
  "step": 20
55
  },
56
  {
57
- "epoch": 0.07,
58
- "grad_norm": 10.096989474079606,
59
- "learning_rate": 3.75e-07,
60
- "logits/chosen": -0.2575300931930542,
61
- "logits/rejected": -0.4580558240413666,
62
- "logps/chosen": -300.87896728515625,
63
- "logps/rejected": -255.5655517578125,
64
- "loss": 0.5805,
65
- "rewards/accuracies": 0.7749999761581421,
66
- "rewards/chosen": 0.14600001275539398,
67
- "rewards/margins": 0.2884979844093323,
68
- "rewards/rejected": -0.14249801635742188,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.1,
73
- "grad_norm": 9.68944337059453,
74
- "learning_rate": 5e-07,
75
- "logits/chosen": -0.6759181022644043,
76
- "logits/rejected": -0.6345951557159424,
77
- "logps/chosen": -317.50872802734375,
78
- "logps/rejected": -302.39630126953125,
79
- "loss": 0.4819,
80
- "rewards/accuracies": 0.840624988079071,
81
- "rewards/chosen": 0.07102981209754944,
82
- "rewards/margins": 0.6418195366859436,
83
- "rewards/rejected": -0.5707896947860718,
84
  "step": 40
85
  },
86
  {
87
- "epoch": 0.12,
88
- "grad_norm": 13.329379682299182,
89
- "learning_rate": 4.990486745229364e-07,
90
- "logits/chosen": -0.12263472378253937,
91
- "logits/rejected": 0.44540151953697205,
92
- "logps/chosen": -374.64556884765625,
93
- "logps/rejected": -388.1717224121094,
94
- "loss": 0.3966,
95
- "rewards/accuracies": 0.800000011920929,
96
- "rewards/chosen": -0.4549541473388672,
97
- "rewards/margins": 1.0250240564346313,
98
- "rewards/rejected": -1.4799782037734985,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.15,
103
- "grad_norm": 17.333516248641253,
104
- "learning_rate": 4.96201938253052e-07,
105
- "logits/chosen": -0.30300790071487427,
106
- "logits/rejected": 0.3122316002845764,
107
- "logps/chosen": -394.78106689453125,
108
- "logps/rejected": -432.4813537597656,
109
- "loss": 0.3861,
110
- "rewards/accuracies": 0.8187500238418579,
111
- "rewards/chosen": -0.7015730142593384,
112
- "rewards/margins": 1.1719900369644165,
113
- "rewards/rejected": -1.8735630512237549,
114
  "step": 60
115
  },
116
  {
117
- "epoch": 0.17,
118
- "grad_norm": 15.677534908750197,
119
- "learning_rate": 4.91481456572267e-07,
120
- "logits/chosen": 0.7395630478858948,
121
- "logits/rejected": 1.5376254320144653,
122
- "logps/chosen": -425.17236328125,
123
- "logps/rejected": -448.2694396972656,
124
- "loss": 0.3474,
125
- "rewards/accuracies": 0.831250011920929,
126
- "rewards/chosen": -0.8609533309936523,
127
- "rewards/margins": 1.3486477136611938,
128
- "rewards/rejected": -2.2096011638641357,
129
  "step": 70
130
  },
131
  {
132
  "epoch": 0.2,
133
- "grad_norm": 17.182808543364636,
134
- "learning_rate": 4.849231551964771e-07,
135
- "logits/chosen": 2.598942995071411,
136
- "logits/rejected": 3.4538092613220215,
137
- "logps/chosen": -448.8929748535156,
138
- "logps/rejected": -540.0630493164062,
139
- "loss": 0.3215,
140
- "rewards/accuracies": 0.890625,
141
- "rewards/chosen": -1.3736767768859863,
142
- "rewards/margins": 1.7528272867202759,
143
- "rewards/rejected": -3.126504421234131,
144
  "step": 80
145
  },
146
  {
147
  "epoch": 0.23,
148
- "grad_norm": 16.648755569621386,
149
- "learning_rate": 4.7657694675916247e-07,
150
- "logits/chosen": 2.8463895320892334,
151
- "logits/rejected": 3.732513427734375,
152
- "logps/chosen": -496.74005126953125,
153
- "logps/rejected": -623.58984375,
154
- "loss": 0.3048,
155
- "rewards/accuracies": 0.878125011920929,
156
- "rewards/chosen": -1.7601783275604248,
157
- "rewards/margins": 1.9939384460449219,
158
- "rewards/rejected": -3.7541167736053467,
159
  "step": 90
160
  },
161
  {
162
- "epoch": 0.25,
163
- "grad_norm": 15.972608527062494,
164
- "learning_rate": 4.6650635094610966e-07,
165
- "logits/chosen": 2.0133347511291504,
166
- "logits/rejected": 3.3279690742492676,
167
- "logps/chosen": -554.5970458984375,
168
- "logps/rejected": -683.0777587890625,
169
- "loss": 0.2797,
170
- "rewards/accuracies": 0.859375,
171
- "rewards/chosen": -2.21871018409729,
172
- "rewards/margins": 2.000453233718872,
173
- "rewards/rejected": -4.219162940979004,
174
  "step": 100
175
  },
176
  {
177
  "epoch": 0.28,
178
- "grad_norm": 16.95927334748175,
179
- "learning_rate": 4.5478801107224794e-07,
180
- "logits/chosen": 2.1293346881866455,
181
- "logits/rejected": 3.9433817863464355,
182
- "logps/chosen": -545.55078125,
183
- "logps/rejected": -698.3030395507812,
184
- "loss": 0.2718,
185
- "rewards/accuracies": 0.887499988079071,
186
- "rewards/chosen": -2.248697280883789,
187
- "rewards/margins": 2.459144353866577,
188
- "rewards/rejected": -4.707841873168945,
189
  "step": 110
190
  },
191
  {
192
- "epoch": 0.3,
193
- "grad_norm": 15.769838259410646,
194
- "learning_rate": 4.415111107797445e-07,
195
- "logits/chosen": 2.2328364849090576,
196
- "logits/rejected": 3.943868637084961,
197
- "logps/chosen": -547.4822998046875,
198
- "logps/rejected": -709.2218017578125,
199
- "loss": 0.2597,
200
- "rewards/accuracies": 0.859375,
201
- "rewards/chosen": -2.386432409286499,
202
- "rewards/margins": 2.306048631668091,
203
- "rewards/rejected": -4.692481517791748,
204
  "step": 120
205
  },
206
  {
207
  "epoch": 0.33,
208
- "grad_norm": 16.240997635455848,
209
- "learning_rate": 4.2677669529663686e-07,
210
- "logits/chosen": 3.3713316917419434,
211
- "logits/rejected": 4.970644950866699,
212
- "logps/chosen": -669.5197143554688,
213
- "logps/rejected": -839.8416748046875,
214
- "loss": 0.2523,
215
- "rewards/accuracies": 0.8687499761581421,
216
- "rewards/chosen": -3.3710944652557373,
217
- "rewards/margins": 2.5790421962738037,
218
- "rewards/rejected": -5.950136661529541,
219
  "step": 130
220
  },
221
  {
222
- "epoch": 0.35,
223
- "grad_norm": 16.664869807154886,
224
- "learning_rate": 4.106969024216348e-07,
225
- "logits/chosen": 3.0220611095428467,
226
- "logits/rejected": 4.610594749450684,
227
- "logps/chosen": -647.0032958984375,
228
- "logps/rejected": -834.1439208984375,
229
- "loss": 0.2514,
230
- "rewards/accuracies": 0.90625,
231
- "rewards/chosen": -3.284292221069336,
232
- "rewards/margins": 2.7818052768707275,
233
- "rewards/rejected": -6.066097259521484,
234
  "step": 140
235
  },
236
  {
237
  "epoch": 0.38,
238
- "grad_norm": 17.103959159416473,
239
- "learning_rate": 3.933941090877615e-07,
240
- "logits/chosen": 1.9788957834243774,
241
- "logits/rejected": 3.797266721725464,
242
- "logps/chosen": -657.1544799804688,
243
- "logps/rejected": -866.92236328125,
244
- "loss": 0.2465,
245
- "rewards/accuracies": 0.887499988079071,
246
- "rewards/chosen": -3.3205840587615967,
247
- "rewards/margins": 2.8848683834075928,
248
- "rewards/rejected": -6.205452919006348,
249
  "step": 150
250
  },
251
  {
252
- "epoch": 0.4,
253
- "grad_norm": 22.71759647433438,
254
- "learning_rate": 3.75e-07,
255
- "logits/chosen": 2.598877429962158,
256
- "logits/rejected": 3.922821044921875,
257
- "logps/chosen": -650.6119995117188,
258
- "logps/rejected": -860.2496337890625,
259
- "loss": 0.2424,
260
- "rewards/accuracies": 0.893750011920929,
261
- "rewards/chosen": -3.585509777069092,
262
- "rewards/margins": 2.7439045906066895,
263
- "rewards/rejected": -6.3294148445129395,
264
  "step": 160
265
  },
266
  {
267
- "epoch": 0.42,
268
- "grad_norm": 17.154680074008297,
269
- "learning_rate": 3.5565456543517485e-07,
270
- "logits/chosen": 1.2129310369491577,
271
- "logits/rejected": 3.644993543624878,
272
- "logps/chosen": -650.40576171875,
273
- "logps/rejected": -869.5897216796875,
274
- "loss": 0.245,
275
- "rewards/accuracies": 0.8656250238418579,
276
- "rewards/chosen": -3.348802089691162,
277
- "rewards/margins": 3.0448169708251953,
278
- "rewards/rejected": -6.393619537353516,
279
  "step": 170
280
  },
281
  {
282
- "epoch": 0.45,
283
- "grad_norm": 15.288786440112402,
284
- "learning_rate": 3.355050358314172e-07,
285
- "logits/chosen": 2.0979018211364746,
286
- "logits/rejected": 3.6165339946746826,
287
- "logps/chosen": -733.31298828125,
288
- "logps/rejected": -946.0720825195312,
289
- "loss": 0.225,
290
- "rewards/accuracies": 0.909375011920929,
291
- "rewards/chosen": -3.9488494396209717,
292
- "rewards/margins": 3.014504909515381,
293
- "rewards/rejected": -6.963354587554932,
294
  "step": 180
295
  },
296
  {
297
- "epoch": 0.47,
298
- "grad_norm": 15.20089211524797,
299
- "learning_rate": 3.147047612756302e-07,
300
- "logits/chosen": 1.049578309059143,
301
- "logits/rejected": 3.2230868339538574,
302
- "logps/chosen": -655.8287963867188,
303
- "logps/rejected": -909.56787109375,
304
- "loss": 0.2177,
305
- "rewards/accuracies": 0.8999999761581421,
306
- "rewards/chosen": -3.187329053878784,
307
- "rewards/margins": 3.4261791706085205,
308
- "rewards/rejected": -6.613508701324463,
309
  "step": 190
310
  },
311
  {
312
- "epoch": 0.5,
313
- "grad_norm": 19.21517389497067,
314
- "learning_rate": 2.934120444167326e-07,
315
- "logits/chosen": 2.0917961597442627,
316
- "logits/rejected": 4.381856918334961,
317
- "logps/chosen": -707.9210205078125,
318
- "logps/rejected": -967.8511962890625,
319
- "loss": 0.2291,
320
  "rewards/accuracies": 0.875,
321
- "rewards/chosen": -4.029782772064209,
322
- "rewards/margins": 3.380286455154419,
323
- "rewards/rejected": -7.410069465637207,
324
  "step": 200
325
  },
326
  {
327
- "epoch": 0.53,
328
- "grad_norm": 17.876619392703006,
329
- "learning_rate": 2.717889356869146e-07,
330
- "logits/chosen": 2.075894832611084,
331
- "logits/rejected": 3.812873363494873,
332
- "logps/chosen": -664.9110717773438,
333
- "logps/rejected": -898.7711791992188,
334
- "loss": 0.2335,
335
- "rewards/accuracies": 0.90625,
336
- "rewards/chosen": -3.6399245262145996,
337
- "rewards/margins": 3.0123069286346436,
338
- "rewards/rejected": -6.652230739593506,
339
  "step": 210
340
  },
341
  {
342
- "epoch": 0.55,
343
- "grad_norm": 16.42311250323521,
344
- "learning_rate": 2.5e-07,
345
- "logits/chosen": 2.077141523361206,
346
- "logits/rejected": 4.0336527824401855,
347
- "logps/chosen": -714.0510864257812,
348
- "logps/rejected": -951.3372802734375,
349
- "loss": 0.2163,
350
- "rewards/accuracies": 0.893750011920929,
351
- "rewards/chosen": -4.000287055969238,
352
- "rewards/margins": 3.160945415496826,
353
- "rewards/rejected": -7.161231994628906,
354
  "step": 220
355
  },
356
  {
357
- "epoch": 0.57,
358
- "grad_norm": 21.86260854020408,
359
- "learning_rate": 2.2821106431308543e-07,
360
- "logits/chosen": 1.8970081806182861,
361
- "logits/rejected": 3.8517441749572754,
362
- "logps/chosen": -711.104248046875,
363
- "logps/rejected": -952.9786987304688,
364
- "loss": 0.2307,
365
  "rewards/accuracies": 0.875,
366
- "rewards/chosen": -4.102777481079102,
367
- "rewards/margins": 3.2044379711151123,
368
- "rewards/rejected": -7.307215213775635,
369
  "step": 230
370
  },
371
  {
372
- "epoch": 0.6,
373
- "grad_norm": 17.415535830140726,
374
- "learning_rate": 2.065879555832674e-07,
375
- "logits/chosen": 1.93063485622406,
376
- "logits/rejected": 3.716691255569458,
377
- "logps/chosen": -733.0238037109375,
378
- "logps/rejected": -995.0330200195312,
379
- "loss": 0.2135,
380
- "rewards/accuracies": 0.8968750238418579,
381
- "rewards/chosen": -4.205197811126709,
382
- "rewards/margins": 3.2796833515167236,
383
- "rewards/rejected": -7.4848809242248535,
384
  "step": 240
385
  },
386
  {
387
- "epoch": 0.62,
388
- "grad_norm": 23.48694643420195,
389
- "learning_rate": 1.8529523872436977e-07,
390
- "logits/chosen": 1.754500150680542,
391
- "logits/rejected": 3.7942306995391846,
392
- "logps/chosen": -756.137939453125,
393
- "logps/rejected": -1002.9494018554688,
394
- "loss": 0.2284,
395
- "rewards/accuracies": 0.887499988079071,
396
- "rewards/chosen": -4.309305667877197,
397
- "rewards/margins": 3.231421947479248,
398
- "rewards/rejected": -7.5407280921936035,
399
  "step": 250
400
  },
401
  {
402
- "epoch": 0.65,
403
- "grad_norm": 16.97115932824073,
404
- "learning_rate": 1.6449496416858282e-07,
405
- "logits/chosen": 2.1520519256591797,
406
- "logits/rejected": 4.258932590484619,
407
- "logps/chosen": -750.8233642578125,
408
- "logps/rejected": -1022.2374267578125,
409
- "loss": 0.2136,
410
- "rewards/accuracies": 0.8999999761581421,
411
- "rewards/chosen": -4.413332939147949,
412
- "rewards/margins": 3.5410499572753906,
413
- "rewards/rejected": -7.954381465911865,
414
  "step": 260
415
  },
416
  {
417
- "epoch": 0.68,
418
- "grad_norm": 20.50346815073402,
419
- "learning_rate": 1.4434543456482518e-07,
420
- "logits/chosen": 1.4445512294769287,
421
- "logits/rejected": 3.574235439300537,
422
- "logps/chosen": -682.8297729492188,
423
- "logps/rejected": -979.3341674804688,
424
- "loss": 0.1978,
425
- "rewards/accuracies": 0.9156249761581421,
426
- "rewards/chosen": -3.6296730041503906,
427
- "rewards/margins": 3.7298316955566406,
428
- "rewards/rejected": -7.359505653381348,
429
  "step": 270
430
  },
431
  {
432
- "epoch": 0.7,
433
- "grad_norm": 16.82073433691609,
434
- "learning_rate": 1.2500000000000005e-07,
435
- "logits/chosen": 1.172753930091858,
436
- "logits/rejected": 3.4942619800567627,
437
- "logps/chosen": -675.3418579101562,
438
- "logps/rejected": -974.0270385742188,
439
- "loss": 0.2189,
440
- "rewards/accuracies": 0.934374988079071,
441
- "rewards/chosen": -3.6095943450927734,
442
- "rewards/margins": 3.8261497020721436,
443
- "rewards/rejected": -7.435744285583496,
444
  "step": 280
445
  },
446
  {
447
- "epoch": 0.72,
448
- "grad_norm": 27.148115499609514,
449
- "learning_rate": 1.0660589091223854e-07,
450
- "logits/chosen": 1.3045436143875122,
451
- "logits/rejected": 3.874147891998291,
452
- "logps/chosen": -724.1644287109375,
453
- "logps/rejected": -1020.5606689453125,
454
- "loss": 0.2159,
455
- "rewards/accuracies": 0.918749988079071,
456
- "rewards/chosen": -3.9626336097717285,
457
- "rewards/margins": 3.8746466636657715,
458
- "rewards/rejected": -7.837281227111816,
459
  "step": 290
460
  },
461
  {
462
- "epoch": 0.75,
463
- "grad_norm": 18.896563678409045,
464
- "learning_rate": 8.930309757836516e-08,
465
- "logits/chosen": 1.5070204734802246,
466
- "logits/rejected": 3.8179619312286377,
467
- "logps/chosen": -756.2379150390625,
468
- "logps/rejected": -1010.8900146484375,
469
- "loss": 0.2118,
470
- "rewards/accuracies": 0.90625,
471
- "rewards/chosen": -4.342662811279297,
472
- "rewards/margins": 3.381243944168091,
473
- "rewards/rejected": -7.72390604019165,
474
  "step": 300
475
  },
476
  {
477
- "epoch": 0.78,
478
- "grad_norm": 18.302278636631012,
479
- "learning_rate": 7.322330470336313e-08,
480
- "logits/chosen": 2.013995885848999,
481
- "logits/rejected": 4.006863117218018,
482
- "logps/chosen": -755.84326171875,
483
- "logps/rejected": -1039.4290771484375,
484
- "loss": 0.2007,
485
- "rewards/accuracies": 0.925000011920929,
486
- "rewards/chosen": -4.390773296356201,
487
- "rewards/margins": 3.617499589920044,
488
- "rewards/rejected": -8.008273124694824,
489
  "step": 310
490
  },
491
  {
492
- "epoch": 0.8,
493
- "grad_norm": 19.269057964941258,
494
- "learning_rate": 5.848888922025552e-08,
495
- "logits/chosen": 1.4602447748184204,
496
- "logits/rejected": 3.709857940673828,
497
- "logps/chosen": -771.56640625,
498
- "logps/rejected": -1046.9013671875,
499
- "loss": 0.2211,
500
- "rewards/accuracies": 0.903124988079071,
501
- "rewards/chosen": -4.392203330993652,
502
- "rewards/margins": 3.6757659912109375,
503
- "rewards/rejected": -8.067970275878906,
504
  "step": 320
505
  },
506
  {
507
- "epoch": 0.82,
508
- "grad_norm": 16.212116381856944,
509
- "learning_rate": 4.521198892775202e-08,
510
- "logits/chosen": 1.5877026319503784,
511
- "logits/rejected": 3.5275306701660156,
512
- "logps/chosen": -748.9736328125,
513
- "logps/rejected": -1030.6241455078125,
514
- "loss": 0.1902,
515
- "rewards/accuracies": 0.909375011920929,
516
- "rewards/chosen": -4.198099613189697,
517
- "rewards/margins": 3.724585771560669,
518
- "rewards/rejected": -7.922685146331787,
519
  "step": 330
520
  },
521
  {
522
- "epoch": 0.85,
523
- "grad_norm": 23.048514547738275,
524
- "learning_rate": 3.349364905389032e-08,
525
- "logits/chosen": 1.2227389812469482,
526
- "logits/rejected": 3.192277193069458,
527
- "logps/chosen": -744.6568603515625,
528
- "logps/rejected": -1009.3792724609375,
529
- "loss": 0.2061,
530
- "rewards/accuracies": 0.893750011920929,
531
- "rewards/chosen": -4.285494804382324,
532
- "rewards/margins": 3.405372142791748,
533
- "rewards/rejected": -7.690866947174072,
534
  "step": 340
535
  },
536
  {
537
- "epoch": 0.88,
538
- "grad_norm": 18.4900810885199,
539
- "learning_rate": 2.3423053240837514e-08,
540
- "logits/chosen": 1.2598426342010498,
541
- "logits/rejected": 3.358072280883789,
542
- "logps/chosen": -737.2872314453125,
543
- "logps/rejected": -1013.8024291992188,
544
- "loss": 0.2239,
545
- "rewards/accuracies": 0.871874988079071,
546
- "rewards/chosen": -4.232865333557129,
547
- "rewards/margins": 3.510840892791748,
548
- "rewards/rejected": -7.743706703186035,
549
  "step": 350
550
  },
551
  {
552
- "epoch": 0.9,
553
- "grad_norm": 16.48305540217272,
554
- "learning_rate": 1.507684480352292e-08,
555
- "logits/chosen": 1.188299298286438,
556
- "logits/rejected": 3.3616530895233154,
557
- "logps/chosen": -726.2432861328125,
558
- "logps/rejected": -1018.1263427734375,
559
- "loss": 0.1958,
560
  "rewards/accuracies": 0.90625,
561
- "rewards/chosen": -4.075113773345947,
562
- "rewards/margins": 3.6948330402374268,
563
- "rewards/rejected": -7.769946098327637,
564
  "step": 360
565
  },
566
- {
567
- "epoch": 0.93,
568
- "grad_norm": 20.27854050236199,
569
- "learning_rate": 8.518543427732949e-09,
570
- "logits/chosen": 1.1218559741973877,
571
- "logits/rejected": 3.376429319381714,
572
- "logps/chosen": -706.4921875,
573
- "logps/rejected": -1009.96533203125,
574
- "loss": 0.199,
575
- "rewards/accuracies": 0.9156249761581421,
576
- "rewards/chosen": -4.078927516937256,
577
- "rewards/margins": 3.7843894958496094,
578
- "rewards/rejected": -7.863317966461182,
579
- "step": 370
580
- },
581
  {
582
  "epoch": 0.95,
583
- "grad_norm": 21.44365501017452,
584
- "learning_rate": 3.798061746947995e-09,
585
- "logits/chosen": 1.2741193771362305,
586
- "logits/rejected": 3.5388190746307373,
587
- "logps/chosen": -727.7400512695312,
588
- "logps/rejected": -996.4246215820312,
589
- "loss": 0.2143,
590
- "rewards/accuracies": 0.9281250238418579,
591
- "rewards/chosen": -4.139595985412598,
592
- "rewards/margins": 3.5642218589782715,
593
- "rewards/rejected": -7.703817844390869,
594
- "step": 380
595
  },
596
  {
597
  "epoch": 0.97,
598
- "grad_norm": 19.980658149789488,
599
- "learning_rate": 9.513254770636137e-10,
600
- "logits/chosen": 1.3589586019515991,
601
- "logits/rejected": 3.3931171894073486,
602
- "logps/chosen": -751.2428588867188,
603
- "logps/rejected": -1019.2142333984375,
604
- "loss": 0.2117,
605
- "rewards/accuracies": 0.893750011920929,
606
- "rewards/chosen": -4.217888832092285,
607
- "rewards/margins": 3.668452739715576,
608
- "rewards/rejected": -7.8863420486450195,
609
- "step": 390
610
  },
611
  {
612
  "epoch": 1.0,
613
- "grad_norm": 19.223012927780225,
614
  "learning_rate": 0.0,
615
- "logits/chosen": 1.3647325038909912,
616
- "logits/rejected": 3.651308536529541,
617
- "logps/chosen": -756.8760986328125,
618
- "logps/rejected": -1033.67333984375,
619
- "loss": 0.2003,
620
- "rewards/accuracies": 0.9125000238418579,
621
- "rewards/chosen": -4.446074485778809,
622
- "rewards/margins": 3.6355972290039062,
623
- "rewards/rejected": -8.081671714782715,
624
- "step": 400
625
  },
626
  {
627
  "epoch": 1.0,
628
- "step": 400,
629
  "total_flos": 0.0,
630
- "train_loss": 0.2785977178812027,
631
- "train_runtime": 11929.9898,
632
- "train_samples_per_second": 8.58,
633
- "train_steps_per_second": 0.034
634
  }
635
  ],
636
  "logging_steps": 10,
637
- "max_steps": 400,
638
  "num_input_tokens_seen": 0,
639
  "num_train_epochs": 1,
640
  "save_steps": 100,
641
  "total_flos": 0.0,
642
- "train_batch_size": 8,
643
  "trial_name": null,
644
  "trial_params": null
645
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9980806142034548,
5
+ "eval_steps": 10000000,
6
+ "global_step": 390,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
+ "grad_norm": 851.8646963671724,
14
+ "learning_rate": 1.282051282051282e-08,
15
+ "logits/chosen": -2.5583817958831787,
16
+ "logits/rejected": -2.4487552642822266,
17
+ "logps/chosen": -258.1644592285156,
18
+ "logps/rejected": -216.25729370117188,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
 
25
  },
26
  {
27
  "epoch": 0.03,
28
+ "grad_norm": 827.2359528491244,
29
+ "learning_rate": 1.2820512820512818e-07,
30
+ "logits/chosen": -2.605931282043457,
31
+ "logits/rejected": -2.552781105041504,
32
+ "logps/chosen": -267.6236267089844,
33
+ "logps/rejected": -217.6671905517578,
34
+ "loss": 0.6967,
35
+ "rewards/accuracies": 0.4097222089767456,
36
+ "rewards/chosen": -0.03547710180282593,
37
+ "rewards/margins": -0.018225612118840218,
38
+ "rewards/rejected": -0.01725148782134056,
39
  "step": 10
40
  },
41
  {
42
  "epoch": 0.05,
43
+ "grad_norm": 598.05709697014,
44
+ "learning_rate": 2.5641025641025636e-07,
45
+ "logits/chosen": -2.6306538581848145,
46
+ "logits/rejected": -2.5675768852233887,
47
+ "logps/chosen": -260.528564453125,
48
+ "logps/rejected": -207.09140014648438,
49
+ "loss": 0.5351,
50
  "rewards/accuracies": 0.737500011920929,
51
+ "rewards/chosen": 0.4653858244419098,
52
+ "rewards/margins": 0.4877452850341797,
53
+ "rewards/rejected": -0.022359488531947136,
54
  "step": 20
55
  },
56
  {
57
+ "epoch": 0.08,
58
+ "grad_norm": 777.4500662657566,
59
+ "learning_rate": 3.8461538461538463e-07,
60
+ "logits/chosen": -2.651550769805908,
61
+ "logits/rejected": -2.5767629146575928,
62
+ "logps/chosen": -250.84542846679688,
63
+ "logps/rejected": -198.71180725097656,
64
+ "loss": 0.3391,
65
+ "rewards/accuracies": 0.862500011920929,
66
+ "rewards/chosen": 2.5810751914978027,
67
+ "rewards/margins": 2.4870572090148926,
68
+ "rewards/rejected": 0.09401801228523254,
69
  "step": 30
70
  },
71
  {
72
  "epoch": 0.1,
73
+ "grad_norm": 344.88554576974366,
74
+ "learning_rate": 4.99989986344963e-07,
75
+ "logits/chosen": -2.6472256183624268,
76
+ "logits/rejected": -2.5672099590301514,
77
+ "logps/chosen": -243.0611114501953,
78
+ "logps/rejected": -193.21621704101562,
79
+ "loss": 0.3063,
80
+ "rewards/accuracies": 0.831250011920929,
81
+ "rewards/chosen": 4.372925758361816,
82
+ "rewards/margins": 4.090872287750244,
83
+ "rewards/rejected": 0.28205329179763794,
84
  "step": 40
85
  },
86
  {
87
+ "epoch": 0.13,
88
+ "grad_norm": 481.831925522066,
89
+ "learning_rate": 4.987893180827479e-07,
90
+ "logits/chosen": -2.658193588256836,
91
+ "logits/rejected": -2.5846261978149414,
92
+ "logps/chosen": -256.97283935546875,
93
+ "logps/rejected": -203.25177001953125,
94
+ "loss": 0.342,
95
+ "rewards/accuracies": 0.862500011920929,
96
+ "rewards/chosen": 6.758584499359131,
97
+ "rewards/margins": 5.620272636413574,
98
+ "rewards/rejected": 1.1383121013641357,
99
  "step": 50
100
  },
101
  {
102
  "epoch": 0.15,
103
+ "grad_norm": 507.1864724110635,
104
+ "learning_rate": 4.955969343539162e-07,
105
+ "logits/chosen": -2.6098527908325195,
106
+ "logits/rejected": -2.5346760749816895,
107
+ "logps/chosen": -260.9481506347656,
108
+ "logps/rejected": -208.94192504882812,
109
+ "loss": 0.3189,
110
+ "rewards/accuracies": 0.8687499761581421,
111
+ "rewards/chosen": 4.727741241455078,
112
+ "rewards/margins": 5.257795333862305,
113
+ "rewards/rejected": -0.5300543308258057,
114
  "step": 60
115
  },
116
  {
117
+ "epoch": 0.18,
118
+ "grad_norm": 347.65616298975976,
119
+ "learning_rate": 4.90438392204474e-07,
120
+ "logits/chosen": -2.5738308429718018,
121
+ "logits/rejected": -2.496386766433716,
122
+ "logps/chosen": -291.3803405761719,
123
+ "logps/rejected": -228.85986328125,
124
+ "loss": 0.3314,
125
+ "rewards/accuracies": 0.90625,
126
+ "rewards/chosen": 3.918160915374756,
127
+ "rewards/margins": 6.1196980476379395,
128
+ "rewards/rejected": -2.201537609100342,
129
  "step": 70
130
  },
131
  {
132
  "epoch": 0.2,
133
+ "grad_norm": 548.6777756471274,
134
+ "learning_rate": 4.83354989019146e-07,
135
+ "logits/chosen": -2.5354433059692383,
136
+ "logits/rejected": -2.4591267108917236,
137
+ "logps/chosen": -259.90399169921875,
138
+ "logps/rejected": -205.8795623779297,
139
+ "loss": 0.2963,
140
+ "rewards/accuracies": 0.893750011920929,
141
+ "rewards/chosen": 4.5924973487854,
142
+ "rewards/margins": 6.274462699890137,
143
+ "rewards/rejected": -1.6819652318954468,
144
  "step": 80
145
  },
146
  {
147
  "epoch": 0.23,
148
+ "grad_norm": 746.9613110272085,
149
+ "learning_rate": 4.7440343190975353e-07,
150
+ "logits/chosen": -2.572143316268921,
151
+ "logits/rejected": -2.514286518096924,
152
+ "logps/chosen": -256.91656494140625,
153
+ "logps/rejected": -218.2171630859375,
154
+ "loss": 0.3067,
155
+ "rewards/accuracies": 0.862500011920929,
156
+ "rewards/chosen": 2.8972291946411133,
157
+ "rewards/margins": 5.130236625671387,
158
+ "rewards/rejected": -2.2330079078674316,
159
  "step": 90
160
  },
161
  {
162
+ "epoch": 0.26,
163
+ "grad_norm": 292.7469627770502,
164
+ "learning_rate": 4.6365538373900506e-07,
165
+ "logits/chosen": -2.6255955696105957,
166
+ "logits/rejected": -2.5519518852233887,
167
+ "logps/chosen": -235.6707000732422,
168
+ "logps/rejected": -201.62594604492188,
169
+ "loss": 0.4639,
170
+ "rewards/accuracies": 0.862500011920929,
171
+ "rewards/chosen": 3.8989346027374268,
172
+ "rewards/margins": 5.672036170959473,
173
+ "rewards/rejected": -1.773101806640625,
174
  "step": 100
175
  },
176
  {
177
  "epoch": 0.28,
178
+ "grad_norm": 462.4159148288855,
179
+ "learning_rate": 4.5119688941406386e-07,
180
+ "logits/chosen": -2.6220152378082275,
181
+ "logits/rejected": -2.5428764820098877,
182
+ "logps/chosen": -256.5944519042969,
183
+ "logps/rejected": -210.29629516601562,
184
+ "loss": 0.3852,
185
+ "rewards/accuracies": 0.856249988079071,
186
+ "rewards/chosen": 5.447988033294678,
187
+ "rewards/margins": 6.571375370025635,
188
+ "rewards/rejected": -1.1233874559402466,
189
  "step": 110
190
  },
191
  {
192
+ "epoch": 0.31,
193
+ "grad_norm": 684.7134637036779,
194
+ "learning_rate": 4.3712768704277524e-07,
195
+ "logits/chosen": -2.5924530029296875,
196
+ "logits/rejected": -2.523179769515991,
197
+ "logps/chosen": -261.6965026855469,
198
+ "logps/rejected": -209.29666137695312,
199
+ "loss": 0.3703,
200
+ "rewards/accuracies": 0.893750011920929,
201
+ "rewards/chosen": 4.659531116485596,
202
+ "rewards/margins": 6.417691707611084,
203
+ "rewards/rejected": -1.758161187171936,
204
  "step": 120
205
  },
206
  {
207
  "epoch": 0.33,
208
+ "grad_norm": 363.69452359861873,
209
+ "learning_rate": 4.2156040946718343e-07,
210
+ "logits/chosen": -2.5601067543029785,
211
+ "logits/rejected": -2.4928698539733887,
212
+ "logps/chosen": -251.21463012695312,
213
+ "logps/rejected": -198.01596069335938,
214
+ "loss": 0.3349,
215
+ "rewards/accuracies": 0.8500000238418579,
216
+ "rewards/chosen": 3.6531822681427,
217
+ "rewards/margins": 6.222817897796631,
218
+ "rewards/rejected": -2.569636583328247,
219
  "step": 130
220
  },
221
  {
222
+ "epoch": 0.36,
223
+ "grad_norm": 388.6865732944718,
224
+ "learning_rate": 4.046196825665637e-07,
225
+ "logits/chosen": -2.58420729637146,
226
+ "logits/rejected": -2.5159454345703125,
227
+ "logps/chosen": -269.7867431640625,
228
+ "logps/rejected": -217.6550750732422,
229
+ "loss": 0.3662,
230
+ "rewards/accuracies": 0.8374999761581421,
231
+ "rewards/chosen": 3.2337214946746826,
232
+ "rewards/margins": 5.692519187927246,
233
+ "rewards/rejected": -2.4587976932525635,
234
  "step": 140
235
  },
236
  {
237
  "epoch": 0.38,
238
+ "grad_norm": 392.10882995088383,
239
+ "learning_rate": 3.864411275486261e-07,
240
+ "logits/chosen": -2.568713426589966,
241
+ "logits/rejected": -2.5009543895721436,
242
+ "logps/chosen": -263.0735778808594,
243
+ "logps/rejected": -213.8876495361328,
244
+ "loss": 0.3955,
245
+ "rewards/accuracies": 0.856249988079071,
246
+ "rewards/chosen": 4.510898590087891,
247
+ "rewards/margins": 6.606268882751465,
248
+ "rewards/rejected": -2.095369815826416,
249
  "step": 150
250
  },
251
  {
252
+ "epoch": 0.41,
253
+ "grad_norm": 466.21844913881523,
254
+ "learning_rate": 3.671702752161759e-07,
255
+ "logits/chosen": -2.567312717437744,
256
+ "logits/rejected": -2.497217893600464,
257
+ "logps/chosen": -244.8683624267578,
258
+ "logps/rejected": -200.1343536376953,
259
+ "loss": 0.3547,
260
+ "rewards/accuracies": 0.862500011920929,
261
+ "rewards/chosen": 2.441751480102539,
262
+ "rewards/margins": 6.709475040435791,
263
+ "rewards/rejected": -4.267723083496094,
264
  "step": 160
265
  },
266
  {
267
+ "epoch": 0.44,
268
+ "grad_norm": 681.7420438130807,
269
+ "learning_rate": 3.4696140090121375e-07,
270
+ "logits/chosen": -2.5749361515045166,
271
+ "logits/rejected": -2.509636640548706,
272
+ "logps/chosen": -266.73236083984375,
273
+ "logps/rejected": -213.6396026611328,
274
+ "loss": 0.3163,
275
+ "rewards/accuracies": 0.8812500238418579,
276
+ "rewards/chosen": 2.3756332397460938,
277
+ "rewards/margins": 6.7328901290893555,
278
+ "rewards/rejected": -4.357257843017578,
279
  "step": 170
280
  },
281
  {
282
+ "epoch": 0.46,
283
+ "grad_norm": 450.6259386496744,
284
+ "learning_rate": 3.259762893935617e-07,
285
+ "logits/chosen": -2.635408878326416,
286
+ "logits/rejected": -2.547847270965576,
287
+ "logps/chosen": -237.7583465576172,
288
+ "logps/rejected": -188.80947875976562,
289
+ "loss": 0.3732,
290
+ "rewards/accuracies": 0.875,
291
+ "rewards/chosen": 2.5650315284729004,
292
+ "rewards/margins": 5.900453090667725,
293
+ "rewards/rejected": -3.335421323776245,
294
  "step": 180
295
  },
296
  {
297
+ "epoch": 0.49,
298
+ "grad_norm": 393.57273774416257,
299
+ "learning_rate": 3.0438293975154184e-07,
300
+ "logits/chosen": -2.5968377590179443,
301
+ "logits/rejected": -2.520313024520874,
302
+ "logps/chosen": -259.823974609375,
303
+ "logps/rejected": -205.4634246826172,
304
+ "loss": 0.3175,
305
+ "rewards/accuracies": 0.887499988079071,
306
+ "rewards/chosen": 2.9979615211486816,
307
+ "rewards/margins": 6.678049564361572,
308
+ "rewards/rejected": -3.6800880432128906,
309
  "step": 190
310
  },
311
  {
312
+ "epoch": 0.51,
313
+ "grad_norm": 569.8752439327117,
314
+ "learning_rate": 2.823542203635138e-07,
315
+ "logits/chosen": -2.62797212600708,
316
+ "logits/rejected": -2.5373997688293457,
317
+ "logps/chosen": -274.06500244140625,
318
+ "logps/rejected": -219.7933807373047,
319
+ "loss": 0.36,
320
  "rewards/accuracies": 0.875,
321
+ "rewards/chosen": 3.8748526573181152,
322
+ "rewards/margins": 7.579891204833984,
323
+ "rewards/rejected": -3.705038070678711,
324
  "step": 200
325
  },
326
  {
327
+ "epoch": 0.54,
328
+ "grad_norm": 528.723840543068,
329
+ "learning_rate": 2.600664850273538e-07,
330
+ "logits/chosen": -2.612349033355713,
331
+ "logits/rejected": -2.537264347076416,
332
+ "logps/chosen": -266.8662109375,
333
+ "logps/rejected": -212.73904418945312,
334
+ "loss": 0.4949,
335
+ "rewards/accuracies": 0.8500000238418579,
336
+ "rewards/chosen": 2.512554883956909,
337
+ "rewards/margins": 6.339517116546631,
338
+ "rewards/rejected": -3.8269622325897217,
339
  "step": 210
340
  },
341
  {
342
+ "epoch": 0.56,
343
+ "grad_norm": 499.14391994879634,
344
+ "learning_rate": 2.3769816112703045e-07,
345
+ "logits/chosen": -2.6316843032836914,
346
+ "logits/rejected": -2.567474842071533,
347
+ "logps/chosen": -257.08892822265625,
348
+ "logps/rejected": -214.96658325195312,
349
+ "loss": 0.3779,
350
+ "rewards/accuracies": 0.887499988079071,
351
+ "rewards/chosen": 2.75276517868042,
352
+ "rewards/margins": 5.872605323791504,
353
+ "rewards/rejected": -3.119840145111084,
354
  "step": 220
355
  },
356
  {
357
+ "epoch": 0.59,
358
+ "grad_norm": 411.504863411568,
359
+ "learning_rate": 2.1542832120881677e-07,
360
+ "logits/chosen": -2.6690242290496826,
361
+ "logits/rejected": -2.5815441608428955,
362
+ "logps/chosen": -266.91094970703125,
363
+ "logps/rejected": -217.81674194335938,
364
+ "loss": 0.33,
365
  "rewards/accuracies": 0.875,
366
+ "rewards/chosen": 3.8953208923339844,
367
+ "rewards/margins": 6.836805820465088,
368
+ "rewards/rejected": -2.9414849281311035,
369
  "step": 230
370
  },
371
  {
372
+ "epoch": 0.61,
373
+ "grad_norm": 592.3030101879839,
374
+ "learning_rate": 1.934352493925695e-07,
375
+ "logits/chosen": -2.65104341506958,
376
+ "logits/rejected": -2.6025872230529785,
377
+ "logps/chosen": -262.87799072265625,
378
+ "logps/rejected": -222.07803344726562,
379
+ "loss": 0.3337,
380
+ "rewards/accuracies": 0.887499988079071,
381
+ "rewards/chosen": 2.9119772911071777,
382
+ "rewards/margins": 8.015997886657715,
383
+ "rewards/rejected": -5.104020118713379,
384
  "step": 240
385
  },
386
  {
387
+ "epoch": 0.64,
388
+ "grad_norm": 484.1153834763422,
389
+ "learning_rate": 1.7189501409486059e-07,
390
+ "logits/chosen": -2.6601402759552,
391
+ "logits/rejected": -2.5887746810913086,
392
+ "logps/chosen": -267.6604309082031,
393
+ "logps/rejected": -223.6746826171875,
394
+ "loss": 0.3508,
395
+ "rewards/accuracies": 0.8500000238418579,
396
+ "rewards/chosen": 2.4709649085998535,
397
+ "rewards/margins": 6.8468828201293945,
398
+ "rewards/rejected": -4.375916957855225,
399
  "step": 250
400
  },
401
  {
402
+ "epoch": 0.67,
403
+ "grad_norm": 369.72495233490616,
404
+ "learning_rate": 1.5098005849021078e-07,
405
+ "logits/chosen": -2.6492981910705566,
406
+ "logits/rejected": -2.5904128551483154,
407
+ "logps/chosen": -261.8894958496094,
408
+ "logps/rejected": -209.9693145751953,
409
+ "loss": 0.3216,
410
+ "rewards/accuracies": 0.8812500238418579,
411
+ "rewards/chosen": 2.382293224334717,
412
+ "rewards/margins": 6.440347194671631,
413
+ "rewards/rejected": -4.058054447174072,
414
  "step": 260
415
  },
416
  {
417
+ "epoch": 0.69,
418
+ "grad_norm": 423.81065150105087,
419
+ "learning_rate": 1.30857819994673e-07,
420
+ "logits/chosen": -2.626788377761841,
421
+ "logits/rejected": -2.541728973388672,
422
+ "logps/chosen": -270.8172912597656,
423
+ "logps/rejected": -231.0417022705078,
424
+ "loss": 0.4168,
425
+ "rewards/accuracies": 0.90625,
426
+ "rewards/chosen": 3.459994077682495,
427
+ "rewards/margins": 9.697509765625,
428
+ "rewards/rejected": -6.237515926361084,
429
  "step": 270
430
  },
431
  {
432
+ "epoch": 0.72,
433
+ "grad_norm": 328.5138374152781,
434
+ "learning_rate": 1.116893898236716e-07,
435
+ "logits/chosen": -2.656203508377075,
436
+ "logits/rejected": -2.6001851558685303,
437
+ "logps/chosen": -269.64996337890625,
438
+ "logps/rejected": -220.4279327392578,
439
+ "loss": 0.3201,
440
+ "rewards/accuracies": 0.875,
441
+ "rewards/chosen": 2.4338252544403076,
442
+ "rewards/margins": 6.923757076263428,
443
+ "rewards/rejected": -4.489931583404541,
444
  "step": 280
445
  },
446
  {
447
+ "epoch": 0.74,
448
+ "grad_norm": 382.1224954400422,
449
+ "learning_rate": 9.362822335518062e-08,
450
+ "logits/chosen": -2.6185498237609863,
451
+ "logits/rejected": -2.572524309158325,
452
+ "logps/chosen": -267.46600341796875,
453
+ "logps/rejected": -217.64089965820312,
454
+ "loss": 0.319,
455
+ "rewards/accuracies": 0.875,
456
+ "rewards/chosen": 2.6880507469177246,
457
+ "rewards/margins": 6.601712703704834,
458
+ "rewards/rejected": -3.9136624336242676,
459
  "step": 290
460
  },
461
  {
462
+ "epoch": 0.77,
463
+ "grad_norm": 546.200687023906,
464
+ "learning_rate": 7.681891162260015e-08,
465
+ "logits/chosen": -2.6388938426971436,
466
+ "logits/rejected": -2.5842864513397217,
467
+ "logps/chosen": -273.81378173828125,
468
+ "logps/rejected": -221.57211303710938,
469
+ "loss": 0.3443,
470
+ "rewards/accuracies": 0.8812500238418579,
471
+ "rewards/chosen": 2.8565659523010254,
472
+ "rewards/margins": 6.786903381347656,
473
+ "rewards/rejected": -3.9303371906280518,
474
  "step": 300
475
  },
476
  {
477
+ "epoch": 0.79,
478
+ "grad_norm": 307.4415182886001,
479
+ "learning_rate": 6.139602377230247e-08,
480
+ "logits/chosen": -2.604128360748291,
481
+ "logits/rejected": -2.5361759662628174,
482
+ "logps/chosen": -277.45074462890625,
483
+ "logps/rejected": -216.7763671875,
484
+ "loss": 0.3778,
485
+ "rewards/accuracies": 0.8687499761581421,
486
+ "rewards/chosen": 2.8939478397369385,
487
+ "rewards/margins": 7.20609188079834,
488
+ "rewards/rejected": -4.3121442794799805,
489
  "step": 310
490
  },
491
  {
492
+ "epoch": 0.82,
493
+ "grad_norm": 438.0667457692161,
494
+ "learning_rate": 4.748302975270837e-08,
495
+ "logits/chosen": -2.6282718181610107,
496
+ "logits/rejected": -2.5831058025360107,
497
+ "logps/chosen": -260.8962707519531,
498
+ "logps/rejected": -205.3025665283203,
499
+ "loss": 0.3343,
500
+ "rewards/accuracies": 0.8999999761581421,
501
+ "rewards/chosen": 2.4093146324157715,
502
+ "rewards/margins": 6.292300701141357,
503
+ "rewards/rejected": -3.882986068725586,
504
  "step": 320
505
  },
506
  {
507
+ "epoch": 0.84,
508
+ "grad_norm": 997.9385970550169,
509
+ "learning_rate": 3.5191311859445795e-08,
510
+ "logits/chosen": -2.6487433910369873,
511
+ "logits/rejected": -2.5906193256378174,
512
+ "logps/chosen": -264.0924377441406,
513
+ "logps/rejected": -218.33743286132812,
514
+ "loss": 0.3351,
515
+ "rewards/accuracies": 0.8999999761581421,
516
+ "rewards/chosen": 3.3637747764587402,
517
+ "rewards/margins": 6.7286200523376465,
518
+ "rewards/rejected": -3.364844799041748,
519
  "step": 330
520
  },
521
  {
522
+ "epoch": 0.87,
523
+ "grad_norm": 381.15409194754363,
524
+ "learning_rate": 2.4619273049795996e-08,
525
+ "logits/chosen": -2.6333811283111572,
526
+ "logits/rejected": -2.577167510986328,
527
+ "logps/chosen": -259.5494689941406,
528
+ "logps/rejected": -211.8303680419922,
529
+ "loss": 0.2972,
530
+ "rewards/accuracies": 0.9125000238418579,
531
+ "rewards/chosen": 3.1570422649383545,
532
+ "rewards/margins": 7.332770347595215,
533
+ "rewards/rejected": -4.175727844238281,
534
  "step": 340
535
  },
536
  {
537
+ "epoch": 0.9,
538
+ "grad_norm": 832.2502586276644,
539
+ "learning_rate": 1.5851549164932115e-08,
540
+ "logits/chosen": -2.6442418098449707,
541
+ "logits/rejected": -2.595158576965332,
542
+ "logps/chosen": -269.0829162597656,
543
+ "logps/rejected": -227.48281860351562,
544
+ "loss": 0.3277,
545
+ "rewards/accuracies": 0.8812500238418579,
546
+ "rewards/chosen": 3.1007628440856934,
547
+ "rewards/margins": 6.698919773101807,
548
+ "rewards/rejected": -3.598156452178955,
549
  "step": 350
550
  },
551
  {
552
+ "epoch": 0.92,
553
+ "grad_norm": 372.63658491482676,
554
+ "learning_rate": 8.958331366609423e-09,
555
+ "logits/chosen": -2.6463985443115234,
556
+ "logits/rejected": -2.5795321464538574,
557
+ "logps/chosen": -274.5993957519531,
558
+ "logps/rejected": -220.4040069580078,
559
+ "loss": 0.3835,
560
  "rewards/accuracies": 0.90625,
561
+ "rewards/chosen": 2.770512819290161,
562
+ "rewards/margins": 6.775577545166016,
563
+ "rewards/rejected": -4.005064964294434,
564
  "step": 360
565
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
566
  {
567
  "epoch": 0.95,
568
+ "grad_norm": 695.4821075832409,
569
+ "learning_rate": 3.994804212627461e-09,
570
+ "logits/chosen": -2.6053659915924072,
571
+ "logits/rejected": -2.5696444511413574,
572
+ "logps/chosen": -273.0481262207031,
573
+ "logps/rejected": -229.8125762939453,
574
+ "loss": 0.3476,
575
+ "rewards/accuracies": 0.8687499761581421,
576
+ "rewards/chosen": 3.7336528301239014,
577
+ "rewards/margins": 6.8774590492248535,
578
+ "rewards/rejected": -3.143805980682373,
579
+ "step": 370
580
  },
581
  {
582
  "epoch": 0.97,
583
+ "grad_norm": 284.5109598747135,
584
+ "learning_rate": 1.0007038696262516e-09,
585
+ "logits/chosen": -2.65440034866333,
586
+ "logits/rejected": -2.61323618888855,
587
+ "logps/chosen": -262.66998291015625,
588
+ "logps/rejected": -231.3791961669922,
589
+ "loss": 0.3209,
590
+ "rewards/accuracies": 0.875,
591
+ "rewards/chosen": 3.4292311668395996,
592
+ "rewards/margins": 6.774644374847412,
593
+ "rewards/rejected": -3.3454127311706543,
594
+ "step": 380
595
  },
596
  {
597
  "epoch": 1.0,
598
+ "grad_norm": 490.88652867728223,
599
  "learning_rate": 0.0,
600
+ "logits/chosen": -2.6619739532470703,
601
+ "logits/rejected": -2.6010966300964355,
602
+ "logps/chosen": -250.47537231445312,
603
+ "logps/rejected": -211.33810424804688,
604
+ "loss": 0.3426,
605
+ "rewards/accuracies": 0.862500011920929,
606
+ "rewards/chosen": 2.711453914642334,
607
+ "rewards/margins": 6.628802299499512,
608
+ "rewards/rejected": -3.917348861694336,
609
+ "step": 390
610
  },
611
  {
612
  "epoch": 1.0,
613
+ "step": 390,
614
  "total_flos": 0.0,
615
+ "train_loss": 0.36373490798167696,
616
+ "train_runtime": 5835.088,
617
+ "train_samples_per_second": 8.569,
618
+ "train_steps_per_second": 0.067
619
  }
620
  ],
621
  "logging_steps": 10,
622
+ "max_steps": 390,
623
  "num_input_tokens_seen": 0,
624
  "num_train_epochs": 1,
625
  "save_steps": 100,
626
  "total_flos": 0.0,
627
+ "train_batch_size": 4,
628
  "trial_name": null,
629
  "trial_params": null
630
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c46a043ab303535917560fd89542c405187453fe88a04c8c789ec363f76563bc
3
- size 6328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:527970ec9a0d5db7ce1f62ba6ec3a4699cc73881b2dacbb36f612ad48ab1823d
3
+ size 6264