RikkiXu commited on
Commit
2868cb6
1 Parent(s): e63789a

Model save

Browse files
README.md CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # zephyr-7b-dpo-full
15
 
16
- This model was trained from scratch on the None dataset.
17
 
18
  ## Model description
19
 
@@ -52,7 +52,7 @@ The following hyperparameters were used during training:
52
 
53
  ### Framework versions
54
 
55
- - Transformers 4.41.1
56
  - Pytorch 2.1.2+cu118
57
- - Datasets 2.16.1
58
- - Tokenizers 0.19.1
 
13
 
14
  # zephyr-7b-dpo-full
15
 
16
+ This model was trained from scratch on an unknown dataset.
17
 
18
  ## Model description
19
 
 
52
 
53
  ### Framework versions
54
 
55
+ - Transformers 4.39.3
56
  - Pytorch 2.1.2+cu118
57
+ - Datasets 2.19.1
58
+ - Tokenizers 0.15.2
all_results.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "epoch": 0.9980806142034548,
3
- "total_flos": 0.0,
4
- "train_loss": 0.2947830175742125,
5
- "train_runtime": 6391.2384,
6
  "train_samples": 50000,
7
- "train_samples_per_second": 7.823,
8
- "train_steps_per_second": 0.061
9
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.19850983711389394,
4
+ "train_runtime": 5891.9709,
 
5
  "train_samples": 50000,
6
+ "train_samples_per_second": 8.486,
7
+ "train_steps_per_second": 0.066
8
  }
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.41.1",
24
  "use_cache": false,
25
  "vocab_size": 32000
26
  }
 
20
  "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.3",
24
  "use_cache": false,
25
  "vocab_size": 32000
26
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.41.1"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.39.3"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c879c438925d1255e6bb62e95443c7b9be4c54b554c5661246d16917a1112924
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54769bdbecf4786a63d2238e3ee62adeef4bf0b3137da2a6064f572ac502b5ba
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c2dd72f78dcf556e6528d803c589c6105e3a837532325d354a805e21db0d419
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afed13f2fb9196ba7a439af3fcd67e27fa09aae99b793638e19395a092d9769c
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a55211455915c4c6cc8aa0486c54d0d029dd088874b044c923db3b11eb37773
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c7933f77356443d56777d8b1904dbda1adb1d3912d8ea0bbc0ed445713ff355
3
  size 4540516344
runs/Jun22_03-50-02_n136-112-146/events.out.tfevents.1719000110.n136-112-146.3291321.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c93e77188d1967c15a538723893f88d78684fcc7088c9505aba293d7dd533805
3
- size 26098
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbb5b59084b4db0c5f28e66e720f82f6617c2874d0658cfef0fd6b7c4dfc8342
3
+ size 32644
tokenizer.json CHANGED
@@ -134,7 +134,6 @@
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
137
- "ignore_merges": false,
138
  "vocab": {
139
  "<unk>": 0,
140
  "<s>": 1,
 
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
 
137
  "vocab": {
138
  "<unk>": 0,
139
  "<s>": 1,
train_results.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "epoch": 0.9980806142034548,
3
- "total_flos": 0.0,
4
- "train_loss": 0.2947830175742125,
5
- "train_runtime": 6391.2384,
6
  "train_samples": 50000,
7
- "train_samples_per_second": 7.823,
8
- "train_steps_per_second": 0.061
9
  }
 
1
  {
2
+ "epoch": 1.0,
3
+ "train_loss": 0.19850983711389394,
4
+ "train_runtime": 5891.9709,
 
5
  "train_samples": 50000,
6
+ "train_samples_per_second": 8.486,
7
+ "train_steps_per_second": 0.066
8
  }
trainer_state.json CHANGED
@@ -9,13 +9,13 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0025591810620601407,
13
- "grad_norm": 709.6319834867974,
14
  "learning_rate": 5.128205128205128e-09,
15
- "logits/chosen": -2.5583817958831787,
16
- "logits/rejected": -2.4487552642822266,
17
  "logps/chosen": -258.1644592285156,
18
- "logps/rejected": -216.25729370117188,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
@@ -24,598 +24,598 @@
24
  "step": 1
25
  },
26
  {
27
- "epoch": 0.025591810620601407,
28
- "grad_norm": 670.9725237886468,
29
  "learning_rate": 5.128205128205127e-08,
30
- "logits/chosen": -2.606004238128662,
31
- "logits/rejected": -2.553126573562622,
32
- "logps/chosen": -267.5910339355469,
33
- "logps/rejected": -217.67140197753906,
34
- "loss": 0.6976,
35
- "rewards/accuracies": 0.4513888955116272,
36
- "rewards/chosen": -0.013270225375890732,
37
- "rewards/margins": 0.0032119054812937975,
38
- "rewards/rejected": -0.01648213155567646,
39
  "step": 10
40
  },
41
  {
42
- "epoch": 0.05118362124120281,
43
- "grad_norm": 590.5735450897452,
44
  "learning_rate": 1.0256410256410255e-07,
45
- "logits/chosen": -2.6245453357696533,
46
- "logits/rejected": -2.562206745147705,
47
- "logps/chosen": -261.142333984375,
48
- "logps/rejected": -207.0738067626953,
49
- "loss": 0.6676,
50
- "rewards/accuracies": 0.6312500238418579,
51
- "rewards/chosen": 0.08092932403087616,
52
- "rewards/margins": 0.09077399969100952,
53
- "rewards/rejected": -0.009844672866165638,
54
  "step": 20
55
  },
56
  {
57
- "epoch": 0.07677543186180422,
58
- "grad_norm": 644.2505273786248,
59
  "learning_rate": 1.5384615384615385e-07,
60
- "logits/chosen": -2.6289420127868652,
61
- "logits/rejected": -2.5566790103912354,
62
- "logps/chosen": -253.9628143310547,
63
- "logps/rejected": -198.890380859375,
64
- "loss": 0.5156,
65
- "rewards/accuracies": 0.800000011920929,
66
- "rewards/chosen": 0.5922147035598755,
67
- "rewards/margins": 0.6031575798988342,
68
- "rewards/rejected": -0.010942881926894188,
69
  "step": 30
70
  },
71
  {
72
- "epoch": 0.10236724248240563,
73
- "grad_norm": 428.1589759637874,
74
  "learning_rate": 1.999959945379852e-07,
75
- "logits/chosen": -2.637456178665161,
76
- "logits/rejected": -2.5609130859375,
77
- "logps/chosen": -247.48666381835938,
78
- "logps/rejected": -193.59280395507812,
79
- "loss": 0.3866,
80
- "rewards/accuracies": 0.8062499761581421,
81
- "rewards/chosen": 1.4313427209854126,
82
- "rewards/margins": 1.3846019506454468,
83
- "rewards/rejected": 0.04674070328474045,
84
  "step": 40
85
  },
86
  {
87
- "epoch": 0.12795905310300704,
88
- "grad_norm": 366.0325785868742,
89
  "learning_rate": 1.9951572723309917e-07,
90
- "logits/chosen": -2.684983730316162,
91
- "logits/rejected": -2.618076801300049,
92
- "logps/chosen": -262.3794250488281,
93
- "logps/rejected": -204.0382843017578,
94
- "loss": 0.3488,
95
- "rewards/accuracies": 0.8500000238418579,
96
- "rewards/chosen": 2.9288697242736816,
97
- "rewards/margins": 2.3735203742980957,
98
- "rewards/rejected": 0.5553494095802307,
99
  "step": 50
100
  },
101
  {
102
- "epoch": 0.15355086372360843,
103
- "grad_norm": 375.0458504104559,
104
  "learning_rate": 1.9823877374156647e-07,
105
- "logits/chosen": -2.669403553009033,
106
- "logits/rejected": -2.6018548011779785,
107
- "logps/chosen": -262.41534423828125,
108
- "logps/rejected": -206.5387420654297,
109
- "loss": 0.311,
110
- "rewards/accuracies": 0.824999988079071,
111
- "rewards/chosen": 3.2061855792999268,
112
- "rewards/margins": 2.446310520172119,
113
- "rewards/rejected": 0.7598745822906494,
114
  "step": 60
115
  },
116
  {
117
- "epoch": 0.17914267434420986,
118
- "grad_norm": 371.9136724850062,
119
  "learning_rate": 1.9617535688178958e-07,
120
- "logits/chosen": -2.6535415649414062,
121
- "logits/rejected": -2.590367078781128,
122
- "logps/chosen": -290.74871826171875,
123
- "logps/rejected": -224.1929168701172,
124
- "loss": 0.3051,
125
- "rewards/accuracies": 0.875,
126
- "rewards/chosen": 3.5809733867645264,
127
- "rewards/margins": 3.0821049213409424,
128
- "rewards/rejected": 0.4988683760166168,
129
  "step": 70
130
  },
131
  {
132
- "epoch": 0.20473448496481125,
133
- "grad_norm": 381.95411975681657,
134
  "learning_rate": 1.9334199560765839e-07,
135
- "logits/chosen": -2.6085634231567383,
136
- "logits/rejected": -2.5423429012298584,
137
- "logps/chosen": -260.9337463378906,
138
- "logps/rejected": -203.12725830078125,
139
- "loss": 0.2569,
140
- "rewards/accuracies": 0.8687499761581421,
141
- "rewards/chosen": 3.3122222423553467,
142
- "rewards/margins": 3.3377063274383545,
143
- "rewards/rejected": -0.025484371930360794,
144
  "step": 80
145
  },
146
  {
147
- "epoch": 0.23032629558541268,
148
- "grad_norm": 331.9831865016541,
149
  "learning_rate": 1.897613727639014e-07,
150
- "logits/chosen": -2.610903739929199,
151
- "logits/rejected": -2.556959867477417,
152
- "logps/chosen": -256.61138916015625,
153
- "logps/rejected": -215.09701538085938,
154
- "loss": 0.2658,
155
- "rewards/accuracies": 0.8687499761581421,
156
- "rewards/chosen": 2.5669326782226562,
157
- "rewards/margins": 2.867710590362549,
158
- "rewards/rejected": -0.3007778227329254,
159
  "step": 90
160
  },
161
  {
162
- "epoch": 0.2559181062060141,
163
- "grad_norm": 634.1345493995642,
164
  "learning_rate": 1.8546215349560202e-07,
165
- "logits/chosen": -2.6384997367858887,
166
- "logits/rejected": -2.563520908355713,
167
- "logps/chosen": -236.5784912109375,
168
- "logps/rejected": -199.6844940185547,
169
- "loss": 0.3117,
170
- "rewards/accuracies": 0.8500000238418579,
171
- "rewards/chosen": 2.7952113151550293,
172
- "rewards/margins": 3.302074909210205,
173
- "rewards/rejected": -0.5068637132644653,
174
  "step": 100
175
  },
176
  {
177
- "epoch": 0.28150991682661547,
178
- "grad_norm": 319.85079714910466,
179
  "learning_rate": 1.8047875576562553e-07,
180
- "logits/chosen": -2.630192518234253,
181
- "logits/rejected": -2.549891710281372,
182
- "logps/chosen": -258.26971435546875,
183
- "logps/rejected": -208.4285430908203,
184
- "loss": 0.282,
185
- "rewards/accuracies": 0.862500011920929,
186
- "rewards/chosen": 3.702361583709717,
187
- "rewards/margins": 3.704643726348877,
188
- "rewards/rejected": -0.002281466033309698,
189
  "step": 110
190
  },
191
  {
192
- "epoch": 0.30710172744721687,
193
- "grad_norm": 376.18943633320504,
194
  "learning_rate": 1.748510748171101e-07,
195
- "logits/chosen": -2.6130287647247314,
196
- "logits/rejected": -2.543454170227051,
197
- "logps/chosen": -262.194580078125,
198
- "logps/rejected": -206.72488403320312,
199
- "loss": 0.2802,
200
- "rewards/accuracies": 0.8812500238418579,
201
- "rewards/chosen": 3.633894443511963,
202
- "rewards/margins": 3.8131375312805176,
203
- "rewards/rejected": -0.17924347519874573,
204
  "step": 120
205
  },
206
  {
207
- "epoch": 0.3326935380678183,
208
- "grad_norm": 258.08224292776333,
209
  "learning_rate": 1.6862416378687336e-07,
210
- "logits/chosen": -2.587902069091797,
211
- "logits/rejected": -2.5218377113342285,
212
- "logps/chosen": -250.2627410888672,
213
- "logps/rejected": -193.98731994628906,
214
- "loss": 0.2644,
215
- "rewards/accuracies": 0.856249988079071,
216
- "rewards/chosen": 3.5202510356903076,
217
- "rewards/margins": 3.647292375564575,
218
- "rewards/rejected": -0.12704159319400787,
219
  "step": 130
220
  },
221
  {
222
- "epoch": 0.3582853486884197,
223
- "grad_norm": 312.9816016514414,
224
  "learning_rate": 1.6184787302662547e-07,
225
- "logits/chosen": -2.6076598167419434,
226
- "logits/rejected": -2.539602279663086,
227
- "logps/chosen": -267.7660217285156,
228
- "logps/rejected": -213.330078125,
229
- "loss": 0.3055,
230
- "rewards/accuracies": 0.875,
231
- "rewards/chosen": 3.705143690109253,
232
- "rewards/margins": 3.5916450023651123,
233
- "rewards/rejected": 0.11349865049123764,
234
  "step": 140
235
  },
236
  {
237
- "epoch": 0.3838771593090211,
238
- "grad_norm": 298.63492430036257,
239
  "learning_rate": 1.5457645101945046e-07,
240
- "logits/chosen": -2.588223457336426,
241
- "logits/rejected": -2.521327495574951,
242
- "logps/chosen": -262.39776611328125,
243
- "logps/rejected": -210.2900848388672,
244
- "loss": 0.2675,
245
- "rewards/accuracies": 0.887499988079071,
246
- "rewards/chosen": 4.096975803375244,
247
- "rewards/margins": 4.044339656829834,
248
- "rewards/rejected": 0.05263558775186539,
249
  "step": 150
250
  },
251
  {
252
- "epoch": 0.4094689699296225,
253
- "grad_norm": 277.3876418755778,
254
  "learning_rate": 1.4686811008647035e-07,
255
- "logits/chosen": -2.5860092639923096,
256
- "logits/rejected": -2.5171661376953125,
257
- "logps/chosen": -241.29501342773438,
258
- "logps/rejected": -192.94955444335938,
259
- "loss": 0.2544,
260
- "rewards/accuracies": 0.887499988079071,
261
- "rewards/chosen": 3.8214523792266846,
262
- "rewards/margins": 3.785486936569214,
263
- "rewards/rejected": 0.0359656922519207,
264
  "step": 160
265
  },
266
  {
267
- "epoch": 0.4350607805502239,
268
- "grad_norm": 339.61682629601955,
269
  "learning_rate": 1.387845603604855e-07,
270
- "logits/chosen": -2.5826542377471924,
271
- "logits/rejected": -2.51751446723938,
272
- "logps/chosen": -262.99298095703125,
273
- "logps/rejected": -206.5457000732422,
274
- "loss": 0.2334,
275
- "rewards/accuracies": 0.875,
276
- "rewards/chosen": 3.8493824005126953,
277
- "rewards/margins": 3.933476209640503,
278
- "rewards/rejected": -0.08409398794174194,
279
  "step": 170
280
  },
281
  {
282
- "epoch": 0.46065259117082535,
283
- "grad_norm": 293.5783730044579,
284
  "learning_rate": 1.3039051575742468e-07,
285
- "logits/chosen": -2.6330299377441406,
286
- "logits/rejected": -2.547426700592041,
287
- "logps/chosen": -235.4442596435547,
288
- "logps/rejected": -183.63687133789062,
289
- "loss": 0.2871,
290
- "rewards/accuracies": 0.862500011920929,
291
- "rewards/chosen": 3.294564723968506,
292
- "rewards/margins": 3.4877796173095703,
293
- "rewards/rejected": -0.193215012550354,
294
  "step": 180
295
  },
296
  {
297
- "epoch": 0.48624440179142675,
298
- "grad_norm": 298.0737945063753,
299
  "learning_rate": 1.2175317590061675e-07,
300
- "logits/chosen": -2.5894675254821777,
301
- "logits/rejected": -2.51374888420105,
302
- "logps/chosen": -257.42132568359375,
303
- "logps/rejected": -199.62106323242188,
304
- "loss": 0.239,
305
- "rewards/accuracies": 0.887499988079071,
306
- "rewards/chosen": 3.699636936187744,
307
- "rewards/margins": 3.845207929611206,
308
- "rewards/rejected": -0.1455712914466858,
309
  "step": 190
310
  },
311
  {
312
- "epoch": 0.5118362124120281,
313
- "grad_norm": 326.3771289852159,
314
  "learning_rate": 1.1294168814540553e-07,
315
- "logits/chosen": -2.625087261199951,
316
- "logits/rejected": -2.5350747108459473,
317
- "logps/chosen": -271.9561462402344,
318
- "logps/rejected": -214.1556854248047,
319
- "loss": 0.2317,
320
- "rewards/accuracies": 0.90625,
321
- "rewards/chosen": 4.283473014831543,
322
- "rewards/margins": 4.552158832550049,
323
- "rewards/rejected": -0.2686860263347626,
324
  "step": 200
325
  },
326
  {
327
- "epoch": 0.5374280230326296,
328
- "grad_norm": 510.1343443372893,
329
  "learning_rate": 1.0402659401094151e-07,
330
- "logits/chosen": -2.602785587310791,
331
- "logits/rejected": -2.5288333892822266,
332
- "logps/chosen": -264.21099853515625,
333
- "logps/rejected": -207.3443145751953,
334
- "loss": 0.3153,
335
- "rewards/accuracies": 0.856249988079071,
336
- "rewards/chosen": 3.421428680419922,
337
- "rewards/margins": 3.9131996631622314,
338
- "rewards/rejected": -0.4917708933353424,
339
  "step": 210
340
  },
341
  {
342
- "epoch": 0.5630198336532309,
343
- "grad_norm": 362.61099626331173,
344
  "learning_rate": 9.507926445081218e-08,
345
- "logits/chosen": -2.607333183288574,
346
- "logits/rejected": -2.5410573482513428,
347
- "logps/chosen": -255.5386199951172,
348
- "logps/rejected": -210.58261108398438,
349
- "loss": 0.2657,
350
- "rewards/accuracies": 0.893750011920929,
351
- "rewards/chosen": 3.069124698638916,
352
- "rewards/margins": 3.4770140647888184,
353
- "rewards/rejected": -0.4078896641731262,
354
  "step": 220
355
  },
356
  {
357
- "epoch": 0.5886116442738324,
358
- "grad_norm": 270.86067735708986,
359
  "learning_rate": 8.61713284835267e-08,
360
- "logits/chosen": -2.6318671703338623,
361
- "logits/rejected": -2.5431408882141113,
362
- "logps/chosen": -265.3260803222656,
363
- "logps/rejected": -212.6390380859375,
364
- "loss": 0.249,
365
- "rewards/accuracies": 0.8500000238418579,
366
- "rewards/chosen": 4.038520812988281,
367
- "rewards/margins": 3.9009087085723877,
368
- "rewards/rejected": 0.1376120150089264,
369
  "step": 230
370
  },
371
  {
372
- "epoch": 0.6142034548944337,
373
- "grad_norm": 496.26361322510223,
374
  "learning_rate": 7.73740997570278e-08,
375
- "logits/chosen": -2.6172757148742676,
376
- "logits/rejected": -2.566117763519287,
377
- "logps/chosen": -259.39569091796875,
378
- "logps/rejected": -214.13717651367188,
379
- "loss": 0.2362,
380
- "rewards/accuracies": 0.8999999761581421,
381
- "rewards/chosen": 4.167778968811035,
382
- "rewards/margins": 4.450715065002441,
383
- "rewards/rejected": -0.28293582797050476,
384
  "step": 240
385
  },
386
  {
387
- "epoch": 0.6397952655150352,
388
- "grad_norm": 385.39306248287744,
389
  "learning_rate": 6.875800563794424e-08,
390
- "logits/chosen": -2.6301958560943604,
391
- "logits/rejected": -2.5562362670898438,
392
- "logps/chosen": -264.5022277832031,
393
- "logps/rejected": -217.06472778320312,
394
- "loss": 0.2541,
395
- "rewards/accuracies": 0.862500011920929,
396
- "rewards/chosen": 3.638233184814453,
397
- "rewards/margins": 3.9798378944396973,
398
- "rewards/rejected": -0.34160494804382324,
399
  "step": 250
400
  },
401
  {
402
- "epoch": 0.6653870761356366,
403
- "grad_norm": 246.0647695237683,
404
  "learning_rate": 6.039202339608431e-08,
405
- "logits/chosen": -2.6222901344299316,
406
- "logits/rejected": -2.5577120780944824,
407
- "logps/chosen": -258.84637451171875,
408
- "logps/rejected": -203.7962188720703,
409
- "loss": 0.2317,
410
- "rewards/accuracies": 0.893750011920929,
411
- "rewards/chosen": 3.5067996978759766,
412
- "rewards/margins": 3.801957607269287,
413
- "rewards/rejected": -0.295158326625824,
414
  "step": 260
415
  },
416
  {
417
- "epoch": 0.690978886756238,
418
- "grad_norm": 382.6551210350461,
419
  "learning_rate": 5.2343127997869205e-08,
420
- "logits/chosen": -2.602179765701294,
421
- "logits/rejected": -2.5085558891296387,
422
- "logps/chosen": -270.6271057128906,
423
- "logps/rejected": -222.81655883789062,
424
- "loss": 0.2926,
425
- "rewards/accuracies": 0.8999999761581421,
426
- "rewards/chosen": 2.9784107208251953,
427
- "rewards/margins": 4.063758850097656,
428
- "rewards/rejected": -1.085348129272461,
429
  "step": 270
430
  },
431
  {
432
- "epoch": 0.7165706973768394,
433
- "grad_norm": 291.15530026271324,
434
  "learning_rate": 4.4675755929468636e-08,
435
- "logits/chosen": -2.627225160598755,
436
- "logits/rejected": -2.565744638442993,
437
- "logps/chosen": -266.7916259765625,
438
- "logps/rejected": -214.30709838867188,
439
- "loss": 0.2375,
440
- "rewards/accuracies": 0.90625,
441
- "rewards/chosen": 3.4573512077331543,
442
- "rewards/margins": 4.138546943664551,
443
- "rewards/rejected": -0.6811951398849487,
444
  "step": 280
445
  },
446
  {
447
- "epoch": 0.7421625079974408,
448
- "grad_norm": 302.9770657235692,
449
  "learning_rate": 3.745128934207224e-08,
450
- "logits/chosen": -2.590620279312134,
451
- "logits/rejected": -2.5388922691345215,
452
- "logps/chosen": -265.51971435546875,
453
- "logps/rejected": -212.6427764892578,
454
- "loss": 0.2286,
455
- "rewards/accuracies": 0.893750011920929,
456
- "rewards/chosen": 3.2131683826446533,
457
- "rewards/margins": 3.975499391555786,
458
- "rewards/rejected": -0.7623313665390015,
459
  "step": 290
460
  },
461
  {
462
- "epoch": 0.7677543186180422,
463
- "grad_norm": 339.1306333767976,
464
  "learning_rate": 3.0727564649040063e-08,
465
- "logits/chosen": -2.6111063957214355,
466
- "logits/rejected": -2.5518836975097656,
467
- "logps/chosen": -271.9001159667969,
468
- "logps/rejected": -216.2710723876953,
469
- "loss": 0.2446,
470
- "rewards/accuracies": 0.9125000238418579,
471
- "rewards/chosen": 3.3372998237609863,
472
- "rewards/margins": 3.9620718955993652,
473
- "rewards/rejected": -0.6247718930244446,
474
  "step": 300
475
  },
476
  {
477
- "epoch": 0.7933461292386437,
478
- "grad_norm": 235.34557136252968,
479
  "learning_rate": 2.4558409508920985e-08,
480
- "logits/chosen": -2.5742857456207275,
481
- "logits/rejected": -2.499851703643799,
482
- "logps/chosen": -275.28167724609375,
483
- "logps/rejected": -211.0914764404297,
484
- "loss": 0.2598,
485
- "rewards/accuracies": 0.862500011920929,
486
- "rewards/chosen": 3.4961440563201904,
487
- "rewards/margins": 4.247152805328369,
488
- "rewards/rejected": -0.7510083913803101,
489
  "step": 310
490
  },
491
  {
492
- "epoch": 0.818937939859245,
493
- "grad_norm": 368.2414347720992,
494
  "learning_rate": 1.899321190108335e-08,
495
- "logits/chosen": -2.59656023979187,
496
- "logits/rejected": -2.545151472091675,
497
- "logps/chosen": -258.7973327636719,
498
- "logps/rejected": -200.56219482421875,
499
- "loss": 0.2329,
500
- "rewards/accuracies": 0.8999999761581421,
501
- "rewards/chosen": 3.0572078227996826,
502
- "rewards/margins": 3.9228408336639404,
503
- "rewards/rejected": -0.8656327128410339,
504
  "step": 320
505
  },
506
  {
507
- "epoch": 0.8445297504798465,
508
- "grad_norm": 412.2800478150093,
509
  "learning_rate": 1.4076524743778317e-08,
510
- "logits/chosen": -2.6195101737976074,
511
- "logits/rejected": -2.55584716796875,
512
- "logps/chosen": -262.2277526855469,
513
- "logps/rejected": -213.35403442382812,
514
- "loss": 0.231,
515
- "rewards/accuracies": 0.925000011920929,
516
- "rewards/chosen": 3.7354748249053955,
517
- "rewards/margins": 4.0478105545043945,
518
- "rewards/rejected": -0.31233564019203186,
519
  "step": 330
520
  },
521
  {
522
- "epoch": 0.8701215611004478,
523
- "grad_norm": 291.1936087568274,
524
  "learning_rate": 9.847709219918398e-09,
525
- "logits/chosen": -2.6071531772613525,
526
- "logits/rejected": -2.5460381507873535,
527
- "logps/chosen": -257.75164794921875,
528
- "logps/rejected": -206.60986328125,
529
- "loss": 0.221,
530
- "rewards/accuracies": 0.949999988079071,
531
- "rewards/chosen": 3.5297904014587402,
532
- "rewards/margins": 4.399306297302246,
533
- "rewards/rejected": -0.8695155382156372,
534
  "step": 340
535
  },
536
  {
537
- "epoch": 0.8957133717210493,
538
- "grad_norm": 301.99507828681374,
539
  "learning_rate": 6.340619665972846e-09,
540
- "logits/chosen": -2.617502212524414,
541
- "logits/rejected": -2.5671114921569824,
542
- "logps/chosen": -267.4635009765625,
543
- "logps/rejected": -222.5786895751953,
544
- "loss": 0.2579,
545
- "rewards/accuracies": 0.8687499761581421,
546
- "rewards/chosen": 3.393681287765503,
547
- "rewards/margins": 3.9400668144226074,
548
- "rewards/rejected": -0.546385645866394,
549
  "step": 350
550
  },
551
  {
552
- "epoch": 0.9213051823416507,
553
- "grad_norm": 348.99526280716225,
554
  "learning_rate": 3.583332546643769e-09,
555
- "logits/chosen": -2.6180646419525146,
556
- "logits/rejected": -2.5445971488952637,
557
- "logps/chosen": -272.35406494140625,
558
- "logps/rejected": -214.8975067138672,
559
- "loss": 0.2825,
560
- "rewards/accuracies": 0.893750011920929,
561
- "rewards/chosen": 3.431410551071167,
562
- "rewards/margins": 4.0157270431518555,
563
- "rewards/rejected": -0.584316074848175,
564
  "step": 360
565
  },
566
  {
567
- "epoch": 0.946896992962252,
568
- "grad_norm": 251.0460885716357,
569
  "learning_rate": 1.5979216850509847e-09,
570
- "logits/chosen": -2.5794007778167725,
571
- "logits/rejected": -2.5391972064971924,
572
- "logps/chosen": -271.82513427734375,
573
- "logps/rejected": -225.1968231201172,
574
- "loss": 0.2529,
575
- "rewards/accuracies": 0.862500011920929,
576
- "rewards/chosen": 3.722885847091675,
577
- "rewards/margins": 4.0348358154296875,
578
- "rewards/rejected": -0.3119499981403351,
579
  "step": 370
580
  },
581
  {
582
- "epoch": 0.9724888035828535,
583
- "grad_norm": 201.79043052952954,
584
  "learning_rate": 4.002815478505006e-10,
585
- "logits/chosen": -2.6285715103149414,
586
- "logits/rejected": -2.585494041442871,
587
- "logps/chosen": -261.1199951171875,
588
- "logps/rejected": -226.61105346679688,
589
- "loss": 0.255,
590
- "rewards/accuracies": 0.8500000238418579,
591
- "rewards/chosen": 3.6326937675476074,
592
- "rewards/margins": 4.036477088928223,
593
- "rewards/rejected": -0.40378379821777344,
594
  "step": 380
595
  },
596
  {
597
- "epoch": 0.9980806142034548,
598
- "grad_norm": 458.24518941114127,
599
  "learning_rate": 0.0,
600
- "logits/chosen": -2.641213893890381,
601
- "logits/rejected": -2.575150966644287,
602
- "logps/chosen": -248.17904663085938,
603
- "logps/rejected": -206.1395721435547,
604
- "loss": 0.2365,
605
- "rewards/accuracies": 0.918749988079071,
606
- "rewards/chosen": 3.4077067375183105,
607
- "rewards/margins": 4.0728960037231445,
608
- "rewards/rejected": -0.6651893854141235,
609
  "step": 390
610
  },
611
  {
612
- "epoch": 0.9980806142034548,
613
  "step": 390,
614
  "total_flos": 0.0,
615
- "train_loss": 0.2947830175742125,
616
- "train_runtime": 6391.2384,
617
- "train_samples_per_second": 7.823,
618
- "train_steps_per_second": 0.061
619
  }
620
  ],
621
  "logging_steps": 10,
@@ -623,18 +623,6 @@
623
  "num_input_tokens_seen": 0,
624
  "num_train_epochs": 1,
625
  "save_steps": 100,
626
- "stateful_callbacks": {
627
- "TrainerControl": {
628
- "args": {
629
- "should_epoch_stop": false,
630
- "should_evaluate": false,
631
- "should_log": false,
632
- "should_save": true,
633
- "should_training_stop": false
634
- },
635
- "attributes": {}
636
- }
637
- },
638
  "total_flos": 0.0,
639
  "train_batch_size": 4,
640
  "trial_name": null,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.0,
13
+ "grad_norm": 747.1989420753074,
14
  "learning_rate": 5.128205128205128e-09,
15
+ "logits/chosen": -2.5617921352386475,
16
+ "logits/rejected": -2.415619373321533,
17
  "logps/chosen": -258.1644592285156,
18
+ "logps/rejected": -191.65736389160156,
19
  "loss": 0.6931,
20
  "rewards/accuracies": 0.0,
21
  "rewards/chosen": 0.0,
 
24
  "step": 1
25
  },
26
  {
27
+ "epoch": 0.03,
28
+ "grad_norm": 772.6555707154646,
29
  "learning_rate": 5.128205128205127e-08,
30
+ "logits/chosen": -2.610959053039551,
31
+ "logits/rejected": -2.52398681640625,
32
+ "logps/chosen": -267.3009948730469,
33
+ "logps/rejected": -198.18966674804688,
34
+ "loss": 0.7062,
35
+ "rewards/accuracies": 0.4444444477558136,
36
+ "rewards/chosen": -0.006153374910354614,
37
+ "rewards/margins": -0.0059446971863508224,
38
+ "rewards/rejected": -0.0002086775202769786,
39
  "step": 10
40
  },
41
  {
42
+ "epoch": 0.05,
43
+ "grad_norm": 568.2747762208306,
44
  "learning_rate": 1.0256410256410255e-07,
45
+ "logits/chosen": -2.626389980316162,
46
+ "logits/rejected": -2.5251834392547607,
47
+ "logps/chosen": -261.1226501464844,
48
+ "logps/rejected": -198.20663452148438,
49
+ "loss": 0.6286,
50
+ "rewards/accuracies": 0.643750011920929,
51
+ "rewards/chosen": 0.1013670414686203,
52
+ "rewards/margins": 0.16386722028255463,
53
+ "rewards/rejected": -0.06250017881393433,
54
  "step": 20
55
  },
56
  {
57
+ "epoch": 0.08,
58
+ "grad_norm": 494.0619878155869,
59
  "learning_rate": 1.5384615384615385e-07,
60
+ "logits/chosen": -2.631648302078247,
61
+ "logits/rejected": -2.533083915710449,
62
+ "logps/chosen": -253.81259155273438,
63
+ "logps/rejected": -194.48611450195312,
64
+ "loss": 0.4093,
65
+ "rewards/accuracies": 0.8374999761581421,
66
+ "rewards/chosen": 0.6733725666999817,
67
+ "rewards/margins": 0.9389022588729858,
68
+ "rewards/rejected": -0.2655297815799713,
69
  "step": 30
70
  },
71
  {
72
+ "epoch": 0.1,
73
+ "grad_norm": 479.5871736251968,
74
  "learning_rate": 1.999959945379852e-07,
75
+ "logits/chosen": -2.6485562324523926,
76
+ "logits/rejected": -2.5283126831054688,
77
+ "logps/chosen": -247.01113891601562,
78
+ "logps/rejected": -198.84341430664062,
79
+ "loss": 0.2427,
80
+ "rewards/accuracies": 0.8812500238418579,
81
+ "rewards/chosen": 1.6728582382202148,
82
+ "rewards/margins": 2.513617753982544,
83
+ "rewards/rejected": -0.84075927734375,
84
  "step": 40
85
  },
86
  {
87
+ "epoch": 0.13,
88
+ "grad_norm": 268.0648126747043,
89
  "learning_rate": 1.9951572723309917e-07,
90
+ "logits/chosen": -2.701251983642578,
91
+ "logits/rejected": -2.596193790435791,
92
+ "logps/chosen": -260.8350830078125,
93
+ "logps/rejected": -204.7702178955078,
94
+ "loss": 0.2435,
95
+ "rewards/accuracies": 0.8999999761581421,
96
+ "rewards/chosen": 3.198915481567383,
97
+ "rewards/margins": 4.101343631744385,
98
+ "rewards/rejected": -0.902428150177002,
99
  "step": 50
100
  },
101
  {
102
+ "epoch": 0.15,
103
+ "grad_norm": 321.17307426963725,
104
  "learning_rate": 1.9823877374156647e-07,
105
+ "logits/chosen": -2.6826422214508057,
106
+ "logits/rejected": -2.5699238777160645,
107
+ "logps/chosen": -260.90740966796875,
108
+ "logps/rejected": -190.84542846679688,
109
+ "loss": 0.2062,
110
+ "rewards/accuracies": 0.918749988079071,
111
+ "rewards/chosen": 3.89800763130188,
112
+ "rewards/margins": 4.676665306091309,
113
+ "rewards/rejected": -0.7786582112312317,
114
  "step": 60
115
  },
116
  {
117
+ "epoch": 0.18,
118
+ "grad_norm": 341.66511985659804,
119
  "learning_rate": 1.9617535688178958e-07,
120
+ "logits/chosen": -2.664588451385498,
121
+ "logits/rejected": -2.5252487659454346,
122
+ "logps/chosen": -289.1197204589844,
123
+ "logps/rejected": -219.9818572998047,
124
+ "loss": 0.1912,
125
+ "rewards/accuracies": 0.925000011920929,
126
+ "rewards/chosen": 4.399308204650879,
127
+ "rewards/margins": 5.727428436279297,
128
+ "rewards/rejected": -1.3281205892562866,
129
  "step": 70
130
  },
131
  {
132
+ "epoch": 0.2,
133
+ "grad_norm": 345.50558992425647,
134
  "learning_rate": 1.9334199560765839e-07,
135
+ "logits/chosen": -2.612795352935791,
136
+ "logits/rejected": -2.4848904609680176,
137
+ "logps/chosen": -260.74462890625,
138
+ "logps/rejected": -209.5861053466797,
139
+ "loss": 0.1745,
140
+ "rewards/accuracies": 0.9125000238418579,
141
+ "rewards/chosen": 3.798893690109253,
142
+ "rewards/margins": 5.829946994781494,
143
+ "rewards/rejected": -2.031052827835083,
144
  "step": 80
145
  },
146
  {
147
+ "epoch": 0.23,
148
+ "grad_norm": 623.1485221220804,
149
  "learning_rate": 1.897613727639014e-07,
150
+ "logits/chosen": -2.6385045051574707,
151
+ "logits/rejected": -2.5179295539855957,
152
+ "logps/chosen": -254.0596160888672,
153
+ "logps/rejected": -191.18374633789062,
154
+ "loss": 0.1808,
155
+ "rewards/accuracies": 0.9125000238418579,
156
+ "rewards/chosen": 3.131239414215088,
157
+ "rewards/margins": 5.626199722290039,
158
+ "rewards/rejected": -2.494959831237793,
159
  "step": 90
160
  },
161
  {
162
+ "epoch": 0.26,
163
+ "grad_norm": 253.52286816658938,
164
  "learning_rate": 1.8546215349560202e-07,
165
+ "logits/chosen": -2.6612112522125244,
166
+ "logits/rejected": -2.533451795578003,
167
+ "logps/chosen": -235.6942138671875,
168
+ "logps/rejected": -196.51657104492188,
169
+ "loss": 0.1619,
170
+ "rewards/accuracies": 0.8999999761581421,
171
+ "rewards/chosen": 3.5456454753875732,
172
+ "rewards/margins": 5.960389614105225,
173
+ "rewards/rejected": -2.4147439002990723,
174
  "step": 100
175
  },
176
  {
177
+ "epoch": 0.28,
178
+ "grad_norm": 266.28422663795834,
179
  "learning_rate": 1.8047875576562553e-07,
180
+ "logits/chosen": -2.6504344940185547,
181
+ "logits/rejected": -2.52449369430542,
182
+ "logps/chosen": -254.90847778320312,
183
+ "logps/rejected": -191.41806030273438,
184
+ "loss": 0.1662,
185
+ "rewards/accuracies": 0.925000011920929,
186
+ "rewards/chosen": 4.884920597076416,
187
+ "rewards/margins": 6.528225898742676,
188
+ "rewards/rejected": -1.6433048248291016,
189
  "step": 110
190
  },
191
  {
192
+ "epoch": 0.31,
193
+ "grad_norm": 248.5491633368023,
194
  "learning_rate": 1.748510748171101e-07,
195
+ "logits/chosen": -2.626223564147949,
196
+ "logits/rejected": -2.531078815460205,
197
+ "logps/chosen": -258.3885803222656,
198
+ "logps/rejected": -210.59378051757812,
199
+ "loss": 0.1688,
200
+ "rewards/accuracies": 0.9437500238418579,
201
+ "rewards/chosen": 4.675450801849365,
202
+ "rewards/margins": 6.811753273010254,
203
+ "rewards/rejected": -2.1363017559051514,
204
  "step": 120
205
  },
206
  {
207
+ "epoch": 0.33,
208
+ "grad_norm": 219.45295559501494,
209
  "learning_rate": 1.6862416378687336e-07,
210
+ "logits/chosen": -2.6003150939941406,
211
+ "logits/rejected": -2.465550661087036,
212
+ "logps/chosen": -249.7939910888672,
213
+ "logps/rejected": -186.2041778564453,
214
+ "loss": 0.1744,
215
+ "rewards/accuracies": 0.9125000238418579,
216
+ "rewards/chosen": 3.695958375930786,
217
+ "rewards/margins": 6.534309387207031,
218
+ "rewards/rejected": -2.838351011276245,
219
  "step": 130
220
  },
221
  {
222
+ "epoch": 0.36,
223
+ "grad_norm": 335.11070283406536,
224
  "learning_rate": 1.6184787302662547e-07,
225
+ "logits/chosen": -2.6305205821990967,
226
+ "logits/rejected": -2.519258975982666,
227
+ "logps/chosen": -269.74249267578125,
228
+ "logps/rejected": -205.24441528320312,
229
+ "loss": 0.1847,
230
+ "rewards/accuracies": 0.918749988079071,
231
+ "rewards/chosen": 3.7317748069763184,
232
+ "rewards/margins": 6.535207271575928,
233
+ "rewards/rejected": -2.8034329414367676,
234
  "step": 140
235
  },
236
  {
237
+ "epoch": 0.38,
238
+ "grad_norm": 237.3062963631657,
239
  "learning_rate": 1.5457645101945046e-07,
240
+ "logits/chosen": -2.613662004470825,
241
+ "logits/rejected": -2.50919771194458,
242
+ "logps/chosen": -262.550537109375,
243
+ "logps/rejected": -209.60781860351562,
244
+ "loss": 0.1987,
245
+ "rewards/accuracies": 0.8999999761581421,
246
+ "rewards/chosen": 4.1123762130737305,
247
+ "rewards/margins": 7.0581793785095215,
248
+ "rewards/rejected": -2.945803165435791,
249
  "step": 150
250
  },
251
  {
252
+ "epoch": 0.41,
253
+ "grad_norm": 781.538478360743,
254
  "learning_rate": 1.4686811008647035e-07,
255
+ "logits/chosen": -2.6011874675750732,
256
+ "logits/rejected": -2.4891130924224854,
257
+ "logps/chosen": -241.5615692138672,
258
+ "logps/rejected": -182.00228881835938,
259
+ "loss": 0.1643,
260
+ "rewards/accuracies": 0.9437500238418579,
261
+ "rewards/chosen": 4.244563102722168,
262
+ "rewards/margins": 6.548022270202637,
263
+ "rewards/rejected": -2.3034589290618896,
264
  "step": 160
265
  },
266
  {
267
+ "epoch": 0.44,
268
+ "grad_norm": 420.43337815116814,
269
  "learning_rate": 1.387845603604855e-07,
270
+ "logits/chosen": -2.6014420986175537,
271
+ "logits/rejected": -2.4862489700317383,
272
+ "logps/chosen": -260.2299499511719,
273
+ "logps/rejected": -197.31021118164062,
274
+ "loss": 0.1451,
275
+ "rewards/accuracies": 0.925000011920929,
276
+ "rewards/chosen": 4.481954574584961,
277
+ "rewards/margins": 6.5617804527282715,
278
+ "rewards/rejected": -2.0798258781433105,
279
  "step": 170
280
  },
281
  {
282
+ "epoch": 0.46,
283
+ "grad_norm": 219.52091746145038,
284
  "learning_rate": 1.3039051575742468e-07,
285
+ "logits/chosen": -2.6457467079162598,
286
+ "logits/rejected": -2.5342042446136475,
287
+ "logps/chosen": -234.87063598632812,
288
+ "logps/rejected": -187.8373565673828,
289
+ "loss": 0.1641,
290
+ "rewards/accuracies": 0.9375,
291
+ "rewards/chosen": 3.4876370429992676,
292
+ "rewards/margins": 5.9574689865112305,
293
+ "rewards/rejected": -2.469831943511963,
294
  "step": 180
295
  },
296
  {
297
+ "epoch": 0.49,
298
+ "grad_norm": 391.30420402591136,
299
  "learning_rate": 1.2175317590061675e-07,
300
+ "logits/chosen": -2.598358154296875,
301
+ "logits/rejected": -2.5016226768493652,
302
+ "logps/chosen": -256.7655334472656,
303
+ "logps/rejected": -203.38717651367188,
304
+ "loss": 0.1704,
305
+ "rewards/accuracies": 0.925000011920929,
306
+ "rewards/chosen": 3.634434223175049,
307
+ "rewards/margins": 6.311253547668457,
308
+ "rewards/rejected": -2.6768198013305664,
309
  "step": 190
310
  },
311
  {
312
+ "epoch": 0.51,
313
+ "grad_norm": 573.9951634797832,
314
  "learning_rate": 1.1294168814540553e-07,
315
+ "logits/chosen": -2.640943765640259,
316
+ "logits/rejected": -2.5037968158721924,
317
+ "logps/chosen": -270.6221923828125,
318
+ "logps/rejected": -199.61119079589844,
319
+ "loss": 0.1514,
320
+ "rewards/accuracies": 0.9375,
321
+ "rewards/chosen": 4.9661078453063965,
322
+ "rewards/margins": 7.715418338775635,
323
+ "rewards/rejected": -2.749310255050659,
324
  "step": 200
325
  },
326
  {
327
+ "epoch": 0.54,
328
+ "grad_norm": 383.73670531019474,
329
  "learning_rate": 1.0402659401094151e-07,
330
+ "logits/chosen": -2.6191225051879883,
331
+ "logits/rejected": -2.5026001930236816,
332
+ "logps/chosen": -262.648681640625,
333
+ "logps/rejected": -207.82363891601562,
334
+ "loss": 0.1405,
335
+ "rewards/accuracies": 0.925000011920929,
336
+ "rewards/chosen": 4.20405912399292,
337
+ "rewards/margins": 7.261610507965088,
338
+ "rewards/rejected": -3.057551383972168,
339
  "step": 210
340
  },
341
  {
342
+ "epoch": 0.56,
343
+ "grad_norm": 271.1779091479927,
344
  "learning_rate": 9.507926445081218e-08,
345
+ "logits/chosen": -2.618332862854004,
346
+ "logits/rejected": -2.5180063247680664,
347
+ "logps/chosen": -255.3096160888672,
348
+ "logps/rejected": -204.3534698486328,
349
+ "loss": 0.1701,
350
+ "rewards/accuracies": 0.949999988079071,
351
+ "rewards/chosen": 3.6050515174865723,
352
+ "rewards/margins": 7.0129852294921875,
353
+ "rewards/rejected": -3.4079346656799316,
354
  "step": 220
355
  },
356
  {
357
+ "epoch": 0.59,
358
+ "grad_norm": 290.4603808348808,
359
  "learning_rate": 8.61713284835267e-08,
360
+ "logits/chosen": -2.646491050720215,
361
+ "logits/rejected": -2.51234769821167,
362
+ "logps/chosen": -264.74420166015625,
363
+ "logps/rejected": -205.0327911376953,
364
+ "loss": 0.1425,
365
+ "rewards/accuracies": 0.918749988079071,
366
+ "rewards/chosen": 4.327627182006836,
367
+ "rewards/margins": 7.188753604888916,
368
+ "rewards/rejected": -2.8611254692077637,
369
  "step": 230
370
  },
371
  {
372
+ "epoch": 0.61,
373
+ "grad_norm": 422.65919601372,
374
  "learning_rate": 7.73740997570278e-08,
375
+ "logits/chosen": -2.6252548694610596,
376
+ "logits/rejected": -2.5125999450683594,
377
+ "logps/chosen": -258.0855712890625,
378
+ "logps/rejected": -207.72213745117188,
379
+ "loss": 0.1341,
380
+ "rewards/accuracies": 0.9312499761581421,
381
+ "rewards/chosen": 4.273736953735352,
382
+ "rewards/margins": 7.429440498352051,
383
+ "rewards/rejected": -3.155702829360962,
384
  "step": 240
385
  },
386
  {
387
+ "epoch": 0.64,
388
+ "grad_norm": 236.70780195109265,
389
  "learning_rate": 6.875800563794424e-08,
390
+ "logits/chosen": -2.637636184692383,
391
+ "logits/rejected": -2.522014856338501,
392
+ "logps/chosen": -264.11163330078125,
393
+ "logps/rejected": -213.34262084960938,
394
+ "loss": 0.1561,
395
+ "rewards/accuracies": 0.9375,
396
+ "rewards/chosen": 3.965393543243408,
397
+ "rewards/margins": 7.122384548187256,
398
+ "rewards/rejected": -3.1569907665252686,
399
  "step": 250
400
  },
401
  {
402
+ "epoch": 0.67,
403
+ "grad_norm": 261.4238126511671,
404
  "learning_rate": 6.039202339608431e-08,
405
+ "logits/chosen": -2.6351547241210938,
406
+ "logits/rejected": -2.5343315601348877,
407
+ "logps/chosen": -257.4027099609375,
408
+ "logps/rejected": -206.6591033935547,
409
+ "loss": 0.1287,
410
+ "rewards/accuracies": 0.949999988079071,
411
+ "rewards/chosen": 4.236881732940674,
412
+ "rewards/margins": 7.065675258636475,
413
+ "rewards/rejected": -2.828793525695801,
414
  "step": 260
415
  },
416
  {
417
+ "epoch": 0.69,
418
+ "grad_norm": 290.68123700235606,
419
  "learning_rate": 5.2343127997869205e-08,
420
+ "logits/chosen": -2.610731840133667,
421
+ "logits/rejected": -2.4807095527648926,
422
+ "logps/chosen": -270.530517578125,
423
+ "logps/rejected": -215.2696533203125,
424
+ "loss": 0.2282,
425
+ "rewards/accuracies": 0.9312499761581421,
426
+ "rewards/chosen": 3.1549553871154785,
427
+ "rewards/margins": 7.109915733337402,
428
+ "rewards/rejected": -3.9549612998962402,
429
  "step": 270
430
  },
431
  {
432
+ "epoch": 0.72,
433
+ "grad_norm": 212.4315878069041,
434
  "learning_rate": 4.4675755929468636e-08,
435
+ "logits/chosen": -2.6439692974090576,
436
+ "logits/rejected": -2.521467447280884,
437
+ "logps/chosen": -266.0245056152344,
438
+ "logps/rejected": -204.48606872558594,
439
+ "loss": 0.139,
440
+ "rewards/accuracies": 0.956250011920929,
441
+ "rewards/chosen": 3.838078737258911,
442
+ "rewards/margins": 7.6062164306640625,
443
+ "rewards/rejected": -3.7681381702423096,
444
  "step": 280
445
  },
446
  {
447
+ "epoch": 0.74,
448
+ "grad_norm": 229.50979085481035,
449
  "learning_rate": 3.745128934207224e-08,
450
+ "logits/chosen": -2.598167657852173,
451
+ "logits/rejected": -2.5218276977539062,
452
+ "logps/chosen": -266.5920104980469,
453
+ "logps/rejected": -222.931884765625,
454
+ "loss": 0.1507,
455
+ "rewards/accuracies": 0.9375,
456
+ "rewards/chosen": 3.6127426624298096,
457
+ "rewards/margins": 6.654023170471191,
458
+ "rewards/rejected": -3.0412800312042236,
459
  "step": 290
460
  },
461
  {
462
+ "epoch": 0.77,
463
+ "grad_norm": 321.80727834811256,
464
  "learning_rate": 3.0727564649040063e-08,
465
+ "logits/chosen": -2.6253504753112793,
466
+ "logits/rejected": -2.5061511993408203,
467
+ "logps/chosen": -270.4463806152344,
468
+ "logps/rejected": -213.0215301513672,
469
+ "loss": 0.141,
470
+ "rewards/accuracies": 0.956250011920929,
471
+ "rewards/chosen": 3.923215389251709,
472
+ "rewards/margins": 7.3881683349609375,
473
+ "rewards/rejected": -3.464953660964966,
474
  "step": 300
475
  },
476
  {
477
+ "epoch": 0.79,
478
+ "grad_norm": 171.80295572629132,
479
  "learning_rate": 2.4558409508920985e-08,
480
+ "logits/chosen": -2.5865883827209473,
481
+ "logits/rejected": -2.491208553314209,
482
+ "logps/chosen": -275.579345703125,
483
+ "logps/rejected": -208.02297973632812,
484
+ "loss": 0.1401,
485
+ "rewards/accuracies": 0.9125000238418579,
486
+ "rewards/chosen": 3.985370635986328,
487
+ "rewards/margins": 7.237724304199219,
488
+ "rewards/rejected": -3.2523536682128906,
489
  "step": 310
490
  },
491
  {
492
+ "epoch": 0.82,
493
+ "grad_norm": 294.8015156484179,
494
  "learning_rate": 1.899321190108335e-08,
495
+ "logits/chosen": -2.6083686351776123,
496
+ "logits/rejected": -2.508666515350342,
497
+ "logps/chosen": -258.6110534667969,
498
+ "logps/rejected": -207.57406616210938,
499
+ "loss": 0.1709,
500
+ "rewards/accuracies": 0.8812500238418579,
501
+ "rewards/chosen": 3.153282642364502,
502
+ "rewards/margins": 7.0364227294921875,
503
+ "rewards/rejected": -3.8831400871276855,
504
  "step": 320
505
  },
506
  {
507
+ "epoch": 0.84,
508
+ "grad_norm": 262.1222746234904,
509
  "learning_rate": 1.4076524743778317e-08,
510
+ "logits/chosen": -2.6292643547058105,
511
+ "logits/rejected": -2.531942129135132,
512
+ "logps/chosen": -261.5999450683594,
513
+ "logps/rejected": -207.26699829101562,
514
+ "loss": 0.1437,
515
+ "rewards/accuracies": 0.96875,
516
+ "rewards/chosen": 4.273932456970215,
517
+ "rewards/margins": 7.467469215393066,
518
+ "rewards/rejected": -3.1935369968414307,
519
  "step": 330
520
  },
521
  {
522
+ "epoch": 0.87,
523
+ "grad_norm": 299.7750791753028,
524
  "learning_rate": 9.847709219918398e-09,
525
+ "logits/chosen": -2.6148436069488525,
526
+ "logits/rejected": -2.5132031440734863,
527
+ "logps/chosen": -256.66473388671875,
528
+ "logps/rejected": -194.7910919189453,
529
+ "loss": 0.1894,
530
+ "rewards/accuracies": 0.9624999761581421,
531
+ "rewards/chosen": 4.073337078094482,
532
+ "rewards/margins": 7.587210655212402,
533
+ "rewards/rejected": -3.51387357711792,
534
  "step": 340
535
  },
536
  {
537
+ "epoch": 0.9,
538
+ "grad_norm": 346.7439242044206,
539
  "learning_rate": 6.340619665972846e-09,
540
+ "logits/chosen": -2.630897045135498,
541
+ "logits/rejected": -2.5146021842956543,
542
+ "logps/chosen": -265.88916015625,
543
+ "logps/rejected": -212.77487182617188,
544
+ "loss": 0.1582,
545
+ "rewards/accuracies": 0.9125000238418579,
546
+ "rewards/chosen": 3.5220444202423096,
547
+ "rewards/margins": 7.0233283042907715,
548
+ "rewards/rejected": -3.501283645629883,
549
  "step": 350
550
  },
551
  {
552
+ "epoch": 0.92,
553
+ "grad_norm": 307.153905111391,
554
  "learning_rate": 3.583332546643769e-09,
555
+ "logits/chosen": -2.6361472606658936,
556
+ "logits/rejected": -2.5298829078674316,
557
+ "logps/chosen": -271.5682067871094,
558
+ "logps/rejected": -209.4921417236328,
559
+ "loss": 0.1547,
560
+ "rewards/accuracies": 0.9375,
561
+ "rewards/chosen": 3.9090733528137207,
562
+ "rewards/margins": 7.22186279296875,
563
+ "rewards/rejected": -3.3127894401550293,
564
  "step": 360
565
  },
566
  {
567
+ "epoch": 0.95,
568
+ "grad_norm": 183.78861249224778,
569
  "learning_rate": 1.5979216850509847e-09,
570
+ "logits/chosen": -2.593644618988037,
571
+ "logits/rejected": -2.4859519004821777,
572
+ "logps/chosen": -270.0194396972656,
573
+ "logps/rejected": -212.9616241455078,
574
+ "loss": 0.1314,
575
+ "rewards/accuracies": 0.9312499761581421,
576
+ "rewards/chosen": 3.80389142036438,
577
+ "rewards/margins": 6.737584114074707,
578
+ "rewards/rejected": -2.933692216873169,
579
  "step": 370
580
  },
581
  {
582
+ "epoch": 0.97,
583
+ "grad_norm": 201.1703750461263,
584
  "learning_rate": 4.002815478505006e-10,
585
+ "logits/chosen": -2.6470093727111816,
586
+ "logits/rejected": -2.552563428878784,
587
+ "logps/chosen": -261.07574462890625,
588
+ "logps/rejected": -216.7428741455078,
589
+ "loss": 0.1449,
590
+ "rewards/accuracies": 0.9437500238418579,
591
+ "rewards/chosen": 4.029725074768066,
592
+ "rewards/margins": 6.822300910949707,
593
+ "rewards/rejected": -2.7925755977630615,
594
  "step": 380
595
  },
596
  {
597
+ "epoch": 1.0,
598
+ "grad_norm": 359.6013698550439,
599
  "learning_rate": 0.0,
600
+ "logits/chosen": -2.6531007289886475,
601
+ "logits/rejected": -2.540839195251465,
602
+ "logps/chosen": -249.09298706054688,
603
+ "logps/rejected": -204.24765014648438,
604
+ "loss": 0.146,
605
+ "rewards/accuracies": 0.949999988079071,
606
+ "rewards/chosen": 3.7912139892578125,
607
+ "rewards/margins": 7.412507057189941,
608
+ "rewards/rejected": -3.621293306350708,
609
  "step": 390
610
  },
611
  {
612
+ "epoch": 1.0,
613
  "step": 390,
614
  "total_flos": 0.0,
615
+ "train_loss": 0.19850983711389394,
616
+ "train_runtime": 5891.9709,
617
+ "train_samples_per_second": 8.486,
618
+ "train_steps_per_second": 0.066
619
  }
620
  ],
621
  "logging_steps": 10,
 
623
  "num_input_tokens_seen": 0,
624
  "num_train_epochs": 1,
625
  "save_steps": 100,
 
 
 
 
 
 
 
 
 
 
 
 
626
  "total_flos": 0.0,
627
  "train_batch_size": 4,
628
  "trial_name": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3267756347c2c9afb597e20553862a0ac248b214358676d7dfbabc75346c7d63
3
- size 6520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74d76a49f592a53356dcdced2c2a52df2dee5cb0bad93a9b5c068843a8eccfa4
3
+ size 6328