Model save
Browse files- all_results.json +5 -5
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- train_results.json +5 -5
- trainer_state.json +643 -825
- training_args.bin +1 -1
all_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 2.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
-
"train_samples":
|
6 |
-
"train_samples_per_second": 8.
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 2.0,
|
3 |
+
"train_loss": 0.31343152379167494,
|
4 |
+
"train_runtime": 8630.1644,
|
5 |
+
"train_samples": 37141,
|
6 |
+
"train_samples_per_second": 8.607,
|
7 |
+
"train_steps_per_second": 0.067
|
8 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4976698672
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:336e82446271b1f84cf981b43382c3329448f66e3165029498821a5bd11498d5
|
3 |
size 4976698672
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999802720
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:04920783a09dd11a6700fc1ac2fdc19eacf04737ea07cae97a703a48df72397c
|
3 |
size 4999802720
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4915916176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:578fe292401e4d92076f630656f260d30707a4d6564a9d36b1aa5ef1982b6f52
|
3 |
size 4915916176
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1168138808
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:56f45d6dc596c0804896541624466d9d983344958d85ebc4464d33f48bf36439
|
3 |
size 1168138808
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 2.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
-
"train_samples":
|
6 |
-
"train_samples_per_second": 8.
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 2.0,
|
3 |
+
"train_loss": 0.31343152379167494,
|
4 |
+
"train_runtime": 8630.1644,
|
5 |
+
"train_samples": 37141,
|
6 |
+
"train_samples_per_second": 8.607,
|
7 |
+
"train_steps_per_second": 0.067
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,1019 +1,837 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch": 1.
|
5 |
"eval_steps": 10000,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.03,
|
13 |
-
"learning_rate": 1.
|
14 |
-
"logits/chosen": -0.
|
15 |
-
"logits/rejected": -0.
|
16 |
-
"logps/chosen": -
|
17 |
-
"logps/rejected": -
|
18 |
-
"loss": 0.
|
19 |
-
"rewards/accuracies": 0.
|
20 |
-
"rewards/chosen": -
|
21 |
-
"rewards/margins":
|
22 |
-
"rewards/rejected": -
|
23 |
"step": 10
|
24 |
},
|
25 |
{
|
26 |
-
"epoch": 0.
|
27 |
-
"learning_rate":
|
28 |
-
"logits/chosen": -0.
|
29 |
-
"logits/rejected": -0.
|
30 |
-
"logps/chosen": -
|
31 |
-
"logps/rejected": -
|
32 |
-
"loss": 0.
|
33 |
-
"rewards/accuracies": 0.
|
34 |
-
"rewards/chosen": -0.
|
35 |
-
"rewards/margins": 0.
|
36 |
-
"rewards/rejected": -0.
|
37 |
"step": 20
|
38 |
},
|
39 |
{
|
40 |
-
"epoch": 0.
|
41 |
-
"learning_rate":
|
42 |
-
"logits/chosen":
|
43 |
-
"logits/rejected": 0.
|
44 |
-
"logps/chosen": -
|
45 |
-
"logps/rejected": -
|
46 |
-
"loss": 0.
|
47 |
-
"rewards/accuracies": 0.
|
48 |
-
"rewards/chosen": 0.
|
49 |
-
"rewards/margins": 0.
|
50 |
-
"rewards/rejected": -
|
51 |
"step": 30
|
52 |
},
|
53 |
{
|
54 |
-
"epoch": 0.
|
55 |
-
"learning_rate":
|
56 |
-
"logits/chosen":
|
57 |
-
"logits/rejected": 0.
|
58 |
-
"logps/chosen": -
|
59 |
-
"logps/rejected": -
|
60 |
-
"loss": 0.
|
61 |
-
"rewards/accuracies": 0.
|
62 |
-
"rewards/chosen":
|
63 |
-
"rewards/margins": 0.
|
64 |
-
"rewards/rejected":
|
65 |
"step": 40
|
66 |
},
|
67 |
{
|
68 |
-
"epoch": 0.
|
69 |
-
"learning_rate":
|
70 |
-
"logits/chosen": 0.
|
71 |
-
"logits/rejected": 0.
|
72 |
-
"logps/chosen": -
|
73 |
-
"logps/rejected": -
|
74 |
-
"loss": 0.
|
75 |
-
"rewards/accuracies": 0.
|
76 |
-
"rewards/chosen":
|
77 |
-
"rewards/margins": 0.
|
78 |
-
"rewards/rejected":
|
79 |
"step": 50
|
80 |
},
|
81 |
{
|
82 |
-
"epoch": 0.
|
83 |
-
"learning_rate":
|
84 |
-
"logits/chosen":
|
85 |
-
"logits/rejected": 0.
|
86 |
-
"logps/chosen": -
|
87 |
-
"logps/rejected": -
|
88 |
-
"loss": 0.
|
89 |
-
"rewards/accuracies": 0.
|
90 |
-
"rewards/chosen":
|
91 |
-
"rewards/margins": 0.
|
92 |
-
"rewards/rejected":
|
93 |
"step": 60
|
94 |
},
|
95 |
{
|
96 |
-
"epoch": 0.
|
97 |
-
"learning_rate": 9.
|
98 |
-
"logits/chosen": -0.
|
99 |
-
"logits/rejected": 0.
|
100 |
-
"logps/chosen": -
|
101 |
-
"logps/rejected": -
|
102 |
-
"loss": 0.
|
103 |
-
"rewards/accuracies": 0.
|
104 |
-
"rewards/chosen": -0.
|
105 |
-
"rewards/margins": 0.
|
106 |
-
"rewards/rejected": -0.
|
107 |
"step": 70
|
108 |
},
|
109 |
{
|
110 |
-
"epoch": 0.
|
111 |
-
"learning_rate": 9.
|
112 |
-
"logits/chosen": 0.
|
113 |
-
"logits/rejected": 0.
|
114 |
-
"logps/chosen": -
|
115 |
-
"logps/rejected": -
|
116 |
-
"loss": 0.
|
117 |
-
"rewards/accuracies": 0.
|
118 |
-
"rewards/chosen":
|
119 |
-
"rewards/margins": 0.
|
120 |
-
"rewards/rejected": -0.
|
121 |
"step": 80
|
122 |
},
|
123 |
{
|
124 |
-
"epoch": 0.
|
125 |
-
"learning_rate": 9.
|
126 |
-
"logits/chosen": -0.
|
127 |
-
"logits/rejected": -0.
|
128 |
-
"logps/chosen": -
|
129 |
-
"logps/rejected": -
|
130 |
-
"loss": 0.
|
131 |
-
"rewards/accuracies": 0.
|
132 |
-
"rewards/chosen": 0.
|
133 |
-
"rewards/margins": 0.
|
134 |
-
"rewards/rejected": -0.
|
135 |
"step": 90
|
136 |
},
|
137 |
{
|
138 |
-
"epoch": 0.
|
139 |
-
"learning_rate": 9.
|
140 |
-
"logits/chosen": -0.
|
141 |
-
"logits/rejected": -0.
|
142 |
-
"logps/chosen": -
|
143 |
-
"logps/rejected": -
|
144 |
-
"loss": 0.
|
145 |
-
"rewards/accuracies": 0.
|
146 |
-
"rewards/chosen": 0.
|
147 |
-
"rewards/margins": 0.
|
148 |
-
"rewards/rejected": -0.
|
149 |
"step": 100
|
150 |
},
|
151 |
{
|
152 |
-
"epoch": 0.
|
153 |
-
"learning_rate": 9.
|
154 |
-
"logits/chosen": -0.
|
155 |
-
"logits/rejected": -0.
|
156 |
-
"logps/chosen": -
|
157 |
-
"logps/rejected": -
|
158 |
-
"loss": 0.
|
159 |
-
"rewards/accuracies": 0.
|
160 |
-
"rewards/chosen": 0.
|
161 |
-
"rewards/margins": 0.
|
162 |
-
"rewards/rejected": -0.
|
163 |
"step": 110
|
164 |
},
|
165 |
{
|
166 |
-
"epoch": 0.
|
167 |
-
"learning_rate": 9.
|
168 |
-
"logits/chosen": -0.
|
169 |
-
"logits/rejected": -0.
|
170 |
-
"logps/chosen": -
|
171 |
-
"logps/rejected": -
|
172 |
-
"loss": 0.
|
173 |
-
"rewards/accuracies": 0.
|
174 |
-
"rewards/chosen": 0.
|
175 |
-
"rewards/margins": 0.
|
176 |
-
"rewards/rejected": -0.
|
177 |
"step": 120
|
178 |
},
|
179 |
{
|
180 |
-
"epoch": 0.
|
181 |
-
"learning_rate": 9.
|
182 |
-
"logits/chosen":
|
183 |
-
"logits/rejected": -0.
|
184 |
-
"logps/chosen": -
|
185 |
-
"logps/rejected": -
|
186 |
-
"loss": 0.
|
187 |
-
"rewards/accuracies": 0.
|
188 |
-
"rewards/chosen": 0.
|
189 |
-
"rewards/margins": 0.
|
190 |
-
"rewards/rejected": -0.
|
191 |
"step": 130
|
192 |
},
|
193 |
{
|
194 |
-
"epoch": 0.
|
195 |
-
"learning_rate": 9.
|
196 |
-
"logits/chosen": -0.
|
197 |
-
"logits/rejected": -0.
|
198 |
-
"logps/chosen": -
|
199 |
-
"logps/rejected": -
|
200 |
-
"loss": 0.
|
201 |
-
"rewards/accuracies": 0.
|
202 |
-
"rewards/chosen": -0.
|
203 |
-
"rewards/margins": 0.
|
204 |
-
"rewards/rejected": -0.
|
205 |
"step": 140
|
206 |
},
|
207 |
{
|
208 |
-
"epoch": 0.
|
209 |
-
"learning_rate": 9.
|
210 |
-
"logits/chosen": -0.
|
211 |
-
"logits/rejected":
|
212 |
-
"logps/chosen": -
|
213 |
-
"logps/rejected": -
|
214 |
-
"loss": 0.
|
215 |
-
"rewards/accuracies": 0.
|
216 |
-
"rewards/chosen": -0.
|
217 |
-
"rewards/margins": 0.
|
218 |
-
"rewards/rejected": -0.
|
219 |
"step": 150
|
220 |
},
|
221 |
{
|
222 |
-
"epoch": 0.
|
223 |
-
"learning_rate": 9.
|
224 |
-
"logits/chosen": -0.
|
225 |
-
"logits/rejected": -0.
|
226 |
-
"logps/chosen": -
|
227 |
-
"logps/rejected": -
|
228 |
-
"loss": 0.
|
229 |
-
"rewards/accuracies": 0.
|
230 |
-
"rewards/chosen": -0.
|
231 |
-
"rewards/margins": 0.
|
232 |
-
"rewards/rejected": -0.
|
233 |
"step": 160
|
234 |
},
|
235 |
{
|
236 |
-
"epoch": 0.
|
237 |
-
"learning_rate":
|
238 |
-
"logits/chosen": -0.
|
239 |
-
"logits/rejected": -0.
|
240 |
-
"logps/chosen": -
|
241 |
-
"logps/rejected": -
|
242 |
-
"loss": 0.
|
243 |
-
"rewards/accuracies": 0.
|
244 |
-
"rewards/chosen": -0.
|
245 |
-
"rewards/margins": 0.
|
246 |
-
"rewards/rejected": -0.
|
247 |
"step": 170
|
248 |
},
|
249 |
{
|
250 |
-
"epoch": 0.
|
251 |
-
"learning_rate":
|
252 |
-
"logits/chosen": -0.
|
253 |
-
"logits/rejected":
|
254 |
-
"logps/chosen": -
|
255 |
-
"logps/rejected": -
|
256 |
-
"loss": 0.
|
257 |
-
"rewards/accuracies": 0.
|
258 |
-
"rewards/chosen": -0.
|
259 |
-
"rewards/margins": 0.
|
260 |
-
"rewards/rejected": -0.
|
261 |
"step": 180
|
262 |
},
|
263 |
{
|
264 |
-
"epoch": 0.
|
265 |
-
"learning_rate":
|
266 |
-
"logits/chosen": -0.
|
267 |
-
"logits/rejected":
|
268 |
-
"logps/chosen": -
|
269 |
-
"logps/rejected": -
|
270 |
-
"loss": 0.
|
271 |
-
"rewards/accuracies": 0.
|
272 |
-
"rewards/chosen": -0.
|
273 |
-
"rewards/margins": 0.
|
274 |
-
"rewards/rejected": -0.
|
275 |
"step": 190
|
276 |
},
|
277 |
{
|
278 |
-
"epoch": 0.
|
279 |
-
"learning_rate":
|
280 |
-
"logits/chosen": -0.
|
281 |
-
"logits/rejected": -0.
|
282 |
-
"logps/chosen": -
|
283 |
-
"logps/rejected": -
|
284 |
-
"loss": 0.
|
285 |
-
"rewards/accuracies": 0.
|
286 |
-
"rewards/chosen": -0.
|
287 |
-
"rewards/margins": 0.
|
288 |
-
"rewards/rejected": -0.
|
289 |
"step": 200
|
290 |
},
|
291 |
{
|
292 |
-
"epoch": 0.
|
293 |
-
"learning_rate": 8.
|
294 |
-
"logits/chosen":
|
295 |
-
"logits/rejected":
|
296 |
-
"logps/chosen": -
|
297 |
-
"logps/rejected": -
|
298 |
-
"loss": 0.
|
299 |
-
"rewards/accuracies": 0.
|
300 |
-
"rewards/chosen": -0.
|
301 |
-
"rewards/margins": 0.
|
302 |
-
"rewards/rejected": -0.
|
303 |
"step": 210
|
304 |
},
|
305 |
{
|
306 |
-
"epoch": 0.
|
307 |
-
"learning_rate":
|
308 |
-
"logits/chosen": -0.
|
309 |
-
"logits/rejected":
|
310 |
-
"logps/chosen": -
|
311 |
-
"logps/rejected": -
|
312 |
-
"loss": 0.
|
313 |
-
"rewards/accuracies": 0.
|
314 |
-
"rewards/chosen": -0.
|
315 |
-
"rewards/margins": 0.
|
316 |
-
"rewards/rejected": -0.
|
317 |
"step": 220
|
318 |
},
|
319 |
{
|
320 |
-
"epoch": 0.
|
321 |
-
"learning_rate":
|
322 |
-
"logits/chosen":
|
323 |
-
"logits/rejected":
|
324 |
-
"logps/chosen": -
|
325 |
-
"logps/rejected": -
|
326 |
-
"loss": 0.
|
327 |
-
"rewards/accuracies": 0.
|
328 |
-
"rewards/chosen": -0.
|
329 |
-
"rewards/margins": 0.
|
330 |
-
"rewards/rejected": -0.
|
331 |
"step": 230
|
332 |
},
|
333 |
{
|
334 |
-
"epoch": 0.
|
335 |
-
"learning_rate":
|
336 |
-
"logits/chosen":
|
337 |
-
"logits/rejected":
|
338 |
-
"logps/chosen": -
|
339 |
-
"logps/rejected": -
|
340 |
-
"loss": 0.
|
341 |
-
"rewards/accuracies": 0.
|
342 |
-
"rewards/chosen": -0.
|
343 |
-
"rewards/margins": 0.
|
344 |
-
"rewards/rejected": -0.
|
345 |
"step": 240
|
346 |
},
|
347 |
{
|
348 |
-
"epoch": 0.
|
349 |
-
"learning_rate":
|
350 |
-
"logits/chosen": -0.
|
351 |
-
"logits/rejected":
|
352 |
-
"logps/chosen": -
|
353 |
-
"logps/rejected": -
|
354 |
-
"loss": 0.
|
355 |
-
"rewards/accuracies": 0.
|
356 |
-
"rewards/chosen": -0.
|
357 |
-
"rewards/margins": 0.
|
358 |
-
"rewards/rejected": -0.
|
359 |
"step": 250
|
360 |
},
|
361 |
{
|
362 |
-
"epoch": 0.
|
363 |
-
"learning_rate":
|
364 |
-
"logits/chosen": -0.
|
365 |
-
"logits/rejected":
|
366 |
-
"logps/chosen": -
|
367 |
-
"logps/rejected": -
|
368 |
-
"loss": 0.
|
369 |
-
"rewards/accuracies": 0.
|
370 |
-
"rewards/chosen": -0.
|
371 |
-
"rewards/margins": 0.
|
372 |
-
"rewards/rejected": -0.
|
373 |
"step": 260
|
374 |
},
|
375 |
{
|
376 |
-
"epoch": 0.
|
377 |
-
"learning_rate":
|
378 |
-
"logits/chosen":
|
379 |
-
"logits/rejected":
|
380 |
-
"logps/chosen": -
|
381 |
-
"logps/rejected": -
|
382 |
-
"loss": 0.
|
383 |
-
"rewards/accuracies": 0.
|
384 |
-
"rewards/chosen": -0.
|
385 |
-
"rewards/margins": 0.
|
386 |
-
"rewards/rejected": -0.
|
387 |
"step": 270
|
388 |
},
|
389 |
{
|
390 |
-
"epoch": 0.
|
391 |
-
"learning_rate":
|
392 |
-
"logits/chosen":
|
393 |
-
"logits/rejected":
|
394 |
-
"logps/chosen": -
|
395 |
-
"logps/rejected": -
|
396 |
-
"loss": 0.
|
397 |
-
"rewards/accuracies": 0.
|
398 |
-
"rewards/chosen": -0.
|
399 |
-
"rewards/margins": 0.
|
400 |
-
"rewards/rejected": -0.
|
401 |
"step": 280
|
402 |
},
|
403 |
{
|
404 |
-
"epoch": 0
|
405 |
-
"learning_rate":
|
406 |
-
"logits/chosen":
|
407 |
-
"logits/rejected":
|
408 |
-
"logps/chosen": -
|
409 |
-
"logps/rejected": -
|
410 |
-
"loss": 0.
|
411 |
-
"rewards/accuracies": 0.
|
412 |
-
"rewards/chosen": -0.
|
413 |
-
"rewards/margins": 0.
|
414 |
-
"rewards/rejected": -0.
|
415 |
"step": 290
|
416 |
},
|
417 |
{
|
418 |
-
"epoch":
|
419 |
-
"learning_rate":
|
420 |
-
"logits/chosen":
|
421 |
-
"logits/rejected":
|
422 |
-
"logps/chosen": -
|
423 |
-
"logps/rejected": -
|
424 |
-
"loss": 0.
|
425 |
-
"rewards/accuracies": 0.
|
426 |
-
"rewards/chosen": -0.
|
427 |
-
"rewards/margins": 0.
|
428 |
-
"rewards/rejected": -0.
|
429 |
"step": 300
|
430 |
},
|
431 |
{
|
432 |
-
"epoch":
|
433 |
-
"learning_rate":
|
434 |
-
"logits/chosen":
|
435 |
-
"logits/rejected":
|
436 |
-
"logps/chosen": -
|
437 |
-
"logps/rejected": -
|
438 |
-
"loss": 0.
|
439 |
-
"rewards/accuracies": 0.
|
440 |
-
"rewards/chosen": -0.
|
441 |
-
"rewards/margins": 0.
|
442 |
-
"rewards/rejected": -
|
443 |
"step": 310
|
444 |
},
|
445 |
{
|
446 |
-
"epoch":
|
447 |
-
"learning_rate":
|
448 |
-
"logits/chosen":
|
449 |
-
"logits/rejected":
|
450 |
-
"logps/chosen": -
|
451 |
-
"logps/rejected": -
|
452 |
-
"loss": 0.
|
453 |
-
"rewards/accuracies": 0.
|
454 |
-
"rewards/chosen": -0.
|
455 |
-
"rewards/margins": 0.
|
456 |
-
"rewards/rejected": -
|
457 |
"step": 320
|
458 |
},
|
459 |
{
|
460 |
-
"epoch":
|
461 |
-
"learning_rate":
|
462 |
-
"logits/chosen":
|
463 |
-
"logits/rejected":
|
464 |
-
"logps/chosen": -
|
465 |
-
"logps/rejected": -
|
466 |
-
"loss": 0.
|
467 |
-
"rewards/accuracies": 0.
|
468 |
-
"rewards/chosen": -
|
469 |
-
"rewards/margins": 0.
|
470 |
-
"rewards/rejected": -
|
471 |
"step": 330
|
472 |
},
|
473 |
{
|
474 |
-
"epoch":
|
475 |
-
"learning_rate":
|
476 |
-
"logits/chosen":
|
477 |
-
"logits/rejected":
|
478 |
-
"logps/chosen": -
|
479 |
-
"logps/rejected": -
|
480 |
-
"loss": 0.
|
481 |
-
"rewards/accuracies": 0.
|
482 |
-
"rewards/chosen": -
|
483 |
-
"rewards/margins": 0.
|
484 |
-
"rewards/rejected": -
|
485 |
"step": 340
|
486 |
},
|
487 |
{
|
488 |
-
"epoch":
|
489 |
-
"learning_rate":
|
490 |
-
"logits/chosen":
|
491 |
-
"logits/rejected":
|
492 |
-
"logps/chosen": -
|
493 |
-
"logps/rejected": -
|
494 |
-
"loss": 0.
|
495 |
-
"rewards/accuracies": 0.
|
496 |
-
"rewards/chosen": -
|
497 |
-
"rewards/margins": 0.
|
498 |
-
"rewards/rejected": -
|
499 |
"step": 350
|
500 |
},
|
501 |
{
|
502 |
-
"epoch": 1.
|
503 |
-
"learning_rate":
|
504 |
-
"logits/chosen":
|
505 |
-
"logits/rejected":
|
506 |
-
"logps/chosen": -
|
507 |
-
"logps/rejected": -
|
508 |
-
"loss": 0.
|
509 |
-
"rewards/accuracies": 0.
|
510 |
-
"rewards/chosen": -
|
511 |
-
"rewards/margins": 0.
|
512 |
-
"rewards/rejected": -
|
513 |
"step": 360
|
514 |
},
|
515 |
{
|
516 |
-
"epoch": 1.
|
517 |
-
"learning_rate":
|
518 |
-
"logits/chosen":
|
519 |
-
"logits/rejected":
|
520 |
-
"logps/chosen": -
|
521 |
-
"logps/rejected": -
|
522 |
-
"loss": 0.
|
523 |
-
"rewards/accuracies": 0.
|
524 |
-
"rewards/chosen": -
|
525 |
-
"rewards/margins": 0.
|
526 |
-
"rewards/rejected": -
|
527 |
"step": 370
|
528 |
},
|
529 |
{
|
530 |
-
"epoch": 1.
|
531 |
-
"learning_rate":
|
532 |
-
"logits/chosen":
|
533 |
-
"logits/rejected":
|
534 |
-
"logps/chosen": -
|
535 |
-
"logps/rejected": -
|
536 |
-
"loss": 0.
|
537 |
-
"rewards/accuracies": 0.
|
538 |
-
"rewards/chosen": -
|
539 |
-
"rewards/margins": 0.
|
540 |
-
"rewards/rejected": -
|
541 |
"step": 380
|
542 |
},
|
543 |
{
|
544 |
-
"epoch": 1.
|
545 |
-
"learning_rate":
|
546 |
-
"logits/chosen":
|
547 |
-
"logits/rejected":
|
548 |
-
"logps/chosen": -
|
549 |
-
"logps/rejected": -
|
550 |
-
"loss": 0.
|
551 |
-
"rewards/accuracies": 0.
|
552 |
-
"rewards/chosen": -
|
553 |
-
"rewards/margins": 0.
|
554 |
-
"rewards/rejected": -
|
555 |
"step": 390
|
556 |
},
|
557 |
{
|
558 |
-
"epoch": 1.
|
559 |
-
"learning_rate":
|
560 |
-
"logits/chosen":
|
561 |
-
"logits/rejected":
|
562 |
-
"logps/chosen": -
|
563 |
-
"logps/rejected": -
|
564 |
-
"loss": 0.
|
565 |
-
"rewards/accuracies": 0.
|
566 |
-
"rewards/chosen": -
|
567 |
-
"rewards/margins": 0.
|
568 |
-
"rewards/rejected": -
|
569 |
"step": 400
|
570 |
},
|
571 |
{
|
572 |
-
"epoch": 1.
|
573 |
-
"learning_rate":
|
574 |
-
"logits/chosen":
|
575 |
-
"logits/rejected":
|
576 |
-
"logps/chosen": -
|
577 |
-
"logps/rejected": -
|
578 |
-
"loss": 0.
|
579 |
-
"rewards/accuracies": 0.
|
580 |
-
"rewards/chosen": -
|
581 |
-
"rewards/margins": 0.
|
582 |
-
"rewards/rejected": -
|
583 |
"step": 410
|
584 |
},
|
585 |
{
|
586 |
-
"epoch": 1.
|
587 |
-
"learning_rate":
|
588 |
-
"logits/chosen":
|
589 |
-
"logits/rejected":
|
590 |
-
"logps/chosen": -
|
591 |
-
"logps/rejected": -
|
592 |
-
"loss": 0.
|
593 |
-
"rewards/accuracies": 0.
|
594 |
-
"rewards/chosen": -
|
595 |
-
"rewards/margins": 0.
|
596 |
-
"rewards/rejected": -
|
597 |
"step": 420
|
598 |
},
|
599 |
{
|
600 |
-
"epoch": 1.
|
601 |
-
"learning_rate":
|
602 |
-
"logits/chosen":
|
603 |
-
"logits/rejected":
|
604 |
-
"logps/chosen": -
|
605 |
-
"logps/rejected": -
|
606 |
-
"loss": 0.
|
607 |
-
"rewards/accuracies": 0.
|
608 |
-
"rewards/chosen": -
|
609 |
-
"rewards/margins": 0.
|
610 |
-
"rewards/rejected": -
|
611 |
"step": 430
|
612 |
},
|
613 |
{
|
614 |
-
"epoch": 1.
|
615 |
-
"learning_rate":
|
616 |
-
"logits/chosen":
|
617 |
-
"logits/rejected":
|
618 |
-
"logps/chosen": -
|
619 |
-
"logps/rejected": -
|
620 |
-
"loss": 0.
|
621 |
-
"rewards/accuracies": 0.
|
622 |
-
"rewards/chosen": -
|
623 |
-
"rewards/margins": 0.
|
624 |
-
"rewards/rejected": -
|
625 |
"step": 440
|
626 |
},
|
627 |
{
|
628 |
-
"epoch": 1.
|
629 |
-
"learning_rate":
|
630 |
-
"logits/chosen":
|
631 |
-
"logits/rejected":
|
632 |
-
"logps/chosen": -
|
633 |
-
"logps/rejected": -
|
634 |
-
"loss": 0.
|
635 |
-
"rewards/accuracies": 0.
|
636 |
-
"rewards/chosen": -
|
637 |
-
"rewards/margins": 0.
|
638 |
-
"rewards/rejected": -
|
639 |
"step": 450
|
640 |
},
|
641 |
{
|
642 |
-
"epoch": 1.
|
643 |
-
"learning_rate":
|
644 |
-
"logits/chosen":
|
645 |
-
"logits/rejected":
|
646 |
-
"logps/chosen": -
|
647 |
-
"logps/rejected": -
|
648 |
-
"loss": 0.
|
649 |
"rewards/accuracies": 0.8125,
|
650 |
-
"rewards/chosen": -
|
651 |
-
"rewards/margins": 0.
|
652 |
-
"rewards/rejected": -
|
653 |
"step": 460
|
654 |
},
|
655 |
{
|
656 |
-
"epoch": 1.
|
657 |
-
"learning_rate":
|
658 |
-
"logits/chosen":
|
659 |
-
"logits/rejected":
|
660 |
-
"logps/chosen": -
|
661 |
-
"logps/rejected": -
|
662 |
-
"loss": 0.
|
663 |
-
"rewards/accuracies": 0.
|
664 |
-
"rewards/chosen": -
|
665 |
-
"rewards/margins": 0.
|
666 |
-
"rewards/rejected": -
|
667 |
"step": 470
|
668 |
},
|
669 |
{
|
670 |
-
"epoch": 1.
|
671 |
-
"learning_rate":
|
672 |
-
"logits/chosen":
|
673 |
-
"logits/rejected":
|
674 |
-
"logps/chosen": -
|
675 |
-
"logps/rejected": -
|
676 |
-
"loss": 0.
|
677 |
-
"rewards/accuracies": 0.
|
678 |
-
"rewards/chosen": -
|
679 |
-
"rewards/margins": 0.
|
680 |
-
"rewards/rejected": -
|
681 |
"step": 480
|
682 |
},
|
683 |
{
|
684 |
-
"epoch": 1.
|
685 |
-
"learning_rate":
|
686 |
-
"logits/chosen":
|
687 |
-
"logits/rejected":
|
688 |
-
"logps/chosen": -
|
689 |
-
"logps/rejected": -
|
690 |
-
"loss": 0.
|
691 |
-
"rewards/accuracies": 0.
|
692 |
-
"rewards/chosen": -
|
693 |
-
"rewards/margins": 0.
|
694 |
-
"rewards/rejected": -2.
|
695 |
"step": 490
|
696 |
},
|
697 |
{
|
698 |
-
"epoch": 1.
|
699 |
-
"learning_rate":
|
700 |
-
"logits/chosen":
|
701 |
-
"logits/rejected":
|
702 |
-
"logps/chosen": -
|
703 |
-
"logps/rejected": -
|
704 |
-
"loss": 0.
|
705 |
-
"rewards/accuracies": 0.
|
706 |
-
"rewards/chosen": -
|
707 |
-
"rewards/margins":
|
708 |
-
"rewards/rejected": -
|
709 |
"step": 500
|
710 |
},
|
711 |
{
|
712 |
-
"epoch": 1.
|
713 |
-
"learning_rate":
|
714 |
-
"logits/chosen":
|
715 |
-
"logits/rejected":
|
716 |
-
"logps/chosen": -
|
717 |
-
"logps/rejected": -
|
718 |
-
"loss": 0.
|
719 |
-
"rewards/accuracies": 0.
|
720 |
-
"rewards/chosen": -
|
721 |
-
"rewards/margins": 0.
|
722 |
-
"rewards/rejected": -
|
723 |
"step": 510
|
724 |
},
|
725 |
{
|
726 |
-
"epoch": 1.
|
727 |
-
"learning_rate":
|
728 |
-
"logits/chosen":
|
729 |
-
"logits/rejected":
|
730 |
-
"logps/chosen": -
|
731 |
-
"logps/rejected": -
|
732 |
-
"loss": 0.
|
733 |
-
"rewards/accuracies": 0.
|
734 |
-
"rewards/chosen": -
|
735 |
-
"rewards/margins": 0.
|
736 |
-
"rewards/rejected": -
|
737 |
"step": 520
|
738 |
},
|
739 |
{
|
740 |
-
"epoch": 1.
|
741 |
-
"learning_rate":
|
742 |
-
"logits/chosen":
|
743 |
-
"logits/rejected": 0.
|
744 |
-
"logps/chosen": -
|
745 |
-
"logps/rejected": -
|
746 |
-
"loss": 0.
|
747 |
-
"rewards/accuracies": 0.
|
748 |
-
"rewards/chosen": -
|
749 |
-
"rewards/margins": 0.
|
750 |
-
"rewards/rejected": -
|
751 |
"step": 530
|
752 |
},
|
753 |
{
|
754 |
-
"epoch": 1.
|
755 |
-
"learning_rate": 1.
|
756 |
-
"logits/chosen":
|
757 |
-
"logits/rejected": 0.
|
758 |
-
"logps/chosen": -
|
759 |
-
"logps/rejected": -
|
760 |
-
"loss": 0.
|
761 |
-
"rewards/accuracies": 0.
|
762 |
-
"rewards/chosen": -
|
763 |
-
"rewards/margins": 0.
|
764 |
-
"rewards/rejected": -
|
765 |
"step": 540
|
766 |
},
|
767 |
{
|
768 |
-
"epoch": 1.
|
769 |
-
"learning_rate":
|
770 |
-
"logits/chosen":
|
771 |
-
"logits/rejected": 0.
|
772 |
-
"logps/chosen": -
|
773 |
-
"logps/rejected": -
|
774 |
-
"loss": 0.
|
775 |
-
"rewards/accuracies": 0.
|
776 |
-
"rewards/chosen": -
|
777 |
-
"rewards/margins": 0.
|
778 |
-
"rewards/rejected": -2.
|
779 |
"step": 550
|
780 |
},
|
781 |
{
|
782 |
-
"epoch": 1.
|
783 |
-
"learning_rate":
|
784 |
-
"logits/chosen":
|
785 |
-
"logits/rejected": 0.
|
786 |
-
"logps/chosen": -
|
787 |
-
"logps/rejected": -
|
788 |
-
"loss": 0.
|
789 |
-
"rewards/accuracies": 0.
|
790 |
-
"rewards/chosen": -
|
791 |
-
"rewards/margins": 0.
|
792 |
-
"rewards/rejected": -
|
793 |
"step": 560
|
794 |
},
|
795 |
{
|
796 |
-
"epoch": 1.
|
797 |
-
"learning_rate":
|
798 |
-
"logits/chosen":
|
799 |
-
"logits/rejected": 0.
|
800 |
-
"logps/chosen": -
|
801 |
-
"logps/rejected": -
|
802 |
-
"loss": 0.
|
803 |
-
"rewards/accuracies": 0.
|
804 |
-
"rewards/chosen": -
|
805 |
-
"rewards/margins":
|
806 |
-
"rewards/rejected": -
|
807 |
"step": 570
|
808 |
},
|
809 |
-
{
|
810 |
-
"epoch": 1.63,
|
811 |
-
"learning_rate": 9.869388139903495e-08,
|
812 |
-
"logits/chosen": -0.04067504033446312,
|
813 |
-
"logits/rejected": -0.04381244629621506,
|
814 |
-
"logps/chosen": -403.59710693359375,
|
815 |
-
"logps/rejected": -443.3935546875,
|
816 |
-
"loss": 0.4333,
|
817 |
-
"rewards/accuracies": 0.8500000238418579,
|
818 |
-
"rewards/chosen": -1.2790496349334717,
|
819 |
-
"rewards/margins": 0.8235654830932617,
|
820 |
-
"rewards/rejected": -2.1026148796081543,
|
821 |
-
"step": 580
|
822 |
-
},
|
823 |
-
{
|
824 |
-
"epoch": 1.66,
|
825 |
-
"learning_rate": 8.452144078061818e-08,
|
826 |
-
"logits/chosen": -0.06195799633860588,
|
827 |
-
"logits/rejected": 0.00023120641708374023,
|
828 |
-
"logps/chosen": -473.6097717285156,
|
829 |
-
"logps/rejected": -475.43475341796875,
|
830 |
-
"loss": 0.4288,
|
831 |
-
"rewards/accuracies": 0.8187500238418579,
|
832 |
-
"rewards/chosen": -1.3708075284957886,
|
833 |
-
"rewards/margins": 1.0009411573410034,
|
834 |
-
"rewards/rejected": -2.371748447418213,
|
835 |
-
"step": 590
|
836 |
-
},
|
837 |
-
{
|
838 |
-
"epoch": 1.69,
|
839 |
-
"learning_rate": 7.135305900598321e-08,
|
840 |
-
"logits/chosen": -0.07134322822093964,
|
841 |
-
"logits/rejected": 0.04816692695021629,
|
842 |
-
"logps/chosen": -485.545166015625,
|
843 |
-
"logps/rejected": -423.0818786621094,
|
844 |
-
"loss": 0.422,
|
845 |
-
"rewards/accuracies": 0.8687499761581421,
|
846 |
-
"rewards/chosen": -1.2954851388931274,
|
847 |
-
"rewards/margins": 0.9698074460029602,
|
848 |
-
"rewards/rejected": -2.2652926445007324,
|
849 |
-
"step": 600
|
850 |
-
},
|
851 |
-
{
|
852 |
-
"epoch": 1.71,
|
853 |
-
"learning_rate": 5.9220559209888166e-08,
|
854 |
-
"logits/chosen": -0.0719083845615387,
|
855 |
-
"logits/rejected": -0.051315806806087494,
|
856 |
-
"logps/chosen": -447.87518310546875,
|
857 |
-
"logps/rejected": -441.8182678222656,
|
858 |
-
"loss": 0.4367,
|
859 |
-
"rewards/accuracies": 0.8125,
|
860 |
-
"rewards/chosen": -1.3435814380645752,
|
861 |
-
"rewards/margins": 0.8819317817687988,
|
862 |
-
"rewards/rejected": -2.225512981414795,
|
863 |
-
"step": 610
|
864 |
-
},
|
865 |
-
{
|
866 |
-
"epoch": 1.74,
|
867 |
-
"learning_rate": 4.815326118139812e-08,
|
868 |
-
"logits/chosen": -0.03743662312626839,
|
869 |
-
"logits/rejected": 0.02814921736717224,
|
870 |
-
"logps/chosen": -447.53411865234375,
|
871 |
-
"logps/rejected": -402.9500427246094,
|
872 |
-
"loss": 0.425,
|
873 |
-
"rewards/accuracies": 0.8500000238418579,
|
874 |
-
"rewards/chosen": -1.2780933380126953,
|
875 |
-
"rewards/margins": 0.9174167513847351,
|
876 |
-
"rewards/rejected": -2.195510149002075,
|
877 |
-
"step": 620
|
878 |
-
},
|
879 |
-
{
|
880 |
-
"epoch": 1.77,
|
881 |
-
"learning_rate": 3.81779105087407e-08,
|
882 |
-
"logits/chosen": -0.12636077404022217,
|
883 |
-
"logits/rejected": 0.03690079599618912,
|
884 |
-
"logps/chosen": -432.38177490234375,
|
885 |
-
"logps/rejected": -432.5584411621094,
|
886 |
-
"loss": 0.4391,
|
887 |
-
"rewards/accuracies": 0.831250011920929,
|
888 |
-
"rewards/chosen": -1.4293471574783325,
|
889 |
-
"rewards/margins": 0.8169783353805542,
|
890 |
-
"rewards/rejected": -2.2463252544403076,
|
891 |
-
"step": 630
|
892 |
-
},
|
893 |
-
{
|
894 |
-
"epoch": 1.8,
|
895 |
-
"learning_rate": 2.9318613945057637e-08,
|
896 |
-
"logits/chosen": -0.1318369209766388,
|
897 |
-
"logits/rejected": -0.046459414064884186,
|
898 |
-
"logps/chosen": -508.367431640625,
|
899 |
-
"logps/rejected": -467.4312438964844,
|
900 |
-
"loss": 0.4208,
|
901 |
-
"rewards/accuracies": 0.8812500238418579,
|
902 |
-
"rewards/chosen": -1.3762736320495605,
|
903 |
-
"rewards/margins": 1.1156460046768188,
|
904 |
-
"rewards/rejected": -2.49191951751709,
|
905 |
-
"step": 640
|
906 |
-
},
|
907 |
-
{
|
908 |
-
"epoch": 1.83,
|
909 |
-
"learning_rate": 2.1596781151249523e-08,
|
910 |
-
"logits/chosen": 0.05323456600308418,
|
911 |
-
"logits/rejected": 0.11673180758953094,
|
912 |
-
"logps/chosen": -460.07366943359375,
|
913 |
-
"logps/rejected": -404.09979248046875,
|
914 |
-
"loss": 0.4408,
|
915 |
-
"rewards/accuracies": 0.831250011920929,
|
916 |
-
"rewards/chosen": -1.3242613077163696,
|
917 |
-
"rewards/margins": 0.9012883305549622,
|
918 |
-
"rewards/rejected": -2.2255492210388184,
|
919 |
-
"step": 650
|
920 |
-
},
|
921 |
-
{
|
922 |
-
"epoch": 1.85,
|
923 |
-
"learning_rate": 1.5031072956701695e-08,
|
924 |
-
"logits/chosen": -0.09114273637533188,
|
925 |
-
"logits/rejected": -0.030487876385450363,
|
926 |
-
"logps/chosen": -523.2674560546875,
|
927 |
-
"logps/rejected": -505.872802734375,
|
928 |
-
"loss": 0.4338,
|
929 |
-
"rewards/accuracies": 0.8812500238418579,
|
930 |
-
"rewards/chosen": -1.4451870918273926,
|
931 |
-
"rewards/margins": 1.1003198623657227,
|
932 |
-
"rewards/rejected": -2.5455069541931152,
|
933 |
-
"step": 660
|
934 |
-
},
|
935 |
-
{
|
936 |
-
"epoch": 1.88,
|
937 |
-
"learning_rate": 9.637356262923723e-09,
|
938 |
-
"logits/chosen": -0.018620457500219345,
|
939 |
-
"logits/rejected": 0.09086041152477264,
|
940 |
-
"logps/chosen": -455.99371337890625,
|
941 |
-
"logps/rejected": -447.4140625,
|
942 |
-
"loss": 0.4191,
|
943 |
-
"rewards/accuracies": 0.856249988079071,
|
944 |
-
"rewards/chosen": -1.377251386642456,
|
945 |
-
"rewards/margins": 0.9726886749267578,
|
946 |
-
"rewards/rejected": -2.3499398231506348,
|
947 |
-
"step": 670
|
948 |
-
},
|
949 |
-
{
|
950 |
-
"epoch": 1.91,
|
951 |
-
"learning_rate": 5.428665699084789e-09,
|
952 |
-
"logits/chosen": -0.15896455943584442,
|
953 |
-
"logits/rejected": 0.003171744989231229,
|
954 |
-
"logps/chosen": -446.88519287109375,
|
955 |
-
"logps/rejected": -409.6089782714844,
|
956 |
-
"loss": 0.417,
|
957 |
-
"rewards/accuracies": 0.862500011920929,
|
958 |
-
"rewards/chosen": -1.368312120437622,
|
959 |
-
"rewards/margins": 0.9707133173942566,
|
960 |
-
"rewards/rejected": -2.3390254974365234,
|
961 |
-
"step": 680
|
962 |
-
},
|
963 |
-
{
|
964 |
-
"epoch": 1.94,
|
965 |
-
"learning_rate": 2.415172122110343e-09,
|
966 |
-
"logits/chosen": -0.183979794383049,
|
967 |
-
"logits/rejected": -0.08970221877098083,
|
968 |
-
"logps/chosen": -522.4932861328125,
|
969 |
-
"logps/rejected": -475.81060791015625,
|
970 |
-
"loss": 0.4177,
|
971 |
-
"rewards/accuracies": 0.8374999761581421,
|
972 |
-
"rewards/chosen": -1.4173470735549927,
|
973 |
-
"rewards/margins": 0.9275003671646118,
|
974 |
-
"rewards/rejected": -2.3448474407196045,
|
975 |
-
"step": 690
|
976 |
-
},
|
977 |
-
{
|
978 |
-
"epoch": 1.97,
|
979 |
-
"learning_rate": 6.041580374618327e-10,
|
980 |
-
"logits/chosen": -0.12467072904109955,
|
981 |
-
"logits/rejected": 0.008299602195620537,
|
982 |
-
"logps/chosen": -485.9302673339844,
|
983 |
-
"logps/rejected": -479.005126953125,
|
984 |
-
"loss": 0.4332,
|
985 |
-
"rewards/accuracies": 0.856249988079071,
|
986 |
-
"rewards/chosen": -1.5329062938690186,
|
987 |
-
"rewards/margins": 0.9002411961555481,
|
988 |
-
"rewards/rejected": -2.433147430419922,
|
989 |
-
"step": 700
|
990 |
-
},
|
991 |
{
|
992 |
"epoch": 2.0,
|
993 |
"learning_rate": 0.0,
|
994 |
-
"logits/chosen": 0.
|
995 |
-
"logits/rejected": 0.
|
996 |
-
"logps/chosen": -
|
997 |
-
"logps/rejected": -
|
998 |
-
"loss": 0.
|
999 |
-
"rewards/accuracies": 0.
|
1000 |
-
"rewards/chosen": -
|
1001 |
-
"rewards/margins": 0.
|
1002 |
-
"rewards/rejected": -2.
|
1003 |
-
"step":
|
1004 |
},
|
1005 |
{
|
1006 |
"epoch": 2.0,
|
1007 |
-
"step":
|
1008 |
"total_flos": 0.0,
|
1009 |
-
"train_loss": 0.
|
1010 |
-
"train_runtime":
|
1011 |
-
"train_samples_per_second": 8.
|
1012 |
-
"train_steps_per_second": 0.
|
1013 |
}
|
1014 |
],
|
1015 |
"logging_steps": 10,
|
1016 |
-
"max_steps":
|
1017 |
"num_train_epochs": 2,
|
1018 |
"save_steps": 10000,
|
1019 |
"total_flos": 0.0,
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 1.9982773471145565,
|
5 |
"eval_steps": 10000,
|
6 |
+
"global_step": 580,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.03,
|
13 |
+
"learning_rate": 1.7241379310344828e-07,
|
14 |
+
"logits/chosen": -0.04663122445344925,
|
15 |
+
"logits/rejected": -0.04833713918924332,
|
16 |
+
"logps/chosen": -182.5131072998047,
|
17 |
+
"logps/rejected": -183.75173950195312,
|
18 |
+
"loss": 0.6541,
|
19 |
+
"rewards/accuracies": 0.39375001192092896,
|
20 |
+
"rewards/chosen": 1.5550864191027358e-05,
|
21 |
+
"rewards/margins": 4.5274849981069565e-05,
|
22 |
+
"rewards/rejected": -2.9723969419137575e-05,
|
23 |
"step": 10
|
24 |
},
|
25 |
{
|
26 |
+
"epoch": 0.07,
|
27 |
+
"learning_rate": 3.4482758620689656e-07,
|
28 |
+
"logits/chosen": -0.04483399540185928,
|
29 |
+
"logits/rejected": -0.040133703500032425,
|
30 |
+
"logps/chosen": -175.80125427246094,
|
31 |
+
"logps/rejected": -183.30113220214844,
|
32 |
+
"loss": 0.6465,
|
33 |
+
"rewards/accuracies": 0.4625000059604645,
|
34 |
+
"rewards/chosen": -0.0010477929608896375,
|
35 |
+
"rewards/margins": -0.0001325715275015682,
|
36 |
+
"rewards/rejected": -0.000915221287868917,
|
37 |
"step": 20
|
38 |
},
|
39 |
{
|
40 |
+
"epoch": 0.1,
|
41 |
+
"learning_rate": 5.172413793103448e-07,
|
42 |
+
"logits/chosen": 0.03275300934910774,
|
43 |
+
"logits/rejected": 0.05599850416183472,
|
44 |
+
"logps/chosen": -188.52236938476562,
|
45 |
+
"logps/rejected": -177.93060302734375,
|
46 |
+
"loss": 0.6521,
|
47 |
+
"rewards/accuracies": 0.48124998807907104,
|
48 |
+
"rewards/chosen": -0.0008330008713528514,
|
49 |
+
"rewards/margins": -0.000871331081725657,
|
50 |
+
"rewards/rejected": 3.833030496025458e-05,
|
51 |
"step": 30
|
52 |
},
|
53 |
{
|
54 |
+
"epoch": 0.14,
|
55 |
+
"learning_rate": 6.896551724137931e-07,
|
56 |
+
"logits/chosen": 0.0016825118800625205,
|
57 |
+
"logits/rejected": 0.0037804923485964537,
|
58 |
+
"logps/chosen": -163.9655303955078,
|
59 |
+
"logps/rejected": -169.550048828125,
|
60 |
+
"loss": 0.6509,
|
61 |
+
"rewards/accuracies": 0.4375,
|
62 |
+
"rewards/chosen": 0.0026731090620160103,
|
63 |
+
"rewards/margins": -0.0011917275842279196,
|
64 |
+
"rewards/rejected": 0.00386483664624393,
|
65 |
"step": 40
|
66 |
},
|
67 |
{
|
68 |
+
"epoch": 0.17,
|
69 |
+
"learning_rate": 8.620689655172412e-07,
|
70 |
+
"logits/chosen": -0.002694876166060567,
|
71 |
+
"logits/rejected": -0.017489472404122353,
|
72 |
+
"logps/chosen": -177.7022705078125,
|
73 |
+
"logps/rejected": -171.62660217285156,
|
74 |
+
"loss": 0.6509,
|
75 |
+
"rewards/accuracies": 0.5562499761581421,
|
76 |
+
"rewards/chosen": 0.004064758308231831,
|
77 |
+
"rewards/margins": 0.001050427439622581,
|
78 |
+
"rewards/rejected": 0.0030143309850245714,
|
79 |
"step": 50
|
80 |
},
|
81 |
{
|
82 |
+
"epoch": 0.21,
|
83 |
+
"learning_rate": 9.999637795788383e-07,
|
84 |
+
"logits/chosen": 0.04514119774103165,
|
85 |
+
"logits/rejected": 0.0163094662129879,
|
86 |
+
"logps/chosen": -154.91921997070312,
|
87 |
+
"logps/rejected": -172.60826110839844,
|
88 |
+
"loss": 0.6471,
|
89 |
+
"rewards/accuracies": 0.4937500059604645,
|
90 |
+
"rewards/chosen": 0.006378197111189365,
|
91 |
+
"rewards/margins": -0.00023460984812118113,
|
92 |
+
"rewards/rejected": 0.006612806580960751,
|
93 |
"step": 60
|
94 |
},
|
95 |
{
|
96 |
+
"epoch": 0.24,
|
97 |
+
"learning_rate": 9.98696615758975e-07,
|
98 |
+
"logits/chosen": -0.01900678686797619,
|
99 |
+
"logits/rejected": 0.04069110006093979,
|
100 |
+
"logps/chosen": -155.2123565673828,
|
101 |
+
"logps/rejected": -151.2286834716797,
|
102 |
+
"loss": 0.6484,
|
103 |
+
"rewards/accuracies": 0.581250011920929,
|
104 |
+
"rewards/chosen": -0.005074621178209782,
|
105 |
+
"rewards/margins": 0.0047805835492908955,
|
106 |
+
"rewards/rejected": -0.009855205193161964,
|
107 |
"step": 70
|
108 |
},
|
109 |
{
|
110 |
+
"epoch": 0.28,
|
111 |
+
"learning_rate": 9.95623675140179e-07,
|
112 |
+
"logits/chosen": 0.014727266505360603,
|
113 |
+
"logits/rejected": 0.035828378051519394,
|
114 |
+
"logps/chosen": -157.910400390625,
|
115 |
+
"logps/rejected": -170.37710571289062,
|
116 |
+
"loss": 0.6375,
|
117 |
+
"rewards/accuracies": 0.574999988079071,
|
118 |
+
"rewards/chosen": 0.009923619218170643,
|
119 |
+
"rewards/margins": 0.011426478624343872,
|
120 |
+
"rewards/rejected": -0.0015028597554191947,
|
121 |
"step": 80
|
122 |
},
|
123 |
{
|
124 |
+
"epoch": 0.31,
|
125 |
+
"learning_rate": 9.907560848179605e-07,
|
126 |
+
"logits/chosen": -0.06330171972513199,
|
127 |
+
"logits/rejected": -0.05511590838432312,
|
128 |
+
"logps/chosen": -170.8775634765625,
|
129 |
+
"logps/rejected": -178.7204132080078,
|
130 |
+
"loss": 0.6279,
|
131 |
+
"rewards/accuracies": 0.5625,
|
132 |
+
"rewards/chosen": -0.0003136039595119655,
|
133 |
+
"rewards/margins": 0.006801925599575043,
|
134 |
+
"rewards/rejected": -0.0071155303157866,
|
135 |
"step": 90
|
136 |
},
|
137 |
{
|
138 |
+
"epoch": 0.34,
|
139 |
+
"learning_rate": 9.841114703012817e-07,
|
140 |
+
"logits/chosen": -0.08814354985952377,
|
141 |
+
"logits/rejected": -0.053762950003147125,
|
142 |
+
"logps/chosen": -177.0261993408203,
|
143 |
+
"logps/rejected": -170.6063690185547,
|
144 |
+
"loss": 0.6162,
|
145 |
+
"rewards/accuracies": 0.550000011920929,
|
146 |
+
"rewards/chosen": -0.01691436395049095,
|
147 |
+
"rewards/margins": 0.0037760145496577024,
|
148 |
+
"rewards/rejected": -0.020690377801656723,
|
149 |
"step": 100
|
150 |
},
|
151 |
{
|
152 |
+
"epoch": 0.38,
|
153 |
+
"learning_rate": 9.757138916907184e-07,
|
154 |
+
"logits/chosen": -0.058143556118011475,
|
155 |
+
"logits/rejected": -0.0431932732462883,
|
156 |
+
"logps/chosen": -171.3651885986328,
|
157 |
+
"logps/rejected": -169.97354125976562,
|
158 |
+
"loss": 0.5997,
|
159 |
+
"rewards/accuracies": 0.59375,
|
160 |
+
"rewards/chosen": -0.02524380013346672,
|
161 |
+
"rewards/margins": 0.02074708789587021,
|
162 |
+
"rewards/rejected": -0.04599089175462723,
|
163 |
"step": 110
|
164 |
},
|
165 |
{
|
166 |
+
"epoch": 0.41,
|
167 |
+
"learning_rate": 9.655937565570122e-07,
|
168 |
+
"logits/chosen": -0.037062786519527435,
|
169 |
+
"logits/rejected": -0.0404152050614357,
|
170 |
+
"logps/chosen": -160.91107177734375,
|
171 |
+
"logps/rejected": -170.44813537597656,
|
172 |
+
"loss": 0.5917,
|
173 |
+
"rewards/accuracies": 0.5874999761581421,
|
174 |
+
"rewards/chosen": -0.07813913375139236,
|
175 |
+
"rewards/margins": 0.017309704795479774,
|
176 |
+
"rewards/rejected": -0.09544883668422699,
|
177 |
"step": 120
|
178 |
},
|
179 |
{
|
180 |
+
"epoch": 0.45,
|
181 |
+
"learning_rate": 9.537877098354784e-07,
|
182 |
+
"logits/chosen": 0.01685944013297558,
|
183 |
+
"logits/rejected": -0.009876708500087261,
|
184 |
+
"logps/chosen": -193.56317138671875,
|
185 |
+
"logps/rejected": -199.8650360107422,
|
186 |
+
"loss": 0.5874,
|
187 |
+
"rewards/accuracies": 0.5874999761581421,
|
188 |
+
"rewards/chosen": -0.1443156749010086,
|
189 |
+
"rewards/margins": 0.01459985226392746,
|
190 |
+
"rewards/rejected": -0.15891551971435547,
|
191 |
"step": 130
|
192 |
},
|
193 |
{
|
194 |
+
"epoch": 0.48,
|
195 |
+
"learning_rate": 9.403385011349638e-07,
|
196 |
+
"logits/chosen": -0.08319505304098129,
|
197 |
+
"logits/rejected": -0.04671981930732727,
|
198 |
+
"logps/chosen": -184.2541961669922,
|
199 |
+
"logps/rejected": -192.4634552001953,
|
200 |
+
"loss": 0.574,
|
201 |
+
"rewards/accuracies": 0.6000000238418579,
|
202 |
+
"rewards/chosen": -0.1738511472940445,
|
203 |
+
"rewards/margins": 0.02675568498671055,
|
204 |
+
"rewards/rejected": -0.2006068229675293,
|
205 |
"step": 140
|
206 |
},
|
207 |
{
|
208 |
+
"epoch": 0.52,
|
209 |
+
"learning_rate": 9.252948299418253e-07,
|
210 |
+
"logits/chosen": -0.029505770653486252,
|
211 |
+
"logits/rejected": 0.007474844343960285,
|
212 |
+
"logps/chosen": -201.96896362304688,
|
213 |
+
"logps/rejected": -209.3897247314453,
|
214 |
+
"loss": 0.5615,
|
215 |
+
"rewards/accuracies": 0.5625,
|
216 |
+
"rewards/chosen": -0.2171604186296463,
|
217 |
+
"rewards/margins": 0.049543753266334534,
|
218 |
+
"rewards/rejected": -0.26670414209365845,
|
219 |
"step": 150
|
220 |
},
|
221 |
{
|
222 |
+
"epoch": 0.55,
|
223 |
+
"learning_rate": 9.087111692794459e-07,
|
224 |
+
"logits/chosen": -0.0830586701631546,
|
225 |
+
"logits/rejected": -0.05916399508714676,
|
226 |
+
"logps/chosen": -188.63107299804688,
|
227 |
+
"logps/rejected": -191.2251434326172,
|
228 |
+
"loss": 0.5454,
|
229 |
+
"rewards/accuracies": 0.543749988079071,
|
230 |
+
"rewards/chosen": -0.2526788115501404,
|
231 |
+
"rewards/margins": 0.013631664216518402,
|
232 |
+
"rewards/rejected": -0.2663104832172394,
|
233 |
"step": 160
|
234 |
},
|
235 |
{
|
236 |
+
"epoch": 0.59,
|
237 |
+
"learning_rate": 8.906475684618158e-07,
|
238 |
+
"logits/chosen": -0.0008124202722683549,
|
239 |
+
"logits/rejected": -0.014014181680977345,
|
240 |
+
"logps/chosen": -185.88009643554688,
|
241 |
+
"logps/rejected": -198.26937866210938,
|
242 |
+
"loss": 0.5244,
|
243 |
+
"rewards/accuracies": 0.53125,
|
244 |
+
"rewards/chosen": -0.2561153471469879,
|
245 |
+
"rewards/margins": 0.03350779041647911,
|
246 |
+
"rewards/rejected": -0.28962308168411255,
|
247 |
"step": 170
|
248 |
},
|
249 |
{
|
250 |
+
"epoch": 0.62,
|
251 |
+
"learning_rate": 8.711694356554049e-07,
|
252 |
+
"logits/chosen": -0.02522989548742771,
|
253 |
+
"logits/rejected": 0.017963331192731857,
|
254 |
+
"logps/chosen": -186.06103515625,
|
255 |
+
"logps/rejected": -192.4676971435547,
|
256 |
+
"loss": 0.5036,
|
257 |
+
"rewards/accuracies": 0.6625000238418579,
|
258 |
+
"rewards/chosen": -0.26322150230407715,
|
259 |
+
"rewards/margins": 0.049817826598882675,
|
260 |
+
"rewards/rejected": -0.3130393326282501,
|
261 |
"step": 180
|
262 |
},
|
263 |
{
|
264 |
+
"epoch": 0.65,
|
265 |
+
"learning_rate": 8.503473010366712e-07,
|
266 |
+
"logits/chosen": -0.03940283879637718,
|
267 |
+
"logits/rejected": 0.012682870030403137,
|
268 |
+
"logps/chosen": -197.4193572998047,
|
269 |
+
"logps/rejected": -196.120849609375,
|
270 |
+
"loss": 0.4869,
|
271 |
+
"rewards/accuracies": 0.5874999761581421,
|
272 |
+
"rewards/chosen": -0.30093440413475037,
|
273 |
+
"rewards/margins": 0.04106402397155762,
|
274 |
+
"rewards/rejected": -0.341998428106308,
|
275 |
"step": 190
|
276 |
},
|
277 |
{
|
278 |
+
"epoch": 0.69,
|
279 |
+
"learning_rate": 8.282565614028067e-07,
|
280 |
+
"logits/chosen": -0.011284901760518551,
|
281 |
+
"logits/rejected": -0.018011104315519333,
|
282 |
+
"logps/chosen": -197.7900848388672,
|
283 |
+
"logps/rejected": -208.4326171875,
|
284 |
+
"loss": 0.4659,
|
285 |
+
"rewards/accuracies": 0.581250011920929,
|
286 |
+
"rewards/chosen": -0.30557769536972046,
|
287 |
+
"rewards/margins": 0.04261254146695137,
|
288 |
+
"rewards/rejected": -0.34819021821022034,
|
289 |
"step": 200
|
290 |
},
|
291 |
{
|
292 |
+
"epoch": 0.72,
|
293 |
+
"learning_rate": 8.049772071604863e-07,
|
294 |
+
"logits/chosen": 0.025003302842378616,
|
295 |
+
"logits/rejected": 0.056857384741306305,
|
296 |
+
"logps/chosen": -208.2877960205078,
|
297 |
+
"logps/rejected": -204.39129638671875,
|
298 |
+
"loss": 0.4466,
|
299 |
+
"rewards/accuracies": 0.6187499761581421,
|
300 |
+
"rewards/chosen": -0.3003290891647339,
|
301 |
+
"rewards/margins": 0.05031699687242508,
|
302 |
+
"rewards/rejected": -0.35064607858657837,
|
303 |
"step": 210
|
304 |
},
|
305 |
{
|
306 |
+
"epoch": 0.76,
|
307 |
+
"learning_rate": 7.805935326811912e-07,
|
308 |
+
"logits/chosen": -0.040947750210762024,
|
309 |
+
"logits/rejected": 0.005281107500195503,
|
310 |
+
"logps/chosen": -198.28762817382812,
|
311 |
+
"logps/rejected": -201.60983276367188,
|
312 |
+
"loss": 0.4448,
|
313 |
+
"rewards/accuracies": 0.5375000238418579,
|
314 |
+
"rewards/chosen": -0.3105833828449249,
|
315 |
+
"rewards/margins": 0.04565637558698654,
|
316 |
+
"rewards/rejected": -0.3562397360801697,
|
317 |
"step": 220
|
318 |
},
|
319 |
{
|
320 |
+
"epoch": 0.79,
|
321 |
+
"learning_rate": 7.551938310719042e-07,
|
322 |
+
"logits/chosen": 0.004909530282020569,
|
323 |
+
"logits/rejected": 0.02475576661527157,
|
324 |
+
"logps/chosen": -217.17431640625,
|
325 |
+
"logps/rejected": -222.0946044921875,
|
326 |
+
"loss": 0.4414,
|
327 |
+
"rewards/accuracies": 0.543749988079071,
|
328 |
+
"rewards/chosen": -0.3833041191101074,
|
329 |
+
"rewards/margins": 0.03259982913732529,
|
330 |
+
"rewards/rejected": -0.4159039556980133,
|
331 |
"step": 230
|
332 |
},
|
333 |
{
|
334 |
+
"epoch": 0.83,
|
335 |
+
"learning_rate": 7.288700744664167e-07,
|
336 |
+
"logits/chosen": 0.020644571632146835,
|
337 |
+
"logits/rejected": 0.05471833795309067,
|
338 |
+
"logps/chosen": -204.6312713623047,
|
339 |
+
"logps/rejected": -212.8608856201172,
|
340 |
+
"loss": 0.4231,
|
341 |
+
"rewards/accuracies": 0.612500011920929,
|
342 |
+
"rewards/chosen": -0.40746060013771057,
|
343 |
+
"rewards/margins": 0.06532532721757889,
|
344 |
+
"rewards/rejected": -0.47278594970703125,
|
345 |
"step": 240
|
346 |
},
|
347 |
{
|
348 |
+
"epoch": 0.86,
|
349 |
+
"learning_rate": 7.017175809949043e-07,
|
350 |
+
"logits/chosen": -0.00010571479651844129,
|
351 |
+
"logits/rejected": 0.024410605430603027,
|
352 |
+
"logps/chosen": -214.3887481689453,
|
353 |
+
"logps/rejected": -221.22830200195312,
|
354 |
+
"loss": 0.4142,
|
355 |
+
"rewards/accuracies": 0.606249988079071,
|
356 |
+
"rewards/chosen": -0.41884541511535645,
|
357 |
+
"rewards/margins": 0.052488796412944794,
|
358 |
+
"rewards/rejected": -0.47133421897888184,
|
359 |
"step": 250
|
360 |
},
|
361 |
{
|
362 |
+
"epoch": 0.9,
|
363 |
+
"learning_rate": 6.738346696376738e-07,
|
364 |
+
"logits/chosen": -0.027939433231949806,
|
365 |
+
"logits/rejected": 0.005430951714515686,
|
366 |
+
"logps/chosen": -215.529296875,
|
367 |
+
"logps/rejected": -218.4357452392578,
|
368 |
+
"loss": 0.418,
|
369 |
+
"rewards/accuracies": 0.637499988079071,
|
370 |
+
"rewards/chosen": -0.45902904868125916,
|
371 |
+
"rewards/margins": 0.07481794059276581,
|
372 |
+
"rewards/rejected": -0.5338469743728638,
|
373 |
"step": 260
|
374 |
},
|
375 |
{
|
376 |
+
"epoch": 0.93,
|
377 |
+
"learning_rate": 6.453223042128555e-07,
|
378 |
+
"logits/chosen": 0.020816553384065628,
|
379 |
+
"logits/rejected": 0.034314144402742386,
|
380 |
+
"logps/chosen": -229.1612548828125,
|
381 |
+
"logps/rejected": -241.5037078857422,
|
382 |
+
"loss": 0.4108,
|
383 |
+
"rewards/accuracies": 0.6499999761581421,
|
384 |
+
"rewards/chosen": -0.48783358931541443,
|
385 |
+
"rewards/margins": 0.08682344853878021,
|
386 |
+
"rewards/rejected": -0.5746570229530334,
|
387 |
"step": 270
|
388 |
},
|
389 |
{
|
390 |
+
"epoch": 0.96,
|
391 |
+
"learning_rate": 6.162837277871552e-07,
|
392 |
+
"logits/chosen": 0.009753930382430553,
|
393 |
+
"logits/rejected": 0.015020027756690979,
|
394 |
+
"logps/chosen": -243.6470184326172,
|
395 |
+
"logps/rejected": -253.1077117919922,
|
396 |
+
"loss": 0.4073,
|
397 |
+
"rewards/accuracies": 0.625,
|
398 |
+
"rewards/chosen": -0.5220585465431213,
|
399 |
+
"rewards/margins": 0.08642096817493439,
|
400 |
+
"rewards/rejected": -0.6084794998168945,
|
401 |
"step": 280
|
402 |
},
|
403 |
{
|
404 |
+
"epoch": 1.0,
|
405 |
+
"learning_rate": 5.868240888334652e-07,
|
406 |
+
"logits/chosen": 0.04999474436044693,
|
407 |
+
"logits/rejected": 0.09121865034103394,
|
408 |
+
"logps/chosen": -241.92050170898438,
|
409 |
+
"logps/rejected": -245.2644805908203,
|
410 |
+
"loss": 0.3966,
|
411 |
+
"rewards/accuracies": 0.518750011920929,
|
412 |
+
"rewards/chosen": -0.5882852673530579,
|
413 |
+
"rewards/margins": 0.03265915438532829,
|
414 |
+
"rewards/rejected": -0.6209444403648376,
|
415 |
"step": 290
|
416 |
},
|
417 |
{
|
418 |
+
"epoch": 1.03,
|
419 |
+
"learning_rate": 5.570500604890124e-07,
|
420 |
+
"logits/chosen": 0.004443037323653698,
|
421 |
+
"logits/rejected": 0.03949873894453049,
|
422 |
+
"logps/chosen": -231.2201385498047,
|
423 |
+
"logps/rejected": -267.2882385253906,
|
424 |
+
"loss": 0.296,
|
425 |
+
"rewards/accuracies": 0.84375,
|
426 |
+
"rewards/chosen": -0.4168514311313629,
|
427 |
+
"rewards/margins": 0.3946831226348877,
|
428 |
+
"rewards/rejected": -0.811534583568573,
|
429 |
"step": 300
|
430 |
},
|
431 |
{
|
432 |
+
"epoch": 1.07,
|
433 |
+
"learning_rate": 5.270694542927088e-07,
|
434 |
+
"logits/chosen": 0.04513490945100784,
|
435 |
+
"logits/rejected": 0.04028277471661568,
|
436 |
+
"logps/chosen": -236.4629669189453,
|
437 |
+
"logps/rejected": -286.91162109375,
|
438 |
+
"loss": 0.2439,
|
439 |
+
"rewards/accuracies": 0.831250011920929,
|
440 |
+
"rewards/chosen": -0.6017101407051086,
|
441 |
+
"rewards/margins": 0.4379204213619232,
|
442 |
+
"rewards/rejected": -1.03963041305542,
|
443 |
"step": 310
|
444 |
},
|
445 |
{
|
446 |
+
"epoch": 1.1,
|
447 |
+
"learning_rate": 4.969908298003572e-07,
|
448 |
+
"logits/chosen": 0.08646970987319946,
|
449 |
+
"logits/rejected": 0.07388903200626373,
|
450 |
+
"logps/chosen": -243.06735229492188,
|
451 |
+
"logps/rejected": -291.15106201171875,
|
452 |
+
"loss": 0.1796,
|
453 |
+
"rewards/accuracies": 0.78125,
|
454 |
+
"rewards/chosen": -0.9106127023696899,
|
455 |
+
"rewards/margins": 0.41714897751808167,
|
456 |
+
"rewards/rejected": -1.3277615308761597,
|
457 |
"step": 320
|
458 |
},
|
459 |
{
|
460 |
+
"epoch": 1.14,
|
461 |
+
"learning_rate": 4.669231014912942e-07,
|
462 |
+
"logits/chosen": 0.14787010848522186,
|
463 |
+
"logits/rejected": 0.15399041771888733,
|
464 |
+
"logps/chosen": -333.15264892578125,
|
465 |
+
"logps/rejected": -407.36944580078125,
|
466 |
+
"loss": 0.1341,
|
467 |
+
"rewards/accuracies": 0.8187500238418579,
|
468 |
+
"rewards/chosen": -1.4717319011688232,
|
469 |
+
"rewards/margins": 0.6132253408432007,
|
470 |
+
"rewards/rejected": -2.0849571228027344,
|
471 |
"step": 330
|
472 |
},
|
473 |
{
|
474 |
+
"epoch": 1.17,
|
475 |
+
"learning_rate": 4.3697514438985533e-07,
|
476 |
+
"logits/chosen": 0.2917874753475189,
|
477 |
+
"logits/rejected": 0.3329147696495056,
|
478 |
+
"logps/chosen": -340.0228576660156,
|
479 |
+
"logps/rejected": -400.05853271484375,
|
480 |
+
"loss": 0.1011,
|
481 |
+
"rewards/accuracies": 0.737500011920929,
|
482 |
+
"rewards/chosen": -1.6781260967254639,
|
483 |
+
"rewards/margins": 0.6559454202651978,
|
484 |
+
"rewards/rejected": -2.334071636199951,
|
485 |
"step": 340
|
486 |
},
|
487 |
{
|
488 |
+
"epoch": 1.21,
|
489 |
+
"learning_rate": 4.0725539982971027e-07,
|
490 |
+
"logits/chosen": 0.22039537131786346,
|
491 |
+
"logits/rejected": 0.22784602642059326,
|
492 |
+
"logps/chosen": -332.1070556640625,
|
493 |
+
"logps/rejected": -402.21575927734375,
|
494 |
+
"loss": 0.0947,
|
495 |
+
"rewards/accuracies": 0.793749988079071,
|
496 |
+
"rewards/chosen": -1.765737533569336,
|
497 |
+
"rewards/margins": 0.6189795732498169,
|
498 |
+
"rewards/rejected": -2.384716749191284,
|
499 |
"step": 350
|
500 |
},
|
501 |
{
|
502 |
+
"epoch": 1.24,
|
503 |
+
"learning_rate": 3.778714827885845e-07,
|
504 |
+
"logits/chosen": 0.2562865614891052,
|
505 |
+
"logits/rejected": 0.28801122307777405,
|
506 |
+
"logps/chosen": -332.604248046875,
|
507 |
+
"logps/rejected": -385.0444030761719,
|
508 |
+
"loss": 0.0888,
|
509 |
+
"rewards/accuracies": 0.78125,
|
510 |
+
"rewards/chosen": -1.6409450769424438,
|
511 |
+
"rewards/margins": 0.5473521947860718,
|
512 |
+
"rewards/rejected": -2.1882972717285156,
|
513 |
"step": 360
|
514 |
},
|
515 |
{
|
516 |
+
"epoch": 1.27,
|
517 |
+
"learning_rate": 3.489297922152136e-07,
|
518 |
+
"logits/chosen": 0.29671263694763184,
|
519 |
+
"logits/rejected": 0.2544638216495514,
|
520 |
+
"logps/chosen": -340.67266845703125,
|
521 |
+
"logps/rejected": -433.8108825683594,
|
522 |
+
"loss": 0.0848,
|
523 |
+
"rewards/accuracies": 0.793749988079071,
|
524 |
+
"rewards/chosen": -1.7547416687011719,
|
525 |
+
"rewards/margins": 0.7966595888137817,
|
526 |
+
"rewards/rejected": -2.551401138305664,
|
527 |
"step": 370
|
528 |
},
|
529 |
{
|
530 |
+
"epoch": 1.31,
|
531 |
+
"learning_rate": 3.205351257595271e-07,
|
532 |
+
"logits/chosen": 0.3430071473121643,
|
533 |
+
"logits/rejected": 0.3394010663032532,
|
534 |
+
"logps/chosen": -339.52386474609375,
|
535 |
+
"logps/rejected": -415.0911560058594,
|
536 |
+
"loss": 0.075,
|
537 |
+
"rewards/accuracies": 0.762499988079071,
|
538 |
+
"rewards/chosen": -1.862532377243042,
|
539 |
+
"rewards/margins": 0.7082081437110901,
|
540 |
+
"rewards/rejected": -2.5707404613494873,
|
541 |
"step": 380
|
542 |
},
|
543 |
{
|
544 |
+
"epoch": 1.34,
|
545 |
+
"learning_rate": 2.9279030030112407e-07,
|
546 |
+
"logits/chosen": 0.4035106599330902,
|
547 |
+
"logits/rejected": 0.34052735567092896,
|
548 |
+
"logps/chosen": -378.32696533203125,
|
549 |
+
"logps/rejected": -467.59075927734375,
|
550 |
+
"loss": 0.0663,
|
551 |
+
"rewards/accuracies": 0.8374999761581421,
|
552 |
+
"rewards/chosen": -2.0151519775390625,
|
553 |
+
"rewards/margins": 0.8498624563217163,
|
554 |
+
"rewards/rejected": -2.8650143146514893,
|
555 |
"step": 390
|
556 |
},
|
557 |
{
|
558 |
+
"epoch": 1.38,
|
559 |
+
"learning_rate": 2.65795779650105e-07,
|
560 |
+
"logits/chosen": 0.6060320138931274,
|
561 |
+
"logits/rejected": 0.5839828252792358,
|
562 |
+
"logps/chosen": -379.7890625,
|
563 |
+
"logps/rejected": -466.09075927734375,
|
564 |
+
"loss": 0.0648,
|
565 |
+
"rewards/accuracies": 0.8187500238418579,
|
566 |
+
"rewards/chosen": -2.1252145767211914,
|
567 |
+
"rewards/margins": 0.841218113899231,
|
568 |
+
"rewards/rejected": -2.966432809829712,
|
569 |
"step": 400
|
570 |
},
|
571 |
{
|
572 |
+
"epoch": 1.41,
|
573 |
+
"learning_rate": 2.3964931076834876e-07,
|
574 |
+
"logits/chosen": 0.5519194602966309,
|
575 |
+
"logits/rejected": 0.6366934776306152,
|
576 |
+
"logps/chosen": -366.50872802734375,
|
577 |
+
"logps/rejected": -428.6471252441406,
|
578 |
+
"loss": 0.0666,
|
579 |
+
"rewards/accuracies": 0.7437499761581421,
|
580 |
+
"rewards/chosen": -2.167186737060547,
|
581 |
+
"rewards/margins": 0.6175845861434937,
|
582 |
+
"rewards/rejected": -2.78477144241333,
|
583 |
"step": 410
|
584 |
},
|
585 |
{
|
586 |
+
"epoch": 1.45,
|
587 |
+
"learning_rate": 2.1444556982847995e-07,
|
588 |
+
"logits/chosen": 0.6113961338996887,
|
589 |
+
"logits/rejected": 0.6564093828201294,
|
590 |
+
"logps/chosen": -409.9825134277344,
|
591 |
+
"logps/rejected": -469.96697998046875,
|
592 |
+
"loss": 0.06,
|
593 |
+
"rewards/accuracies": 0.706250011920929,
|
594 |
+
"rewards/chosen": -2.302380323410034,
|
595 |
+
"rewards/margins": 0.7064192295074463,
|
596 |
+
"rewards/rejected": -3.0087997913360596,
|
597 |
"step": 420
|
598 |
},
|
599 |
{
|
600 |
+
"epoch": 1.48,
|
601 |
+
"learning_rate": 1.902758193921385e-07,
|
602 |
+
"logits/chosen": 0.5821201205253601,
|
603 |
+
"logits/rejected": 0.6013978123664856,
|
604 |
+
"logps/chosen": -398.11602783203125,
|
605 |
+
"logps/rejected": -498.4593811035156,
|
606 |
+
"loss": 0.0628,
|
607 |
+
"rewards/accuracies": 0.824999988079071,
|
608 |
+
"rewards/chosen": -2.178372859954834,
|
609 |
+
"rewards/margins": 0.9250782132148743,
|
610 |
+
"rewards/rejected": -3.1034510135650635,
|
611 |
"step": 430
|
612 |
},
|
613 |
{
|
614 |
+
"epoch": 1.52,
|
615 |
+
"learning_rate": 1.6722757794891285e-07,
|
616 |
+
"logits/chosen": 0.7665516138076782,
|
617 |
+
"logits/rejected": 0.7662865519523621,
|
618 |
+
"logps/chosen": -377.22186279296875,
|
619 |
+
"logps/rejected": -439.4972229003906,
|
620 |
+
"loss": 0.0685,
|
621 |
+
"rewards/accuracies": 0.793749988079071,
|
622 |
+
"rewards/chosen": -2.1608002185821533,
|
623 |
+
"rewards/margins": 0.674599289894104,
|
624 |
+
"rewards/rejected": -2.835399866104126,
|
625 |
"step": 440
|
626 |
},
|
627 |
{
|
628 |
+
"epoch": 1.55,
|
629 |
+
"learning_rate": 1.453843030125278e-07,
|
630 |
+
"logits/chosen": 0.6384638547897339,
|
631 |
+
"logits/rejected": 0.6626998782157898,
|
632 |
+
"logps/chosen": -406.65338134765625,
|
633 |
+
"logps/rejected": -464.52374267578125,
|
634 |
+
"loss": 0.065,
|
635 |
+
"rewards/accuracies": 0.7875000238418579,
|
636 |
+
"rewards/chosen": -2.227891683578491,
|
637 |
+
"rewards/margins": 0.6393032073974609,
|
638 |
+
"rewards/rejected": -2.8671953678131104,
|
639 |
"step": 450
|
640 |
},
|
641 |
{
|
642 |
+
"epoch": 1.58,
|
643 |
+
"learning_rate": 1.2482508892179883e-07,
|
644 |
+
"logits/chosen": 0.6284786462783813,
|
645 |
+
"logits/rejected": 0.5958371162414551,
|
646 |
+
"logps/chosen": -404.72796630859375,
|
647 |
+
"logps/rejected": -500.9256286621094,
|
648 |
+
"loss": 0.0628,
|
649 |
"rewards/accuracies": 0.8125,
|
650 |
+
"rewards/chosen": -2.1921005249023438,
|
651 |
+
"rewards/margins": 0.9022598266601562,
|
652 |
+
"rewards/rejected": -3.094360113143921,
|
653 |
"step": 460
|
654 |
},
|
655 |
{
|
656 |
+
"epoch": 1.62,
|
657 |
+
"learning_rate": 1.0562438044060846e-07,
|
658 |
+
"logits/chosen": 0.6781089901924133,
|
659 |
+
"logits/rejected": 0.6540604829788208,
|
660 |
+
"logps/chosen": -373.59014892578125,
|
661 |
+
"logps/rejected": -451.7066345214844,
|
662 |
+
"loss": 0.0619,
|
663 |
+
"rewards/accuracies": 0.7875000238418579,
|
664 |
+
"rewards/chosen": -2.1520633697509766,
|
665 |
+
"rewards/margins": 0.7549977898597717,
|
666 |
+
"rewards/rejected": -2.9070611000061035,
|
667 |
"step": 470
|
668 |
},
|
669 |
{
|
670 |
+
"epoch": 1.65,
|
671 |
+
"learning_rate": 8.785170319396174e-08,
|
672 |
+
"logits/chosen": 0.6746602058410645,
|
673 |
+
"logits/rejected": 0.6840069890022278,
|
674 |
+
"logps/chosen": -407.03350830078125,
|
675 |
+
"logps/rejected": -509.2193298339844,
|
676 |
+
"loss": 0.0583,
|
677 |
+
"rewards/accuracies": 0.831250011920929,
|
678 |
+
"rewards/chosen": -2.3068630695343018,
|
679 |
+
"rewards/margins": 0.9839746356010437,
|
680 |
+
"rewards/rejected": -3.2908377647399902,
|
681 |
"step": 480
|
682 |
},
|
683 |
{
|
684 |
+
"epoch": 1.69,
|
685 |
+
"learning_rate": 7.157141191620548e-08,
|
686 |
+
"logits/chosen": 0.7601083517074585,
|
687 |
+
"logits/rejected": 0.8036757707595825,
|
688 |
+
"logps/chosen": -383.3311462402344,
|
689 |
+
"logps/rejected": -455.77447509765625,
|
690 |
+
"loss": 0.0588,
|
691 |
+
"rewards/accuracies": 0.7875000238418579,
|
692 |
+
"rewards/chosen": -2.1426267623901367,
|
693 |
+
"rewards/margins": 0.7608777284622192,
|
694 |
+
"rewards/rejected": -2.9035048484802246,
|
695 |
"step": 490
|
696 |
},
|
697 |
{
|
698 |
+
"epoch": 1.72,
|
699 |
+
"learning_rate": 5.684245742300625e-08,
|
700 |
+
"logits/chosen": 0.7304960489273071,
|
701 |
+
"logits/rejected": 0.7055848240852356,
|
702 |
+
"logps/chosen": -382.73931884765625,
|
703 |
+
"logps/rejected": -500.84515380859375,
|
704 |
+
"loss": 0.058,
|
705 |
+
"rewards/accuracies": 0.8374999761581421,
|
706 |
+
"rewards/chosen": -2.192786931991577,
|
707 |
+
"rewards/margins": 1.0639842748641968,
|
708 |
+
"rewards/rejected": -3.2567710876464844,
|
709 |
"step": 500
|
710 |
},
|
711 |
{
|
712 |
+
"epoch": 1.76,
|
713 |
+
"learning_rate": 4.3718173150878446e-08,
|
714 |
+
"logits/chosen": 0.9688064455986023,
|
715 |
+
"logits/rejected": 1.001210331916809,
|
716 |
+
"logps/chosen": -388.23193359375,
|
717 |
+
"logps/rejected": -464.37371826171875,
|
718 |
+
"loss": 0.0555,
|
719 |
+
"rewards/accuracies": 0.800000011920929,
|
720 |
+
"rewards/chosen": -2.178377866744995,
|
721 |
+
"rewards/margins": 0.7957491874694824,
|
722 |
+
"rewards/rejected": -2.9741270542144775,
|
723 |
"step": 510
|
724 |
},
|
725 |
{
|
726 |
+
"epoch": 1.79,
|
727 |
+
"learning_rate": 3.2246082037199526e-08,
|
728 |
+
"logits/chosen": 0.8018981218338013,
|
729 |
+
"logits/rejected": 0.8608972430229187,
|
730 |
+
"logps/chosen": -382.80157470703125,
|
731 |
+
"logps/rejected": -468.1249084472656,
|
732 |
+
"loss": 0.0604,
|
733 |
+
"rewards/accuracies": 0.8374999761581421,
|
734 |
+
"rewards/chosen": -2.2035393714904785,
|
735 |
+
"rewards/margins": 0.8492420315742493,
|
736 |
+
"rewards/rejected": -3.052781343460083,
|
737 |
"step": 520
|
738 |
},
|
739 |
{
|
740 |
+
"epoch": 1.83,
|
741 |
+
"learning_rate": 2.2467724440002333e-08,
|
742 |
+
"logits/chosen": 0.7628926038742065,
|
743 |
+
"logits/rejected": 0.7621763944625854,
|
744 |
+
"logps/chosen": -394.536376953125,
|
745 |
+
"logps/rejected": -478.94598388671875,
|
746 |
+
"loss": 0.0564,
|
747 |
+
"rewards/accuracies": 0.78125,
|
748 |
+
"rewards/chosen": -2.241116523742676,
|
749 |
+
"rewards/margins": 0.7960697412490845,
|
750 |
+
"rewards/rejected": -3.0371861457824707,
|
751 |
"step": 530
|
752 |
},
|
753 |
{
|
754 |
+
"epoch": 1.86,
|
755 |
+
"learning_rate": 1.4418507720641793e-08,
|
756 |
+
"logits/chosen": 0.8032611012458801,
|
757 |
+
"logits/rejected": 0.8268760442733765,
|
758 |
+
"logps/chosen": -413.1278381347656,
|
759 |
+
"logps/rejected": -507.7405700683594,
|
760 |
+
"loss": 0.0575,
|
761 |
+
"rewards/accuracies": 0.793749988079071,
|
762 |
+
"rewards/chosen": -2.3634586334228516,
|
763 |
+
"rewards/margins": 0.9177207946777344,
|
764 |
+
"rewards/rejected": -3.281179428100586,
|
765 |
"step": 540
|
766 |
},
|
767 |
{
|
768 |
+
"epoch": 1.89,
|
769 |
+
"learning_rate": 8.127578033998661e-09,
|
770 |
+
"logits/chosen": 0.8457974195480347,
|
771 |
+
"logits/rejected": 0.8470609784126282,
|
772 |
+
"logps/chosen": -351.919677734375,
|
773 |
+
"logps/rejected": -441.13238525390625,
|
774 |
+
"loss": 0.0589,
|
775 |
+
"rewards/accuracies": 0.8125,
|
776 |
+
"rewards/chosen": -2.1192312240600586,
|
777 |
+
"rewards/margins": 0.815728485584259,
|
778 |
+
"rewards/rejected": -2.934959888458252,
|
779 |
"step": 550
|
780 |
},
|
781 |
{
|
782 |
+
"epoch": 1.93,
|
783 |
+
"learning_rate": 3.6177147904655757e-09,
|
784 |
+
"logits/chosen": 0.8455848693847656,
|
785 |
+
"logits/rejected": 0.8221101760864258,
|
786 |
+
"logps/chosen": -420.8208923339844,
|
787 |
+
"logps/rejected": -514.7294311523438,
|
788 |
+
"loss": 0.0538,
|
789 |
+
"rewards/accuracies": 0.8187500238418579,
|
790 |
+
"rewards/chosen": -2.299787998199463,
|
791 |
+
"rewards/margins": 0.9560101628303528,
|
792 |
+
"rewards/rejected": -3.255797863006592,
|
793 |
"step": 560
|
794 |
},
|
795 |
{
|
796 |
+
"epoch": 1.96,
|
797 |
+
"learning_rate": 9.052481718690997e-10,
|
798 |
+
"logits/chosen": 0.8840595483779907,
|
799 |
+
"logits/rejected": 0.8795121908187866,
|
800 |
+
"logps/chosen": -424.8587951660156,
|
801 |
+
"logps/rejected": -513.06591796875,
|
802 |
+
"loss": 0.0555,
|
803 |
+
"rewards/accuracies": 0.824999988079071,
|
804 |
+
"rewards/chosen": -2.4113454818725586,
|
805 |
+
"rewards/margins": 0.8990417718887329,
|
806 |
+
"rewards/rejected": -3.310387372970581,
|
807 |
"step": 570
|
808 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
809 |
{
|
810 |
"epoch": 2.0,
|
811 |
"learning_rate": 0.0,
|
812 |
+
"logits/chosen": 0.9808861017227173,
|
813 |
+
"logits/rejected": 0.9314306378364563,
|
814 |
+
"logps/chosen": -372.04034423828125,
|
815 |
+
"logps/rejected": -452.5663146972656,
|
816 |
+
"loss": 0.0544,
|
817 |
+
"rewards/accuracies": 0.831250011920929,
|
818 |
+
"rewards/chosen": -2.2187602519989014,
|
819 |
+
"rewards/margins": 0.7363004684448242,
|
820 |
+
"rewards/rejected": -2.9550607204437256,
|
821 |
+
"step": 580
|
822 |
},
|
823 |
{
|
824 |
"epoch": 2.0,
|
825 |
+
"step": 580,
|
826 |
"total_flos": 0.0,
|
827 |
+
"train_loss": 0.31343152379167494,
|
828 |
+
"train_runtime": 8630.1644,
|
829 |
+
"train_samples_per_second": 8.607,
|
830 |
+
"train_steps_per_second": 0.067
|
831 |
}
|
832 |
],
|
833 |
"logging_steps": 10,
|
834 |
+
"max_steps": 580,
|
835 |
"num_train_epochs": 2,
|
836 |
"save_steps": 10000,
|
837 |
"total_flos": 0.0,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6648
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc9d7c7d0bacc27c393dcf6748242363d5a7ffc1ed7d1bdf238aa7708ab53c34
|
3 |
size 6648
|