Model save
Browse files- README.md +19 -3
- all_results.json +4 -4
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- train_results.json +4 -4
- trainer_state.json +491 -427
- training_args.bin +1 -1
README.md
CHANGED
@@ -14,6 +14,16 @@ should probably proofread and complete it, then remove this comment. -->
|
|
14 |
# zephyr-7b-dpo-full
|
15 |
|
16 |
This model was trained from scratch on the None dataset.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
## Model description
|
19 |
|
@@ -33,12 +43,12 @@ More information needed
|
|
33 |
|
34 |
The following hyperparameters were used during training:
|
35 |
- learning_rate: 3e-06
|
36 |
-
- train_batch_size:
|
37 |
- eval_batch_size: 8
|
38 |
-
- seed:
|
39 |
- distributed_type: multi-GPU
|
40 |
- num_devices: 8
|
41 |
-
- gradient_accumulation_steps:
|
42 |
- total_train_batch_size: 128
|
43 |
- total_eval_batch_size: 64
|
44 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
@@ -48,6 +58,12 @@ The following hyperparameters were used during training:
|
|
48 |
|
49 |
### Training results
|
50 |
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
|
52 |
|
53 |
### Framework versions
|
|
|
14 |
# zephyr-7b-dpo-full
|
15 |
|
16 |
This model was trained from scratch on the None dataset.
|
17 |
+
It achieves the following results on the evaluation set:
|
18 |
+
- Loss: 0.0427
|
19 |
+
- Rewards/chosen: -1.1712
|
20 |
+
- Rewards/rejected: -2.0556
|
21 |
+
- Rewards/accuracies: 0.7266
|
22 |
+
- Rewards/margins: 0.8844
|
23 |
+
- Logps/rejected: -517.1834
|
24 |
+
- Logps/chosen: -420.7032
|
25 |
+
- Logits/rejected: -0.0713
|
26 |
+
- Logits/chosen: -0.0801
|
27 |
|
28 |
## Model description
|
29 |
|
|
|
43 |
|
44 |
The following hyperparameters were used during training:
|
45 |
- learning_rate: 3e-06
|
46 |
+
- train_batch_size: 4
|
47 |
- eval_batch_size: 8
|
48 |
+
- seed: 2
|
49 |
- distributed_type: multi-GPU
|
50 |
- num_devices: 8
|
51 |
+
- gradient_accumulation_steps: 4
|
52 |
- total_train_batch_size: 128
|
53 |
- total_eval_batch_size: 64
|
54 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
|
|
58 |
|
59 |
### Training results
|
60 |
|
61 |
+
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
62 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
63 |
+
| 0.0642 | 0.21 | 100 | 0.0487 | -1.3984 | -2.4508 | 0.7188 | 1.0524 | -556.7064 | -443.4212 | 0.4579 | 0.4301 |
|
64 |
+
| 0.047 | 0.42 | 200 | 0.0461 | -1.1146 | -1.8461 | 0.7422 | 0.7314 | -496.2327 | -415.0494 | 0.1274 | 0.1123 |
|
65 |
+
| 0.0401 | 0.63 | 300 | 0.0408 | -1.2816 | -2.1650 | 0.7148 | 0.8834 | -528.1252 | -431.7439 | -0.1765 | -0.1928 |
|
66 |
+
| 0.0435 | 0.84 | 400 | 0.0427 | -1.1712 | -2.0556 | 0.7266 | 0.8844 | -517.1834 | -420.7032 | -0.0713 | -0.0801 |
|
67 |
|
68 |
|
69 |
### Framework versions
|
all_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 61134,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
+
"train_loss": 0.05794826206171288,
|
4 |
+
"train_runtime": 4397.3222,
|
5 |
"train_samples": 61134,
|
6 |
+
"train_samples_per_second": 13.903,
|
7 |
+
"train_steps_per_second": 0.108
|
8 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4976698672
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09b30f17a60c9620148cbd5eb76ec7aa8992e70e6d7a4b2040af7a3e47fd1c42
|
3 |
size 4976698672
|
model-00002-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999802720
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a2c540e131df728dc7e661af77405e10eb54b3db4281b7eed1e89377b4dc105
|
3 |
size 4999802720
|
model-00003-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4915916176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:980ee41ac40131d5176d7e7821579b166f4636c9f8b54a41d57709f2dbcd51da
|
3 |
size 4915916176
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1168138808
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:23f4c41ec0c959598b2ab2b7e53a1d3f3200d53942ef2e6305032dea475d5633
|
3 |
size 1168138808
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 61134,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 1.0,
|
3 |
+
"train_loss": 0.05794826206171288,
|
4 |
+
"train_runtime": 4397.3222,
|
5 |
"train_samples": 61134,
|
6 |
+
"train_samples_per_second": 13.903,
|
7 |
+
"train_steps_per_second": 0.108
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch": 0.
|
5 |
-
"eval_steps":
|
6 |
"global_step": 477,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
@@ -11,675 +11,739 @@
|
|
11 |
{
|
12 |
"epoch": 0.02,
|
13 |
"learning_rate": 6.25e-07,
|
14 |
-
"logits/chosen": 0.
|
15 |
-
"logits/rejected": 0.
|
16 |
-
"logps/chosen": -
|
17 |
-
"logps/rejected": -
|
18 |
-
"loss": 0.
|
19 |
-
"rewards/accuracies": 0.
|
20 |
-
"rewards/chosen":
|
21 |
-
"rewards/margins":
|
22 |
-
"rewards/rejected":
|
23 |
"step": 10
|
24 |
},
|
25 |
{
|
26 |
"epoch": 0.04,
|
27 |
"learning_rate": 1.25e-06,
|
28 |
-
"logits/chosen": 0.
|
29 |
-
"logits/rejected": 0.
|
30 |
-
"logps/chosen": -
|
31 |
-
"logps/rejected": -
|
32 |
-
"loss": 0.
|
33 |
-
"rewards/accuracies": 0.
|
34 |
-
"rewards/chosen": 0.
|
35 |
-
"rewards/margins": 0.
|
36 |
-
"rewards/rejected": -0.
|
37 |
"step": 20
|
38 |
},
|
39 |
{
|
40 |
"epoch": 0.06,
|
41 |
"learning_rate": 1.875e-06,
|
42 |
-
"logits/chosen": 0.
|
43 |
-
"logits/rejected": 0.
|
44 |
-
"logps/chosen": -
|
45 |
-
"logps/rejected": -
|
46 |
-
"loss": 0.
|
47 |
"rewards/accuracies": 0.6812499761581421,
|
48 |
-
"rewards/chosen": -0.
|
49 |
-
"rewards/margins": 0.
|
50 |
-
"rewards/rejected": -0.
|
51 |
"step": 30
|
52 |
},
|
53 |
{
|
54 |
"epoch": 0.08,
|
55 |
"learning_rate": 2.5e-06,
|
56 |
-
"logits/chosen": 0.
|
57 |
-
"logits/rejected": 0.
|
58 |
-
"logps/chosen": -
|
59 |
-
"logps/rejected": -
|
60 |
-
"loss": 0.
|
61 |
"rewards/accuracies": 0.668749988079071,
|
62 |
-
"rewards/chosen": -0.
|
63 |
-
"rewards/margins": 0.
|
64 |
-
"rewards/rejected": -0.
|
65 |
"step": 40
|
66 |
},
|
67 |
{
|
68 |
"epoch": 0.1,
|
69 |
"learning_rate": 2.999839121261416e-06,
|
70 |
-
"logits/chosen": 0.
|
71 |
-
"logits/rejected": 0.
|
72 |
-
"logps/chosen": -
|
73 |
-
"logps/rejected": -
|
74 |
-
"loss": 0.
|
75 |
-
"rewards/accuracies": 0.
|
76 |
-
"rewards/chosen": -0.
|
77 |
-
"rewards/margins": 0.
|
78 |
-
"rewards/rejected": -0.
|
79 |
"step": 50
|
80 |
},
|
81 |
{
|
82 |
"epoch": 0.13,
|
83 |
"learning_rate": 2.994211988057582e-06,
|
84 |
-
"logits/chosen": 0.
|
85 |
-
"logits/rejected": 0.
|
86 |
-
"logps/chosen": -
|
87 |
-
"logps/rejected": -
|
88 |
-
"loss": 0.
|
89 |
-
"rewards/accuracies": 0.
|
90 |
-
"rewards/chosen": -0.
|
91 |
-
"rewards/margins": 0.
|
92 |
-
"rewards/rejected": -0.
|
93 |
"step": 60
|
94 |
},
|
95 |
{
|
96 |
"epoch": 0.15,
|
97 |
"learning_rate": 2.9805753939568693e-06,
|
98 |
-
"logits/chosen": 0.
|
99 |
-
"logits/rejected": 0.
|
100 |
-
"logps/chosen": -
|
101 |
-
"logps/rejected": -
|
102 |
-
"loss": 0.
|
103 |
-
"rewards/accuracies": 0.
|
104 |
-
"rewards/chosen": -0.
|
105 |
-
"rewards/margins": 0.
|
106 |
-
"rewards/rejected": -
|
107 |
"step": 70
|
108 |
},
|
109 |
{
|
110 |
"epoch": 0.17,
|
111 |
"learning_rate": 2.959002435526626e-06,
|
112 |
-
"logits/chosen": 0.
|
113 |
-
"logits/rejected": 0.
|
114 |
-
"logps/chosen": -
|
115 |
-
"logps/rejected": -
|
116 |
-
"loss": 0.
|
117 |
-
"rewards/accuracies": 0.
|
118 |
-
"rewards/chosen": -0.
|
119 |
-
"rewards/margins": 0.
|
120 |
-
"rewards/rejected": -1.
|
121 |
"step": 80
|
122 |
},
|
123 |
{
|
124 |
"epoch": 0.19,
|
125 |
"learning_rate": 2.929608750821129e-06,
|
126 |
-
"logits/chosen": 0.
|
127 |
-
"logits/rejected": 0.
|
128 |
-
"logps/chosen": -
|
129 |
-
"logps/rejected": -
|
130 |
-
"loss": 0.
|
131 |
-
"rewards/accuracies": 0.
|
132 |
-
"rewards/chosen": -0.
|
133 |
-
"rewards/margins": 0.
|
134 |
-
"rewards/rejected": -1.
|
135 |
"step": 90
|
136 |
},
|
137 |
{
|
138 |
"epoch": 0.21,
|
139 |
"learning_rate": 2.892551899524109e-06,
|
140 |
-
"logits/chosen": 0.
|
141 |
-
"logits/rejected": 0.
|
142 |
-
"logps/chosen": -
|
143 |
-
"logps/rejected": -
|
144 |
-
"loss": 0.
|
145 |
-
"rewards/accuracies": 0.
|
146 |
-
"rewards/chosen": -
|
147 |
-
"rewards/margins": 0.
|
148 |
-
"rewards/rejected": -1.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
149 |
"step": 100
|
150 |
},
|
151 |
{
|
152 |
"epoch": 0.23,
|
153 |
"learning_rate": 2.848030518377739e-06,
|
154 |
-
"logits/chosen": 0.
|
155 |
-
"logits/rejected": 0.
|
156 |
-
"logps/chosen": -
|
157 |
-
"logps/rejected": -
|
158 |
-
"loss": 0.
|
159 |
-
"rewards/accuracies": 0.
|
160 |
-
"rewards/chosen": -
|
161 |
-
"rewards/margins": 0.
|
162 |
-
"rewards/rejected": -
|
163 |
"step": 110
|
164 |
},
|
165 |
{
|
166 |
"epoch": 0.25,
|
167 |
"learning_rate": 2.7962832564252724e-06,
|
168 |
-
"logits/chosen": 0.
|
169 |
-
"logits/rejected": 0.
|
170 |
-
"logps/chosen": -
|
171 |
-
"logps/rejected": -
|
172 |
-
"loss": 0.
|
173 |
-
"rewards/accuracies": 0.
|
174 |
-
"rewards/chosen": -
|
175 |
-
"rewards/margins": 0.
|
176 |
-
"rewards/rejected": -1.
|
177 |
"step": 120
|
178 |
},
|
179 |
{
|
180 |
"epoch": 0.27,
|
181 |
"learning_rate": 2.7375874957747644e-06,
|
182 |
-
"logits/chosen": 0.
|
183 |
-
"logits/rejected": 0.
|
184 |
-
"logps/chosen": -
|
185 |
-
"logps/rejected": -
|
186 |
-
"loss": 0.
|
187 |
-
"rewards/accuracies": 0.
|
188 |
-
"rewards/chosen": -0.
|
189 |
-
"rewards/margins": 0.
|
190 |
-
"rewards/rejected": -1.
|
191 |
"step": 130
|
192 |
},
|
193 |
{
|
194 |
"epoch": 0.29,
|
195 |
"learning_rate": 2.672257864741005e-06,
|
196 |
-
"logits/chosen": 0.
|
197 |
-
"logits/rejected": 0.
|
198 |
-
"logps/chosen": -
|
199 |
-
"logps/rejected": -
|
200 |
-
"loss": 0.
|
201 |
-
"rewards/accuracies": 0.
|
202 |
-
"rewards/chosen": -0.
|
203 |
-
"rewards/margins": 0.
|
204 |
-
"rewards/rejected": -1.
|
205 |
"step": 140
|
206 |
},
|
207 |
{
|
208 |
"epoch": 0.31,
|
209 |
"learning_rate": 2.600644551335706e-06,
|
210 |
-
"logits/chosen": 0.
|
211 |
-
"logits/rejected": 0.
|
212 |
-
"logps/chosen": -
|
213 |
-
"logps/rejected": -
|
214 |
-
"loss": 0.
|
215 |
-
"rewards/accuracies": 0.
|
216 |
-
"rewards/chosen": -
|
217 |
-
"rewards/margins": 0.
|
218 |
-
"rewards/rejected": -1.
|
219 |
"step": 150
|
220 |
},
|
221 |
{
|
222 |
"epoch": 0.33,
|
223 |
"learning_rate": 2.5231314261461732e-06,
|
224 |
-
"logits/chosen": 0.
|
225 |
-
"logits/rejected": 0.
|
226 |
-
"logps/chosen": -
|
227 |
-
"logps/rejected": -
|
228 |
-
"loss": 0.
|
229 |
-
"rewards/accuracies": 0.
|
230 |
-
"rewards/chosen": -
|
231 |
-
"rewards/margins": 0.
|
232 |
-
"rewards/rejected": -
|
233 |
"step": 160
|
234 |
},
|
235 |
{
|
236 |
"epoch": 0.36,
|
237 |
"learning_rate": 2.440133984664454e-06,
|
238 |
-
"logits/chosen":
|
239 |
-
"logits/rejected":
|
240 |
-
"logps/chosen": -
|
241 |
-
"logps/rejected": -
|
242 |
-
"loss": 0.
|
243 |
-
"rewards/accuracies": 0.
|
244 |
-
"rewards/chosen": -
|
245 |
-
"rewards/margins": 0.
|
246 |
-
"rewards/rejected": -1.
|
247 |
"step": 170
|
248 |
},
|
249 |
{
|
250 |
"epoch": 0.38,
|
251 |
"learning_rate": 2.3520971200967337e-06,
|
252 |
-
"logits/chosen":
|
253 |
-
"logits/rejected":
|
254 |
-
"logps/chosen": -
|
255 |
-
"logps/rejected": -
|
256 |
-
"loss": 0.
|
257 |
-
"rewards/accuracies": 0.
|
258 |
-
"rewards/chosen": -
|
259 |
-
"rewards/margins": 0.
|
260 |
-
"rewards/rejected": -1.
|
261 |
"step": 180
|
262 |
},
|
263 |
{
|
264 |
"epoch": 0.4,
|
265 |
"learning_rate": 2.2594927385914546e-06,
|
266 |
-
"logits/chosen":
|
267 |
-
"logits/rejected":
|
268 |
-
"logps/chosen": -
|
269 |
-
"logps/rejected": -
|
270 |
-
"loss": 0.
|
271 |
-
"rewards/accuracies": 0.
|
272 |
-
"rewards/chosen": -
|
273 |
-
"rewards/margins": 0.
|
274 |
-
"rewards/rejected": -1.
|
275 |
"step": 190
|
276 |
},
|
277 |
{
|
278 |
"epoch": 0.42,
|
279 |
"learning_rate": 2.1628172296692954e-06,
|
280 |
-
"logits/chosen":
|
281 |
-
"logits/rejected":
|
282 |
-
"logps/chosen": -
|
283 |
-
"logps/rejected": -
|
284 |
-
"loss": 0.
|
285 |
-
"rewards/accuracies": 0.
|
286 |
-
"rewards/chosen": -
|
287 |
-
"rewards/margins": 0.
|
288 |
-
"rewards/rejected": -1.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
289 |
"step": 200
|
290 |
},
|
291 |
{
|
292 |
"epoch": 0.44,
|
293 |
"learning_rate": 2.062588805414343e-06,
|
294 |
-
"logits/chosen": -0.
|
295 |
-
"logits/rejected": 0.
|
296 |
-
"logps/chosen": -
|
297 |
-
"logps/rejected": -
|
298 |
-
"loss": 0.
|
299 |
-
"rewards/accuracies": 0.
|
300 |
-
"rewards/chosen": -
|
301 |
-
"rewards/margins": 0.
|
302 |
-
"rewards/rejected": -1.
|
303 |
"step": 210
|
304 |
},
|
305 |
{
|
306 |
"epoch": 0.46,
|
307 |
"learning_rate": 1.9593447226892386e-06,
|
308 |
-
"logits/chosen": -0.
|
309 |
-
"logits/rejected": -0.
|
310 |
-
"logps/chosen": -
|
311 |
-
"logps/rejected": -
|
312 |
-
"loss": 0.
|
313 |
-
"rewards/accuracies": 0.
|
314 |
-
"rewards/chosen": -
|
315 |
-
"rewards/margins": 0.
|
316 |
-
"rewards/rejected": -
|
317 |
"step": 220
|
318 |
},
|
319 |
{
|
320 |
"epoch": 0.48,
|
321 |
"learning_rate": 1.853638403264141e-06,
|
322 |
-
"logits/chosen": -0.
|
323 |
-
"logits/rejected": -0.
|
324 |
-
"logps/chosen": -
|
325 |
-
"logps/rejected": -
|
326 |
-
"loss": 0.
|
327 |
-
"rewards/accuracies": 0.
|
328 |
-
"rewards/chosen": -
|
329 |
-
"rewards/margins": 0.
|
330 |
-
"rewards/rejected": -
|
331 |
"step": 230
|
332 |
},
|
333 |
{
|
334 |
"epoch": 0.5,
|
335 |
"learning_rate": 1.7460364672965328e-06,
|
336 |
-
"logits/chosen": -0.
|
337 |
-
"logits/rejected": -0.
|
338 |
-
"logps/chosen": -
|
339 |
-
"logps/rejected": -
|
340 |
-
"loss": 0.
|
341 |
-
"rewards/accuracies": 0.
|
342 |
-
"rewards/chosen": -
|
343 |
-
"rewards/margins": 0.
|
344 |
-
"rewards/rejected": -1.
|
345 |
"step": 240
|
346 |
},
|
347 |
{
|
348 |
"epoch": 0.52,
|
349 |
"learning_rate": 1.637115696063402e-06,
|
350 |
-
"logits/chosen": -0.
|
351 |
-
"logits/rejected": -0.
|
352 |
-
"logps/chosen": -
|
353 |
-
"logps/rejected": -
|
354 |
-
"loss": 0.
|
355 |
-
"rewards/accuracies": 0.
|
356 |
-
"rewards/chosen": -
|
357 |
-
"rewards/margins": 0.
|
358 |
-
"rewards/rejected": -1.
|
359 |
"step": 250
|
360 |
},
|
361 |
{
|
362 |
"epoch": 0.54,
|
363 |
"learning_rate": 1.5274599402265162e-06,
|
364 |
-
"logits/chosen": -0.
|
365 |
-
"logits/rejected": -0.
|
366 |
-
"logps/chosen": -
|
367 |
-
"logps/rejected": -
|
368 |
-
"loss": 0.
|
369 |
-
"rewards/accuracies": 0.
|
370 |
-
"rewards/chosen": -
|
371 |
-
"rewards/margins": 0.
|
372 |
-
"rewards/rejected": -
|
373 |
"step": 260
|
374 |
},
|
375 |
{
|
376 |
"epoch": 0.57,
|
377 |
"learning_rate": 1.4176569902035088e-06,
|
378 |
-
"logits/chosen": -0.
|
379 |
-
"logits/rejected": -0.
|
380 |
-
"logps/chosen": -
|
381 |
-
"logps/rejected": -
|
382 |
-
"loss": 0.
|
383 |
-
"rewards/accuracies": 0.
|
384 |
-
"rewards/chosen": -1.
|
385 |
-
"rewards/margins": 0.
|
386 |
-
"rewards/rejected": -1.
|
387 |
"step": 270
|
388 |
},
|
389 |
{
|
390 |
"epoch": 0.59,
|
391 |
"learning_rate": 1.308295425420593e-06,
|
392 |
-
"logits/chosen": -0.
|
393 |
-
"logits/rejected": -0.
|
394 |
-
"logps/chosen": -
|
395 |
-
"logps/rejected": -
|
396 |
-
"loss": 0.
|
397 |
-
"rewards/accuracies": 0.
|
398 |
-
"rewards/chosen": -1.
|
399 |
-
"rewards/margins": 0.
|
400 |
-
"rewards/rejected": -
|
401 |
"step": 280
|
402 |
},
|
403 |
{
|
404 |
"epoch": 0.61,
|
405 |
"learning_rate": 1.1999614593359337e-06,
|
406 |
-
"logits/chosen": -0.
|
407 |
-
"logits/rejected": -0.
|
408 |
-
"logps/chosen": -
|
409 |
-
"logps/rejected": -
|
410 |
-
"loss": 0.
|
411 |
-
"rewards/accuracies": 0.
|
412 |
-
"rewards/chosen": -1.
|
413 |
-
"rewards/margins": 0.
|
414 |
-
"rewards/rejected": -
|
415 |
"step": 290
|
416 |
},
|
417 |
{
|
418 |
"epoch": 0.63,
|
419 |
"learning_rate": 1.0932357971453745e-06,
|
420 |
-
"logits/chosen": -0.
|
421 |
-
"logits/rejected": -0.
|
422 |
-
"logps/chosen": -
|
423 |
-
"logps/rejected": -
|
424 |
-
"loss": 0.
|
425 |
-
"rewards/accuracies": 0.
|
426 |
-
"rewards/chosen": -1.
|
427 |
-
"rewards/margins": 0.
|
428 |
-
"rewards/rejected": -2.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
429 |
"step": 300
|
430 |
},
|
431 |
{
|
432 |
"epoch": 0.65,
|
433 |
"learning_rate": 9.886905230142433e-07,
|
434 |
-
"logits/chosen": -0.
|
435 |
-
"logits/rejected": -0.
|
436 |
-
"logps/chosen": -
|
437 |
-
"logps/rejected": -
|
438 |
-
"loss": 0.
|
439 |
-
"rewards/accuracies": 0.
|
440 |
-
"rewards/chosen": -1.
|
441 |
-
"rewards/margins": 0.
|
442 |
-
"rewards/rejected": -2.
|
443 |
"step": 310
|
444 |
},
|
445 |
{
|
446 |
"epoch": 0.67,
|
447 |
"learning_rate": 8.868860335206678e-07,
|
448 |
-
"logits/chosen": -0.
|
449 |
-
"logits/rejected": -0.
|
450 |
-
"logps/chosen": -
|
451 |
-
"logps/rejected": -
|
452 |
-
"loss": 0.
|
453 |
-
"rewards/accuracies": 0.
|
454 |
-
"rewards/chosen": -1.
|
455 |
-
"rewards/margins":
|
456 |
-
"rewards/rejected": -
|
457 |
"step": 320
|
458 |
},
|
459 |
{
|
460 |
"epoch": 0.69,
|
461 |
"learning_rate": 7.883680337481599e-07,
|
462 |
-
"logits/chosen": -0.
|
463 |
-
"logits/rejected": -0.
|
464 |
-
"logps/chosen": -
|
465 |
-
"logps/rejected": -
|
466 |
-
"loss": 0.
|
467 |
-
"rewards/accuracies": 0.
|
468 |
-
"rewards/chosen": -1.
|
469 |
-
"rewards/margins":
|
470 |
-
"rewards/rejected": -
|
471 |
"step": 330
|
472 |
},
|
473 |
{
|
474 |
"epoch": 0.71,
|
475 |
"learning_rate": 6.936646121293654e-07,
|
476 |
-
"logits/chosen": -0.
|
477 |
-
"logits/rejected": -0.
|
478 |
-
"logps/chosen": -
|
479 |
-
"logps/rejected": -
|
480 |
-
"loss": 0.
|
481 |
-
"rewards/accuracies": 0.
|
482 |
-
"rewards/chosen": -1.
|
483 |
-
"rewards/margins": 0.
|
484 |
-
"rewards/rejected": -2.
|
485 |
"step": 340
|
486 |
},
|
487 |
{
|
488 |
"epoch": 0.73,
|
489 |
"learning_rate": 6.032834097207889e-07,
|
490 |
-
"logits/chosen": -0.
|
491 |
-
"logits/rejected": -0.
|
492 |
-
"logps/chosen": -
|
493 |
-
"logps/rejected": -
|
494 |
-
"loss": 0.
|
495 |
-
"rewards/accuracies": 0.
|
496 |
-
"rewards/chosen": -1.
|
497 |
-
"rewards/margins": 0.
|
498 |
-
"rewards/rejected": -2.
|
499 |
"step": 350
|
500 |
},
|
501 |
{
|
502 |
"epoch": 0.75,
|
503 |
"learning_rate": 5.177088990820725e-07,
|
504 |
-
"logits/chosen": -0.
|
505 |
-
"logits/rejected": -0.
|
506 |
-
"logps/chosen": -
|
507 |
-
"logps/rejected": -
|
508 |
-
"loss": 0.
|
509 |
-
"rewards/accuracies": 0.
|
510 |
-
"rewards/chosen": -1.
|
511 |
-
"rewards/margins": 0.
|
512 |
-
"rewards/rejected": -2.
|
513 |
"step": 360
|
514 |
},
|
515 |
{
|
516 |
"epoch": 0.77,
|
517 |
"learning_rate": 4.3739978734594494e-07,
|
518 |
-
"logits/chosen": -0.
|
519 |
-
"logits/rejected": -0.
|
520 |
-
"logps/chosen": -
|
521 |
-
"logps/rejected": -
|
522 |
-
"loss": 0.
|
523 |
-
"rewards/accuracies": 0.
|
524 |
-
"rewards/chosen": -1.
|
525 |
-
"rewards/margins": 0.
|
526 |
-
"rewards/rejected": -2.
|
527 |
"step": 370
|
528 |
},
|
529 |
{
|
530 |
"epoch": 0.8,
|
531 |
"learning_rate": 3.627865573992087e-07,
|
532 |
-
"logits/chosen": -0.
|
533 |
-
"logits/rejected": -0.
|
534 |
-
"logps/chosen": -
|
535 |
-
"logps/rejected": -
|
536 |
-
"loss": 0.
|
537 |
-
"rewards/accuracies": 0.
|
538 |
-
"rewards/chosen": -1.
|
539 |
-
"rewards/margins": 0.
|
540 |
-
"rewards/rejected": -2.
|
541 |
"step": 380
|
542 |
},
|
543 |
{
|
544 |
"epoch": 0.82,
|
545 |
"learning_rate": 2.9426916035484166e-07,
|
546 |
-
"logits/chosen": -0.
|
547 |
-
"logits/rejected": -0.
|
548 |
-
"logps/chosen": -
|
549 |
-
"logps/rejected": -
|
550 |
-
"loss": 0.
|
551 |
-
"rewards/accuracies": 0.
|
552 |
-
"rewards/chosen": -1.
|
553 |
-
"rewards/margins": 0.
|
554 |
-
"rewards/rejected": -2.
|
555 |
"step": 390
|
556 |
},
|
557 |
{
|
558 |
"epoch": 0.84,
|
559 |
"learning_rate": 2.322148716843081e-07,
|
560 |
-
"logits/chosen": -0.
|
561 |
-
"logits/rejected": -0.
|
562 |
-
"logps/chosen": -
|
563 |
-
"logps/rejected": -
|
564 |
-
"loss": 0.
|
565 |
-
"rewards/accuracies": 0.
|
566 |
-
"rewards/chosen": -1.
|
567 |
-
"rewards/margins": 0.
|
568 |
-
"rewards/rejected": -2.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
569 |
"step": 400
|
570 |
},
|
571 |
{
|
572 |
"epoch": 0.86,
|
573 |
"learning_rate": 1.7695632250191002e-07,
|
574 |
-
"logits/chosen": -0.
|
575 |
-
"logits/rejected": -0.
|
576 |
-
"logps/chosen": -
|
577 |
-
"logps/rejected": -
|
578 |
-
"loss": 0.
|
579 |
-
"rewards/accuracies": 0.
|
580 |
-
"rewards/chosen": -1.
|
581 |
-
"rewards/margins": 0.
|
582 |
-
"rewards/rejected": -
|
583 |
"step": 410
|
584 |
},
|
585 |
{
|
586 |
"epoch": 0.88,
|
587 |
"learning_rate": 1.2878971655412515e-07,
|
588 |
-
"logits/chosen": -0.
|
589 |
-
"logits/rejected": -0.
|
590 |
-
"logps/chosen": -
|
591 |
-
"logps/rejected": -
|
592 |
-
"loss": 0.
|
593 |
-
"rewards/accuracies": 0.
|
594 |
-
"rewards/chosen": -1.
|
595 |
-
"rewards/margins": 0.
|
596 |
-
"rewards/rejected": -
|
597 |
"step": 420
|
598 |
},
|
599 |
{
|
600 |
"epoch": 0.9,
|
601 |
"learning_rate": 8.797324247145411e-08,
|
602 |
-
"logits/chosen": -0.
|
603 |
-
"logits/rejected": -0.
|
604 |
-
"logps/chosen": -
|
605 |
-
"logps/rejected": -
|
606 |
-
"loss": 0.
|
607 |
"rewards/accuracies": 0.7250000238418579,
|
608 |
-
"rewards/chosen": -1.
|
609 |
-
"rewards/margins": 0.
|
610 |
-
"rewards/rejected": -2.
|
611 |
"step": 430
|
612 |
},
|
613 |
{
|
614 |
"epoch": 0.92,
|
615 |
"learning_rate": 5.472568979361853e-08,
|
616 |
-
"logits/chosen": -0.
|
617 |
-
"logits/rejected": -0.
|
618 |
-
"logps/chosen": -
|
619 |
-
"logps/rejected": -
|
620 |
-
"loss": 0.
|
621 |
-
"rewards/accuracies": 0.
|
622 |
-
"rewards/chosen": -1.
|
623 |
-
"rewards/margins": 0.
|
624 |
-
"rewards/rejected": -2.
|
625 |
"step": 440
|
626 |
},
|
627 |
{
|
628 |
"epoch": 0.94,
|
629 |
"learning_rate": 2.922527618666465e-08,
|
630 |
-
"logits/chosen": -0.
|
631 |
-
"logits/rejected": -0.
|
632 |
-
"logps/chosen": -
|
633 |
-
"logps/rejected": -
|
634 |
-
"loss": 0.
|
635 |
-
"rewards/accuracies": 0.
|
636 |
-
"rewards/chosen": -1.
|
637 |
-
"rewards/margins":
|
638 |
-
"rewards/rejected": -
|
639 |
"step": 450
|
640 |
},
|
641 |
{
|
642 |
"epoch": 0.96,
|
643 |
"learning_rate": 1.1608692138469379e-08,
|
644 |
-
"logits/chosen": -0.
|
645 |
-
"logits/rejected": -0.
|
646 |
-
"logps/chosen": -
|
647 |
-
"logps/rejected": -
|
648 |
-
"loss": 0.
|
649 |
-
"rewards/accuracies": 0.
|
650 |
-
"rewards/chosen": -1.
|
651 |
-
"rewards/margins":
|
652 |
-
"rewards/rejected": -
|
653 |
"step": 460
|
654 |
},
|
655 |
{
|
656 |
"epoch": 0.98,
|
657 |
"learning_rate": 1.970368253390198e-09,
|
658 |
-
"logits/chosen": -0.
|
659 |
-
"logits/rejected": -0.
|
660 |
-
"logps/chosen": -
|
661 |
-
"logps/rejected": -
|
662 |
-
"loss": 0.
|
663 |
-
"rewards/accuracies": 0.
|
664 |
-
"rewards/chosen": -1.
|
665 |
-
"rewards/margins":
|
666 |
-
"rewards/rejected": -
|
667 |
"step": 470
|
668 |
},
|
669 |
{
|
670 |
"epoch": 1.0,
|
671 |
"step": 477,
|
672 |
"total_flos": 0.0,
|
673 |
-
"train_loss": 0.
|
674 |
-
"train_runtime":
|
675 |
-
"train_samples_per_second":
|
676 |
-
"train_steps_per_second": 0.
|
677 |
}
|
678 |
],
|
679 |
"logging_steps": 10,
|
680 |
"max_steps": 477,
|
681 |
"num_train_epochs": 1,
|
682 |
-
"save_steps":
|
683 |
"total_flos": 0.0,
|
684 |
"trial_name": null,
|
685 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 0.9984301412872841,
|
5 |
+
"eval_steps": 100,
|
6 |
"global_step": 477,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
|
|
11 |
{
|
12 |
"epoch": 0.02,
|
13 |
"learning_rate": 6.25e-07,
|
14 |
+
"logits/chosen": 0.2953376770019531,
|
15 |
+
"logits/rejected": 0.3199407160282135,
|
16 |
+
"logps/chosen": -295.2257080078125,
|
17 |
+
"logps/rejected": -290.8670959472656,
|
18 |
+
"loss": 0.1859,
|
19 |
+
"rewards/accuracies": 0.4000000059604645,
|
20 |
+
"rewards/chosen": 0.00015557199367322028,
|
21 |
+
"rewards/margins": 0.0003326966252643615,
|
22 |
+
"rewards/rejected": -0.00017712473345454782,
|
23 |
"step": 10
|
24 |
},
|
25 |
{
|
26 |
"epoch": 0.04,
|
27 |
"learning_rate": 1.25e-06,
|
28 |
+
"logits/chosen": 0.26482802629470825,
|
29 |
+
"logits/rejected": 0.33679550886154175,
|
30 |
+
"logps/chosen": -318.49652099609375,
|
31 |
+
"logps/rejected": -291.0345458984375,
|
32 |
+
"loss": 0.1869,
|
33 |
+
"rewards/accuracies": 0.612500011920929,
|
34 |
+
"rewards/chosen": -0.002573266625404358,
|
35 |
+
"rewards/margins": 0.011565858498215675,
|
36 |
+
"rewards/rejected": -0.014139125123620033,
|
37 |
"step": 20
|
38 |
},
|
39 |
{
|
40 |
"epoch": 0.06,
|
41 |
"learning_rate": 1.875e-06,
|
42 |
+
"logits/chosen": 0.3993307650089264,
|
43 |
+
"logits/rejected": 0.45834311842918396,
|
44 |
+
"logps/chosen": -301.6546936035156,
|
45 |
+
"logps/rejected": -295.427734375,
|
46 |
+
"loss": 0.1703,
|
47 |
"rewards/accuracies": 0.6812499761581421,
|
48 |
+
"rewards/chosen": -0.05365907400846481,
|
49 |
+
"rewards/margins": 0.07345211505889893,
|
50 |
+
"rewards/rejected": -0.12711118161678314,
|
51 |
"step": 30
|
52 |
},
|
53 |
{
|
54 |
"epoch": 0.08,
|
55 |
"learning_rate": 2.5e-06,
|
56 |
+
"logits/chosen": 0.35423707962036133,
|
57 |
+
"logits/rejected": 0.4464724063873291,
|
58 |
+
"logps/chosen": -333.11279296875,
|
59 |
+
"logps/rejected": -334.0306091308594,
|
60 |
+
"loss": 0.1389,
|
61 |
"rewards/accuracies": 0.668749988079071,
|
62 |
+
"rewards/chosen": -0.2748720049858093,
|
63 |
+
"rewards/margins": 0.14294137060642242,
|
64 |
+
"rewards/rejected": -0.41781339049339294,
|
65 |
"step": 40
|
66 |
},
|
67 |
{
|
68 |
"epoch": 0.1,
|
69 |
"learning_rate": 2.999839121261416e-06,
|
70 |
+
"logits/chosen": 0.35561248660087585,
|
71 |
+
"logits/rejected": 0.3796083331108093,
|
72 |
+
"logps/chosen": -374.3013916015625,
|
73 |
+
"logps/rejected": -392.155517578125,
|
74 |
+
"loss": 0.0917,
|
75 |
+
"rewards/accuracies": 0.6499999761581421,
|
76 |
+
"rewards/chosen": -0.6061999201774597,
|
77 |
+
"rewards/margins": 0.2705405652523041,
|
78 |
+
"rewards/rejected": -0.8767404556274414,
|
79 |
"step": 50
|
80 |
},
|
81 |
{
|
82 |
"epoch": 0.13,
|
83 |
"learning_rate": 2.994211988057582e-06,
|
84 |
+
"logits/chosen": 0.17622683942317963,
|
85 |
+
"logits/rejected": 0.313021719455719,
|
86 |
+
"logps/chosen": -346.9971618652344,
|
87 |
+
"logps/rejected": -349.66387939453125,
|
88 |
+
"loss": 0.0903,
|
89 |
+
"rewards/accuracies": 0.59375,
|
90 |
+
"rewards/chosen": -0.5485752820968628,
|
91 |
+
"rewards/margins": 0.28624793887138367,
|
92 |
+
"rewards/rejected": -0.8348232507705688,
|
93 |
"step": 60
|
94 |
},
|
95 |
{
|
96 |
"epoch": 0.15,
|
97 |
"learning_rate": 2.9805753939568693e-06,
|
98 |
+
"logits/chosen": 0.1983145773410797,
|
99 |
+
"logits/rejected": 0.23167672753334045,
|
100 |
+
"logps/chosen": -356.7235412597656,
|
101 |
+
"logps/rejected": -386.5905456542969,
|
102 |
+
"loss": 0.0815,
|
103 |
+
"rewards/accuracies": 0.6312500238418579,
|
104 |
+
"rewards/chosen": -0.5771406888961792,
|
105 |
+
"rewards/margins": 0.28289300203323364,
|
106 |
+
"rewards/rejected": -0.8600338101387024,
|
107 |
"step": 70
|
108 |
},
|
109 |
{
|
110 |
"epoch": 0.17,
|
111 |
"learning_rate": 2.959002435526626e-06,
|
112 |
+
"logits/chosen": 0.35488566756248474,
|
113 |
+
"logits/rejected": 0.4059143662452698,
|
114 |
+
"logps/chosen": -371.07623291015625,
|
115 |
+
"logps/rejected": -389.4164123535156,
|
116 |
+
"loss": 0.0655,
|
117 |
+
"rewards/accuracies": 0.643750011920929,
|
118 |
+
"rewards/chosen": -0.755893349647522,
|
119 |
+
"rewards/margins": 0.5531964302062988,
|
120 |
+
"rewards/rejected": -1.3090897798538208,
|
121 |
"step": 80
|
122 |
},
|
123 |
{
|
124 |
"epoch": 0.19,
|
125 |
"learning_rate": 2.929608750821129e-06,
|
126 |
+
"logits/chosen": 0.2614609897136688,
|
127 |
+
"logits/rejected": 0.4048990309238434,
|
128 |
+
"logps/chosen": -387.1451110839844,
|
129 |
+
"logps/rejected": -373.1450500488281,
|
130 |
+
"loss": 0.0699,
|
131 |
+
"rewards/accuracies": 0.7250000238418579,
|
132 |
+
"rewards/chosen": -0.7152177095413208,
|
133 |
+
"rewards/margins": 0.39531436562538147,
|
134 |
+
"rewards/rejected": -1.1105320453643799,
|
135 |
"step": 90
|
136 |
},
|
137 |
{
|
138 |
"epoch": 0.21,
|
139 |
"learning_rate": 2.892551899524109e-06,
|
140 |
+
"logits/chosen": 0.35849034786224365,
|
141 |
+
"logits/rejected": 0.3978693187236786,
|
142 |
+
"logps/chosen": -395.839599609375,
|
143 |
+
"logps/rejected": -456.36016845703125,
|
144 |
+
"loss": 0.0642,
|
145 |
+
"rewards/accuracies": 0.668749988079071,
|
146 |
+
"rewards/chosen": -1.0831067562103271,
|
147 |
+
"rewards/margins": 0.505523681640625,
|
148 |
+
"rewards/rejected": -1.5886304378509521,
|
149 |
+
"step": 100
|
150 |
+
},
|
151 |
+
{
|
152 |
+
"epoch": 0.21,
|
153 |
+
"eval_logits/chosen": 0.4300641715526581,
|
154 |
+
"eval_logits/rejected": 0.45788079500198364,
|
155 |
+
"eval_logps/chosen": -443.42120361328125,
|
156 |
+
"eval_logps/rejected": -556.7063598632812,
|
157 |
+
"eval_loss": 0.04867542162537575,
|
158 |
+
"eval_rewards/accuracies": 0.71875,
|
159 |
+
"eval_rewards/chosen": -1.3983652591705322,
|
160 |
+
"eval_rewards/margins": 1.0524473190307617,
|
161 |
+
"eval_rewards/rejected": -2.450812578201294,
|
162 |
+
"eval_runtime": 73.3058,
|
163 |
+
"eval_samples_per_second": 27.283,
|
164 |
+
"eval_steps_per_second": 0.437,
|
165 |
"step": 100
|
166 |
},
|
167 |
{
|
168 |
"epoch": 0.23,
|
169 |
"learning_rate": 2.848030518377739e-06,
|
170 |
+
"logits/chosen": 0.2927325367927551,
|
171 |
+
"logits/rejected": 0.3830544054508209,
|
172 |
+
"logps/chosen": -459.0276794433594,
|
173 |
+
"logps/rejected": -510.6993103027344,
|
174 |
+
"loss": 0.0405,
|
175 |
+
"rewards/accuracies": 0.699999988079071,
|
176 |
+
"rewards/chosen": -1.4556727409362793,
|
177 |
+
"rewards/margins": 0.8043819665908813,
|
178 |
+
"rewards/rejected": -2.260054588317871,
|
179 |
"step": 110
|
180 |
},
|
181 |
{
|
182 |
"epoch": 0.25,
|
183 |
"learning_rate": 2.7962832564252724e-06,
|
184 |
+
"logits/chosen": 0.15991483628749847,
|
185 |
+
"logits/rejected": 0.2417908012866974,
|
186 |
+
"logps/chosen": -477.8702087402344,
|
187 |
+
"logps/rejected": -537.2095947265625,
|
188 |
+
"loss": 0.0489,
|
189 |
+
"rewards/accuracies": 0.7124999761581421,
|
190 |
+
"rewards/chosen": -1.243250846862793,
|
191 |
+
"rewards/margins": 0.6618646383285522,
|
192 |
+
"rewards/rejected": -1.9051154851913452,
|
193 |
"step": 120
|
194 |
},
|
195 |
{
|
196 |
"epoch": 0.27,
|
197 |
"learning_rate": 2.7375874957747644e-06,
|
198 |
+
"logits/chosen": 0.06199001148343086,
|
199 |
+
"logits/rejected": 0.19189774990081787,
|
200 |
+
"logps/chosen": -411.61602783203125,
|
201 |
+
"logps/rejected": -418.01019287109375,
|
202 |
+
"loss": 0.0665,
|
203 |
+
"rewards/accuracies": 0.668749988079071,
|
204 |
+
"rewards/chosen": -0.8180139660835266,
|
205 |
+
"rewards/margins": 0.39107373356819153,
|
206 |
+
"rewards/rejected": -1.209087610244751,
|
207 |
"step": 130
|
208 |
},
|
209 |
{
|
210 |
"epoch": 0.29,
|
211 |
"learning_rate": 2.672257864741005e-06,
|
212 |
+
"logits/chosen": 0.16753675043582916,
|
213 |
+
"logits/rejected": 0.2094758003950119,
|
214 |
+
"logps/chosen": -402.02484130859375,
|
215 |
+
"logps/rejected": -429.792724609375,
|
216 |
+
"loss": 0.0608,
|
217 |
+
"rewards/accuracies": 0.7437499761581421,
|
218 |
+
"rewards/chosen": -0.8829413652420044,
|
219 |
+
"rewards/margins": 0.4748268127441406,
|
220 |
+
"rewards/rejected": -1.3577684164047241,
|
221 |
"step": 140
|
222 |
},
|
223 |
{
|
224 |
"epoch": 0.31,
|
225 |
"learning_rate": 2.600644551335706e-06,
|
226 |
+
"logits/chosen": 0.12230317294597626,
|
227 |
+
"logits/rejected": 0.2780611515045166,
|
228 |
+
"logps/chosen": -399.07012939453125,
|
229 |
+
"logps/rejected": -447.2755432128906,
|
230 |
+
"loss": 0.0409,
|
231 |
+
"rewards/accuracies": 0.6875,
|
232 |
+
"rewards/chosen": -1.1799192428588867,
|
233 |
+
"rewards/margins": 0.7245117425918579,
|
234 |
+
"rewards/rejected": -1.9044307470321655,
|
235 |
"step": 150
|
236 |
},
|
237 |
{
|
238 |
"epoch": 0.33,
|
239 |
"learning_rate": 2.5231314261461732e-06,
|
240 |
+
"logits/chosen": 0.15223169326782227,
|
241 |
+
"logits/rejected": 0.2353695183992386,
|
242 |
+
"logps/chosen": -423.8455505371094,
|
243 |
+
"logps/rejected": -493.4649353027344,
|
244 |
+
"loss": 0.0439,
|
245 |
+
"rewards/accuracies": 0.7437499761581421,
|
246 |
+
"rewards/chosen": -1.3035107851028442,
|
247 |
+
"rewards/margins": 0.7260502576828003,
|
248 |
+
"rewards/rejected": -2.0295610427856445,
|
249 |
"step": 160
|
250 |
},
|
251 |
{
|
252 |
"epoch": 0.36,
|
253 |
"learning_rate": 2.440133984664454e-06,
|
254 |
+
"logits/chosen": 0.07280103117227554,
|
255 |
+
"logits/rejected": 0.15081152319908142,
|
256 |
+
"logps/chosen": -472.6494140625,
|
257 |
+
"logps/rejected": -495.68768310546875,
|
258 |
+
"loss": 0.0448,
|
259 |
+
"rewards/accuracies": 0.65625,
|
260 |
+
"rewards/chosen": -1.1380369663238525,
|
261 |
+
"rewards/margins": 0.558665931224823,
|
262 |
+
"rewards/rejected": -1.6967029571533203,
|
263 |
"step": 170
|
264 |
},
|
265 |
{
|
266 |
"epoch": 0.38,
|
267 |
"learning_rate": 2.3520971200967337e-06,
|
268 |
+
"logits/chosen": 0.1431509107351303,
|
269 |
+
"logits/rejected": 0.12718148529529572,
|
270 |
+
"logps/chosen": -421.31591796875,
|
271 |
+
"logps/rejected": -481.0022888183594,
|
272 |
+
"loss": 0.0398,
|
273 |
+
"rewards/accuracies": 0.6875,
|
274 |
+
"rewards/chosen": -1.277320146560669,
|
275 |
+
"rewards/margins": 0.7172085642814636,
|
276 |
+
"rewards/rejected": -1.9945287704467773,
|
277 |
"step": 180
|
278 |
},
|
279 |
{
|
280 |
"epoch": 0.4,
|
281 |
"learning_rate": 2.2594927385914546e-06,
|
282 |
+
"logits/chosen": 0.1423339694738388,
|
283 |
+
"logits/rejected": 0.20066264271736145,
|
284 |
+
"logps/chosen": -385.1085205078125,
|
285 |
+
"logps/rejected": -413.58660888671875,
|
286 |
+
"loss": 0.044,
|
287 |
+
"rewards/accuracies": 0.6875,
|
288 |
+
"rewards/chosen": -1.1722856760025024,
|
289 |
+
"rewards/margins": 0.5436893105506897,
|
290 |
+
"rewards/rejected": -1.715975046157837,
|
291 |
"step": 190
|
292 |
},
|
293 |
{
|
294 |
"epoch": 0.42,
|
295 |
"learning_rate": 2.1628172296692954e-06,
|
296 |
+
"logits/chosen": 0.08901094645261765,
|
297 |
+
"logits/rejected": 0.13271097838878632,
|
298 |
+
"logps/chosen": -385.1289367675781,
|
299 |
+
"logps/rejected": -454.6835021972656,
|
300 |
+
"loss": 0.047,
|
301 |
+
"rewards/accuracies": 0.6875,
|
302 |
+
"rewards/chosen": -1.0956504344940186,
|
303 |
+
"rewards/margins": 0.6720448732376099,
|
304 |
+
"rewards/rejected": -1.767695426940918,
|
305 |
+
"step": 200
|
306 |
+
},
|
307 |
+
{
|
308 |
+
"epoch": 0.42,
|
309 |
+
"eval_logits/chosen": 0.11227821558713913,
|
310 |
+
"eval_logits/rejected": 0.12742303311824799,
|
311 |
+
"eval_logps/chosen": -415.04937744140625,
|
312 |
+
"eval_logps/rejected": -496.232666015625,
|
313 |
+
"eval_loss": 0.0460943840444088,
|
314 |
+
"eval_rewards/accuracies": 0.7421875,
|
315 |
+
"eval_rewards/chosen": -1.1146469116210938,
|
316 |
+
"eval_rewards/margins": 0.7314289212226868,
|
317 |
+
"eval_rewards/rejected": -1.8460756540298462,
|
318 |
+
"eval_runtime": 73.4078,
|
319 |
+
"eval_samples_per_second": 27.245,
|
320 |
+
"eval_steps_per_second": 0.436,
|
321 |
"step": 200
|
322 |
},
|
323 |
{
|
324 |
"epoch": 0.44,
|
325 |
"learning_rate": 2.062588805414343e-06,
|
326 |
+
"logits/chosen": -0.006222477648407221,
|
327 |
+
"logits/rejected": 0.02312941662967205,
|
328 |
+
"logps/chosen": -435.7500915527344,
|
329 |
+
"logps/rejected": -491.56561279296875,
|
330 |
+
"loss": 0.0379,
|
331 |
+
"rewards/accuracies": 0.7437499761581421,
|
332 |
+
"rewards/chosen": -1.1337292194366455,
|
333 |
+
"rewards/margins": 0.7197057008743286,
|
334 |
+
"rewards/rejected": -1.8534349203109741,
|
335 |
"step": 210
|
336 |
},
|
337 |
{
|
338 |
"epoch": 0.46,
|
339 |
"learning_rate": 1.9593447226892386e-06,
|
340 |
+
"logits/chosen": -0.17459270358085632,
|
341 |
+
"logits/rejected": -0.05563253164291382,
|
342 |
+
"logps/chosen": -466.9278259277344,
|
343 |
+
"logps/rejected": -490.50567626953125,
|
344 |
+
"loss": 0.033,
|
345 |
+
"rewards/accuracies": 0.6937500238418579,
|
346 |
+
"rewards/chosen": -1.4599158763885498,
|
347 |
+
"rewards/margins": 0.615450382232666,
|
348 |
+
"rewards/rejected": -2.075366497039795,
|
349 |
"step": 220
|
350 |
},
|
351 |
{
|
352 |
"epoch": 0.48,
|
353 |
"learning_rate": 1.853638403264141e-06,
|
354 |
+
"logits/chosen": -0.037995658814907074,
|
355 |
+
"logits/rejected": -0.011381956748664379,
|
356 |
+
"logps/chosen": -448.68939208984375,
|
357 |
+
"logps/rejected": -532.2055053710938,
|
358 |
+
"loss": 0.0365,
|
359 |
+
"rewards/accuracies": 0.699999988079071,
|
360 |
+
"rewards/chosen": -1.6667238473892212,
|
361 |
+
"rewards/margins": 0.8385307192802429,
|
362 |
+
"rewards/rejected": -2.5052547454833984,
|
363 |
"step": 230
|
364 |
},
|
365 |
{
|
366 |
"epoch": 0.5,
|
367 |
"learning_rate": 1.7460364672965328e-06,
|
368 |
+
"logits/chosen": -0.0721597820520401,
|
369 |
+
"logits/rejected": -0.08736585080623627,
|
370 |
+
"logps/chosen": -470.03045654296875,
|
371 |
+
"logps/rejected": -542.5087280273438,
|
372 |
+
"loss": 0.0484,
|
373 |
+
"rewards/accuracies": 0.7124999761581421,
|
374 |
+
"rewards/chosen": -1.167931318283081,
|
375 |
+
"rewards/margins": 0.7865667939186096,
|
376 |
+
"rewards/rejected": -1.954498052597046,
|
377 |
"step": 240
|
378 |
},
|
379 |
{
|
380 |
"epoch": 0.52,
|
381 |
"learning_rate": 1.637115696063402e-06,
|
382 |
+
"logits/chosen": -0.07914544641971588,
|
383 |
+
"logits/rejected": -0.05659206956624985,
|
384 |
+
"logps/chosen": -430.75726318359375,
|
385 |
+
"logps/rejected": -472.98284912109375,
|
386 |
+
"loss": 0.0475,
|
387 |
+
"rewards/accuracies": 0.6875,
|
388 |
+
"rewards/chosen": -1.1496978998184204,
|
389 |
+
"rewards/margins": 0.6665709018707275,
|
390 |
+
"rewards/rejected": -1.8162685632705688,
|
391 |
"step": 250
|
392 |
},
|
393 |
{
|
394 |
"epoch": 0.54,
|
395 |
"learning_rate": 1.5274599402265162e-06,
|
396 |
+
"logits/chosen": -0.23472614586353302,
|
397 |
+
"logits/rejected": -0.11506778001785278,
|
398 |
+
"logps/chosen": -493.5909118652344,
|
399 |
+
"logps/rejected": -528.9447021484375,
|
400 |
+
"loss": 0.0421,
|
401 |
+
"rewards/accuracies": 0.699999988079071,
|
402 |
+
"rewards/chosen": -1.5179877281188965,
|
403 |
+
"rewards/margins": 0.8330425024032593,
|
404 |
+
"rewards/rejected": -2.3510303497314453,
|
405 |
"step": 260
|
406 |
},
|
407 |
{
|
408 |
"epoch": 0.57,
|
409 |
"learning_rate": 1.4176569902035088e-06,
|
410 |
+
"logits/chosen": -0.17047128081321716,
|
411 |
+
"logits/rejected": -0.09592470526695251,
|
412 |
+
"logps/chosen": -456.50042724609375,
|
413 |
+
"logps/rejected": -494.05059814453125,
|
414 |
+
"loss": 0.0513,
|
415 |
+
"rewards/accuracies": 0.7562500238418579,
|
416 |
+
"rewards/chosen": -1.2792171239852905,
|
417 |
+
"rewards/margins": 0.6969503164291382,
|
418 |
+
"rewards/rejected": -1.9761674404144287,
|
419 |
"step": 270
|
420 |
},
|
421 |
{
|
422 |
"epoch": 0.59,
|
423 |
"learning_rate": 1.308295425420593e-06,
|
424 |
+
"logits/chosen": -0.24125584959983826,
|
425 |
+
"logits/rejected": -0.21810472011566162,
|
426 |
+
"logps/chosen": -445.63916015625,
|
427 |
+
"logps/rejected": -531.8533325195312,
|
428 |
+
"loss": 0.0463,
|
429 |
+
"rewards/accuracies": 0.6937500238418579,
|
430 |
+
"rewards/chosen": -1.3838244676589966,
|
431 |
+
"rewards/margins": 0.7514439225196838,
|
432 |
+
"rewards/rejected": -2.135268211364746,
|
433 |
"step": 280
|
434 |
},
|
435 |
{
|
436 |
"epoch": 0.61,
|
437 |
"learning_rate": 1.1999614593359337e-06,
|
438 |
+
"logits/chosen": -0.2246803492307663,
|
439 |
+
"logits/rejected": -0.19805817306041718,
|
440 |
+
"logps/chosen": -426.7688903808594,
|
441 |
+
"logps/rejected": -472.63134765625,
|
442 |
+
"loss": 0.0407,
|
443 |
+
"rewards/accuracies": 0.6625000238418579,
|
444 |
+
"rewards/chosen": -1.2982797622680664,
|
445 |
+
"rewards/margins": 0.6589063405990601,
|
446 |
+
"rewards/rejected": -1.9571861028671265,
|
447 |
"step": 290
|
448 |
},
|
449 |
{
|
450 |
"epoch": 0.63,
|
451 |
"learning_rate": 1.0932357971453745e-06,
|
452 |
+
"logits/chosen": -0.28935036063194275,
|
453 |
+
"logits/rejected": -0.16529107093811035,
|
454 |
+
"logps/chosen": -460.61138916015625,
|
455 |
+
"logps/rejected": -460.7236328125,
|
456 |
+
"loss": 0.0401,
|
457 |
+
"rewards/accuracies": 0.6625000238418579,
|
458 |
+
"rewards/chosen": -1.4895880222320557,
|
459 |
+
"rewards/margins": 0.5250975489616394,
|
460 |
+
"rewards/rejected": -2.0146853923797607,
|
461 |
+
"step": 300
|
462 |
+
},
|
463 |
+
{
|
464 |
+
"epoch": 0.63,
|
465 |
+
"eval_logits/chosen": -0.1928369402885437,
|
466 |
+
"eval_logits/rejected": -0.17653387784957886,
|
467 |
+
"eval_logps/chosen": -431.743896484375,
|
468 |
+
"eval_logps/rejected": -528.1251831054688,
|
469 |
+
"eval_loss": 0.04077731445431709,
|
470 |
+
"eval_rewards/accuracies": 0.71484375,
|
471 |
+
"eval_rewards/chosen": -1.2815920114517212,
|
472 |
+
"eval_rewards/margins": 0.8834086656570435,
|
473 |
+
"eval_rewards/rejected": -2.1650009155273438,
|
474 |
+
"eval_runtime": 75.4051,
|
475 |
+
"eval_samples_per_second": 26.523,
|
476 |
+
"eval_steps_per_second": 0.424,
|
477 |
"step": 300
|
478 |
},
|
479 |
{
|
480 |
"epoch": 0.65,
|
481 |
"learning_rate": 9.886905230142433e-07,
|
482 |
+
"logits/chosen": -0.19448669254779816,
|
483 |
+
"logits/rejected": -0.2653646469116211,
|
484 |
+
"logps/chosen": -470.20654296875,
|
485 |
+
"logps/rejected": -532.0836181640625,
|
486 |
+
"loss": 0.0436,
|
487 |
+
"rewards/accuracies": 0.6625000238418579,
|
488 |
+
"rewards/chosen": -1.4866724014282227,
|
489 |
+
"rewards/margins": 0.6113816499710083,
|
490 |
+
"rewards/rejected": -2.0980541706085205,
|
491 |
"step": 310
|
492 |
},
|
493 |
{
|
494 |
"epoch": 0.67,
|
495 |
"learning_rate": 8.868860335206678e-07,
|
496 |
+
"logits/chosen": -0.274607390165329,
|
497 |
+
"logits/rejected": -0.18271437287330627,
|
498 |
+
"logps/chosen": -476.87591552734375,
|
499 |
+
"logps/rejected": -476.43304443359375,
|
500 |
+
"loss": 0.0457,
|
501 |
+
"rewards/accuracies": 0.6499999761581421,
|
502 |
+
"rewards/chosen": -1.2201156616210938,
|
503 |
+
"rewards/margins": 0.4418094754219055,
|
504 |
+
"rewards/rejected": -1.6619250774383545,
|
505 |
"step": 320
|
506 |
},
|
507 |
{
|
508 |
"epoch": 0.69,
|
509 |
"learning_rate": 7.883680337481599e-07,
|
510 |
+
"logits/chosen": -0.17901699244976044,
|
511 |
+
"logits/rejected": -0.19661930203437805,
|
512 |
+
"logps/chosen": -414.8687438964844,
|
513 |
+
"logps/rejected": -532.6778564453125,
|
514 |
+
"loss": 0.0437,
|
515 |
+
"rewards/accuracies": 0.75,
|
516 |
+
"rewards/chosen": -1.1104294061660767,
|
517 |
+
"rewards/margins": 0.8232589960098267,
|
518 |
+
"rewards/rejected": -1.9336884021759033,
|
519 |
"step": 330
|
520 |
},
|
521 |
{
|
522 |
"epoch": 0.71,
|
523 |
"learning_rate": 6.936646121293654e-07,
|
524 |
+
"logits/chosen": -0.21059110760688782,
|
525 |
+
"logits/rejected": -0.22660131752490997,
|
526 |
+
"logps/chosen": -434.880859375,
|
527 |
+
"logps/rejected": -541.180908203125,
|
528 |
+
"loss": 0.0366,
|
529 |
+
"rewards/accuracies": 0.637499988079071,
|
530 |
+
"rewards/chosen": -1.4544843435287476,
|
531 |
+
"rewards/margins": 0.8150385618209839,
|
532 |
+
"rewards/rejected": -2.2695229053497314,
|
533 |
"step": 340
|
534 |
},
|
535 |
{
|
536 |
"epoch": 0.73,
|
537 |
"learning_rate": 6.032834097207889e-07,
|
538 |
+
"logits/chosen": -0.17869403958320618,
|
539 |
+
"logits/rejected": -0.1880768984556198,
|
540 |
+
"logps/chosen": -464.3429260253906,
|
541 |
+
"logps/rejected": -565.3632202148438,
|
542 |
+
"loss": 0.0355,
|
543 |
+
"rewards/accuracies": 0.706250011920929,
|
544 |
+
"rewards/chosen": -1.5690315961837769,
|
545 |
+
"rewards/margins": 0.8189191818237305,
|
546 |
+
"rewards/rejected": -2.387950897216797,
|
547 |
"step": 350
|
548 |
},
|
549 |
{
|
550 |
"epoch": 0.75,
|
551 |
"learning_rate": 5.177088990820725e-07,
|
552 |
+
"logits/chosen": -0.2201962023973465,
|
553 |
+
"logits/rejected": -0.17389468848705292,
|
554 |
+
"logps/chosen": -494.4137268066406,
|
555 |
+
"logps/rejected": -554.2338256835938,
|
556 |
+
"loss": 0.0348,
|
557 |
+
"rewards/accuracies": 0.7250000238418579,
|
558 |
+
"rewards/chosen": -1.467558741569519,
|
559 |
+
"rewards/margins": 0.9708374738693237,
|
560 |
+
"rewards/rejected": -2.4383959770202637,
|
561 |
"step": 360
|
562 |
},
|
563 |
{
|
564 |
"epoch": 0.77,
|
565 |
"learning_rate": 4.3739978734594494e-07,
|
566 |
+
"logits/chosen": -0.2049982100725174,
|
567 |
+
"logits/rejected": -0.13553449511528015,
|
568 |
+
"logps/chosen": -510.56610107421875,
|
569 |
+
"logps/rejected": -544.7944946289062,
|
570 |
+
"loss": 0.0343,
|
571 |
+
"rewards/accuracies": 0.71875,
|
572 |
+
"rewards/chosen": -1.5760037899017334,
|
573 |
+
"rewards/margins": 0.8068605661392212,
|
574 |
+
"rewards/rejected": -2.382864236831665,
|
575 |
"step": 370
|
576 |
},
|
577 |
{
|
578 |
"epoch": 0.8,
|
579 |
"learning_rate": 3.627865573992087e-07,
|
580 |
+
"logits/chosen": -0.23506097495555878,
|
581 |
+
"logits/rejected": -0.1945139467716217,
|
582 |
+
"logps/chosen": -473.996337890625,
|
583 |
+
"logps/rejected": -546.2703857421875,
|
584 |
+
"loss": 0.0411,
|
585 |
+
"rewards/accuracies": 0.731249988079071,
|
586 |
+
"rewards/chosen": -1.4194319248199463,
|
587 |
+
"rewards/margins": 0.9198926091194153,
|
588 |
+
"rewards/rejected": -2.339324474334717,
|
589 |
"step": 380
|
590 |
},
|
591 |
{
|
592 |
"epoch": 0.82,
|
593 |
"learning_rate": 2.9426916035484166e-07,
|
594 |
+
"logits/chosen": -0.18075546622276306,
|
595 |
+
"logits/rejected": -0.12812264263629913,
|
596 |
+
"logps/chosen": -451.8818359375,
|
597 |
+
"logps/rejected": -511.1201171875,
|
598 |
+
"loss": 0.0428,
|
599 |
+
"rewards/accuracies": 0.6812499761581421,
|
600 |
+
"rewards/chosen": -1.2848039865493774,
|
601 |
+
"rewards/margins": 0.7686089277267456,
|
602 |
+
"rewards/rejected": -2.053412675857544,
|
603 |
"step": 390
|
604 |
},
|
605 |
{
|
606 |
"epoch": 0.84,
|
607 |
"learning_rate": 2.322148716843081e-07,
|
608 |
+
"logits/chosen": -0.15861022472381592,
|
609 |
+
"logits/rejected": -0.1434582620859146,
|
610 |
+
"logps/chosen": -434.2018127441406,
|
611 |
+
"logps/rejected": -514.0694580078125,
|
612 |
+
"loss": 0.0435,
|
613 |
+
"rewards/accuracies": 0.668749988079071,
|
614 |
+
"rewards/chosen": -1.3711330890655518,
|
615 |
+
"rewards/margins": 0.6456871032714844,
|
616 |
+
"rewards/rejected": -2.016820192337036,
|
617 |
+
"step": 400
|
618 |
+
},
|
619 |
+
{
|
620 |
+
"epoch": 0.84,
|
621 |
+
"eval_logits/chosen": -0.08009739220142365,
|
622 |
+
"eval_logits/rejected": -0.07131167501211166,
|
623 |
+
"eval_logps/chosen": -420.7032470703125,
|
624 |
+
"eval_logps/rejected": -517.1834106445312,
|
625 |
+
"eval_loss": 0.04270826652646065,
|
626 |
+
"eval_rewards/accuracies": 0.7265625,
|
627 |
+
"eval_rewards/chosen": -1.1711856126785278,
|
628 |
+
"eval_rewards/margins": 0.8843976259231567,
|
629 |
+
"eval_rewards/rejected": -2.0555832386016846,
|
630 |
+
"eval_runtime": 75.7653,
|
631 |
+
"eval_samples_per_second": 26.397,
|
632 |
+
"eval_steps_per_second": 0.422,
|
633 |
"step": 400
|
634 |
},
|
635 |
{
|
636 |
"epoch": 0.86,
|
637 |
"learning_rate": 1.7695632250191002e-07,
|
638 |
+
"logits/chosen": -0.16302308440208435,
|
639 |
+
"logits/rejected": -0.1545868217945099,
|
640 |
+
"logps/chosen": -431.54083251953125,
|
641 |
+
"logps/rejected": -492.18975830078125,
|
642 |
+
"loss": 0.0419,
|
643 |
+
"rewards/accuracies": 0.75,
|
644 |
+
"rewards/chosen": -1.1856681108474731,
|
645 |
+
"rewards/margins": 0.7893311381340027,
|
646 |
+
"rewards/rejected": -1.9749990701675415,
|
647 |
"step": 410
|
648 |
},
|
649 |
{
|
650 |
"epoch": 0.88,
|
651 |
"learning_rate": 1.2878971655412515e-07,
|
652 |
+
"logits/chosen": -0.15071699023246765,
|
653 |
+
"logits/rejected": -0.1457306295633316,
|
654 |
+
"logps/chosen": -424.42547607421875,
|
655 |
+
"logps/rejected": -474.6815490722656,
|
656 |
+
"loss": 0.0388,
|
657 |
+
"rewards/accuracies": 0.737500011920929,
|
658 |
+
"rewards/chosen": -1.1596481800079346,
|
659 |
+
"rewards/margins": 0.791610836982727,
|
660 |
+
"rewards/rejected": -1.9512590169906616,
|
661 |
"step": 420
|
662 |
},
|
663 |
{
|
664 |
"epoch": 0.9,
|
665 |
"learning_rate": 8.797324247145411e-08,
|
666 |
+
"logits/chosen": -0.16887474060058594,
|
667 |
+
"logits/rejected": -0.11855659633874893,
|
668 |
+
"logps/chosen": -482.5933532714844,
|
669 |
+
"logps/rejected": -524.8614501953125,
|
670 |
+
"loss": 0.0433,
|
671 |
"rewards/accuracies": 0.7250000238418579,
|
672 |
+
"rewards/chosen": -1.3214952945709229,
|
673 |
+
"rewards/margins": 0.7315466403961182,
|
674 |
+
"rewards/rejected": -2.053041934967041,
|
675 |
"step": 430
|
676 |
},
|
677 |
{
|
678 |
"epoch": 0.92,
|
679 |
"learning_rate": 5.472568979361853e-08,
|
680 |
+
"logits/chosen": -0.1303589642047882,
|
681 |
+
"logits/rejected": -0.1282893717288971,
|
682 |
+
"logps/chosen": -428.30877685546875,
|
683 |
+
"logps/rejected": -485.59185791015625,
|
684 |
+
"loss": 0.0406,
|
685 |
+
"rewards/accuracies": 0.6625000238418579,
|
686 |
+
"rewards/chosen": -1.3397411108016968,
|
687 |
+
"rewards/margins": 0.8013665080070496,
|
688 |
+
"rewards/rejected": -2.1411080360412598,
|
689 |
"step": 440
|
690 |
},
|
691 |
{
|
692 |
"epoch": 0.94,
|
693 |
"learning_rate": 2.922527618666465e-08,
|
694 |
+
"logits/chosen": -0.20207130908966064,
|
695 |
+
"logits/rejected": -0.07939404994249344,
|
696 |
+
"logps/chosen": -461.0880432128906,
|
697 |
+
"logps/rejected": -495.314453125,
|
698 |
+
"loss": 0.0414,
|
699 |
+
"rewards/accuracies": 0.668749988079071,
|
700 |
+
"rewards/chosen": -1.184661626815796,
|
701 |
+
"rewards/margins": 0.7377935647964478,
|
702 |
+
"rewards/rejected": -1.922455072402954,
|
703 |
"step": 450
|
704 |
},
|
705 |
{
|
706 |
"epoch": 0.96,
|
707 |
"learning_rate": 1.1608692138469379e-08,
|
708 |
+
"logits/chosen": -0.18077249825000763,
|
709 |
+
"logits/rejected": -0.11532274633646011,
|
710 |
+
"logps/chosen": -430.862060546875,
|
711 |
+
"logps/rejected": -464.39385986328125,
|
712 |
+
"loss": 0.0406,
|
713 |
+
"rewards/accuracies": 0.6937500238418579,
|
714 |
+
"rewards/chosen": -1.2867480516433716,
|
715 |
+
"rewards/margins": 0.6147353053092957,
|
716 |
+
"rewards/rejected": -1.9014835357666016,
|
717 |
"step": 460
|
718 |
},
|
719 |
{
|
720 |
"epoch": 0.98,
|
721 |
"learning_rate": 1.970368253390198e-09,
|
722 |
+
"logits/chosen": -0.21146085858345032,
|
723 |
+
"logits/rejected": -0.20344269275665283,
|
724 |
+
"logps/chosen": -457.4112243652344,
|
725 |
+
"logps/rejected": -528.634521484375,
|
726 |
+
"loss": 0.039,
|
727 |
+
"rewards/accuracies": 0.706250011920929,
|
728 |
+
"rewards/chosen": -1.1956093311309814,
|
729 |
+
"rewards/margins": 0.7480987906455994,
|
730 |
+
"rewards/rejected": -1.943708062171936,
|
731 |
"step": 470
|
732 |
},
|
733 |
{
|
734 |
"epoch": 1.0,
|
735 |
"step": 477,
|
736 |
"total_flos": 0.0,
|
737 |
+
"train_loss": 0.05794826206171288,
|
738 |
+
"train_runtime": 4397.3222,
|
739 |
+
"train_samples_per_second": 13.903,
|
740 |
+
"train_steps_per_second": 0.108
|
741 |
}
|
742 |
],
|
743 |
"logging_steps": 10,
|
744 |
"max_steps": 477,
|
745 |
"num_train_epochs": 1,
|
746 |
+
"save_steps": 1000,
|
747 |
"total_flos": 0.0,
|
748 |
"trial_name": null,
|
749 |
"trial_params": null
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6648
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e48325f90141d2baab96e119b555cd2507689a536047cb496312be1a6efb1d06
|
3 |
size 6648
|