Model save
Browse files- README.md +3 -3
- all_results.json +4 -4
- config.json +1 -1
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +1 -1
- model-00002-of-00003.safetensors +1 -1
- model-00003-of-00003.safetensors +1 -1
- runs/Jun05_23-17-05_n136-082-130/events.out.tfevents.1717600754.n136-082-130.1889466.0 +2 -2
- train_results.json +4 -4
- trainer_state.json +165 -153
- training_args.bin +2 -2
README.md
CHANGED
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
13 |
|
14 |
# zephyr-7b-dpo-full
|
15 |
|
16 |
-
This model was trained from scratch on
|
17 |
|
18 |
## Model description
|
19 |
|
@@ -52,7 +52,7 @@ The following hyperparameters were used during training:
|
|
52 |
|
53 |
### Framework versions
|
54 |
|
55 |
-
- Transformers 4.
|
56 |
- Pytorch 2.1.2+cu118
|
57 |
-
- Datasets 2.
|
58 |
- Tokenizers 0.19.1
|
|
|
13 |
|
14 |
# zephyr-7b-dpo-full
|
15 |
|
16 |
+
This model was trained from scratch on the None dataset.
|
17 |
|
18 |
## Model description
|
19 |
|
|
|
52 |
|
53 |
### Framework versions
|
54 |
|
55 |
+
- Transformers 4.41.1
|
56 |
- Pytorch 2.1.2+cu118
|
57 |
+
- Datasets 2.16.1
|
58 |
- Tokenizers 0.19.1
|
all_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"epoch": 0.9984,
|
3 |
"total_flos": 0.0,
|
4 |
-
"train_loss": 0.
|
5 |
-
"train_runtime":
|
6 |
"train_samples": 39942,
|
7 |
-
"train_samples_per_second":
|
8 |
-
"train_steps_per_second": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
"epoch": 0.9984,
|
3 |
"total_flos": 0.0,
|
4 |
+
"train_loss": 0.4014626894241724,
|
5 |
+
"train_runtime": 5052.0322,
|
6 |
"train_samples": 39942,
|
7 |
+
"train_samples_per_second": 7.906,
|
8 |
+
"train_steps_per_second": 0.031
|
9 |
}
|
config.json
CHANGED
@@ -20,7 +20,7 @@
|
|
20 |
"sliding_window": 4096,
|
21 |
"tie_word_embeddings": false,
|
22 |
"torch_dtype": "bfloat16",
|
23 |
-
"transformers_version": "4.
|
24 |
"use_cache": false,
|
25 |
"vocab_size": 32002
|
26 |
}
|
|
|
20 |
"sliding_window": 4096,
|
21 |
"tie_word_embeddings": false,
|
22 |
"torch_dtype": "bfloat16",
|
23 |
+
"transformers_version": "4.41.1",
|
24 |
"use_cache": false,
|
25 |
"vocab_size": 32002
|
26 |
}
|
generation_config.json
CHANGED
@@ -2,5 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 32000,
|
5 |
-
"transformers_version": "4.
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 32000,
|
5 |
+
"transformers_version": "4.41.1"
|
6 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4943178720
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f512a730a26a61a1c16526c02b35b83222d091cd9296f1da7e71affd7995410
|
3 |
size 4943178720
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4999819336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac0006e40eab93ba96e9c4ac961c6bd73d92585b3f54f427b278b0fb7b9767b2
|
3 |
size 4999819336
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4540532728
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a220e930529efb9b6318941d96516bbbb5fa2c9d14cf80d2860389fe098f2e86
|
3 |
size 4540532728
|
runs/Jun05_23-17-05_n136-082-130/events.out.tfevents.1717600754.n136-082-130.1889466.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38495588af7a3d409c5c24789e209e8c63e45c66ecbc5399988e055f803c60c7
|
3 |
+
size 16361
|
train_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"epoch": 0.9984,
|
3 |
"total_flos": 0.0,
|
4 |
-
"train_loss": 0.
|
5 |
-
"train_runtime":
|
6 |
"train_samples": 39942,
|
7 |
-
"train_samples_per_second":
|
8 |
-
"train_steps_per_second": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
"epoch": 0.9984,
|
3 |
"total_flos": 0.0,
|
4 |
+
"train_loss": 0.4014626894241724,
|
5 |
+
"train_runtime": 5052.0322,
|
6 |
"train_samples": 39942,
|
7 |
+
"train_samples_per_second": 7.906,
|
8 |
+
"train_steps_per_second": 0.031
|
9 |
}
|
trainer_state.json
CHANGED
@@ -10,7 +10,7 @@
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.0064,
|
13 |
-
"grad_norm": 1341.
|
14 |
"learning_rate": 3.125e-09,
|
15 |
"logits/chosen": -3.9499800205230713,
|
16 |
"logits/rejected": -4.237819194793701,
|
@@ -25,237 +25,237 @@
|
|
25 |
},
|
26 |
{
|
27 |
"epoch": 0.064,
|
28 |
-
"grad_norm":
|
29 |
"learning_rate": 3.125e-08,
|
30 |
-
"logits/chosen": -4.
|
31 |
-
"logits/rejected": -4.
|
32 |
-
"logps/chosen": -351.
|
33 |
-
"logps/rejected": -308.
|
34 |
-
"loss": 0.
|
35 |
-
"rewards/accuracies": 0.
|
36 |
-
"rewards/chosen": -0.
|
37 |
-
"rewards/margins": -0.
|
38 |
-
"rewards/rejected": 0.
|
39 |
"step": 10
|
40 |
},
|
41 |
{
|
42 |
"epoch": 0.128,
|
43 |
-
"grad_norm":
|
44 |
"learning_rate": 4.9899357349880975e-08,
|
45 |
-
"logits/chosen": -4.
|
46 |
-
"logits/rejected": -4.
|
47 |
-
"logps/chosen": -334.
|
48 |
-
"logps/rejected": -293.
|
49 |
-
"loss": 0.
|
50 |
-
"rewards/accuracies": 0.
|
51 |
-
"rewards/chosen": 0.
|
52 |
-
"rewards/margins": 0.
|
53 |
-
"rewards/rejected": 0.
|
54 |
"step": 20
|
55 |
},
|
56 |
{
|
57 |
"epoch": 0.192,
|
58 |
-
"grad_norm":
|
59 |
"learning_rate": 4.877641290737884e-08,
|
60 |
-
"logits/chosen": -4.
|
61 |
-
"logits/rejected": -4.
|
62 |
-
"logps/chosen": -327.
|
63 |
-
"logps/rejected": -295.
|
64 |
-
"loss": 0.
|
65 |
-
"rewards/accuracies": 0.
|
66 |
-
"rewards/chosen": 0.
|
67 |
-
"rewards/margins": 0.
|
68 |
-
"rewards/rejected": 0.
|
69 |
"step": 30
|
70 |
},
|
71 |
{
|
72 |
"epoch": 0.256,
|
73 |
-
"grad_norm":
|
74 |
"learning_rate": 4.646121984004665e-08,
|
75 |
-
"logits/chosen": -4.
|
76 |
-
"logits/rejected": -4.
|
77 |
-
"logps/chosen": -330.
|
78 |
-
"logps/rejected": -288.
|
79 |
-
"loss": 0.
|
80 |
"rewards/accuracies": 0.8218749761581421,
|
81 |
-
"rewards/chosen": 1.
|
82 |
-
"rewards/margins": 1.
|
83 |
-
"rewards/rejected": 0.
|
84 |
"step": 40
|
85 |
},
|
86 |
{
|
87 |
"epoch": 0.32,
|
88 |
-
"grad_norm":
|
89 |
"learning_rate": 4.3069871595684784e-08,
|
90 |
-
"logits/chosen": -4.
|
91 |
-
"logits/rejected": -4.
|
92 |
-
"logps/chosen": -329.
|
93 |
-
"logps/rejected": -291.
|
94 |
-
"loss": 0.
|
95 |
-
"rewards/accuracies": 0.
|
96 |
-
"rewards/chosen": 2.
|
97 |
-
"rewards/margins": 1.
|
98 |
-
"rewards/rejected": 1.
|
99 |
"step": 50
|
100 |
},
|
101 |
{
|
102 |
"epoch": 0.384,
|
103 |
-
"grad_norm":
|
104 |
"learning_rate": 3.8772424536302564e-08,
|
105 |
-
"logits/chosen": -4.
|
106 |
-
"logits/rejected": -4.
|
107 |
-
"logps/chosen": -
|
108 |
-
"logps/rejected": -291.
|
109 |
-
"loss": 0.
|
110 |
-
"rewards/accuracies": 0.
|
111 |
-
"rewards/chosen":
|
112 |
-
"rewards/margins": 1.
|
113 |
-
"rewards/rejected": 1.
|
114 |
"step": 60
|
115 |
},
|
116 |
{
|
117 |
"epoch": 0.448,
|
118 |
-
"grad_norm":
|
119 |
"learning_rate": 3.378437060203357e-08,
|
120 |
-
"logits/chosen": -4.
|
121 |
-
"logits/rejected": -4.
|
122 |
-
"logps/chosen": -320.
|
123 |
-
"logps/rejected": -288.
|
124 |
-
"loss": 0.
|
125 |
-
"rewards/accuracies": 0.
|
126 |
-
"rewards/chosen": 3.
|
127 |
-
"rewards/margins":
|
128 |
-
"rewards/rejected": 1.
|
129 |
"step": 70
|
130 |
},
|
131 |
{
|
132 |
"epoch": 0.512,
|
133 |
-
"grad_norm":
|
134 |
"learning_rate": 2.8355831645441387e-08,
|
135 |
-
"logits/chosen": -4.
|
136 |
-
"logits/rejected": -4.
|
137 |
-
"logps/chosen": -
|
138 |
-
"logps/rejected": -307.
|
139 |
-
"loss": 0.
|
140 |
-
"rewards/accuracies": 0.
|
141 |
-
"rewards/chosen": 3.
|
142 |
-
"rewards/margins": 2.
|
143 |
-
"rewards/rejected": 1.
|
144 |
"step": 80
|
145 |
},
|
146 |
{
|
147 |
"epoch": 0.576,
|
148 |
-
"grad_norm":
|
149 |
"learning_rate": 2.2759017277414164e-08,
|
150 |
-
"logits/chosen": -4.
|
151 |
-
"logits/rejected": -4.
|
152 |
-
"logps/chosen": -
|
153 |
-
"logps/rejected": -295.
|
154 |
-
"loss": 0.
|
155 |
-
"rewards/accuracies": 0.
|
156 |
-
"rewards/chosen":
|
157 |
-
"rewards/margins":
|
158 |
-
"rewards/rejected":
|
159 |
"step": 90
|
160 |
},
|
161 |
{
|
162 |
"epoch": 0.64,
|
163 |
-
"grad_norm":
|
164 |
"learning_rate": 1.7274575140626317e-08,
|
165 |
-
"logits/chosen": -4.
|
166 |
-
"logits/rejected": -4.
|
167 |
-
"logps/chosen": -330.
|
168 |
-
"logps/rejected": -285.
|
169 |
-
"loss": 0.
|
170 |
-
"rewards/accuracies": 0.
|
171 |
-
"rewards/chosen": 3.
|
172 |
-
"rewards/margins": 2.
|
173 |
-
"rewards/rejected":
|
174 |
"step": 100
|
175 |
},
|
176 |
{
|
177 |
"epoch": 0.704,
|
178 |
-
"grad_norm":
|
179 |
"learning_rate": 1.217751806485235e-08,
|
180 |
-
"logits/chosen": -4.
|
181 |
-
"logits/rejected": -4.
|
182 |
-
"logps/chosen": -
|
183 |
-
"logps/rejected": -276.
|
184 |
-
"loss": 0.
|
185 |
-
"rewards/accuracies": 0.
|
186 |
-
"rewards/chosen": 3.
|
187 |
-
"rewards/margins": 2.
|
188 |
-
"rewards/rejected":
|
189 |
"step": 110
|
190 |
},
|
191 |
{
|
192 |
"epoch": 0.768,
|
193 |
-
"grad_norm":
|
194 |
"learning_rate": 7.723433775328384e-09,
|
195 |
-
"logits/chosen": -4.
|
196 |
-
"logits/rejected": -4.
|
197 |
-
"logps/chosen": -
|
198 |
-
"logps/rejected": -280.
|
199 |
-
"loss": 0.
|
200 |
-
"rewards/accuracies": 0.
|
201 |
-
"rewards/chosen": 3.
|
202 |
-
"rewards/margins": 2.
|
203 |
-
"rewards/rejected": 1.
|
204 |
"step": 120
|
205 |
},
|
206 |
{
|
207 |
"epoch": 0.832,
|
208 |
-
"grad_norm":
|
209 |
"learning_rate": 4.135668656967433e-09,
|
210 |
-
"logits/chosen": -4.
|
211 |
-
"logits/rejected": -4.
|
212 |
-
"logps/chosen": -331.
|
213 |
-
"logps/rejected": -286.
|
214 |
-
"loss": 0.
|
215 |
"rewards/accuracies": 0.875,
|
216 |
-
"rewards/chosen": 3.
|
217 |
-
"rewards/margins": 2.
|
218 |
-
"rewards/rejected": 1.
|
219 |
"step": 130
|
220 |
},
|
221 |
{
|
222 |
"epoch": 0.896,
|
223 |
-
"grad_norm":
|
224 |
"learning_rate": 1.5941282340065698e-09,
|
225 |
-
"logits/chosen": -4.
|
226 |
-
"logits/rejected": -4.
|
227 |
-
"logps/chosen": -
|
228 |
-
"logps/rejected": -303.
|
229 |
-
"loss": 0.
|
230 |
-
"rewards/accuracies": 0.
|
231 |
-
"rewards/chosen": 3.
|
232 |
-
"rewards/margins": 2.
|
233 |
-
"rewards/rejected":
|
234 |
"step": 140
|
235 |
},
|
236 |
{
|
237 |
"epoch": 0.96,
|
238 |
-
"grad_norm":
|
239 |
"learning_rate": 2.262559558016325e-10,
|
240 |
-
"logits/chosen": -4.
|
241 |
-
"logits/rejected": -4.
|
242 |
-
"logps/chosen": -339.
|
243 |
-
"logps/rejected": -295.
|
244 |
-
"loss": 0.
|
245 |
-
"rewards/accuracies": 0.
|
246 |
-
"rewards/chosen": 3.
|
247 |
-
"rewards/margins": 2.
|
248 |
-
"rewards/rejected": 1.
|
249 |
"step": 150
|
250 |
},
|
251 |
{
|
252 |
"epoch": 0.9984,
|
253 |
"step": 156,
|
254 |
"total_flos": 0.0,
|
255 |
-
"train_loss": 0.
|
256 |
-
"train_runtime":
|
257 |
-
"train_samples_per_second":
|
258 |
-
"train_steps_per_second": 0.
|
259 |
}
|
260 |
],
|
261 |
"logging_steps": 10,
|
@@ -263,6 +263,18 @@
|
|
263 |
"num_input_tokens_seen": 0,
|
264 |
"num_train_epochs": 1,
|
265 |
"save_steps": 100,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
266 |
"total_flos": 0.0,
|
267 |
"train_batch_size": 8,
|
268 |
"trial_name": null,
|
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.0064,
|
13 |
+
"grad_norm": 1341.948853106591,
|
14 |
"learning_rate": 3.125e-09,
|
15 |
"logits/chosen": -3.9499800205230713,
|
16 |
"logits/rejected": -4.237819194793701,
|
|
|
25 |
},
|
26 |
{
|
27 |
"epoch": 0.064,
|
28 |
+
"grad_norm": 1338.549586928997,
|
29 |
"learning_rate": 3.125e-08,
|
30 |
+
"logits/chosen": -4.129465579986572,
|
31 |
+
"logits/rejected": -4.351754188537598,
|
32 |
+
"logps/chosen": -351.4585266113281,
|
33 |
+
"logps/rejected": -308.89202880859375,
|
34 |
+
"loss": 0.7278,
|
35 |
+
"rewards/accuracies": 0.375,
|
36 |
+
"rewards/chosen": -0.016052477061748505,
|
37 |
+
"rewards/margins": -0.04994054138660431,
|
38 |
+
"rewards/rejected": 0.033888060599565506,
|
39 |
"step": 10
|
40 |
},
|
41 |
{
|
42 |
"epoch": 0.128,
|
43 |
+
"grad_norm": 1285.3502810697903,
|
44 |
"learning_rate": 4.9899357349880975e-08,
|
45 |
+
"logits/chosen": -4.195886611938477,
|
46 |
+
"logits/rejected": -4.383217811584473,
|
47 |
+
"logps/chosen": -334.9483337402344,
|
48 |
+
"logps/rejected": -293.82867431640625,
|
49 |
+
"loss": 0.6803,
|
50 |
+
"rewards/accuracies": 0.550000011920929,
|
51 |
+
"rewards/chosen": 0.20189008116722107,
|
52 |
+
"rewards/margins": 0.08841639757156372,
|
53 |
+
"rewards/rejected": 0.11347369104623795,
|
54 |
"step": 20
|
55 |
},
|
56 |
{
|
57 |
"epoch": 0.192,
|
58 |
+
"grad_norm": 938.2936106934807,
|
59 |
"learning_rate": 4.877641290737884e-08,
|
60 |
+
"logits/chosen": -4.2334418296813965,
|
61 |
+
"logits/rejected": -4.3670477867126465,
|
62 |
+
"logps/chosen": -327.77716064453125,
|
63 |
+
"logps/rejected": -295.27203369140625,
|
64 |
+
"loss": 0.5536,
|
65 |
+
"rewards/accuracies": 0.699999988079071,
|
66 |
+
"rewards/chosen": 0.9395688772201538,
|
67 |
+
"rewards/margins": 0.4487342834472656,
|
68 |
+
"rewards/rejected": 0.49083465337753296,
|
69 |
"step": 30
|
70 |
},
|
71 |
{
|
72 |
"epoch": 0.256,
|
73 |
+
"grad_norm": 899.5533271284629,
|
74 |
"learning_rate": 4.646121984004665e-08,
|
75 |
+
"logits/chosen": -4.14696741104126,
|
76 |
+
"logits/rejected": -4.348996162414551,
|
77 |
+
"logps/chosen": -330.3467712402344,
|
78 |
+
"logps/rejected": -288.37066650390625,
|
79 |
+
"loss": 0.4332,
|
80 |
"rewards/accuracies": 0.8218749761581421,
|
81 |
+
"rewards/chosen": 1.8148998022079468,
|
82 |
+
"rewards/margins": 1.0534685850143433,
|
83 |
+
"rewards/rejected": 0.7614310383796692,
|
84 |
"step": 40
|
85 |
},
|
86 |
{
|
87 |
"epoch": 0.32,
|
88 |
+
"grad_norm": 737.601892593307,
|
89 |
"learning_rate": 4.3069871595684784e-08,
|
90 |
+
"logits/chosen": -4.238839149475098,
|
91 |
+
"logits/rejected": -4.417731285095215,
|
92 |
+
"logps/chosen": -329.99560546875,
|
93 |
+
"logps/rejected": -291.286865234375,
|
94 |
+
"loss": 0.3783,
|
95 |
+
"rewards/accuracies": 0.8531249761581421,
|
96 |
+
"rewards/chosen": 2.4285898208618164,
|
97 |
+
"rewards/margins": 1.3909423351287842,
|
98 |
+
"rewards/rejected": 1.0376476049423218,
|
99 |
"step": 50
|
100 |
},
|
101 |
{
|
102 |
"epoch": 0.384,
|
103 |
+
"grad_norm": 725.6613657693058,
|
104 |
"learning_rate": 3.8772424536302564e-08,
|
105 |
+
"logits/chosen": -4.256905555725098,
|
106 |
+
"logits/rejected": -4.42824125289917,
|
107 |
+
"logps/chosen": -321.1641845703125,
|
108 |
+
"logps/rejected": -291.2255859375,
|
109 |
+
"loss": 0.3554,
|
110 |
+
"rewards/accuracies": 0.846875011920929,
|
111 |
+
"rewards/chosen": 2.799910306930542,
|
112 |
+
"rewards/margins": 1.64876389503479,
|
113 |
+
"rewards/rejected": 1.151146411895752,
|
114 |
"step": 60
|
115 |
},
|
116 |
{
|
117 |
"epoch": 0.448,
|
118 |
+
"grad_norm": 650.256008502672,
|
119 |
"learning_rate": 3.378437060203357e-08,
|
120 |
+
"logits/chosen": -4.1877875328063965,
|
121 |
+
"logits/rejected": -4.377414703369141,
|
122 |
+
"logps/chosen": -320.77117919921875,
|
123 |
+
"logps/rejected": -288.512451171875,
|
124 |
+
"loss": 0.3377,
|
125 |
+
"rewards/accuracies": 0.7906249761581421,
|
126 |
+
"rewards/chosen": 3.0348763465881348,
|
127 |
+
"rewards/margins": 1.8613855838775635,
|
128 |
+
"rewards/rejected": 1.1734905242919922,
|
129 |
"step": 70
|
130 |
},
|
131 |
{
|
132 |
"epoch": 0.512,
|
133 |
+
"grad_norm": 643.4652639779088,
|
134 |
"learning_rate": 2.8355831645441387e-08,
|
135 |
+
"logits/chosen": -4.054490089416504,
|
136 |
+
"logits/rejected": -4.3433709144592285,
|
137 |
+
"logps/chosen": -346.53253173828125,
|
138 |
+
"logps/rejected": -307.34930419921875,
|
139 |
+
"loss": 0.3369,
|
140 |
+
"rewards/accuracies": 0.8656250238418579,
|
141 |
+
"rewards/chosen": 3.375593662261963,
|
142 |
+
"rewards/margins": 2.1429190635681152,
|
143 |
+
"rewards/rejected": 1.2326747179031372,
|
144 |
"step": 80
|
145 |
},
|
146 |
{
|
147 |
"epoch": 0.576,
|
148 |
+
"grad_norm": 658.4601656800384,
|
149 |
"learning_rate": 2.2759017277414164e-08,
|
150 |
+
"logits/chosen": -4.182233810424805,
|
151 |
+
"logits/rejected": -4.391860485076904,
|
152 |
+
"logps/chosen": -333.4163513183594,
|
153 |
+
"logps/rejected": -295.03515625,
|
154 |
+
"loss": 0.3301,
|
155 |
+
"rewards/accuracies": 0.859375,
|
156 |
+
"rewards/chosen": 2.9584574699401855,
|
157 |
+
"rewards/margins": 1.9474881887435913,
|
158 |
+
"rewards/rejected": 1.0109691619873047,
|
159 |
"step": 90
|
160 |
},
|
161 |
{
|
162 |
"epoch": 0.64,
|
163 |
+
"grad_norm": 716.6222925151064,
|
164 |
"learning_rate": 1.7274575140626317e-08,
|
165 |
+
"logits/chosen": -4.168593406677246,
|
166 |
+
"logits/rejected": -4.387479305267334,
|
167 |
+
"logps/chosen": -330.75775146484375,
|
168 |
+
"logps/rejected": -285.5531311035156,
|
169 |
+
"loss": 0.3315,
|
170 |
+
"rewards/accuracies": 0.862500011920929,
|
171 |
+
"rewards/chosen": 3.267747163772583,
|
172 |
+
"rewards/margins": 2.245748281478882,
|
173 |
+
"rewards/rejected": 1.0219987630844116,
|
174 |
"step": 100
|
175 |
},
|
176 |
{
|
177 |
"epoch": 0.704,
|
178 |
+
"grad_norm": 660.4496184068971,
|
179 |
"learning_rate": 1.217751806485235e-08,
|
180 |
+
"logits/chosen": -4.1479597091674805,
|
181 |
+
"logits/rejected": -4.388964653015137,
|
182 |
+
"logps/chosen": -312.3727722167969,
|
183 |
+
"logps/rejected": -276.13372802734375,
|
184 |
+
"loss": 0.3225,
|
185 |
+
"rewards/accuracies": 0.8656250238418579,
|
186 |
+
"rewards/chosen": 3.277716875076294,
|
187 |
+
"rewards/margins": 2.2097737789154053,
|
188 |
+
"rewards/rejected": 1.0679429769515991,
|
189 |
"step": 110
|
190 |
},
|
191 |
{
|
192 |
"epoch": 0.768,
|
193 |
+
"grad_norm": 637.7261383555538,
|
194 |
"learning_rate": 7.723433775328384e-09,
|
195 |
+
"logits/chosen": -4.145724296569824,
|
196 |
+
"logits/rejected": -4.354761123657227,
|
197 |
+
"logps/chosen": -326.25091552734375,
|
198 |
+
"logps/rejected": -280.4151916503906,
|
199 |
+
"loss": 0.3134,
|
200 |
+
"rewards/accuracies": 0.8531249761581421,
|
201 |
+
"rewards/chosen": 3.3363654613494873,
|
202 |
+
"rewards/margins": 2.202853202819824,
|
203 |
+
"rewards/rejected": 1.1335121393203735,
|
204 |
"step": 120
|
205 |
},
|
206 |
{
|
207 |
"epoch": 0.832,
|
208 |
+
"grad_norm": 768.5710075733069,
|
209 |
"learning_rate": 4.135668656967433e-09,
|
210 |
+
"logits/chosen": -4.2324018478393555,
|
211 |
+
"logits/rejected": -4.393607139587402,
|
212 |
+
"logps/chosen": -331.71051025390625,
|
213 |
+
"logps/rejected": -286.67938232421875,
|
214 |
+
"loss": 0.3203,
|
215 |
"rewards/accuracies": 0.875,
|
216 |
+
"rewards/chosen": 3.384364604949951,
|
217 |
+
"rewards/margins": 2.2786831855773926,
|
218 |
+
"rewards/rejected": 1.1056816577911377,
|
219 |
"step": 130
|
220 |
},
|
221 |
{
|
222 |
"epoch": 0.896,
|
223 |
+
"grad_norm": 742.2155895768275,
|
224 |
"learning_rate": 1.5941282340065698e-09,
|
225 |
+
"logits/chosen": -4.18549919128418,
|
226 |
+
"logits/rejected": -4.4001359939575195,
|
227 |
+
"logps/chosen": -333.18035888671875,
|
228 |
+
"logps/rejected": -303.45220947265625,
|
229 |
+
"loss": 0.3282,
|
230 |
+
"rewards/accuracies": 0.84375,
|
231 |
+
"rewards/chosen": 3.2540221214294434,
|
232 |
+
"rewards/margins": 2.2057249546051025,
|
233 |
+
"rewards/rejected": 1.048297643661499,
|
234 |
"step": 140
|
235 |
},
|
236 |
{
|
237 |
"epoch": 0.96,
|
238 |
+
"grad_norm": 668.7483012920786,
|
239 |
"learning_rate": 2.262559558016325e-10,
|
240 |
+
"logits/chosen": -4.1221604347229,
|
241 |
+
"logits/rejected": -4.351316452026367,
|
242 |
+
"logps/chosen": -339.6897888183594,
|
243 |
+
"logps/rejected": -295.03692626953125,
|
244 |
+
"loss": 0.3208,
|
245 |
+
"rewards/accuracies": 0.859375,
|
246 |
+
"rewards/chosen": 3.4081547260284424,
|
247 |
+
"rewards/margins": 2.2413132190704346,
|
248 |
+
"rewards/rejected": 1.1668416261672974,
|
249 |
"step": 150
|
250 |
},
|
251 |
{
|
252 |
"epoch": 0.9984,
|
253 |
"step": 156,
|
254 |
"total_flos": 0.0,
|
255 |
+
"train_loss": 0.4014626894241724,
|
256 |
+
"train_runtime": 5052.0322,
|
257 |
+
"train_samples_per_second": 7.906,
|
258 |
+
"train_steps_per_second": 0.031
|
259 |
}
|
260 |
],
|
261 |
"logging_steps": 10,
|
|
|
263 |
"num_input_tokens_seen": 0,
|
264 |
"num_train_epochs": 1,
|
265 |
"save_steps": 100,
|
266 |
+
"stateful_callbacks": {
|
267 |
+
"TrainerControl": {
|
268 |
+
"args": {
|
269 |
+
"should_epoch_stop": false,
|
270 |
+
"should_evaluate": false,
|
271 |
+
"should_log": false,
|
272 |
+
"should_save": true,
|
273 |
+
"should_training_stop": false
|
274 |
+
},
|
275 |
+
"attributes": {}
|
276 |
+
}
|
277 |
+
},
|
278 |
"total_flos": 0.0,
|
279 |
"train_batch_size": 8,
|
280 |
"trial_name": null,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d92a249952e57df29e57ec30d0088226bf91a9212b8928552a333e4b9069864
|
3 |
+
size 6520
|