Muffins987 commited on
Commit
9e27342
1 Parent(s): 0588142

End of training

Browse files
checkpoint-20000/config.json DELETED
@@ -1,28 +0,0 @@
1
- {
2
- "_name_or_path": "roberta-base",
3
- "architectures": [
4
- "RobertaForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
- "classifier_dropout": null,
9
- "eos_token_id": 2,
10
- "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
- "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
- "layer_norm_eps": 1e-05,
16
- "max_position_embeddings": 514,
17
- "model_type": "roberta",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
- "pad_token_id": 1,
21
- "position_embedding_type": "absolute",
22
- "problem_type": "single_label_classification",
23
- "torch_dtype": "float32",
24
- "transformers_version": "4.27.3",
25
- "type_vocab_size": 1,
26
- "use_cache": true,
27
- "vocab_size": 50265
28
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-20000/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa797b2689e4361509ec28ab77b70cfa22dd8c228561850d90b81af595cf76d8
3
- size 997295237
 
 
 
 
checkpoint-20000/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b18778a6e6c620c392907edd95d92b910c41503c00089f4aba849b3a2fe3fc98
3
- size 498662069
 
 
 
 
checkpoint-20000/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:29b021641ff4ccb2ed49b6abd8f12555cbe72f7d3312ff7e914694e2d2832757
3
- size 14575
 
 
 
 
checkpoint-20000/scaler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d273ef501258b360f249686d742a71ab3633a906701652469585df39d0f9f644
3
- size 557
 
 
 
 
checkpoint-20000/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f558146f2de74214a9fa0d946b00f333aa62aa61aac5a1a4c7d6ae9ae01daea4
3
- size 627
 
 
 
 
checkpoint-20000/special_tokens_map.json DELETED
@@ -1,15 +0,0 @@
1
- {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-20000/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/tokenizer_config.json DELETED
@@ -1,15 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": "<s>",
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "errors": "replace",
7
- "mask_token": "<mask>",
8
- "model_max_length": 512,
9
- "pad_token": "<pad>",
10
- "sep_token": "</s>",
11
- "special_tokens_map_file": null,
12
- "tokenizer_class": "RobertaTokenizer",
13
- "trim_offsets": true,
14
- "unk_token": "<unk>"
15
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-20000/trainer_state.json DELETED
@@ -1,266 +0,0 @@
1
- {
2
- "best_metric": 0.7516359309934564,
3
- "best_model_checkpoint": "robertabase-subjectivity-1-actual/checkpoint-20000",
4
- "epoch": 1.0,
5
- "global_step": 20000,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.03,
12
- "learning_rate": 9.875750000000001e-06,
13
- "loss": 0.6845,
14
- "step": 500
15
- },
16
- {
17
- "epoch": 0.05,
18
- "learning_rate": 9.751250000000001e-06,
19
- "loss": 0.6182,
20
- "step": 1000
21
- },
22
- {
23
- "epoch": 0.07,
24
- "learning_rate": 9.6265e-06,
25
- "loss": 0.6045,
26
- "step": 1500
27
- },
28
- {
29
- "epoch": 0.1,
30
- "learning_rate": 9.501500000000002e-06,
31
- "loss": 0.6057,
32
- "step": 2000
33
- },
34
- {
35
- "epoch": 0.12,
36
- "learning_rate": 9.3765e-06,
37
- "loss": 0.5845,
38
- "step": 2500
39
- },
40
- {
41
- "epoch": 0.15,
42
- "learning_rate": 9.251500000000001e-06,
43
- "loss": 0.5819,
44
- "step": 3000
45
- },
46
- {
47
- "epoch": 0.17,
48
- "learning_rate": 9.1265e-06,
49
- "loss": 0.6158,
50
- "step": 3500
51
- },
52
- {
53
- "epoch": 0.2,
54
- "learning_rate": 9.00175e-06,
55
- "loss": 0.5941,
56
- "step": 4000
57
- },
58
- {
59
- "epoch": 0.23,
60
- "learning_rate": 8.876750000000001e-06,
61
- "loss": 0.5953,
62
- "step": 4500
63
- },
64
- {
65
- "epoch": 0.25,
66
- "learning_rate": 8.751750000000001e-06,
67
- "loss": 0.5965,
68
- "step": 5000
69
- },
70
- {
71
- "epoch": 0.28,
72
- "learning_rate": 8.62675e-06,
73
- "loss": 0.6077,
74
- "step": 5500
75
- },
76
- {
77
- "epoch": 0.3,
78
- "learning_rate": 8.50175e-06,
79
- "loss": 0.5663,
80
- "step": 6000
81
- },
82
- {
83
- "epoch": 0.33,
84
- "learning_rate": 8.37675e-06,
85
- "loss": 0.5649,
86
- "step": 6500
87
- },
88
- {
89
- "epoch": 0.35,
90
- "learning_rate": 8.251750000000001e-06,
91
- "loss": 0.6021,
92
- "step": 7000
93
- },
94
- {
95
- "epoch": 0.38,
96
- "learning_rate": 8.126750000000001e-06,
97
- "loss": 0.5803,
98
- "step": 7500
99
- },
100
- {
101
- "epoch": 0.4,
102
- "learning_rate": 8.00175e-06,
103
- "loss": 0.5851,
104
- "step": 8000
105
- },
106
- {
107
- "epoch": 0.42,
108
- "learning_rate": 7.877000000000001e-06,
109
- "loss": 0.5872,
110
- "step": 8500
111
- },
112
- {
113
- "epoch": 0.45,
114
- "learning_rate": 7.752000000000001e-06,
115
- "loss": 0.5597,
116
- "step": 9000
117
- },
118
- {
119
- "epoch": 0.47,
120
- "learning_rate": 7.6272500000000005e-06,
121
- "loss": 0.5764,
122
- "step": 9500
123
- },
124
- {
125
- "epoch": 0.5,
126
- "learning_rate": 7.502250000000001e-06,
127
- "loss": 0.5588,
128
- "step": 10000
129
- },
130
- {
131
- "epoch": 0.53,
132
- "learning_rate": 7.37725e-06,
133
- "loss": 0.5993,
134
- "step": 10500
135
- },
136
- {
137
- "epoch": 0.55,
138
- "learning_rate": 7.252250000000001e-06,
139
- "loss": 0.5873,
140
- "step": 11000
141
- },
142
- {
143
- "epoch": 0.57,
144
- "learning_rate": 7.127250000000001e-06,
145
- "loss": 0.5928,
146
- "step": 11500
147
- },
148
- {
149
- "epoch": 0.6,
150
- "learning_rate": 7.002500000000001e-06,
151
- "loss": 0.5655,
152
- "step": 12000
153
- },
154
- {
155
- "epoch": 0.62,
156
- "learning_rate": 6.877500000000001e-06,
157
- "loss": 0.5705,
158
- "step": 12500
159
- },
160
- {
161
- "epoch": 0.65,
162
- "learning_rate": 6.752500000000001e-06,
163
- "loss": 0.5528,
164
- "step": 13000
165
- },
166
- {
167
- "epoch": 0.68,
168
- "learning_rate": 6.6275e-06,
169
- "loss": 0.59,
170
- "step": 13500
171
- },
172
- {
173
- "epoch": 0.7,
174
- "learning_rate": 6.502500000000001e-06,
175
- "loss": 0.5587,
176
- "step": 14000
177
- },
178
- {
179
- "epoch": 0.72,
180
- "learning_rate": 6.377500000000001e-06,
181
- "loss": 0.5601,
182
- "step": 14500
183
- },
184
- {
185
- "epoch": 0.75,
186
- "learning_rate": 6.2525e-06,
187
- "loss": 0.5804,
188
- "step": 15000
189
- },
190
- {
191
- "epoch": 0.78,
192
- "learning_rate": 6.127500000000001e-06,
193
- "loss": 0.5639,
194
- "step": 15500
195
- },
196
- {
197
- "epoch": 0.8,
198
- "learning_rate": 6.0025e-06,
199
- "loss": 0.5722,
200
- "step": 16000
201
- },
202
- {
203
- "epoch": 0.82,
204
- "learning_rate": 5.8775e-06,
205
- "loss": 0.5601,
206
- "step": 16500
207
- },
208
- {
209
- "epoch": 0.85,
210
- "learning_rate": 5.752500000000001e-06,
211
- "loss": 0.5758,
212
- "step": 17000
213
- },
214
- {
215
- "epoch": 0.88,
216
- "learning_rate": 5.6275e-06,
217
- "loss": 0.5648,
218
- "step": 17500
219
- },
220
- {
221
- "epoch": 0.9,
222
- "learning_rate": 5.5025e-06,
223
- "loss": 0.5723,
224
- "step": 18000
225
- },
226
- {
227
- "epoch": 0.93,
228
- "learning_rate": 5.37775e-06,
229
- "loss": 0.5714,
230
- "step": 18500
231
- },
232
- {
233
- "epoch": 0.95,
234
- "learning_rate": 5.252750000000001e-06,
235
- "loss": 0.5639,
236
- "step": 19000
237
- },
238
- {
239
- "epoch": 0.97,
240
- "learning_rate": 5.128000000000001e-06,
241
- "loss": 0.5499,
242
- "step": 19500
243
- },
244
- {
245
- "epoch": 1.0,
246
- "learning_rate": 5.003e-06,
247
- "loss": 0.5463,
248
- "step": 20000
249
- },
250
- {
251
- "epoch": 1.0,
252
- "eval_accuracy": 0.7495,
253
- "eval_f1": 0.7516359309934564,
254
- "eval_loss": 0.564261257648468,
255
- "eval_runtime": 133.9593,
256
- "eval_samples_per_second": 149.299,
257
- "eval_steps_per_second": 37.325,
258
- "step": 20000
259
- }
260
- ],
261
- "max_steps": 40000,
262
- "num_train_epochs": 2,
263
- "total_flos": 6445345189838880.0,
264
- "trial_name": null,
265
- "trial_params": null
266
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-20000/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6de3c2ba13ce79e3f44a2647ea8e3f91330a0d13b2a72de1dbbc7e86baf262c
3
- size 3707
 
 
 
 
checkpoint-20000/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
runs/Apr02_03-12-09_201fb07c756b/events.out.tfevents.1680405142.201fb07c756b.23.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80d8ae0633b823dd012b9ae64609a5e1c2ecef197bc3a2a3d6297b3eb1157fa9
3
- size 17434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:548333ffd758aabc8ab3678070bc2afb7bb8dab661e5ae5315254db79c7d18a0
3
+ size 17794