Sanjib Narzary commited on
Commit
1fb09ac
1 Parent(s): fd0ac73

after 18 epocs training

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*
README.md CHANGED
@@ -1,3 +1,70 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ base_model: alayaran/bodo-gpt2-clm-setencepiece
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - alayaran/bodo-monolingual-dataset
8
+ metrics:
9
+ - accuracy
10
+ model-index:
11
+ - name: bodo-gpt2-clm-setencepiece
12
+ results:
13
+ - task:
14
+ name: Causal Language Modeling
15
+ type: text-generation
16
+ dataset:
17
+ name: alayaran/bodo-monolingual-dataset
18
+ type: alayaran/bodo-monolingual-dataset
19
+ metrics:
20
+ - name: Accuracy
21
+ type: accuracy
22
+ value: 0.023990075482247693
23
  ---
24
+
25
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
26
+ should probably proofread and complete it, then remove this comment. -->
27
+
28
+ # bodo-gpt2-clm-setencepiece
29
+
30
+ This model is a fine-tuned version of [alayaran/bodo-gpt2-clm-setencepiece](https://huggingface.co/alayaran/bodo-gpt2-clm-setencepiece) on the alayaran/bodo-monolingual-dataset dataset.
31
+ It achieves the following results on the evaluation set:
32
+ - Loss: 8.4768
33
+ - Accuracy: 0.0240
34
+
35
+ ## Model description
36
+
37
+ More information needed
38
+
39
+ ## Intended uses & limitations
40
+
41
+ More information needed
42
+
43
+ ## Training and evaluation data
44
+
45
+ More information needed
46
+
47
+ ## Training procedure
48
+
49
+ ### Training hyperparameters
50
+
51
+ The following hyperparameters were used during training:
52
+ - learning_rate: 0.005
53
+ - train_batch_size: 12
54
+ - eval_batch_size: 12
55
+ - seed: 42
56
+ - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-08
57
+ - lr_scheduler_type: linear
58
+ - lr_scheduler_warmup_steps: 1000
59
+ - num_epochs: 20.0
60
+
61
+ ### Training results
62
+
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - Transformers 4.34.0.dev0
68
+ - Pytorch 2.0.1+cu117
69
+ - Datasets 2.14.5
70
+ - Tokenizers 0.13.3
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_accuracy": 0.023990075482247693,
4
+ "eval_loss": 8.476836204528809,
5
+ "eval_runtime": 33.4965,
6
+ "eval_samples": 784,
7
+ "eval_samples_per_second": 23.405,
8
+ "eval_steps_per_second": 1.97,
9
+ "perplexity": 4802.232555675052,
10
+ "train_loss": 8.536983246027036,
11
+ "train_runtime": 34345.8231,
12
+ "train_samples": 16507,
13
+ "train_samples_per_second": 9.612,
14
+ "train_steps_per_second": 0.801
15
+ }
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "activation_function": "gelu_new",
3
  "architectures": [
4
  "GPT2LMHeadModel"
@@ -31,7 +32,8 @@
31
  "max_length": 50
32
  }
33
  },
 
34
  "transformers_version": "4.34.0.dev0",
35
  "use_cache": true,
36
- "vocab_size": 50257
37
  }
 
1
  {
2
+ "_name_or_path": "alayaran/bodo-gpt2-clm-setencepiece",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
32
  "max_length": 50
33
  }
34
  },
35
+ "torch_dtype": "float32",
36
  "transformers_version": "4.34.0.dev0",
37
  "use_cache": true,
38
+ "vocab_size": 50265
39
  }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_accuracy": 0.023990075482247693,
4
+ "eval_loss": 8.476836204528809,
5
+ "eval_runtime": 33.4965,
6
+ "eval_samples": 784,
7
+ "eval_samples_per_second": 23.405,
8
+ "eval_steps_per_second": 1.97,
9
+ "perplexity": 4802.232555675052
10
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.34.0.dev0"
6
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dba494e67043b2a48956540c1ae6b9ee73f2e378a7b1c0d47360dd646676aa4c
3
+ size 497831773
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "train_loss": 8.536983246027036,
4
+ "train_runtime": 34345.8231,
5
+ "train_samples": 16507,
6
+ "train_samples_per_second": 9.612,
7
+ "train_steps_per_second": 0.801
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,358 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 20.0,
5
+ "eval_steps": 10.0,
6
+ "global_step": 27520,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.36,
13
+ "learning_rate": 0.0025,
14
+ "loss": 8.8155,
15
+ "step": 500
16
+ },
17
+ {
18
+ "epoch": 0.73,
19
+ "learning_rate": 0.005,
20
+ "loss": 8.0292,
21
+ "step": 1000
22
+ },
23
+ {
24
+ "epoch": 1.09,
25
+ "learning_rate": 0.004905731523378582,
26
+ "loss": 8.3372,
27
+ "step": 1500
28
+ },
29
+ {
30
+ "epoch": 1.45,
31
+ "learning_rate": 0.004811463046757164,
32
+ "loss": 8.756,
33
+ "step": 2000
34
+ },
35
+ {
36
+ "epoch": 1.82,
37
+ "learning_rate": 0.004717194570135747,
38
+ "loss": 8.8022,
39
+ "step": 2500
40
+ },
41
+ {
42
+ "epoch": 2.18,
43
+ "learning_rate": 0.004622926093514329,
44
+ "loss": 8.8056,
45
+ "step": 3000
46
+ },
47
+ {
48
+ "epoch": 2.54,
49
+ "learning_rate": 0.004528657616892911,
50
+ "loss": 8.7766,
51
+ "step": 3500
52
+ },
53
+ {
54
+ "epoch": 2.91,
55
+ "learning_rate": 0.004434389140271493,
56
+ "loss": 8.7784,
57
+ "step": 4000
58
+ },
59
+ {
60
+ "epoch": 3.27,
61
+ "learning_rate": 0.004340120663650075,
62
+ "loss": 8.6866,
63
+ "step": 4500
64
+ },
65
+ {
66
+ "epoch": 3.63,
67
+ "learning_rate": 0.004245852187028658,
68
+ "loss": 8.6724,
69
+ "step": 5000
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "learning_rate": 0.00415158371040724,
74
+ "loss": 8.6884,
75
+ "step": 5500
76
+ },
77
+ {
78
+ "epoch": 4.36,
79
+ "learning_rate": 0.004057315233785822,
80
+ "loss": 8.674,
81
+ "step": 6000
82
+ },
83
+ {
84
+ "epoch": 4.72,
85
+ "learning_rate": 0.003963046757164404,
86
+ "loss": 8.6885,
87
+ "step": 6500
88
+ },
89
+ {
90
+ "epoch": 5.09,
91
+ "learning_rate": 0.0038687782805429866,
92
+ "loss": 8.6323,
93
+ "step": 7000
94
+ },
95
+ {
96
+ "epoch": 5.45,
97
+ "learning_rate": 0.0037745098039215687,
98
+ "loss": 8.6416,
99
+ "step": 7500
100
+ },
101
+ {
102
+ "epoch": 5.81,
103
+ "learning_rate": 0.0036802413273001513,
104
+ "loss": 8.6168,
105
+ "step": 8000
106
+ },
107
+ {
108
+ "epoch": 6.18,
109
+ "learning_rate": 0.003585972850678733,
110
+ "loss": 8.5761,
111
+ "step": 8500
112
+ },
113
+ {
114
+ "epoch": 6.54,
115
+ "learning_rate": 0.003491704374057315,
116
+ "loss": 8.5403,
117
+ "step": 9000
118
+ },
119
+ {
120
+ "epoch": 6.9,
121
+ "learning_rate": 0.0033974358974358976,
122
+ "loss": 8.5917,
123
+ "step": 9500
124
+ },
125
+ {
126
+ "epoch": 7.27,
127
+ "learning_rate": 0.0033031674208144797,
128
+ "loss": 8.5902,
129
+ "step": 10000
130
+ },
131
+ {
132
+ "epoch": 7.63,
133
+ "learning_rate": 0.0032088989441930622,
134
+ "loss": 8.5589,
135
+ "step": 10500
136
+ },
137
+ {
138
+ "epoch": 7.99,
139
+ "learning_rate": 0.0031146304675716443,
140
+ "loss": 8.5885,
141
+ "step": 11000
142
+ },
143
+ {
144
+ "epoch": 8.36,
145
+ "learning_rate": 0.003020361990950226,
146
+ "loss": 8.5635,
147
+ "step": 11500
148
+ },
149
+ {
150
+ "epoch": 8.72,
151
+ "learning_rate": 0.0029260935143288085,
152
+ "loss": 8.581,
153
+ "step": 12000
154
+ },
155
+ {
156
+ "epoch": 9.08,
157
+ "learning_rate": 0.0028318250377073906,
158
+ "loss": 8.5647,
159
+ "step": 12500
160
+ },
161
+ {
162
+ "epoch": 9.45,
163
+ "learning_rate": 0.002737556561085973,
164
+ "loss": 8.5448,
165
+ "step": 13000
166
+ },
167
+ {
168
+ "epoch": 9.81,
169
+ "learning_rate": 0.0026432880844645553,
170
+ "loss": 8.5538,
171
+ "step": 13500
172
+ },
173
+ {
174
+ "epoch": 10.17,
175
+ "learning_rate": 0.002549019607843137,
176
+ "loss": 8.5544,
177
+ "step": 14000
178
+ },
179
+ {
180
+ "epoch": 10.54,
181
+ "learning_rate": 0.0024547511312217195,
182
+ "loss": 8.5452,
183
+ "step": 14500
184
+ },
185
+ {
186
+ "epoch": 10.9,
187
+ "learning_rate": 0.0023604826546003016,
188
+ "loss": 8.564,
189
+ "step": 15000
190
+ },
191
+ {
192
+ "epoch": 11.26,
193
+ "learning_rate": 0.0022662141779788837,
194
+ "loss": 8.5346,
195
+ "step": 15500
196
+ },
197
+ {
198
+ "epoch": 11.63,
199
+ "learning_rate": 0.0021719457013574662,
200
+ "loss": 8.5496,
201
+ "step": 16000
202
+ },
203
+ {
204
+ "epoch": 11.99,
205
+ "learning_rate": 0.0020776772247360483,
206
+ "loss": 8.5342,
207
+ "step": 16500
208
+ },
209
+ {
210
+ "epoch": 12.35,
211
+ "learning_rate": 0.0019834087481146304,
212
+ "loss": 8.4924,
213
+ "step": 17000
214
+ },
215
+ {
216
+ "epoch": 12.72,
217
+ "learning_rate": 0.0018891402714932128,
218
+ "loss": 8.499,
219
+ "step": 17500
220
+ },
221
+ {
222
+ "epoch": 13.08,
223
+ "learning_rate": 0.0017948717948717949,
224
+ "loss": 8.4993,
225
+ "step": 18000
226
+ },
227
+ {
228
+ "epoch": 13.44,
229
+ "learning_rate": 0.0017006033182503772,
230
+ "loss": 8.4997,
231
+ "step": 18500
232
+ },
233
+ {
234
+ "epoch": 13.81,
235
+ "learning_rate": 0.0016063348416289595,
236
+ "loss": 8.4866,
237
+ "step": 19000
238
+ },
239
+ {
240
+ "epoch": 14.17,
241
+ "learning_rate": 0.0015120663650075414,
242
+ "loss": 8.4729,
243
+ "step": 19500
244
+ },
245
+ {
246
+ "epoch": 14.53,
247
+ "learning_rate": 0.0014177978883861237,
248
+ "loss": 8.4802,
249
+ "step": 20000
250
+ },
251
+ {
252
+ "epoch": 14.9,
253
+ "learning_rate": 0.0013235294117647058,
254
+ "loss": 8.4606,
255
+ "step": 20500
256
+ },
257
+ {
258
+ "epoch": 15.26,
259
+ "learning_rate": 0.0012292609351432881,
260
+ "loss": 8.4512,
261
+ "step": 21000
262
+ },
263
+ {
264
+ "epoch": 15.62,
265
+ "learning_rate": 0.0011349924585218702,
266
+ "loss": 8.4592,
267
+ "step": 21500
268
+ },
269
+ {
270
+ "epoch": 15.99,
271
+ "learning_rate": 0.0010407239819004526,
272
+ "loss": 8.4624,
273
+ "step": 22000
274
+ },
275
+ {
276
+ "epoch": 16.35,
277
+ "learning_rate": 0.0009464555052790347,
278
+ "loss": 8.4251,
279
+ "step": 22500
280
+ },
281
+ {
282
+ "epoch": 16.72,
283
+ "learning_rate": 0.000852187028657617,
284
+ "loss": 8.4294,
285
+ "step": 23000
286
+ },
287
+ {
288
+ "epoch": 17.08,
289
+ "learning_rate": 0.0007579185520361991,
290
+ "loss": 8.4217,
291
+ "step": 23500
292
+ },
293
+ {
294
+ "epoch": 17.44,
295
+ "learning_rate": 0.0006636500754147813,
296
+ "loss": 8.403,
297
+ "step": 24000
298
+ },
299
+ {
300
+ "epoch": 17.81,
301
+ "learning_rate": 0.0005693815987933635,
302
+ "loss": 8.4101,
303
+ "step": 24500
304
+ },
305
+ {
306
+ "epoch": 18.17,
307
+ "learning_rate": 0.00047511312217194567,
308
+ "loss": 8.3884,
309
+ "step": 25000
310
+ },
311
+ {
312
+ "epoch": 18.53,
313
+ "learning_rate": 0.0003808446455505279,
314
+ "loss": 8.3881,
315
+ "step": 25500
316
+ },
317
+ {
318
+ "epoch": 18.9,
319
+ "learning_rate": 0.0002865761689291101,
320
+ "loss": 8.3882,
321
+ "step": 26000
322
+ },
323
+ {
324
+ "epoch": 19.26,
325
+ "learning_rate": 0.00019230769230769233,
326
+ "loss": 8.3682,
327
+ "step": 26500
328
+ },
329
+ {
330
+ "epoch": 19.62,
331
+ "learning_rate": 9.80392156862745e-05,
332
+ "loss": 8.3625,
333
+ "step": 27000
334
+ },
335
+ {
336
+ "epoch": 19.99,
337
+ "learning_rate": 3.770739064856712e-06,
338
+ "loss": 8.3572,
339
+ "step": 27500
340
+ },
341
+ {
342
+ "epoch": 20.0,
343
+ "step": 27520,
344
+ "total_flos": 8.626295144448e+16,
345
+ "train_loss": 8.536983246027036,
346
+ "train_runtime": 34345.8231,
347
+ "train_samples_per_second": 9.612,
348
+ "train_steps_per_second": 0.801
349
+ }
350
+ ],
351
+ "logging_steps": 500,
352
+ "max_steps": 27520,
353
+ "num_train_epochs": 20,
354
+ "save_steps": 2500,
355
+ "total_flos": 8.626295144448e+16,
356
+ "trial_name": null,
357
+ "trial_params": null
358
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b892bcf6f39d48aa99bc9764f9b915c2749d7dd174482a23688bebeb808d362
3
+ size 4091