manato003 commited on
Commit
ed4a421
·
verified ·
1 Parent(s): cc1bb6f

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ library_name: transformers
4
+ tags:
5
+ - autotrain
6
+ - text-classification
7
+ base_model: studio-ousia/luke-japanese-base-lite
8
+ widget:
9
+ - text: "I love AutoTrain"
10
+ ---
11
+
12
+ # Model Trained Using AutoTrain
13
+
14
+ - Problem type: Text Classification
15
+
16
+ ## Validation Metrics
17
+ loss: 0.5158100724220276
18
+
19
+ f1: 0.7777777777777778
20
+
21
+ precision: 0.6712328767123288
22
+
23
+ recall: 0.9245283018867925
24
+
25
+ auc: 0.8304930006086427
26
+
27
+ accuracy: 0.7565217391304347
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<ent2>": 32771,
3
+ "<ent>": 32770
4
+ }
checkpoint-230/config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "studio-ousia/luke-japanese-base-lite",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "LukeForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "bert_model_name": "models/luke-japanese/hf_xlm_roberta",
9
+ "bos_token_id": 0,
10
+ "classifier_dropout": null,
11
+ "cls_entity_prediction": false,
12
+ "entity_emb_size": 256,
13
+ "entity_vocab_size": 4,
14
+ "eos_token_id": 2,
15
+ "hidden_act": "gelu",
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 768,
18
+ "id2label": {
19
+ "0": "\u60aa\u3044",
20
+ "1": "\u826f\u3044"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "label2id": {
25
+ "\u60aa\u3044": 0,
26
+ "\u826f\u3044": 1
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "luke",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "single_label_classification",
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.48.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": true,
40
+ "use_entity_aware_attention": true,
41
+ "vocab_size": 32772
42
+ }
checkpoint-230/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e875fdd6ff9f1f63a626c3fd95c5d7570e7de8e48c447c151e30ef08bdfcec0
3
+ size 532307280
checkpoint-230/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f75ddb123e777de62e46d6c1ae5261e997113f9585033a45c4cf54be482723ae
3
+ size 889863098
checkpoint-230/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a00ae46a38a73f402956850b524194acd54becd011fcc7d9c07b576676cd3a60
3
+ size 13990
checkpoint-230/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:982f6a76f8ce7001d605ce8f2bf04b6ba28511c7e2e2b30a14b1670adad81285
3
+ size 1064
checkpoint-230/trainer_state.json ADDED
@@ -0,0 +1,390 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.5158100724220276,
3
+ "best_model_checkpoint": "2ClassLarge/checkpoint-230",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 230,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.043478260869565216,
13
+ "grad_norm": 2.4869840145111084,
14
+ "learning_rate": 4.310344827586207e-06,
15
+ "loss": 0.685,
16
+ "step": 5
17
+ },
18
+ {
19
+ "epoch": 0.08695652173913043,
20
+ "grad_norm": 4.935397624969482,
21
+ "learning_rate": 8.620689655172414e-06,
22
+ "loss": 0.6875,
23
+ "step": 10
24
+ },
25
+ {
26
+ "epoch": 0.13043478260869565,
27
+ "grad_norm": 4.38919734954834,
28
+ "learning_rate": 1.2931034482758622e-05,
29
+ "loss": 0.6703,
30
+ "step": 15
31
+ },
32
+ {
33
+ "epoch": 0.17391304347826086,
34
+ "grad_norm": 2.8880741596221924,
35
+ "learning_rate": 1.7241379310344828e-05,
36
+ "loss": 0.7051,
37
+ "step": 20
38
+ },
39
+ {
40
+ "epoch": 0.21739130434782608,
41
+ "grad_norm": 2.1519901752471924,
42
+ "learning_rate": 2.1551724137931033e-05,
43
+ "loss": 0.696,
44
+ "step": 25
45
+ },
46
+ {
47
+ "epoch": 0.2608695652173913,
48
+ "grad_norm": 2.6489181518554688,
49
+ "learning_rate": 2.5862068965517244e-05,
50
+ "loss": 0.6522,
51
+ "step": 30
52
+ },
53
+ {
54
+ "epoch": 0.30434782608695654,
55
+ "grad_norm": 8.696531295776367,
56
+ "learning_rate": 3.017241379310345e-05,
57
+ "loss": 0.7208,
58
+ "step": 35
59
+ },
60
+ {
61
+ "epoch": 0.34782608695652173,
62
+ "grad_norm": 3.1740002632141113,
63
+ "learning_rate": 3.4482758620689657e-05,
64
+ "loss": 0.6975,
65
+ "step": 40
66
+ },
67
+ {
68
+ "epoch": 0.391304347826087,
69
+ "grad_norm": 4.449029445648193,
70
+ "learning_rate": 3.8793103448275865e-05,
71
+ "loss": 0.7015,
72
+ "step": 45
73
+ },
74
+ {
75
+ "epoch": 0.43478260869565216,
76
+ "grad_norm": 3.1051077842712402,
77
+ "learning_rate": 4.3103448275862066e-05,
78
+ "loss": 0.8118,
79
+ "step": 50
80
+ },
81
+ {
82
+ "epoch": 0.4782608695652174,
83
+ "grad_norm": 2.311675786972046,
84
+ "learning_rate": 4.741379310344828e-05,
85
+ "loss": 0.6846,
86
+ "step": 55
87
+ },
88
+ {
89
+ "epoch": 0.5217391304347826,
90
+ "grad_norm": 4.207972526550293,
91
+ "learning_rate": 4.9806576402321086e-05,
92
+ "loss": 0.6656,
93
+ "step": 60
94
+ },
95
+ {
96
+ "epoch": 0.5652173913043478,
97
+ "grad_norm": 1.8182612657546997,
98
+ "learning_rate": 4.932301740812379e-05,
99
+ "loss": 0.6918,
100
+ "step": 65
101
+ },
102
+ {
103
+ "epoch": 0.6086956521739131,
104
+ "grad_norm": 6.007029056549072,
105
+ "learning_rate": 4.8839458413926506e-05,
106
+ "loss": 0.6666,
107
+ "step": 70
108
+ },
109
+ {
110
+ "epoch": 0.6521739130434783,
111
+ "grad_norm": 3.8037335872650146,
112
+ "learning_rate": 4.835589941972921e-05,
113
+ "loss": 0.6205,
114
+ "step": 75
115
+ },
116
+ {
117
+ "epoch": 0.6956521739130435,
118
+ "grad_norm": 4.0360822677612305,
119
+ "learning_rate": 4.787234042553192e-05,
120
+ "loss": 0.6171,
121
+ "step": 80
122
+ },
123
+ {
124
+ "epoch": 0.7391304347826086,
125
+ "grad_norm": 4.280089855194092,
126
+ "learning_rate": 4.738878143133462e-05,
127
+ "loss": 0.7091,
128
+ "step": 85
129
+ },
130
+ {
131
+ "epoch": 0.782608695652174,
132
+ "grad_norm": 3.4874963760375977,
133
+ "learning_rate": 4.690522243713733e-05,
134
+ "loss": 0.7116,
135
+ "step": 90
136
+ },
137
+ {
138
+ "epoch": 0.8260869565217391,
139
+ "grad_norm": 2.0585179328918457,
140
+ "learning_rate": 4.642166344294004e-05,
141
+ "loss": 0.7636,
142
+ "step": 95
143
+ },
144
+ {
145
+ "epoch": 0.8695652173913043,
146
+ "grad_norm": 2.7472143173217773,
147
+ "learning_rate": 4.5938104448742745e-05,
148
+ "loss": 0.516,
149
+ "step": 100
150
+ },
151
+ {
152
+ "epoch": 0.9130434782608695,
153
+ "grad_norm": 5.185933589935303,
154
+ "learning_rate": 4.545454545454546e-05,
155
+ "loss": 0.8768,
156
+ "step": 105
157
+ },
158
+ {
159
+ "epoch": 0.9565217391304348,
160
+ "grad_norm": 6.4161481857299805,
161
+ "learning_rate": 4.4970986460348165e-05,
162
+ "loss": 0.6838,
163
+ "step": 110
164
+ },
165
+ {
166
+ "epoch": 1.0,
167
+ "grad_norm": 6.210981369018555,
168
+ "learning_rate": 4.448742746615087e-05,
169
+ "loss": 0.654,
170
+ "step": 115
171
+ },
172
+ {
173
+ "epoch": 1.0,
174
+ "eval_accuracy": 0.7304347826086957,
175
+ "eval_auc": 0.8009738283627511,
176
+ "eval_f1": 0.7394957983193278,
177
+ "eval_loss": 0.5908015370368958,
178
+ "eval_precision": 0.6666666666666666,
179
+ "eval_recall": 0.8301886792452831,
180
+ "eval_runtime": 146.397,
181
+ "eval_samples_per_second": 1.571,
182
+ "eval_steps_per_second": 0.102,
183
+ "step": 115
184
+ },
185
+ {
186
+ "epoch": 1.0434782608695652,
187
+ "grad_norm": 2.2397806644439697,
188
+ "learning_rate": 4.400386847195358e-05,
189
+ "loss": 0.6489,
190
+ "step": 120
191
+ },
192
+ {
193
+ "epoch": 1.0869565217391304,
194
+ "grad_norm": 2.275467872619629,
195
+ "learning_rate": 4.352030947775629e-05,
196
+ "loss": 0.6314,
197
+ "step": 125
198
+ },
199
+ {
200
+ "epoch": 1.1304347826086956,
201
+ "grad_norm": 2.6709554195404053,
202
+ "learning_rate": 4.3036750483559e-05,
203
+ "loss": 0.5121,
204
+ "step": 130
205
+ },
206
+ {
207
+ "epoch": 1.1739130434782608,
208
+ "grad_norm": 6.607785701751709,
209
+ "learning_rate": 4.2553191489361704e-05,
210
+ "loss": 0.4302,
211
+ "step": 135
212
+ },
213
+ {
214
+ "epoch": 1.2173913043478262,
215
+ "grad_norm": 18.070392608642578,
216
+ "learning_rate": 4.206963249516441e-05,
217
+ "loss": 0.6407,
218
+ "step": 140
219
+ },
220
+ {
221
+ "epoch": 1.2608695652173914,
222
+ "grad_norm": 3.231846570968628,
223
+ "learning_rate": 4.1586073500967124e-05,
224
+ "loss": 0.4142,
225
+ "step": 145
226
+ },
227
+ {
228
+ "epoch": 1.3043478260869565,
229
+ "grad_norm": 14.255928993225098,
230
+ "learning_rate": 4.110251450676983e-05,
231
+ "loss": 0.5749,
232
+ "step": 150
233
+ },
234
+ {
235
+ "epoch": 1.3478260869565217,
236
+ "grad_norm": 3.993816614151001,
237
+ "learning_rate": 4.061895551257254e-05,
238
+ "loss": 0.6971,
239
+ "step": 155
240
+ },
241
+ {
242
+ "epoch": 1.391304347826087,
243
+ "grad_norm": 6.085445404052734,
244
+ "learning_rate": 4.0135396518375244e-05,
245
+ "loss": 0.6003,
246
+ "step": 160
247
+ },
248
+ {
249
+ "epoch": 1.434782608695652,
250
+ "grad_norm": 4.129803657531738,
251
+ "learning_rate": 3.965183752417796e-05,
252
+ "loss": 0.6509,
253
+ "step": 165
254
+ },
255
+ {
256
+ "epoch": 1.4782608695652173,
257
+ "grad_norm": 4.400495529174805,
258
+ "learning_rate": 3.9168278529980664e-05,
259
+ "loss": 0.523,
260
+ "step": 170
261
+ },
262
+ {
263
+ "epoch": 1.5217391304347827,
264
+ "grad_norm": 4.820560932159424,
265
+ "learning_rate": 3.868471953578336e-05,
266
+ "loss": 0.4522,
267
+ "step": 175
268
+ },
269
+ {
270
+ "epoch": 1.5652173913043477,
271
+ "grad_norm": 13.549372673034668,
272
+ "learning_rate": 3.820116054158607e-05,
273
+ "loss": 0.5863,
274
+ "step": 180
275
+ },
276
+ {
277
+ "epoch": 1.608695652173913,
278
+ "grad_norm": 3.1626596450805664,
279
+ "learning_rate": 3.771760154738878e-05,
280
+ "loss": 0.4827,
281
+ "step": 185
282
+ },
283
+ {
284
+ "epoch": 1.6521739130434783,
285
+ "grad_norm": 3.709228992462158,
286
+ "learning_rate": 3.723404255319149e-05,
287
+ "loss": 0.5505,
288
+ "step": 190
289
+ },
290
+ {
291
+ "epoch": 1.6956521739130435,
292
+ "grad_norm": 4.613496780395508,
293
+ "learning_rate": 3.6750483558994196e-05,
294
+ "loss": 0.6062,
295
+ "step": 195
296
+ },
297
+ {
298
+ "epoch": 1.7391304347826086,
299
+ "grad_norm": 8.096386909484863,
300
+ "learning_rate": 3.62669245647969e-05,
301
+ "loss": 0.657,
302
+ "step": 200
303
+ },
304
+ {
305
+ "epoch": 1.7826086956521738,
306
+ "grad_norm": 7.13643741607666,
307
+ "learning_rate": 3.5783365570599616e-05,
308
+ "loss": 0.7452,
309
+ "step": 205
310
+ },
311
+ {
312
+ "epoch": 1.8260869565217392,
313
+ "grad_norm": 6.733407020568848,
314
+ "learning_rate": 3.529980657640232e-05,
315
+ "loss": 0.3675,
316
+ "step": 210
317
+ },
318
+ {
319
+ "epoch": 1.8695652173913042,
320
+ "grad_norm": 3.988879442214966,
321
+ "learning_rate": 3.481624758220503e-05,
322
+ "loss": 0.5059,
323
+ "step": 215
324
+ },
325
+ {
326
+ "epoch": 1.9130434782608696,
327
+ "grad_norm": 13.215219497680664,
328
+ "learning_rate": 3.4332688588007736e-05,
329
+ "loss": 0.5753,
330
+ "step": 220
331
+ },
332
+ {
333
+ "epoch": 1.9565217391304348,
334
+ "grad_norm": 20.60140037536621,
335
+ "learning_rate": 3.384912959381045e-05,
336
+ "loss": 0.6854,
337
+ "step": 225
338
+ },
339
+ {
340
+ "epoch": 2.0,
341
+ "grad_norm": 7.606060981750488,
342
+ "learning_rate": 3.3365570599613155e-05,
343
+ "loss": 0.4706,
344
+ "step": 230
345
+ },
346
+ {
347
+ "epoch": 2.0,
348
+ "eval_accuracy": 0.7565217391304347,
349
+ "eval_auc": 0.8304930006086427,
350
+ "eval_f1": 0.7777777777777778,
351
+ "eval_loss": 0.5158100724220276,
352
+ "eval_precision": 0.6712328767123288,
353
+ "eval_recall": 0.9245283018867925,
354
+ "eval_runtime": 179.768,
355
+ "eval_samples_per_second": 1.279,
356
+ "eval_steps_per_second": 0.083,
357
+ "step": 230
358
+ }
359
+ ],
360
+ "logging_steps": 5,
361
+ "max_steps": 575,
362
+ "num_input_tokens_seen": 0,
363
+ "num_train_epochs": 5,
364
+ "save_steps": 500,
365
+ "stateful_callbacks": {
366
+ "EarlyStoppingCallback": {
367
+ "args": {
368
+ "early_stopping_patience": 5,
369
+ "early_stopping_threshold": 0.01
370
+ },
371
+ "attributes": {
372
+ "early_stopping_patience_counter": 0
373
+ }
374
+ },
375
+ "TrainerControl": {
376
+ "args": {
377
+ "should_epoch_stop": false,
378
+ "should_evaluate": false,
379
+ "should_log": false,
380
+ "should_save": true,
381
+ "should_training_stop": false
382
+ },
383
+ "attributes": {}
384
+ }
385
+ },
386
+ "total_flos": 602791211679744.0,
387
+ "train_batch_size": 8,
388
+ "trial_name": null,
389
+ "trial_params": null
390
+ }
checkpoint-230/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:247bb778f59afd588c7d469992fb16aa065e3c43260f1c36d02137c15be47d63
3
+ size 5368
config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "studio-ousia/luke-japanese-base-lite",
3
+ "_num_labels": 2,
4
+ "architectures": [
5
+ "LukeForSequenceClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "bert_model_name": "models/luke-japanese/hf_xlm_roberta",
9
+ "bos_token_id": 0,
10
+ "classifier_dropout": null,
11
+ "cls_entity_prediction": false,
12
+ "entity_emb_size": 256,
13
+ "entity_vocab_size": 4,
14
+ "eos_token_id": 2,
15
+ "hidden_act": "gelu",
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 768,
18
+ "id2label": {
19
+ "0": "\u60aa\u3044",
20
+ "1": "\u826f\u3044"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "label2id": {
25
+ "\u60aa\u3044": 0,
26
+ "\u826f\u3044": 1
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "luke",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "single_label_classification",
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.48.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": true,
40
+ "use_entity_aware_attention": true,
41
+ "vocab_size": 32772
42
+ }
entity_vocab.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "[MASK2]": 3,
3
+ "[MASK]": 0,
4
+ "[PAD]": 2,
5
+ "[UNK]": 1
6
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e875fdd6ff9f1f63a626c3fd95c5d7570e7de8e48c447c151e30ef08bdfcec0
3
+ size 532307280
runs/Sep22_02-26-55_r-manato003-dlsite-voice-rec-5mtoqgub-ebe78-a8qm5/events.out.tfevents.1758508017.r-manato003-dlsite-voice-rec-5mtoqgub-ebe78-a8qm5.63.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e60dee4e2ad56421f4c4057c8631751956bc9fa45355490aa082439c3d5856c9
3
- size 31254
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b349a1b735166cd1a05e421783c243767cfe22bcaf6e3ac51f73801026c3194
3
+ size 32549
runs/Sep22_02-26-55_r-manato003-dlsite-voice-rec-5mtoqgub-ebe78-a8qm5/events.out.tfevents.1758521866.r-manato003-dlsite-voice-rec-5mtoqgub-ebe78-a8qm5.63.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b292d489a9c9e2157a47fc1ceed68265ce57f7776d1dab16387b985f0739927
3
+ size 607
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8b73a5e054936c920cf5b7d1ec21ce9c281977078269963beb821c6c86fbff7
3
+ size 841889
special_tokens_map.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<ent>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<ent2>",
12
+ "lstrip": false,
13
+ "normalized": true,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<ent>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "<ent2>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<ent>",
33
+ "lstrip": false,
34
+ "normalized": true,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "<ent2>",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ }
45
+ ],
46
+ "bos_token": {
47
+ "content": "<s>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "cls_token": {
54
+ "content": "<s>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false
59
+ },
60
+ "eos_token": {
61
+ "content": "</s>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false
66
+ },
67
+ "mask_token": {
68
+ "content": "<mask>",
69
+ "lstrip": true,
70
+ "normalized": true,
71
+ "rstrip": false,
72
+ "single_word": false
73
+ },
74
+ "pad_token": {
75
+ "content": "<pad>",
76
+ "lstrip": false,
77
+ "normalized": false,
78
+ "rstrip": false,
79
+ "single_word": false
80
+ },
81
+ "sep_token": {
82
+ "content": "</s>",
83
+ "lstrip": false,
84
+ "normalized": false,
85
+ "rstrip": false,
86
+ "single_word": false
87
+ },
88
+ "unk_token": {
89
+ "content": "<unk>",
90
+ "lstrip": false,
91
+ "normalized": false,
92
+ "rstrip": false,
93
+ "single_word": false
94
+ }
95
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "32769": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": true,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "32770": {
44
+ "content": "<ent>",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "32771": {
52
+ "content": "<ent2>",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ }
59
+ },
60
+ "additional_special_tokens": [
61
+ "<ent>",
62
+ "<ent2>",
63
+ "<ent>",
64
+ "<ent2>",
65
+ "<ent>",
66
+ "<ent2>"
67
+ ],
68
+ "bos_token": "<s>",
69
+ "clean_up_tokenization_spaces": false,
70
+ "cls_token": "<s>",
71
+ "entity_mask2_token": "[MASK2]",
72
+ "entity_mask_token": "[MASK]",
73
+ "entity_pad_token": "[PAD]",
74
+ "entity_token_1": {
75
+ "__type": "AddedToken",
76
+ "content": "<ent>",
77
+ "lstrip": false,
78
+ "normalized": true,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": false
82
+ },
83
+ "entity_token_2": {
84
+ "__type": "AddedToken",
85
+ "content": "<ent2>",
86
+ "lstrip": false,
87
+ "normalized": true,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": false
91
+ },
92
+ "entity_unk_token": "[UNK]",
93
+ "eos_token": "</s>",
94
+ "extra_special_tokens": {},
95
+ "mask_token": "<mask>",
96
+ "max_entity_length": 32,
97
+ "max_mention_length": 30,
98
+ "model_max_length": 512,
99
+ "pad_token": "<pad>",
100
+ "sep_token": "</s>",
101
+ "sp_model_kwargs": {},
102
+ "task": null,
103
+ "tokenizer_class": "MLukeTokenizer",
104
+ "unk_token": "<unk>"
105
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:247bb778f59afd588c7d469992fb16aa065e3c43260f1c36d02137c15be47d63
3
+ size 5368
training_params.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "2ClassLarge/autotrain-data",
3
+ "model": "studio-ousia/luke-japanese-base-lite",
4
+ "lr": 5e-05,
5
+ "epochs": 5,
6
+ "max_seq_length": 512,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "text_column": "autotrain_text",
18
+ "target_column": "autotrain_label",
19
+ "logging_steps": -1,
20
+ "project_name": "2ClassLarge",
21
+ "auto_find_batch_size": false,
22
+ "mixed_precision": "fp16",
23
+ "save_total_limit": 1,
24
+ "push_to_hub": true,
25
+ "eval_strategy": "epoch",
26
+ "username": "manato003",
27
+ "log": "tensorboard",
28
+ "early_stopping_patience": 5,
29
+ "early_stopping_threshold": 0.01
30
+ }