PuxAI commited on
Commit
3c0790b
·
verified ·
1 Parent(s): 286e5c2

Upload folder using huggingface_hub

Browse files
Files changed (29) hide show
  1. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/config.json +120 -0
  2. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/model.safetensors +3 -0
  3. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/optimizer.pt +3 -0
  4. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/rng_state.pth +3 -0
  5. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/scheduler.pt +3 -0
  6. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/tokenizer.json +0 -0
  7. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/tokenizer_config.json +15 -0
  8. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/trainer_state.json +245 -0
  9. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/training_args.bin +3 -0
  10. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/config.json +120 -0
  11. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/model.safetensors +3 -0
  12. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/optimizer.pt +3 -0
  13. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/rng_state.pth +3 -0
  14. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/scheduler.pt +3 -0
  15. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/tokenizer.json +0 -0
  16. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/tokenizer_config.json +15 -0
  17. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/trainer_state.json +456 -0
  18. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/training_args.bin +3 -0
  19. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/config.json +120 -0
  20. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/model.safetensors +3 -0
  21. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/optimizer.pt +3 -0
  22. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/rng_state.pth +3 -0
  23. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/scheduler.pt +3 -0
  24. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/tokenizer.json +0 -0
  25. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/tokenizer_config.json +15 -0
  26. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/trainer_state.json +667 -0
  27. open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/training_args.bin +3 -0
  28. open-pii-masking-500k-ai4privacy/TokenBased-BERT/model.safetensors +1 -1
  29. open-pii-masking-500k-ai4privacy/TokenBased-BERT/training_args.bin +1 -1
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "directionality": "bidi",
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 768,
15
+ "id2label": {
16
+ "0": "O",
17
+ "1": "B-AGE",
18
+ "2": "I-AGE",
19
+ "3": "B-BUILDINGNUM",
20
+ "4": "I-BUILDINGNUM",
21
+ "5": "B-CITY",
22
+ "6": "I-CITY",
23
+ "7": "B-CREDITCARDNUMBER",
24
+ "8": "I-CREDITCARDNUMBER",
25
+ "9": "B-DATE",
26
+ "10": "I-DATE",
27
+ "11": "B-DRIVERLICENSENUM",
28
+ "12": "I-DRIVERLICENSENUM",
29
+ "13": "B-EMAIL",
30
+ "14": "I-EMAIL",
31
+ "15": "B-GENDER",
32
+ "16": "I-GENDER",
33
+ "17": "B-GIVENNAME",
34
+ "18": "I-GIVENNAME",
35
+ "19": "B-IDCARDNUM",
36
+ "20": "I-IDCARDNUM",
37
+ "21": "B-PASSPORTNUM",
38
+ "22": "I-PASSPORTNUM",
39
+ "23": "B-SEX",
40
+ "24": "I-SEX",
41
+ "25": "B-SOCIALNUM",
42
+ "26": "I-SOCIALNUM",
43
+ "27": "B-STREET",
44
+ "28": "I-STREET",
45
+ "29": "B-SURNAME",
46
+ "30": "I-SURNAME",
47
+ "31": "B-TAXNUM",
48
+ "32": "I-TAXNUM",
49
+ "33": "B-TELEPHONENUM",
50
+ "34": "I-TELEPHONENUM",
51
+ "35": "B-TIME",
52
+ "36": "I-TIME",
53
+ "37": "B-TITLE",
54
+ "38": "I-TITLE",
55
+ "39": "B-ZIPCODE",
56
+ "40": "I-ZIPCODE"
57
+ },
58
+ "initializer_range": 0.02,
59
+ "intermediate_size": 3072,
60
+ "is_decoder": false,
61
+ "label2id": {
62
+ "B-AGE": 1,
63
+ "B-BUILDINGNUM": 3,
64
+ "B-CITY": 5,
65
+ "B-CREDITCARDNUMBER": 7,
66
+ "B-DATE": 9,
67
+ "B-DRIVERLICENSENUM": 11,
68
+ "B-EMAIL": 13,
69
+ "B-GENDER": 15,
70
+ "B-GIVENNAME": 17,
71
+ "B-IDCARDNUM": 19,
72
+ "B-PASSPORTNUM": 21,
73
+ "B-SEX": 23,
74
+ "B-SOCIALNUM": 25,
75
+ "B-STREET": 27,
76
+ "B-SURNAME": 29,
77
+ "B-TAXNUM": 31,
78
+ "B-TELEPHONENUM": 33,
79
+ "B-TIME": 35,
80
+ "B-TITLE": 37,
81
+ "B-ZIPCODE": 39,
82
+ "I-AGE": 2,
83
+ "I-BUILDINGNUM": 4,
84
+ "I-CITY": 6,
85
+ "I-CREDITCARDNUMBER": 8,
86
+ "I-DATE": 10,
87
+ "I-DRIVERLICENSENUM": 12,
88
+ "I-EMAIL": 14,
89
+ "I-GENDER": 16,
90
+ "I-GIVENNAME": 18,
91
+ "I-IDCARDNUM": 20,
92
+ "I-PASSPORTNUM": 22,
93
+ "I-SEX": 24,
94
+ "I-SOCIALNUM": 26,
95
+ "I-STREET": 28,
96
+ "I-SURNAME": 30,
97
+ "I-TAXNUM": 32,
98
+ "I-TELEPHONENUM": 34,
99
+ "I-TIME": 36,
100
+ "I-TITLE": 38,
101
+ "I-ZIPCODE": 40,
102
+ "O": 0
103
+ },
104
+ "layer_norm_eps": 1e-12,
105
+ "max_position_embeddings": 512,
106
+ "model_type": "bert",
107
+ "num_attention_heads": 12,
108
+ "num_hidden_layers": 12,
109
+ "pad_token_id": 0,
110
+ "pooler_fc_size": 768,
111
+ "pooler_num_attention_heads": 12,
112
+ "pooler_num_fc_layers": 3,
113
+ "pooler_size_per_head": 128,
114
+ "pooler_type": "first_token_transform",
115
+ "tie_word_embeddings": true,
116
+ "transformers_version": "5.3.0",
117
+ "type_vocab_size": 2,
118
+ "use_cache": false,
119
+ "vocab_size": 119547
120
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a43ae655d890b0d9de8146183a3bc738617fc4122b83faa5e8ddd133a777fa
3
+ size 709200844
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72fb3a75b29e354b99adb41127a5109839e55b953b8892cd5da3b5da8580936a
3
+ size 1418524683
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49a8f6a97b57d6094d1779566cf055143a4430538197a0c59c1f6fbed018e892
3
+ size 14645
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab6b4a28e8f835f500b1343c5f903f234bad931ebbad7203d75a94cc90f81574
3
+ size 1465
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "is_local": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/trainer_state.json ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 14505,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03447087211306446,
14
+ "grad_norm": 1.2918020486831665,
15
+ "learning_rate": 1.9770653797541078e-05,
16
+ "loss": 0.21463787841796875,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.06894174422612892,
21
+ "grad_norm": 1.4572890996932983,
22
+ "learning_rate": 1.9540847983453983e-05,
23
+ "loss": 0.06676451110839844,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.10341261633919338,
28
+ "grad_norm": 0.7736029624938965,
29
+ "learning_rate": 1.9311042169366887e-05,
30
+ "loss": 0.05756806182861328,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.13788348845225784,
35
+ "grad_norm": 0.5729684829711914,
36
+ "learning_rate": 1.9081236355279792e-05,
37
+ "loss": 0.05027279663085937,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.1723543605653223,
42
+ "grad_norm": 0.6462647914886475,
43
+ "learning_rate": 1.8851430541192693e-05,
44
+ "loss": 0.04522799301147461,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.20682523267838676,
49
+ "grad_norm": 1.5296399593353271,
50
+ "learning_rate": 1.8621624727105598e-05,
51
+ "loss": 0.04301478195190429,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.24129610479145122,
56
+ "grad_norm": 0.7765299081802368,
57
+ "learning_rate": 1.83918189130185e-05,
58
+ "loss": 0.041502281188964846,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.2757669769045157,
63
+ "grad_norm": 0.44941186904907227,
64
+ "learning_rate": 1.8162013098931404e-05,
65
+ "loss": 0.036275325775146486,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.31023784901758017,
70
+ "grad_norm": 0.6415440440177917,
71
+ "learning_rate": 1.793220728484431e-05,
72
+ "loss": 0.038519672393798825,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.3447087211306446,
77
+ "grad_norm": 1.7180463075637817,
78
+ "learning_rate": 1.770240147075721e-05,
79
+ "loss": 0.03590824508666992,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.3791795932437091,
84
+ "grad_norm": 1.706147313117981,
85
+ "learning_rate": 1.7472595656670115e-05,
86
+ "loss": 0.03493925476074219,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.4136504653567735,
91
+ "grad_norm": 0.20170536637306213,
92
+ "learning_rate": 1.724278984258302e-05,
93
+ "loss": 0.03378963088989258,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 0.448121337469838,
98
+ "grad_norm": 0.32674115896224976,
99
+ "learning_rate": 1.7012984028495924e-05,
100
+ "loss": 0.031928113937377926,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 0.48259220958290244,
105
+ "grad_norm": 0.3478796184062958,
106
+ "learning_rate": 1.6783178214408825e-05,
107
+ "loss": 0.033053131103515626,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 0.5170630816959669,
112
+ "grad_norm": 0.6997560858726501,
113
+ "learning_rate": 1.655337240032173e-05,
114
+ "loss": 0.03176578521728515,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 0.5515339538090314,
119
+ "grad_norm": 0.34584298729896545,
120
+ "learning_rate": 1.6323566586234634e-05,
121
+ "loss": 0.03126089859008789,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 0.5860048259220958,
126
+ "grad_norm": 0.7542563080787659,
127
+ "learning_rate": 1.609376077214754e-05,
128
+ "loss": 0.030911777496337892,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 0.6204756980351603,
133
+ "grad_norm": 0.7195969223976135,
134
+ "learning_rate": 1.586395495806044e-05,
135
+ "loss": 0.029189849853515627,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 0.6549465701482248,
140
+ "grad_norm": 1.3274503946304321,
141
+ "learning_rate": 1.563414914397334e-05,
142
+ "loss": 0.03025291061401367,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 0.6894174422612892,
147
+ "grad_norm": 0.6382108926773071,
148
+ "learning_rate": 1.5404343329886246e-05,
149
+ "loss": 0.028342636108398437,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 0.7238883143743536,
154
+ "grad_norm": 0.2607400715351105,
155
+ "learning_rate": 1.5174537515799151e-05,
156
+ "loss": 0.028958717346191407,
157
+ "step": 10500
158
+ },
159
+ {
160
+ "epoch": 0.7583591864874182,
161
+ "grad_norm": 1.1773029565811157,
162
+ "learning_rate": 1.4944731701712054e-05,
163
+ "loss": 0.027841657638549806,
164
+ "step": 11000
165
+ },
166
+ {
167
+ "epoch": 0.7928300586004826,
168
+ "grad_norm": 1.8503401279449463,
169
+ "learning_rate": 1.4714925887624959e-05,
170
+ "loss": 0.02816273307800293,
171
+ "step": 11500
172
+ },
173
+ {
174
+ "epoch": 0.827300930713547,
175
+ "grad_norm": 1.0574482679367065,
176
+ "learning_rate": 1.4485120073537862e-05,
177
+ "loss": 0.027129316329956056,
178
+ "step": 12000
179
+ },
180
+ {
181
+ "epoch": 0.8617718028266115,
182
+ "grad_norm": 0.2772712707519531,
183
+ "learning_rate": 1.4255314259450766e-05,
184
+ "loss": 0.02708342361450195,
185
+ "step": 12500
186
+ },
187
+ {
188
+ "epoch": 0.896242674939676,
189
+ "grad_norm": 0.20870168507099152,
190
+ "learning_rate": 1.402550844536367e-05,
191
+ "loss": 0.024648666381835938,
192
+ "step": 13000
193
+ },
194
+ {
195
+ "epoch": 0.9307135470527405,
196
+ "grad_norm": 0.5534266829490662,
197
+ "learning_rate": 1.3795702631276572e-05,
198
+ "loss": 0.026792591094970702,
199
+ "step": 13500
200
+ },
201
+ {
202
+ "epoch": 0.9651844191658049,
203
+ "grad_norm": 0.5170403718948364,
204
+ "learning_rate": 1.3565896817189477e-05,
205
+ "loss": 0.02854232597351074,
206
+ "step": 14000
207
+ },
208
+ {
209
+ "epoch": 0.9996552912788693,
210
+ "grad_norm": 0.27890703082084656,
211
+ "learning_rate": 1.333609100310238e-05,
212
+ "loss": 0.026233253479003907,
213
+ "step": 14500
214
+ },
215
+ {
216
+ "epoch": 1.0,
217
+ "eval_loss": 0.02844008058309555,
218
+ "eval_runtime": 294.9235,
219
+ "eval_samples_per_second": 393.583,
220
+ "eval_steps_per_second": 49.199,
221
+ "step": 14505
222
+ }
223
+ ],
224
+ "logging_steps": 500,
225
+ "max_steps": 43515,
226
+ "num_input_tokens_seen": 0,
227
+ "num_train_epochs": 3,
228
+ "save_steps": 500,
229
+ "stateful_callbacks": {
230
+ "TrainerControl": {
231
+ "args": {
232
+ "should_epoch_stop": false,
233
+ "should_evaluate": false,
234
+ "should_log": false,
235
+ "should_save": true,
236
+ "should_training_stop": false
237
+ },
238
+ "attributes": {}
239
+ }
240
+ },
241
+ "total_flos": 1.213236528701952e+17,
242
+ "train_batch_size": 32,
243
+ "trial_name": null,
244
+ "trial_params": null
245
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-14505/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:083f7dd27ac513d7af866e0a8fbcb3a43e357081778fc03b69d05a60485954c3
3
+ size 5201
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "directionality": "bidi",
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 768,
15
+ "id2label": {
16
+ "0": "O",
17
+ "1": "B-AGE",
18
+ "2": "I-AGE",
19
+ "3": "B-BUILDINGNUM",
20
+ "4": "I-BUILDINGNUM",
21
+ "5": "B-CITY",
22
+ "6": "I-CITY",
23
+ "7": "B-CREDITCARDNUMBER",
24
+ "8": "I-CREDITCARDNUMBER",
25
+ "9": "B-DATE",
26
+ "10": "I-DATE",
27
+ "11": "B-DRIVERLICENSENUM",
28
+ "12": "I-DRIVERLICENSENUM",
29
+ "13": "B-EMAIL",
30
+ "14": "I-EMAIL",
31
+ "15": "B-GENDER",
32
+ "16": "I-GENDER",
33
+ "17": "B-GIVENNAME",
34
+ "18": "I-GIVENNAME",
35
+ "19": "B-IDCARDNUM",
36
+ "20": "I-IDCARDNUM",
37
+ "21": "B-PASSPORTNUM",
38
+ "22": "I-PASSPORTNUM",
39
+ "23": "B-SEX",
40
+ "24": "I-SEX",
41
+ "25": "B-SOCIALNUM",
42
+ "26": "I-SOCIALNUM",
43
+ "27": "B-STREET",
44
+ "28": "I-STREET",
45
+ "29": "B-SURNAME",
46
+ "30": "I-SURNAME",
47
+ "31": "B-TAXNUM",
48
+ "32": "I-TAXNUM",
49
+ "33": "B-TELEPHONENUM",
50
+ "34": "I-TELEPHONENUM",
51
+ "35": "B-TIME",
52
+ "36": "I-TIME",
53
+ "37": "B-TITLE",
54
+ "38": "I-TITLE",
55
+ "39": "B-ZIPCODE",
56
+ "40": "I-ZIPCODE"
57
+ },
58
+ "initializer_range": 0.02,
59
+ "intermediate_size": 3072,
60
+ "is_decoder": false,
61
+ "label2id": {
62
+ "B-AGE": 1,
63
+ "B-BUILDINGNUM": 3,
64
+ "B-CITY": 5,
65
+ "B-CREDITCARDNUMBER": 7,
66
+ "B-DATE": 9,
67
+ "B-DRIVERLICENSENUM": 11,
68
+ "B-EMAIL": 13,
69
+ "B-GENDER": 15,
70
+ "B-GIVENNAME": 17,
71
+ "B-IDCARDNUM": 19,
72
+ "B-PASSPORTNUM": 21,
73
+ "B-SEX": 23,
74
+ "B-SOCIALNUM": 25,
75
+ "B-STREET": 27,
76
+ "B-SURNAME": 29,
77
+ "B-TAXNUM": 31,
78
+ "B-TELEPHONENUM": 33,
79
+ "B-TIME": 35,
80
+ "B-TITLE": 37,
81
+ "B-ZIPCODE": 39,
82
+ "I-AGE": 2,
83
+ "I-BUILDINGNUM": 4,
84
+ "I-CITY": 6,
85
+ "I-CREDITCARDNUMBER": 8,
86
+ "I-DATE": 10,
87
+ "I-DRIVERLICENSENUM": 12,
88
+ "I-EMAIL": 14,
89
+ "I-GENDER": 16,
90
+ "I-GIVENNAME": 18,
91
+ "I-IDCARDNUM": 20,
92
+ "I-PASSPORTNUM": 22,
93
+ "I-SEX": 24,
94
+ "I-SOCIALNUM": 26,
95
+ "I-STREET": 28,
96
+ "I-SURNAME": 30,
97
+ "I-TAXNUM": 32,
98
+ "I-TELEPHONENUM": 34,
99
+ "I-TIME": 36,
100
+ "I-TITLE": 38,
101
+ "I-ZIPCODE": 40,
102
+ "O": 0
103
+ },
104
+ "layer_norm_eps": 1e-12,
105
+ "max_position_embeddings": 512,
106
+ "model_type": "bert",
107
+ "num_attention_heads": 12,
108
+ "num_hidden_layers": 12,
109
+ "pad_token_id": 0,
110
+ "pooler_fc_size": 768,
111
+ "pooler_num_attention_heads": 12,
112
+ "pooler_num_fc_layers": 3,
113
+ "pooler_size_per_head": 128,
114
+ "pooler_type": "first_token_transform",
115
+ "tie_word_embeddings": true,
116
+ "transformers_version": "5.3.0",
117
+ "type_vocab_size": 2,
118
+ "use_cache": false,
119
+ "vocab_size": 119547
120
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e5ebbfce1d819c1aca019e11b6b0c8aff6d36993ad330b412166fe15cf7963
3
+ size 709200844
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a952dccad00907569aec6942aa21ad1bb257211f1fe69ba48effd7ee990d323a
3
+ size 1418524683
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:976843b85f994fe9a6c2cec86dc372b7f6422ef12e6dffa965568ce47a163cf8
3
+ size 14645
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8faec6f332b537cbe0ba652579e9c9f37c1fff9506b6c4994c1391dae58fc7f
3
+ size 1465
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "is_local": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/trainer_state.json ADDED
@@ -0,0 +1,456 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 29010,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03447087211306446,
14
+ "grad_norm": 1.2918020486831665,
15
+ "learning_rate": 1.9770653797541078e-05,
16
+ "loss": 0.21463787841796875,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.06894174422612892,
21
+ "grad_norm": 1.4572890996932983,
22
+ "learning_rate": 1.9540847983453983e-05,
23
+ "loss": 0.06676451110839844,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.10341261633919338,
28
+ "grad_norm": 0.7736029624938965,
29
+ "learning_rate": 1.9311042169366887e-05,
30
+ "loss": 0.05756806182861328,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.13788348845225784,
35
+ "grad_norm": 0.5729684829711914,
36
+ "learning_rate": 1.9081236355279792e-05,
37
+ "loss": 0.05027279663085937,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.1723543605653223,
42
+ "grad_norm": 0.6462647914886475,
43
+ "learning_rate": 1.8851430541192693e-05,
44
+ "loss": 0.04522799301147461,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.20682523267838676,
49
+ "grad_norm": 1.5296399593353271,
50
+ "learning_rate": 1.8621624727105598e-05,
51
+ "loss": 0.04301478195190429,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.24129610479145122,
56
+ "grad_norm": 0.7765299081802368,
57
+ "learning_rate": 1.83918189130185e-05,
58
+ "loss": 0.041502281188964846,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.2757669769045157,
63
+ "grad_norm": 0.44941186904907227,
64
+ "learning_rate": 1.8162013098931404e-05,
65
+ "loss": 0.036275325775146486,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.31023784901758017,
70
+ "grad_norm": 0.6415440440177917,
71
+ "learning_rate": 1.793220728484431e-05,
72
+ "loss": 0.038519672393798825,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.3447087211306446,
77
+ "grad_norm": 1.7180463075637817,
78
+ "learning_rate": 1.770240147075721e-05,
79
+ "loss": 0.03590824508666992,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.3791795932437091,
84
+ "grad_norm": 1.706147313117981,
85
+ "learning_rate": 1.7472595656670115e-05,
86
+ "loss": 0.03493925476074219,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.4136504653567735,
91
+ "grad_norm": 0.20170536637306213,
92
+ "learning_rate": 1.724278984258302e-05,
93
+ "loss": 0.03378963088989258,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 0.448121337469838,
98
+ "grad_norm": 0.32674115896224976,
99
+ "learning_rate": 1.7012984028495924e-05,
100
+ "loss": 0.031928113937377926,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 0.48259220958290244,
105
+ "grad_norm": 0.3478796184062958,
106
+ "learning_rate": 1.6783178214408825e-05,
107
+ "loss": 0.033053131103515626,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 0.5170630816959669,
112
+ "grad_norm": 0.6997560858726501,
113
+ "learning_rate": 1.655337240032173e-05,
114
+ "loss": 0.03176578521728515,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 0.5515339538090314,
119
+ "grad_norm": 0.34584298729896545,
120
+ "learning_rate": 1.6323566586234634e-05,
121
+ "loss": 0.03126089859008789,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 0.5860048259220958,
126
+ "grad_norm": 0.7542563080787659,
127
+ "learning_rate": 1.609376077214754e-05,
128
+ "loss": 0.030911777496337892,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 0.6204756980351603,
133
+ "grad_norm": 0.7195969223976135,
134
+ "learning_rate": 1.586395495806044e-05,
135
+ "loss": 0.029189849853515627,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 0.6549465701482248,
140
+ "grad_norm": 1.3274503946304321,
141
+ "learning_rate": 1.563414914397334e-05,
142
+ "loss": 0.03025291061401367,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 0.6894174422612892,
147
+ "grad_norm": 0.6382108926773071,
148
+ "learning_rate": 1.5404343329886246e-05,
149
+ "loss": 0.028342636108398437,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 0.7238883143743536,
154
+ "grad_norm": 0.2607400715351105,
155
+ "learning_rate": 1.5174537515799151e-05,
156
+ "loss": 0.028958717346191407,
157
+ "step": 10500
158
+ },
159
+ {
160
+ "epoch": 0.7583591864874182,
161
+ "grad_norm": 1.1773029565811157,
162
+ "learning_rate": 1.4944731701712054e-05,
163
+ "loss": 0.027841657638549806,
164
+ "step": 11000
165
+ },
166
+ {
167
+ "epoch": 0.7928300586004826,
168
+ "grad_norm": 1.8503401279449463,
169
+ "learning_rate": 1.4714925887624959e-05,
170
+ "loss": 0.02816273307800293,
171
+ "step": 11500
172
+ },
173
+ {
174
+ "epoch": 0.827300930713547,
175
+ "grad_norm": 1.0574482679367065,
176
+ "learning_rate": 1.4485120073537862e-05,
177
+ "loss": 0.027129316329956056,
178
+ "step": 12000
179
+ },
180
+ {
181
+ "epoch": 0.8617718028266115,
182
+ "grad_norm": 0.2772712707519531,
183
+ "learning_rate": 1.4255314259450766e-05,
184
+ "loss": 0.02708342361450195,
185
+ "step": 12500
186
+ },
187
+ {
188
+ "epoch": 0.896242674939676,
189
+ "grad_norm": 0.20870168507099152,
190
+ "learning_rate": 1.402550844536367e-05,
191
+ "loss": 0.024648666381835938,
192
+ "step": 13000
193
+ },
194
+ {
195
+ "epoch": 0.9307135470527405,
196
+ "grad_norm": 0.5534266829490662,
197
+ "learning_rate": 1.3795702631276572e-05,
198
+ "loss": 0.026792591094970702,
199
+ "step": 13500
200
+ },
201
+ {
202
+ "epoch": 0.9651844191658049,
203
+ "grad_norm": 0.5170403718948364,
204
+ "learning_rate": 1.3565896817189477e-05,
205
+ "loss": 0.02854232597351074,
206
+ "step": 14000
207
+ },
208
+ {
209
+ "epoch": 0.9996552912788693,
210
+ "grad_norm": 0.27890703082084656,
211
+ "learning_rate": 1.333609100310238e-05,
212
+ "loss": 0.026233253479003907,
213
+ "step": 14500
214
+ },
215
+ {
216
+ "epoch": 1.0,
217
+ "eval_loss": 0.02844008058309555,
218
+ "eval_runtime": 294.9235,
219
+ "eval_samples_per_second": 393.583,
220
+ "eval_steps_per_second": 49.199,
221
+ "step": 14505
222
+ },
223
+ {
224
+ "epoch": 1.0341261633919339,
225
+ "grad_norm": 1.2330853939056396,
226
+ "learning_rate": 1.3106285189015284e-05,
227
+ "loss": 0.020723657608032228,
228
+ "step": 15000
229
+ },
230
+ {
231
+ "epoch": 1.0685970355049983,
232
+ "grad_norm": 0.3847709596157074,
233
+ "learning_rate": 1.2876479374928186e-05,
234
+ "loss": 0.02110268211364746,
235
+ "step": 15500
236
+ },
237
+ {
238
+ "epoch": 1.1030679076180627,
239
+ "grad_norm": 0.5638656616210938,
240
+ "learning_rate": 1.264667356084109e-05,
241
+ "loss": 0.021719074249267577,
242
+ "step": 16000
243
+ },
244
+ {
245
+ "epoch": 1.1375387797311272,
246
+ "grad_norm": 0.25548455119132996,
247
+ "learning_rate": 1.2416867746753993e-05,
248
+ "loss": 0.020617319107055665,
249
+ "step": 16500
250
+ },
251
+ {
252
+ "epoch": 1.1720096518441916,
253
+ "grad_norm": 0.2979449927806854,
254
+ "learning_rate": 1.2187061932666896e-05,
255
+ "loss": 0.021886722564697265,
256
+ "step": 17000
257
+ },
258
+ {
259
+ "epoch": 1.206480523957256,
260
+ "grad_norm": 0.18911246955394745,
261
+ "learning_rate": 1.1957256118579801e-05,
262
+ "loss": 0.020644411087036133,
263
+ "step": 17500
264
+ },
265
+ {
266
+ "epoch": 1.2409513960703205,
267
+ "grad_norm": 0.3482404351234436,
268
+ "learning_rate": 1.1727450304492704e-05,
269
+ "loss": 0.021316097259521483,
270
+ "step": 18000
271
+ },
272
+ {
273
+ "epoch": 1.275422268183385,
274
+ "grad_norm": 0.570595383644104,
275
+ "learning_rate": 1.1497644490405609e-05,
276
+ "loss": 0.02059372901916504,
277
+ "step": 18500
278
+ },
279
+ {
280
+ "epoch": 1.3098931402964495,
281
+ "grad_norm": 0.2911337614059448,
282
+ "learning_rate": 1.1267838676318512e-05,
283
+ "loss": 0.02033697509765625,
284
+ "step": 19000
285
+ },
286
+ {
287
+ "epoch": 1.344364012409514,
288
+ "grad_norm": 0.426343709230423,
289
+ "learning_rate": 1.1038032862231416e-05,
290
+ "loss": 0.020020153045654295,
291
+ "step": 19500
292
+ },
293
+ {
294
+ "epoch": 1.3788348845225784,
295
+ "grad_norm": 0.5702527761459351,
296
+ "learning_rate": 1.080822704814432e-05,
297
+ "loss": 0.02054849624633789,
298
+ "step": 20000
299
+ },
300
+ {
301
+ "epoch": 1.4133057566356428,
302
+ "grad_norm": 0.38962048292160034,
303
+ "learning_rate": 1.0578421234057224e-05,
304
+ "loss": 0.01970711326599121,
305
+ "step": 20500
306
+ },
307
+ {
308
+ "epoch": 1.4477766287487073,
309
+ "grad_norm": 0.6985688805580139,
310
+ "learning_rate": 1.0348615419970127e-05,
311
+ "loss": 0.020511926651000975,
312
+ "step": 21000
313
+ },
314
+ {
315
+ "epoch": 1.4822475008617717,
316
+ "grad_norm": 0.36666256189346313,
317
+ "learning_rate": 1.0118809605883028e-05,
318
+ "loss": 0.019306087493896486,
319
+ "step": 21500
320
+ },
321
+ {
322
+ "epoch": 1.5167183729748364,
323
+ "grad_norm": 0.3014184534549713,
324
+ "learning_rate": 9.889003791795934e-06,
325
+ "loss": 0.0195778694152832,
326
+ "step": 22000
327
+ },
328
+ {
329
+ "epoch": 1.5511892450879006,
330
+ "grad_norm": 1.1626015901565552,
331
+ "learning_rate": 9.659197977708836e-06,
332
+ "loss": 0.02046485710144043,
333
+ "step": 22500
334
+ },
335
+ {
336
+ "epoch": 1.5856601172009652,
337
+ "grad_norm": 0.2246391624212265,
338
+ "learning_rate": 9.42939216362174e-06,
339
+ "loss": 0.019212373733520506,
340
+ "step": 23000
341
+ },
342
+ {
343
+ "epoch": 1.6201309893140297,
344
+ "grad_norm": 0.7504266500473022,
345
+ "learning_rate": 9.199586349534643e-06,
346
+ "loss": 0.02021506118774414,
347
+ "step": 23500
348
+ },
349
+ {
350
+ "epoch": 1.654601861427094,
351
+ "grad_norm": 0.5071070194244385,
352
+ "learning_rate": 8.969780535447548e-06,
353
+ "loss": 0.01886193084716797,
354
+ "step": 24000
355
+ },
356
+ {
357
+ "epoch": 1.6890727335401585,
358
+ "grad_norm": 0.30986541509628296,
359
+ "learning_rate": 8.739974721360451e-06,
360
+ "loss": 0.01909671974182129,
361
+ "step": 24500
362
+ },
363
+ {
364
+ "epoch": 1.723543605653223,
365
+ "grad_norm": 1.065539836883545,
366
+ "learning_rate": 8.510168907273356e-06,
367
+ "loss": 0.019661422729492187,
368
+ "step": 25000
369
+ },
370
+ {
371
+ "epoch": 1.7580144777662876,
372
+ "grad_norm": 0.3560684621334076,
373
+ "learning_rate": 8.280363093186259e-06,
374
+ "loss": 0.019211139678955078,
375
+ "step": 25500
376
+ },
377
+ {
378
+ "epoch": 1.7924853498793518,
379
+ "grad_norm": 0.264694482088089,
380
+ "learning_rate": 8.050557279099161e-06,
381
+ "loss": 0.018874462127685548,
382
+ "step": 26000
383
+ },
384
+ {
385
+ "epoch": 1.8269562219924165,
386
+ "grad_norm": 0.49240633845329285,
387
+ "learning_rate": 7.820751465012064e-06,
388
+ "loss": 0.01885951042175293,
389
+ "step": 26500
390
+ },
391
+ {
392
+ "epoch": 1.861427094105481,
393
+ "grad_norm": 0.5500588417053223,
394
+ "learning_rate": 7.590945650924969e-06,
395
+ "loss": 0.019577322006225587,
396
+ "step": 27000
397
+ },
398
+ {
399
+ "epoch": 1.8958979662185453,
400
+ "grad_norm": 0.22589892148971558,
401
+ "learning_rate": 7.361139836837873e-06,
402
+ "loss": 0.018141788482666017,
403
+ "step": 27500
404
+ },
405
+ {
406
+ "epoch": 1.9303688383316098,
407
+ "grad_norm": 0.8253551721572876,
408
+ "learning_rate": 7.131334022750777e-06,
409
+ "loss": 0.018833749771118163,
410
+ "step": 28000
411
+ },
412
+ {
413
+ "epoch": 1.9648397104446742,
414
+ "grad_norm": 0.3297463357448578,
415
+ "learning_rate": 6.901528208663679e-06,
416
+ "loss": 0.01926751708984375,
417
+ "step": 28500
418
+ },
419
+ {
420
+ "epoch": 1.9993105825577389,
421
+ "grad_norm": 0.3467053174972534,
422
+ "learning_rate": 6.671722394576583e-06,
423
+ "loss": 0.019064535140991212,
424
+ "step": 29000
425
+ },
426
+ {
427
+ "epoch": 2.0,
428
+ "eval_loss": 0.024620965123176575,
429
+ "eval_runtime": 295.9862,
430
+ "eval_samples_per_second": 392.17,
431
+ "eval_steps_per_second": 49.023,
432
+ "step": 29010
433
+ }
434
+ ],
435
+ "logging_steps": 500,
436
+ "max_steps": 43515,
437
+ "num_input_tokens_seen": 0,
438
+ "num_train_epochs": 3,
439
+ "save_steps": 500,
440
+ "stateful_callbacks": {
441
+ "TrainerControl": {
442
+ "args": {
443
+ "should_epoch_stop": false,
444
+ "should_evaluate": false,
445
+ "should_log": false,
446
+ "should_save": true,
447
+ "should_training_stop": false
448
+ },
449
+ "attributes": {}
450
+ }
451
+ },
452
+ "total_flos": 2.426473057403904e+17,
453
+ "train_batch_size": 32,
454
+ "trial_name": null,
455
+ "trial_params": null
456
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-29010/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:083f7dd27ac513d7af866e0a8fbcb3a43e357081778fc03b69d05a60485954c3
3
+ size 5201
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/config.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "directionality": "bidi",
10
+ "dtype": "float32",
11
+ "eos_token_id": null,
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 768,
15
+ "id2label": {
16
+ "0": "O",
17
+ "1": "B-AGE",
18
+ "2": "I-AGE",
19
+ "3": "B-BUILDINGNUM",
20
+ "4": "I-BUILDINGNUM",
21
+ "5": "B-CITY",
22
+ "6": "I-CITY",
23
+ "7": "B-CREDITCARDNUMBER",
24
+ "8": "I-CREDITCARDNUMBER",
25
+ "9": "B-DATE",
26
+ "10": "I-DATE",
27
+ "11": "B-DRIVERLICENSENUM",
28
+ "12": "I-DRIVERLICENSENUM",
29
+ "13": "B-EMAIL",
30
+ "14": "I-EMAIL",
31
+ "15": "B-GENDER",
32
+ "16": "I-GENDER",
33
+ "17": "B-GIVENNAME",
34
+ "18": "I-GIVENNAME",
35
+ "19": "B-IDCARDNUM",
36
+ "20": "I-IDCARDNUM",
37
+ "21": "B-PASSPORTNUM",
38
+ "22": "I-PASSPORTNUM",
39
+ "23": "B-SEX",
40
+ "24": "I-SEX",
41
+ "25": "B-SOCIALNUM",
42
+ "26": "I-SOCIALNUM",
43
+ "27": "B-STREET",
44
+ "28": "I-STREET",
45
+ "29": "B-SURNAME",
46
+ "30": "I-SURNAME",
47
+ "31": "B-TAXNUM",
48
+ "32": "I-TAXNUM",
49
+ "33": "B-TELEPHONENUM",
50
+ "34": "I-TELEPHONENUM",
51
+ "35": "B-TIME",
52
+ "36": "I-TIME",
53
+ "37": "B-TITLE",
54
+ "38": "I-TITLE",
55
+ "39": "B-ZIPCODE",
56
+ "40": "I-ZIPCODE"
57
+ },
58
+ "initializer_range": 0.02,
59
+ "intermediate_size": 3072,
60
+ "is_decoder": false,
61
+ "label2id": {
62
+ "B-AGE": 1,
63
+ "B-BUILDINGNUM": 3,
64
+ "B-CITY": 5,
65
+ "B-CREDITCARDNUMBER": 7,
66
+ "B-DATE": 9,
67
+ "B-DRIVERLICENSENUM": 11,
68
+ "B-EMAIL": 13,
69
+ "B-GENDER": 15,
70
+ "B-GIVENNAME": 17,
71
+ "B-IDCARDNUM": 19,
72
+ "B-PASSPORTNUM": 21,
73
+ "B-SEX": 23,
74
+ "B-SOCIALNUM": 25,
75
+ "B-STREET": 27,
76
+ "B-SURNAME": 29,
77
+ "B-TAXNUM": 31,
78
+ "B-TELEPHONENUM": 33,
79
+ "B-TIME": 35,
80
+ "B-TITLE": 37,
81
+ "B-ZIPCODE": 39,
82
+ "I-AGE": 2,
83
+ "I-BUILDINGNUM": 4,
84
+ "I-CITY": 6,
85
+ "I-CREDITCARDNUMBER": 8,
86
+ "I-DATE": 10,
87
+ "I-DRIVERLICENSENUM": 12,
88
+ "I-EMAIL": 14,
89
+ "I-GENDER": 16,
90
+ "I-GIVENNAME": 18,
91
+ "I-IDCARDNUM": 20,
92
+ "I-PASSPORTNUM": 22,
93
+ "I-SEX": 24,
94
+ "I-SOCIALNUM": 26,
95
+ "I-STREET": 28,
96
+ "I-SURNAME": 30,
97
+ "I-TAXNUM": 32,
98
+ "I-TELEPHONENUM": 34,
99
+ "I-TIME": 36,
100
+ "I-TITLE": 38,
101
+ "I-ZIPCODE": 40,
102
+ "O": 0
103
+ },
104
+ "layer_norm_eps": 1e-12,
105
+ "max_position_embeddings": 512,
106
+ "model_type": "bert",
107
+ "num_attention_heads": 12,
108
+ "num_hidden_layers": 12,
109
+ "pad_token_id": 0,
110
+ "pooler_fc_size": 768,
111
+ "pooler_num_attention_heads": 12,
112
+ "pooler_num_fc_layers": 3,
113
+ "pooler_size_per_head": 128,
114
+ "pooler_type": "first_token_transform",
115
+ "tie_word_embeddings": true,
116
+ "transformers_version": "5.3.0",
117
+ "type_vocab_size": 2,
118
+ "use_cache": false,
119
+ "vocab_size": 119547
120
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:587279cd5e3390f0868f9816bc510b9ae5d4ad541baa29ef9216740f8beb0dc4
3
+ size 709200844
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e80171abb0a3983948cda7e46f978d8d9bd4e018f68bd480adf3a65e30fef9e6
3
+ size 1418524683
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108380ea659d4b17f4e60ec5cb9d72013e03ee612f7fca615f44a9ba1018d01b
3
+ size 14645
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33f2c548bf6fd4f4d27bc6e2873e91d0cc7f4898c96581b0486972c7500c96df
3
+ size 1465
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "is_local": false,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/trainer_state.json ADDED
@@ -0,0 +1,667 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 43515,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.03447087211306446,
14
+ "grad_norm": 1.2918020486831665,
15
+ "learning_rate": 1.9770653797541078e-05,
16
+ "loss": 0.21463787841796875,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.06894174422612892,
21
+ "grad_norm": 1.4572890996932983,
22
+ "learning_rate": 1.9540847983453983e-05,
23
+ "loss": 0.06676451110839844,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.10341261633919338,
28
+ "grad_norm": 0.7736029624938965,
29
+ "learning_rate": 1.9311042169366887e-05,
30
+ "loss": 0.05756806182861328,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.13788348845225784,
35
+ "grad_norm": 0.5729684829711914,
36
+ "learning_rate": 1.9081236355279792e-05,
37
+ "loss": 0.05027279663085937,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.1723543605653223,
42
+ "grad_norm": 0.6462647914886475,
43
+ "learning_rate": 1.8851430541192693e-05,
44
+ "loss": 0.04522799301147461,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.20682523267838676,
49
+ "grad_norm": 1.5296399593353271,
50
+ "learning_rate": 1.8621624727105598e-05,
51
+ "loss": 0.04301478195190429,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.24129610479145122,
56
+ "grad_norm": 0.7765299081802368,
57
+ "learning_rate": 1.83918189130185e-05,
58
+ "loss": 0.041502281188964846,
59
+ "step": 3500
60
+ },
61
+ {
62
+ "epoch": 0.2757669769045157,
63
+ "grad_norm": 0.44941186904907227,
64
+ "learning_rate": 1.8162013098931404e-05,
65
+ "loss": 0.036275325775146486,
66
+ "step": 4000
67
+ },
68
+ {
69
+ "epoch": 0.31023784901758017,
70
+ "grad_norm": 0.6415440440177917,
71
+ "learning_rate": 1.793220728484431e-05,
72
+ "loss": 0.038519672393798825,
73
+ "step": 4500
74
+ },
75
+ {
76
+ "epoch": 0.3447087211306446,
77
+ "grad_norm": 1.7180463075637817,
78
+ "learning_rate": 1.770240147075721e-05,
79
+ "loss": 0.03590824508666992,
80
+ "step": 5000
81
+ },
82
+ {
83
+ "epoch": 0.3791795932437091,
84
+ "grad_norm": 1.706147313117981,
85
+ "learning_rate": 1.7472595656670115e-05,
86
+ "loss": 0.03493925476074219,
87
+ "step": 5500
88
+ },
89
+ {
90
+ "epoch": 0.4136504653567735,
91
+ "grad_norm": 0.20170536637306213,
92
+ "learning_rate": 1.724278984258302e-05,
93
+ "loss": 0.03378963088989258,
94
+ "step": 6000
95
+ },
96
+ {
97
+ "epoch": 0.448121337469838,
98
+ "grad_norm": 0.32674115896224976,
99
+ "learning_rate": 1.7012984028495924e-05,
100
+ "loss": 0.031928113937377926,
101
+ "step": 6500
102
+ },
103
+ {
104
+ "epoch": 0.48259220958290244,
105
+ "grad_norm": 0.3478796184062958,
106
+ "learning_rate": 1.6783178214408825e-05,
107
+ "loss": 0.033053131103515626,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 0.5170630816959669,
112
+ "grad_norm": 0.6997560858726501,
113
+ "learning_rate": 1.655337240032173e-05,
114
+ "loss": 0.03176578521728515,
115
+ "step": 7500
116
+ },
117
+ {
118
+ "epoch": 0.5515339538090314,
119
+ "grad_norm": 0.34584298729896545,
120
+ "learning_rate": 1.6323566586234634e-05,
121
+ "loss": 0.03126089859008789,
122
+ "step": 8000
123
+ },
124
+ {
125
+ "epoch": 0.5860048259220958,
126
+ "grad_norm": 0.7542563080787659,
127
+ "learning_rate": 1.609376077214754e-05,
128
+ "loss": 0.030911777496337892,
129
+ "step": 8500
130
+ },
131
+ {
132
+ "epoch": 0.6204756980351603,
133
+ "grad_norm": 0.7195969223976135,
134
+ "learning_rate": 1.586395495806044e-05,
135
+ "loss": 0.029189849853515627,
136
+ "step": 9000
137
+ },
138
+ {
139
+ "epoch": 0.6549465701482248,
140
+ "grad_norm": 1.3274503946304321,
141
+ "learning_rate": 1.563414914397334e-05,
142
+ "loss": 0.03025291061401367,
143
+ "step": 9500
144
+ },
145
+ {
146
+ "epoch": 0.6894174422612892,
147
+ "grad_norm": 0.6382108926773071,
148
+ "learning_rate": 1.5404343329886246e-05,
149
+ "loss": 0.028342636108398437,
150
+ "step": 10000
151
+ },
152
+ {
153
+ "epoch": 0.7238883143743536,
154
+ "grad_norm": 0.2607400715351105,
155
+ "learning_rate": 1.5174537515799151e-05,
156
+ "loss": 0.028958717346191407,
157
+ "step": 10500
158
+ },
159
+ {
160
+ "epoch": 0.7583591864874182,
161
+ "grad_norm": 1.1773029565811157,
162
+ "learning_rate": 1.4944731701712054e-05,
163
+ "loss": 0.027841657638549806,
164
+ "step": 11000
165
+ },
166
+ {
167
+ "epoch": 0.7928300586004826,
168
+ "grad_norm": 1.8503401279449463,
169
+ "learning_rate": 1.4714925887624959e-05,
170
+ "loss": 0.02816273307800293,
171
+ "step": 11500
172
+ },
173
+ {
174
+ "epoch": 0.827300930713547,
175
+ "grad_norm": 1.0574482679367065,
176
+ "learning_rate": 1.4485120073537862e-05,
177
+ "loss": 0.027129316329956056,
178
+ "step": 12000
179
+ },
180
+ {
181
+ "epoch": 0.8617718028266115,
182
+ "grad_norm": 0.2772712707519531,
183
+ "learning_rate": 1.4255314259450766e-05,
184
+ "loss": 0.02708342361450195,
185
+ "step": 12500
186
+ },
187
+ {
188
+ "epoch": 0.896242674939676,
189
+ "grad_norm": 0.20870168507099152,
190
+ "learning_rate": 1.402550844536367e-05,
191
+ "loss": 0.024648666381835938,
192
+ "step": 13000
193
+ },
194
+ {
195
+ "epoch": 0.9307135470527405,
196
+ "grad_norm": 0.5534266829490662,
197
+ "learning_rate": 1.3795702631276572e-05,
198
+ "loss": 0.026792591094970702,
199
+ "step": 13500
200
+ },
201
+ {
202
+ "epoch": 0.9651844191658049,
203
+ "grad_norm": 0.5170403718948364,
204
+ "learning_rate": 1.3565896817189477e-05,
205
+ "loss": 0.02854232597351074,
206
+ "step": 14000
207
+ },
208
+ {
209
+ "epoch": 0.9996552912788693,
210
+ "grad_norm": 0.27890703082084656,
211
+ "learning_rate": 1.333609100310238e-05,
212
+ "loss": 0.026233253479003907,
213
+ "step": 14500
214
+ },
215
+ {
216
+ "epoch": 1.0,
217
+ "eval_loss": 0.02844008058309555,
218
+ "eval_runtime": 294.9235,
219
+ "eval_samples_per_second": 393.583,
220
+ "eval_steps_per_second": 49.199,
221
+ "step": 14505
222
+ },
223
+ {
224
+ "epoch": 1.0341261633919339,
225
+ "grad_norm": 1.2330853939056396,
226
+ "learning_rate": 1.3106285189015284e-05,
227
+ "loss": 0.020723657608032228,
228
+ "step": 15000
229
+ },
230
+ {
231
+ "epoch": 1.0685970355049983,
232
+ "grad_norm": 0.3847709596157074,
233
+ "learning_rate": 1.2876479374928186e-05,
234
+ "loss": 0.02110268211364746,
235
+ "step": 15500
236
+ },
237
+ {
238
+ "epoch": 1.1030679076180627,
239
+ "grad_norm": 0.5638656616210938,
240
+ "learning_rate": 1.264667356084109e-05,
241
+ "loss": 0.021719074249267577,
242
+ "step": 16000
243
+ },
244
+ {
245
+ "epoch": 1.1375387797311272,
246
+ "grad_norm": 0.25548455119132996,
247
+ "learning_rate": 1.2416867746753993e-05,
248
+ "loss": 0.020617319107055665,
249
+ "step": 16500
250
+ },
251
+ {
252
+ "epoch": 1.1720096518441916,
253
+ "grad_norm": 0.2979449927806854,
254
+ "learning_rate": 1.2187061932666896e-05,
255
+ "loss": 0.021886722564697265,
256
+ "step": 17000
257
+ },
258
+ {
259
+ "epoch": 1.206480523957256,
260
+ "grad_norm": 0.18911246955394745,
261
+ "learning_rate": 1.1957256118579801e-05,
262
+ "loss": 0.020644411087036133,
263
+ "step": 17500
264
+ },
265
+ {
266
+ "epoch": 1.2409513960703205,
267
+ "grad_norm": 0.3482404351234436,
268
+ "learning_rate": 1.1727450304492704e-05,
269
+ "loss": 0.021316097259521483,
270
+ "step": 18000
271
+ },
272
+ {
273
+ "epoch": 1.275422268183385,
274
+ "grad_norm": 0.570595383644104,
275
+ "learning_rate": 1.1497644490405609e-05,
276
+ "loss": 0.02059372901916504,
277
+ "step": 18500
278
+ },
279
+ {
280
+ "epoch": 1.3098931402964495,
281
+ "grad_norm": 0.2911337614059448,
282
+ "learning_rate": 1.1267838676318512e-05,
283
+ "loss": 0.02033697509765625,
284
+ "step": 19000
285
+ },
286
+ {
287
+ "epoch": 1.344364012409514,
288
+ "grad_norm": 0.426343709230423,
289
+ "learning_rate": 1.1038032862231416e-05,
290
+ "loss": 0.020020153045654295,
291
+ "step": 19500
292
+ },
293
+ {
294
+ "epoch": 1.3788348845225784,
295
+ "grad_norm": 0.5702527761459351,
296
+ "learning_rate": 1.080822704814432e-05,
297
+ "loss": 0.02054849624633789,
298
+ "step": 20000
299
+ },
300
+ {
301
+ "epoch": 1.4133057566356428,
302
+ "grad_norm": 0.38962048292160034,
303
+ "learning_rate": 1.0578421234057224e-05,
304
+ "loss": 0.01970711326599121,
305
+ "step": 20500
306
+ },
307
+ {
308
+ "epoch": 1.4477766287487073,
309
+ "grad_norm": 0.6985688805580139,
310
+ "learning_rate": 1.0348615419970127e-05,
311
+ "loss": 0.020511926651000975,
312
+ "step": 21000
313
+ },
314
+ {
315
+ "epoch": 1.4822475008617717,
316
+ "grad_norm": 0.36666256189346313,
317
+ "learning_rate": 1.0118809605883028e-05,
318
+ "loss": 0.019306087493896486,
319
+ "step": 21500
320
+ },
321
+ {
322
+ "epoch": 1.5167183729748364,
323
+ "grad_norm": 0.3014184534549713,
324
+ "learning_rate": 9.889003791795934e-06,
325
+ "loss": 0.0195778694152832,
326
+ "step": 22000
327
+ },
328
+ {
329
+ "epoch": 1.5511892450879006,
330
+ "grad_norm": 1.1626015901565552,
331
+ "learning_rate": 9.659197977708836e-06,
332
+ "loss": 0.02046485710144043,
333
+ "step": 22500
334
+ },
335
+ {
336
+ "epoch": 1.5856601172009652,
337
+ "grad_norm": 0.2246391624212265,
338
+ "learning_rate": 9.42939216362174e-06,
339
+ "loss": 0.019212373733520506,
340
+ "step": 23000
341
+ },
342
+ {
343
+ "epoch": 1.6201309893140297,
344
+ "grad_norm": 0.7504266500473022,
345
+ "learning_rate": 9.199586349534643e-06,
346
+ "loss": 0.02021506118774414,
347
+ "step": 23500
348
+ },
349
+ {
350
+ "epoch": 1.654601861427094,
351
+ "grad_norm": 0.5071070194244385,
352
+ "learning_rate": 8.969780535447548e-06,
353
+ "loss": 0.01886193084716797,
354
+ "step": 24000
355
+ },
356
+ {
357
+ "epoch": 1.6890727335401585,
358
+ "grad_norm": 0.30986541509628296,
359
+ "learning_rate": 8.739974721360451e-06,
360
+ "loss": 0.01909671974182129,
361
+ "step": 24500
362
+ },
363
+ {
364
+ "epoch": 1.723543605653223,
365
+ "grad_norm": 1.065539836883545,
366
+ "learning_rate": 8.510168907273356e-06,
367
+ "loss": 0.019661422729492187,
368
+ "step": 25000
369
+ },
370
+ {
371
+ "epoch": 1.7580144777662876,
372
+ "grad_norm": 0.3560684621334076,
373
+ "learning_rate": 8.280363093186259e-06,
374
+ "loss": 0.019211139678955078,
375
+ "step": 25500
376
+ },
377
+ {
378
+ "epoch": 1.7924853498793518,
379
+ "grad_norm": 0.264694482088089,
380
+ "learning_rate": 8.050557279099161e-06,
381
+ "loss": 0.018874462127685548,
382
+ "step": 26000
383
+ },
384
+ {
385
+ "epoch": 1.8269562219924165,
386
+ "grad_norm": 0.49240633845329285,
387
+ "learning_rate": 7.820751465012064e-06,
388
+ "loss": 0.01885951042175293,
389
+ "step": 26500
390
+ },
391
+ {
392
+ "epoch": 1.861427094105481,
393
+ "grad_norm": 0.5500588417053223,
394
+ "learning_rate": 7.590945650924969e-06,
395
+ "loss": 0.019577322006225587,
396
+ "step": 27000
397
+ },
398
+ {
399
+ "epoch": 1.8958979662185453,
400
+ "grad_norm": 0.22589892148971558,
401
+ "learning_rate": 7.361139836837873e-06,
402
+ "loss": 0.018141788482666017,
403
+ "step": 27500
404
+ },
405
+ {
406
+ "epoch": 1.9303688383316098,
407
+ "grad_norm": 0.8253551721572876,
408
+ "learning_rate": 7.131334022750777e-06,
409
+ "loss": 0.018833749771118163,
410
+ "step": 28000
411
+ },
412
+ {
413
+ "epoch": 1.9648397104446742,
414
+ "grad_norm": 0.3297463357448578,
415
+ "learning_rate": 6.901528208663679e-06,
416
+ "loss": 0.01926751708984375,
417
+ "step": 28500
418
+ },
419
+ {
420
+ "epoch": 1.9993105825577389,
421
+ "grad_norm": 0.3467053174972534,
422
+ "learning_rate": 6.671722394576583e-06,
423
+ "loss": 0.019064535140991212,
424
+ "step": 29000
425
+ },
426
+ {
427
+ "epoch": 2.0,
428
+ "eval_loss": 0.024620965123176575,
429
+ "eval_runtime": 295.9862,
430
+ "eval_samples_per_second": 392.17,
431
+ "eval_steps_per_second": 49.023,
432
+ "step": 29010
433
+ },
434
+ {
435
+ "epoch": 2.033781454670803,
436
+ "grad_norm": 0.8970562219619751,
437
+ "learning_rate": 6.4419165804894865e-06,
438
+ "loss": 0.014728745460510255,
439
+ "step": 29500
440
+ },
441
+ {
442
+ "epoch": 2.0682523267838677,
443
+ "grad_norm": 0.3860848546028137,
444
+ "learning_rate": 6.21211076640239e-06,
445
+ "loss": 0.014404890060424804,
446
+ "step": 30000
447
+ },
448
+ {
449
+ "epoch": 2.102723198896932,
450
+ "grad_norm": 0.13925616443157196,
451
+ "learning_rate": 5.982304952315294e-06,
452
+ "loss": 0.014260980606079101,
453
+ "step": 30500
454
+ },
455
+ {
456
+ "epoch": 2.1371940710099966,
457
+ "grad_norm": 0.16990777850151062,
458
+ "learning_rate": 5.752499138228198e-06,
459
+ "loss": 0.015296265602111817,
460
+ "step": 31000
461
+ },
462
+ {
463
+ "epoch": 2.1716649431230612,
464
+ "grad_norm": 0.49458077549934387,
465
+ "learning_rate": 5.522693324141101e-06,
466
+ "loss": 0.015134754180908204,
467
+ "step": 31500
468
+ },
469
+ {
470
+ "epoch": 2.2061358152361255,
471
+ "grad_norm": 0.21638576686382294,
472
+ "learning_rate": 5.292887510054005e-06,
473
+ "loss": 0.013905839920043945,
474
+ "step": 32000
475
+ },
476
+ {
477
+ "epoch": 2.24060668734919,
478
+ "grad_norm": 0.1612468808889389,
479
+ "learning_rate": 5.0630816959669085e-06,
480
+ "loss": 0.014391257286071777,
481
+ "step": 32500
482
+ },
483
+ {
484
+ "epoch": 2.2750775594622543,
485
+ "grad_norm": 0.3535423278808594,
486
+ "learning_rate": 4.833275881879812e-06,
487
+ "loss": 0.014306596755981445,
488
+ "step": 33000
489
+ },
490
+ {
491
+ "epoch": 2.309548431575319,
492
+ "grad_norm": 0.19165025651454926,
493
+ "learning_rate": 4.603470067792716e-06,
494
+ "loss": 0.014357097625732422,
495
+ "step": 33500
496
+ },
497
+ {
498
+ "epoch": 2.344019303688383,
499
+ "grad_norm": 0.267946720123291,
500
+ "learning_rate": 4.373664253705619e-06,
501
+ "loss": 0.014472801208496094,
502
+ "step": 34000
503
+ },
504
+ {
505
+ "epoch": 2.378490175801448,
506
+ "grad_norm": 0.5449894070625305,
507
+ "learning_rate": 4.143858439618523e-06,
508
+ "loss": 0.014849995613098145,
509
+ "step": 34500
510
+ },
511
+ {
512
+ "epoch": 2.412961047914512,
513
+ "grad_norm": 0.19216525554656982,
514
+ "learning_rate": 3.914052625531427e-06,
515
+ "loss": 0.014094852447509765,
516
+ "step": 35000
517
+ },
518
+ {
519
+ "epoch": 2.4474319200275767,
520
+ "grad_norm": 0.8792902827262878,
521
+ "learning_rate": 3.6842468114443297e-06,
522
+ "loss": 0.014561490058898926,
523
+ "step": 35500
524
+ },
525
+ {
526
+ "epoch": 2.481902792140641,
527
+ "grad_norm": 0.7690567374229431,
528
+ "learning_rate": 3.4544409973572335e-06,
529
+ "loss": 0.014212981224060058,
530
+ "step": 36000
531
+ },
532
+ {
533
+ "epoch": 2.5163736642537056,
534
+ "grad_norm": 0.1650545746088028,
535
+ "learning_rate": 3.2246351832701373e-06,
536
+ "loss": 0.014171384811401367,
537
+ "step": 36500
538
+ },
539
+ {
540
+ "epoch": 2.55084453636677,
541
+ "grad_norm": 0.24058787524700165,
542
+ "learning_rate": 2.9948293691830407e-06,
543
+ "loss": 0.01404542350769043,
544
+ "step": 37000
545
+ },
546
+ {
547
+ "epoch": 2.5853154084798344,
548
+ "grad_norm": 0.5449634194374084,
549
+ "learning_rate": 2.765023555095944e-06,
550
+ "loss": 0.013911874771118164,
551
+ "step": 37500
552
+ },
553
+ {
554
+ "epoch": 2.619786280592899,
555
+ "grad_norm": 0.22034971415996552,
556
+ "learning_rate": 2.535217741008848e-06,
557
+ "loss": 0.01398819351196289,
558
+ "step": 38000
559
+ },
560
+ {
561
+ "epoch": 2.6542571527059633,
562
+ "grad_norm": 0.7121732234954834,
563
+ "learning_rate": 2.3054119269217513e-06,
564
+ "loss": 0.013505228996276855,
565
+ "step": 38500
566
+ },
567
+ {
568
+ "epoch": 2.688728024819028,
569
+ "grad_norm": 0.2532294988632202,
570
+ "learning_rate": 2.075606112834655e-06,
571
+ "loss": 0.013691899299621581,
572
+ "step": 39000
573
+ },
574
+ {
575
+ "epoch": 2.723198896932092,
576
+ "grad_norm": 0.4231194853782654,
577
+ "learning_rate": 1.8458002987475585e-06,
578
+ "loss": 0.013375855445861817,
579
+ "step": 39500
580
+ },
581
+ {
582
+ "epoch": 2.757669769045157,
583
+ "grad_norm": 0.3513070046901703,
584
+ "learning_rate": 1.6159944846604618e-06,
585
+ "loss": 0.014004316329956055,
586
+ "step": 40000
587
+ },
588
+ {
589
+ "epoch": 2.7921406411582215,
590
+ "grad_norm": 0.4257631301879883,
591
+ "learning_rate": 1.3861886705733656e-06,
592
+ "loss": 0.013363615989685058,
593
+ "step": 40500
594
+ },
595
+ {
596
+ "epoch": 2.8266115132712857,
597
+ "grad_norm": 0.8114942312240601,
598
+ "learning_rate": 1.1563828564862692e-06,
599
+ "loss": 0.013058338165283203,
600
+ "step": 41000
601
+ },
602
+ {
603
+ "epoch": 2.8610823853843503,
604
+ "grad_norm": 0.39406874775886536,
605
+ "learning_rate": 9.265770423991728e-07,
606
+ "loss": 0.013279375076293945,
607
+ "step": 41500
608
+ },
609
+ {
610
+ "epoch": 2.8955532574974145,
611
+ "grad_norm": 0.4345211088657379,
612
+ "learning_rate": 6.967712283120763e-07,
613
+ "loss": 0.012960193634033203,
614
+ "step": 42000
615
+ },
616
+ {
617
+ "epoch": 2.930024129610479,
618
+ "grad_norm": 0.8741120100021362,
619
+ "learning_rate": 4.6696541422497994e-07,
620
+ "loss": 0.013073492050170898,
621
+ "step": 42500
622
+ },
623
+ {
624
+ "epoch": 2.9644950017235434,
625
+ "grad_norm": 0.2609393894672394,
626
+ "learning_rate": 2.371596001378835e-07,
627
+ "loss": 0.01297585678100586,
628
+ "step": 43000
629
+ },
630
+ {
631
+ "epoch": 2.998965873836608,
632
+ "grad_norm": 0.582190990447998,
633
+ "learning_rate": 7.353786050787085e-09,
634
+ "loss": 0.013908697128295898,
635
+ "step": 43500
636
+ },
637
+ {
638
+ "epoch": 3.0,
639
+ "eval_loss": 0.024155057966709137,
640
+ "eval_runtime": 292.2266,
641
+ "eval_samples_per_second": 397.216,
642
+ "eval_steps_per_second": 49.653,
643
+ "step": 43515
644
+ }
645
+ ],
646
+ "logging_steps": 500,
647
+ "max_steps": 43515,
648
+ "num_input_tokens_seen": 0,
649
+ "num_train_epochs": 3,
650
+ "save_steps": 500,
651
+ "stateful_callbacks": {
652
+ "TrainerControl": {
653
+ "args": {
654
+ "should_epoch_stop": false,
655
+ "should_evaluate": false,
656
+ "should_log": false,
657
+ "should_save": true,
658
+ "should_training_stop": true
659
+ },
660
+ "attributes": {}
661
+ }
662
+ },
663
+ "total_flos": 3.639709586105856e+17,
664
+ "train_batch_size": 32,
665
+ "trial_name": null,
666
+ "trial_params": null
667
+ }
open-pii-masking-500k-ai4privacy/TokenBased-BERT/checkpoint-43515/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:083f7dd27ac513d7af866e0a8fbcb3a43e357081778fc03b69d05a60485954c3
3
+ size 5201
open-pii-masking-500k-ai4privacy/TokenBased-BERT/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1dd1302f9fd3c0aee70eb07d38c2dda6398239696a3c93b34d7cdc1f06859988
3
  size 709200844
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:587279cd5e3390f0868f9816bc510b9ae5d4ad541baa29ef9216740f8beb0dc4
3
  size 709200844
open-pii-masking-500k-ai4privacy/TokenBased-BERT/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8119f44807e5e2d6f17933f21f9ce06b8c8fd74063fe8b3427d1478150aa96d6
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:083f7dd27ac513d7af866e0a8fbcb3a43e357081778fc03b69d05a60485954c3
3
  size 5201