hsohn3 commited on
Commit
33579ce
1 Parent(s): f156eee

Training in progress epoch 0

Browse files
Files changed (7) hide show
  1. README.md +52 -0
  2. config.json +25 -0
  3. special_tokens_map.json +7 -0
  4. tf_model.h5 +3 -0
  5. tokenizer.json +504 -0
  6. tokenizer_config.json +13 -0
  7. vocab.json +1 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_keras_callback
5
+ model-index:
6
+ - name: hsohn3/mayo-bert-uncased-wordlevel-block512-ep10
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
+ probably proofread and complete it, then remove this comment. -->
12
+
13
+ # hsohn3/mayo-bert-uncased-wordlevel-block512-ep10
14
+
15
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Train Loss: 3.0885
18
+ - Epoch: 0
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
38
+ - training_precision: float32
39
+
40
+ ### Training results
41
+
42
+ | Train Loss | Epoch |
43
+ |:----------:|:-----:|
44
+ | 3.0885 | 0 |
45
+
46
+
47
+ ### Framework versions
48
+
49
+ - Transformers 4.20.1
50
+ - TensorFlow 2.8.2
51
+ - Datasets 2.3.2
52
+ - Tokenizers 0.12.1
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.20.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dc94a2ad91a9c5216fc60467a1e525ac99a8118ecf72c0435ac568179b7d2e0
3
+ size 533687680
tokenizer.json ADDED
@@ -0,0 +1,504 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "[UNK]",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "[PAD]",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "[CLS]",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ },
33
+ {
34
+ "id": 3,
35
+ "content": "[SEP]",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ },
42
+ {
43
+ "id": 4,
44
+ "content": "[MASK]",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
+ }
51
+ ],
52
+ "normalizer": {
53
+ "type": "BertNormalizer",
54
+ "clean_text": true,
55
+ "handle_chinese_chars": true,
56
+ "strip_accents": null,
57
+ "lowercase": true
58
+ },
59
+ "pre_tokenizer": {
60
+ "type": "WhitespaceSplit"
61
+ },
62
+ "post_processor": {
63
+ "type": "TemplateProcessing",
64
+ "single": [
65
+ {
66
+ "SpecialToken": {
67
+ "id": "[CLS]",
68
+ "type_id": 0
69
+ }
70
+ },
71
+ {
72
+ "Sequence": {
73
+ "id": "A",
74
+ "type_id": 0
75
+ }
76
+ },
77
+ {
78
+ "SpecialToken": {
79
+ "id": "[SEP]",
80
+ "type_id": 0
81
+ }
82
+ }
83
+ ],
84
+ "pair": [
85
+ {
86
+ "SpecialToken": {
87
+ "id": "[CLS]",
88
+ "type_id": 0
89
+ }
90
+ },
91
+ {
92
+ "Sequence": {
93
+ "id": "A",
94
+ "type_id": 0
95
+ }
96
+ },
97
+ {
98
+ "SpecialToken": {
99
+ "id": "[SEP]",
100
+ "type_id": 0
101
+ }
102
+ },
103
+ {
104
+ "Sequence": {
105
+ "id": "B",
106
+ "type_id": 1
107
+ }
108
+ },
109
+ {
110
+ "SpecialToken": {
111
+ "id": "[SEP]",
112
+ "type_id": 1
113
+ }
114
+ }
115
+ ],
116
+ "special_tokens": {
117
+ "[CLS]": {
118
+ "id": "[CLS]",
119
+ "ids": [
120
+ 2
121
+ ],
122
+ "tokens": [
123
+ "[CLS]"
124
+ ]
125
+ },
126
+ "[SEP]": {
127
+ "id": "[SEP]",
128
+ "ids": [
129
+ 3
130
+ ],
131
+ "tokens": [
132
+ "[SEP]"
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "decoder": {
138
+ "type": "WordPiece",
139
+ "prefix": "##",
140
+ "cleanup": true
141
+ },
142
+ "model": {
143
+ "type": "WordLevel",
144
+ "vocab": {
145
+ "[UNK]": 0,
146
+ "[PAD]": 1,
147
+ "[CLS]": 2,
148
+ "[SEP]": 3,
149
+ "[MASK]": 4,
150
+ "hr=n": 5,
151
+ "rr=n": 6,
152
+ "fio2=h": 7,
153
+ "spo2=n": 8,
154
+ "sbp=n": 9,
155
+ "map=n": 10,
156
+ "dbp=n": 11,
157
+ "temp=n": 12,
158
+ "fio2=n": 13,
159
+ "hr=h": 14,
160
+ "rr=h": 15,
161
+ "of=n": 16,
162
+ "hr=l": 17,
163
+ "map=h": 18,
164
+ "dbp=h": 19,
165
+ "hr=vh": 20,
166
+ "sbp=h": 21,
167
+ "spo2=vh": 22,
168
+ "temp=h": 23,
169
+ "fio2=vh": 24,
170
+ "dbp=vh": 25,
171
+ "map=vh": 26,
172
+ "rr=vh": 27,
173
+ "spo2=l": 28,
174
+ "sbp=vh": 29,
175
+ "dbp=l": 30,
176
+ "sbp=l": 31,
177
+ "map=l": 32,
178
+ "dbp=vl": 33,
179
+ "map=vl": 34,
180
+ "temp=vh": 35,
181
+ "hr=vl": 36,
182
+ "sbp=vl": 37,
183
+ "plt=n": 38,
184
+ "spo2=h": 39,
185
+ "cr=n": 40,
186
+ "bun=n": 41,
187
+ "spo2=n>n": 42,
188
+ "wbc=n": 43,
189
+ "hr=n>n": 44,
190
+ "temp=l": 45,
191
+ "rr=n>n": 46,
192
+ "of=h": 47,
193
+ "spo2=vl": 48,
194
+ "rr=vl": 49,
195
+ "spo2=n>l": 50,
196
+ "spo2=l>n": 51,
197
+ "map=n>n": 52,
198
+ "dbp=n>n": 53,
199
+ "sbp=n>n": 54,
200
+ "rr=n>h": 55,
201
+ "hr=n>l": 56,
202
+ "of=vh": 57,
203
+ "rr=h>n": 58,
204
+ "hr=l>n": 59,
205
+ "hr=h>n": 60,
206
+ "hr=n>h": 61,
207
+ "spo2=h>n": 62,
208
+ "spo2=n>h": 63,
209
+ "hr=l>l": 64,
210
+ "map=l>n": 65,
211
+ "map=n>l": 66,
212
+ "sbp=h>n": 67,
213
+ "map=h>n": 68,
214
+ "rr=h>h": 69,
215
+ "dbp=l>n": 70,
216
+ "dbp=n>l": 71,
217
+ "dbp=h>n": 72,
218
+ "sbp=l>n": 73,
219
+ "sbp=n>l": 74,
220
+ "spo2=l>l": 75,
221
+ "spo2=vh>vh": 76,
222
+ "hr=h>h": 77,
223
+ "sbp=n>h": 78,
224
+ "map=n>h": 79,
225
+ "dbp=n>h": 80,
226
+ "spo2=vh>n": 81,
227
+ "spo2=n>vh": 82,
228
+ "bun=h": 83,
229
+ "plt=l": 84,
230
+ "wbc=h": 85,
231
+ "temp=vl": 86,
232
+ "spo2=vh>h": 87,
233
+ "spo2=h>vh": 88,
234
+ "bun=l": 89,
235
+ "cr=h": 90,
236
+ "plt=h": 91,
237
+ "spo2=vl>l": 92,
238
+ "wbc=l": 93,
239
+ "spo2=l>vl": 94,
240
+ "rr=l>n": 95,
241
+ "hr=vl>l": 96,
242
+ "hr=l>vl": 97,
243
+ "spo2=vl>n": 98,
244
+ "spo2=h>h": 99,
245
+ "rr=n>l": 100,
246
+ "rr=h>vh": 101,
247
+ "cr=l": 102,
248
+ "of=l": 103,
249
+ "rr=vh>h": 104,
250
+ "spo2=n>vl": 105,
251
+ "wbc=vh": 106,
252
+ "bun=vh": 107,
253
+ "hr=vh>h": 108,
254
+ "sbp=vh>h": 109,
255
+ "br=n": 110,
256
+ "rr=l": 111,
257
+ "rr=vl>n": 112,
258
+ "map=vh>h": 113,
259
+ "hr=h>vh": 114,
260
+ "cr=vh": 115,
261
+ "plt=vl": 116,
262
+ "hr=vl>vl": 117,
263
+ "dbp=l>l": 118,
264
+ "map=l>l": 119,
265
+ "lt=n": 120,
266
+ "dbp=vh>h": 121,
267
+ "rr=n>vl": 122,
268
+ "dbp=vl>l": 123,
269
+ "sbp=h>vh": 124,
270
+ "dbp=vh>n": 125,
271
+ "dbp=l>vl": 126,
272
+ "map=h>vh": 127,
273
+ "plt=vh": 128,
274
+ "map=vl>l": 129,
275
+ "sbp=h>h": 130,
276
+ "map=h>h": 131,
277
+ "dbp=h>vh": 132,
278
+ "sbp=l>l": 133,
279
+ "wbc=vl": 134,
280
+ "map=vh>n": 135,
281
+ "bun=vl": 136,
282
+ "dbp=h>h": 137,
283
+ "sbp=vh>n": 138,
284
+ "sbp=vl>l": 139,
285
+ "spo2=vl>vl": 140,
286
+ "map=l>vl": 141,
287
+ "rr=vh>vh": 142,
288
+ "dbp=n>vh": 143,
289
+ "sbp=l>vl": 144,
290
+ "hr=vl>n": 145,
291
+ "cr=vl": 146,
292
+ "map=vl>n": 147,
293
+ "rr=l>l": 148,
294
+ "sbp=n>vh": 149,
295
+ "map=n>vh": 150,
296
+ "map=n>vl": 151,
297
+ "rr=vl>l": 152,
298
+ "hr=n>vl": 153,
299
+ "dbp=vl>n": 154,
300
+ "dbp=n>vl": 155,
301
+ "hr=vh>vh": 156,
302
+ "sbp=vl>n": 157,
303
+ "rr=vl>vl": 158,
304
+ "dbp=vl>vl": 159,
305
+ "sbp=n>vl": 160,
306
+ "rr=vh>n": 161,
307
+ "rr=n>vh": 162,
308
+ "rr=l>vl": 163,
309
+ "sbp=vh>vh": 164,
310
+ "dbp=vh>vh": 165,
311
+ "map=vh>vh": 166,
312
+ "map=vl>vl": 167,
313
+ "temp=n>h": 168,
314
+ "sbp=vl>vl": 169,
315
+ "of=vl": 170,
316
+ "fio2=n>h": 171,
317
+ "temp=l>n": 172,
318
+ "bd=n": 173,
319
+ "br=h": 174,
320
+ "hr=n>vh": 175,
321
+ "hr=vh>n": 176,
322
+ "lt=h": 177,
323
+ "temp=h>n": 178,
324
+ "br=l": 179,
325
+ "spo2=vh>l": 180,
326
+ "spo2=h>l": 181,
327
+ "lt=l": 182,
328
+ "spo2=l>vh": 183,
329
+ "spo2=vl>vh": 184,
330
+ "spo2=l>h": 185,
331
+ "lt=vh": 186,
332
+ "temp=h>vh": 187,
333
+ "temp=n>l": 188,
334
+ "spo2=vh>vl": 189,
335
+ "br=vh": 190,
336
+ "rr=h>vl": 191,
337
+ "rr=vl>h": 192,
338
+ "map=h>l": 193,
339
+ "dbp=h>l": 194,
340
+ "spo2=vl>h": 195,
341
+ "temp=vh>h": 196,
342
+ "sbp=h>l": 197,
343
+ "temp=vl>l": 198,
344
+ "map=l>h": 199,
345
+ "rr=h>l": 200,
346
+ "rr=l>h": 201,
347
+ "lt=vl": 202,
348
+ "dbp=l>h": 203,
349
+ "spo2=h>vl": 204,
350
+ "hr=l>h": 205,
351
+ "br=vl": 206,
352
+ "dbp=vh>l": 207,
353
+ "temp=n>n": 208,
354
+ "sbp=l>h": 209,
355
+ "map=vh>l": 210,
356
+ "rr=vh>vl": 211,
357
+ "temp=l>vl": 212,
358
+ "rr=vl>vh": 213,
359
+ "hr=h>l": 214,
360
+ "fio2=h>vh": 215,
361
+ "temp=h>h": 216,
362
+ "crp=n": 217,
363
+ "hr=vl>h": 218,
364
+ "sbp=vh>l": 219,
365
+ "fio2=h>n": 220,
366
+ "temp=vl>n": 221,
367
+ "hr=h>vl": 222,
368
+ "bd=h": 223,
369
+ "map=h>vl": 224,
370
+ "dbp=l>vh": 225,
371
+ "fio2=vh>h": 226,
372
+ "dbp=h>vl": 227,
373
+ "bd=vl": 228,
374
+ "map=l>vh": 229,
375
+ "sr=n": 230,
376
+ "bd=vh": 231,
377
+ "dbp=vh>vl": 232,
378
+ "temp=l>l": 233,
379
+ "map=vh>vl": 234,
380
+ "dbp=vl>h": 235,
381
+ "hr=vl>vh": 236,
382
+ "of=h>n": 237,
383
+ "hr=l>vh": 238,
384
+ "map=vl>h": 239,
385
+ "sbp=l>vh": 240,
386
+ "sbp=h>vl": 241,
387
+ "crp=l": 242,
388
+ "hr=vh>l": 243,
389
+ "temp=n>vh": 244,
390
+ "of=n>h": 245,
391
+ "sbp=vl>h": 246,
392
+ "hr=vh>vl": 247,
393
+ "sbp=vh>vl": 248,
394
+ "temp=vh>n": 249,
395
+ "dbp=vl>vh": 250,
396
+ "fio2=vl": 251,
397
+ "of=vh>h": 252,
398
+ "bd=l": 253,
399
+ "temp=n>vl": 254,
400
+ "rr=vh>l": 255,
401
+ "of=h>vh": 256,
402
+ "map=vl>vh": 257,
403
+ "temp=vh>vh": 258,
404
+ "temp=l>h": 259,
405
+ "crp=h": 260,
406
+ "fio2=n>n": 261,
407
+ "rr=l>vh": 262,
408
+ "sbp=vl>vh": 263,
409
+ "of=n>l": 264,
410
+ "fio2=n>vh": 265,
411
+ "temp=vl>vl": 266,
412
+ "fio2=h>h": 267,
413
+ "sr=l": 268,
414
+ "crp=vh": 269,
415
+ "of=vh>n": 270,
416
+ "of=n>vh": 271,
417
+ "sr=h": 272,
418
+ "sr=vl": 273,
419
+ "temp=vl>h": 274,
420
+ "temp=h>l": 275,
421
+ "fio2=vh>n": 276,
422
+ "of=n>vl": 277,
423
+ "of=l>n": 278,
424
+ "pct=n": 279,
425
+ "sr=vh": 280,
426
+ "of=vl>n": 281,
427
+ "temp=vh>l": 282,
428
+ "temp=h>vl": 283,
429
+ "temp=l>vh": 284,
430
+ "temp=vh>vl": 285,
431
+ "temp=vl>vh": 286,
432
+ "crp=vl": 287,
433
+ "fio2=vh>vh": 288,
434
+ "of=n>n": 289,
435
+ "of=h>h": 290,
436
+ "pct=h": 291,
437
+ "pct=vh": 292,
438
+ "pct=l": 293,
439
+ "plt=l>vl": 294,
440
+ "plt=vl>l": 295,
441
+ "fio2=h>vl": 296,
442
+ "fio2=vl>h": 297,
443
+ "of=h>l": 298,
444
+ "of=vh>vh": 299,
445
+ "fio2=vh>vl": 300,
446
+ "of=l>l": 301,
447
+ "lt=h>n": 302,
448
+ "lt=l>n": 303,
449
+ "lt=n>h": 304,
450
+ "of=h>vl": 305,
451
+ "of=l>vl": 306,
452
+ "plt=n>l": 307,
453
+ "lt=vl>l": 308,
454
+ "plt=l>n": 309,
455
+ "plt=vl>n": 310,
456
+ "fio2=vl>vh": 311,
457
+ "lt=vh>h": 312,
458
+ "of=vl>h": 313,
459
+ "of=vl>l": 314,
460
+ "cr=n>h": 315,
461
+ "lt=h>vh": 316,
462
+ "of=l>h": 317,
463
+ "of=l>vh": 318,
464
+ "bd=l>n": 319,
465
+ "br=n>l": 320,
466
+ "bun=h>n": 321,
467
+ "bun=l>n": 322,
468
+ "bun=n>h": 323,
469
+ "fio2=n>vl": 324,
470
+ "lt=n>l": 325,
471
+ "lt=vl>n": 326,
472
+ "of=vh>vl": 327,
473
+ "plt=n>h": 328,
474
+ "wbc=h>n": 329,
475
+ "wbc=h>vh": 330,
476
+ "wbc=l>vl": 331,
477
+ "wbc=n>l": 332,
478
+ "bd=n>vl": 333,
479
+ "bd=vl>l": 334,
480
+ "br=n>n": 335,
481
+ "bun=h>vh": 336,
482
+ "bun=n>l": 337,
483
+ "cr=h>n": 338,
484
+ "cr=h>vh": 339,
485
+ "cr=l>n": 340,
486
+ "cr=n>l": 341,
487
+ "cr=n>vl": 342,
488
+ "cr=vl>l": 343,
489
+ "lt=h>l": 344,
490
+ "lt=n>vl": 345,
491
+ "of=vh>l": 346,
492
+ "of=vl>vh": 347,
493
+ "plt=h>n": 348,
494
+ "plt=l>l": 349,
495
+ "plt=vh>h": 350,
496
+ "plt=vh>n": 351,
497
+ "wbc=n>h": 352,
498
+ "wbc=vh>h": 353,
499
+ "wbc=vl>l": 354,
500
+ "wbc=vl>vh": 355
501
+ },
502
+ "unk_token": "[UNK]"
503
+ }
504
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "name_or_path": "/content/drive/MyDrive/Experiments/tokenizers/bert-ehr-wordlevel-uncased",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "special_tokens_map_file": "/content/drive/MyDrive/Experiments/tokenizers/bert-ehr-wordlevel-uncased/special_tokens_map.json",
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "BertTokenizer",
12
+ "unk_token": "[UNK]"
13
+ }
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[UNK]":0,"[PAD]":1,"[CLS]":2,"[SEP]":3,"[MASK]":4,"hr=n":5,"rr=n":6,"fio2=h":7,"spo2=n":8,"sbp=n":9,"map=n":10,"dbp=n":11,"temp=n":12,"fio2=n":13,"hr=h":14,"rr=h":15,"of=n":16,"hr=l":17,"map=h":18,"dbp=h":19,"hr=vh":20,"sbp=h":21,"spo2=vh":22,"temp=h":23,"fio2=vh":24,"dbp=vh":25,"map=vh":26,"rr=vh":27,"spo2=l":28,"sbp=vh":29,"dbp=l":30,"sbp=l":31,"map=l":32,"dbp=vl":33,"map=vl":34,"temp=vh":35,"hr=vl":36,"sbp=vl":37,"plt=n":38,"spo2=h":39,"cr=n":40,"bun=n":41,"spo2=n>n":42,"wbc=n":43,"hr=n>n":44,"temp=l":45,"rr=n>n":46,"of=h":47,"spo2=vl":48,"rr=vl":49,"spo2=n>l":50,"spo2=l>n":51,"map=n>n":52,"dbp=n>n":53,"sbp=n>n":54,"rr=n>h":55,"hr=n>l":56,"of=vh":57,"rr=h>n":58,"hr=l>n":59,"hr=h>n":60,"hr=n>h":61,"spo2=h>n":62,"spo2=n>h":63,"hr=l>l":64,"map=l>n":65,"map=n>l":66,"sbp=h>n":67,"map=h>n":68,"rr=h>h":69,"dbp=l>n":70,"dbp=n>l":71,"dbp=h>n":72,"sbp=l>n":73,"sbp=n>l":74,"spo2=l>l":75,"spo2=vh>vh":76,"hr=h>h":77,"sbp=n>h":78,"map=n>h":79,"dbp=n>h":80,"spo2=vh>n":81,"spo2=n>vh":82,"bun=h":83,"plt=l":84,"wbc=h":85,"temp=vl":86,"spo2=vh>h":87,"spo2=h>vh":88,"bun=l":89,"cr=h":90,"plt=h":91,"spo2=vl>l":92,"wbc=l":93,"spo2=l>vl":94,"rr=l>n":95,"hr=vl>l":96,"hr=l>vl":97,"spo2=vl>n":98,"spo2=h>h":99,"rr=n>l":100,"rr=h>vh":101,"cr=l":102,"of=l":103,"rr=vh>h":104,"spo2=n>vl":105,"wbc=vh":106,"bun=vh":107,"hr=vh>h":108,"sbp=vh>h":109,"br=n":110,"rr=l":111,"rr=vl>n":112,"map=vh>h":113,"hr=h>vh":114,"cr=vh":115,"plt=vl":116,"hr=vl>vl":117,"dbp=l>l":118,"map=l>l":119,"lt=n":120,"dbp=vh>h":121,"rr=n>vl":122,"dbp=vl>l":123,"sbp=h>vh":124,"dbp=vh>n":125,"dbp=l>vl":126,"map=h>vh":127,"plt=vh":128,"map=vl>l":129,"sbp=h>h":130,"map=h>h":131,"dbp=h>vh":132,"sbp=l>l":133,"wbc=vl":134,"map=vh>n":135,"bun=vl":136,"dbp=h>h":137,"sbp=vh>n":138,"sbp=vl>l":139,"spo2=vl>vl":140,"map=l>vl":141,"rr=vh>vh":142,"dbp=n>vh":143,"sbp=l>vl":144,"hr=vl>n":145,"cr=vl":146,"map=vl>n":147,"rr=l>l":148,"sbp=n>vh":149,"map=n>vh":150,"map=n>vl":151,"rr=vl>l":152,"hr=n>vl":153,"dbp=vl>n":154,"dbp=n>vl":155,"hr=vh>vh":156,"sbp=vl>n":157,"rr=vl>vl":158,"dbp=vl>vl":159,"sbp=n>vl":160,"rr=vh>n":161,"rr=n>vh":162,"rr=l>vl":163,"sbp=vh>vh":164,"dbp=vh>vh":165,"map=vh>vh":166,"map=vl>vl":167,"temp=n>h":168,"sbp=vl>vl":169,"of=vl":170,"fio2=n>h":171,"temp=l>n":172,"bd=n":173,"br=h":174,"hr=n>vh":175,"hr=vh>n":176,"lt=h":177,"temp=h>n":178,"br=l":179,"spo2=vh>l":180,"spo2=h>l":181,"lt=l":182,"spo2=l>vh":183,"spo2=vl>vh":184,"spo2=l>h":185,"lt=vh":186,"temp=h>vh":187,"temp=n>l":188,"spo2=vh>vl":189,"br=vh":190,"rr=h>vl":191,"rr=vl>h":192,"map=h>l":193,"dbp=h>l":194,"spo2=vl>h":195,"temp=vh>h":196,"sbp=h>l":197,"temp=vl>l":198,"map=l>h":199,"rr=h>l":200,"rr=l>h":201,"lt=vl":202,"dbp=l>h":203,"spo2=h>vl":204,"hr=l>h":205,"br=vl":206,"dbp=vh>l":207,"temp=n>n":208,"sbp=l>h":209,"map=vh>l":210,"rr=vh>vl":211,"temp=l>vl":212,"rr=vl>vh":213,"hr=h>l":214,"fio2=h>vh":215,"temp=h>h":216,"crp=n":217,"hr=vl>h":218,"sbp=vh>l":219,"fio2=h>n":220,"temp=vl>n":221,"hr=h>vl":222,"bd=h":223,"map=h>vl":224,"dbp=l>vh":225,"fio2=vh>h":226,"dbp=h>vl":227,"bd=vl":228,"map=l>vh":229,"sr=n":230,"bd=vh":231,"dbp=vh>vl":232,"temp=l>l":233,"map=vh>vl":234,"dbp=vl>h":235,"hr=vl>vh":236,"of=h>n":237,"hr=l>vh":238,"map=vl>h":239,"sbp=l>vh":240,"sbp=h>vl":241,"crp=l":242,"hr=vh>l":243,"temp=n>vh":244,"of=n>h":245,"sbp=vl>h":246,"hr=vh>vl":247,"sbp=vh>vl":248,"temp=vh>n":249,"dbp=vl>vh":250,"fio2=vl":251,"of=vh>h":252,"bd=l":253,"temp=n>vl":254,"rr=vh>l":255,"of=h>vh":256,"map=vl>vh":257,"temp=vh>vh":258,"temp=l>h":259,"crp=h":260,"fio2=n>n":261,"rr=l>vh":262,"sbp=vl>vh":263,"of=n>l":264,"fio2=n>vh":265,"temp=vl>vl":266,"fio2=h>h":267,"sr=l":268,"crp=vh":269,"of=vh>n":270,"of=n>vh":271,"sr=h":272,"sr=vl":273,"temp=vl>h":274,"temp=h>l":275,"fio2=vh>n":276,"of=n>vl":277,"of=l>n":278,"pct=n":279,"sr=vh":280,"of=vl>n":281,"temp=vh>l":282,"temp=h>vl":283,"temp=l>vh":284,"temp=vh>vl":285,"temp=vl>vh":286,"crp=vl":287,"fio2=vh>vh":288,"of=n>n":289,"of=h>h":290,"pct=h":291,"pct=vh":292,"pct=l":293,"plt=l>vl":294,"plt=vl>l":295,"fio2=h>vl":296,"fio2=vl>h":297,"of=h>l":298,"of=vh>vh":299,"fio2=vh>vl":300,"of=l>l":301,"lt=h>n":302,"lt=l>n":303,"lt=n>h":304,"of=h>vl":305,"of=l>vl":306,"plt=n>l":307,"lt=vl>l":308,"plt=l>n":309,"plt=vl>n":310,"fio2=vl>vh":311,"lt=vh>h":312,"of=vl>h":313,"of=vl>l":314,"cr=n>h":315,"lt=h>vh":316,"of=l>h":317,"of=l>vh":318,"bd=l>n":319,"br=n>l":320,"bun=h>n":321,"bun=l>n":322,"bun=n>h":323,"fio2=n>vl":324,"lt=n>l":325,"lt=vl>n":326,"of=vh>vl":327,"plt=n>h":328,"wbc=h>n":329,"wbc=h>vh":330,"wbc=l>vl":331,"wbc=n>l":332,"bd=n>vl":333,"bd=vl>l":334,"br=n>n":335,"bun=h>vh":336,"bun=n>l":337,"cr=h>n":338,"cr=h>vh":339,"cr=l>n":340,"cr=n>l":341,"cr=n>vl":342,"cr=vl>l":343,"lt=h>l":344,"lt=n>vl":345,"of=vh>l":346,"of=vl>vh":347,"plt=h>n":348,"plt=l>l":349,"plt=vh>h":350,"plt=vh>n":351,"wbc=n>h":352,"wbc=vh>h":353,"wbc=vl>l":354,"wbc=vl>vh":355}