hsohn3 commited on
Commit
286e43a
1 Parent(s): 36dfc6b

Training in progress epoch 0

Browse files
Files changed (7) hide show
  1. README.md +52 -0
  2. config.json +25 -0
  3. special_tokens_map.json +7 -0
  4. tf_model.h5 +3 -0
  5. tokenizer.json +457 -0
  6. tokenizer_config.json +13 -0
  7. vocab.json +1 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_keras_callback
5
+ model-index:
6
+ - name: hsohn3/cchs-bert-visit-uncased-wordlevel-block512-batch4-ep100
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
+ probably proofread and complete it, then remove this comment. -->
12
+
13
+ # hsohn3/cchs-bert-visit-uncased-wordlevel-block512-batch4-ep100
14
+
15
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Train Loss: 3.8730
18
+ - Epoch: 0
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
38
+ - training_precision: float32
39
+
40
+ ### Training results
41
+
42
+ | Train Loss | Epoch |
43
+ |:----------:|:-----:|
44
+ | 3.8730 | 0 |
45
+
46
+
47
+ ### Framework versions
48
+
49
+ - Transformers 4.20.1
50
+ - TensorFlow 2.8.2
51
+ - Datasets 2.3.2
52
+ - Tokenizers 0.12.1
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.20.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b69715329b50a8cced21c0d466af2819538cb03dd87554bc2678a8c4a0b7b6e
3
+ size 533687680
tokenizer.json ADDED
@@ -0,0 +1,457 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "[UNK]",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "[PAD]",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "[CLS]",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ },
33
+ {
34
+ "id": 3,
35
+ "content": "[SEP]",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ },
42
+ {
43
+ "id": 4,
44
+ "content": "[MASK]",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
+ }
51
+ ],
52
+ "normalizer": {
53
+ "type": "BertNormalizer",
54
+ "clean_text": true,
55
+ "handle_chinese_chars": true,
56
+ "strip_accents": null,
57
+ "lowercase": true
58
+ },
59
+ "pre_tokenizer": {
60
+ "type": "WhitespaceSplit"
61
+ },
62
+ "post_processor": {
63
+ "type": "TemplateProcessing",
64
+ "single": [
65
+ {
66
+ "SpecialToken": {
67
+ "id": "[CLS]",
68
+ "type_id": 0
69
+ }
70
+ },
71
+ {
72
+ "Sequence": {
73
+ "id": "A",
74
+ "type_id": 0
75
+ }
76
+ },
77
+ {
78
+ "SpecialToken": {
79
+ "id": "[SEP]",
80
+ "type_id": 0
81
+ }
82
+ }
83
+ ],
84
+ "pair": [
85
+ {
86
+ "SpecialToken": {
87
+ "id": "[CLS]",
88
+ "type_id": 0
89
+ }
90
+ },
91
+ {
92
+ "Sequence": {
93
+ "id": "A",
94
+ "type_id": 0
95
+ }
96
+ },
97
+ {
98
+ "SpecialToken": {
99
+ "id": "[SEP]",
100
+ "type_id": 0
101
+ }
102
+ },
103
+ {
104
+ "Sequence": {
105
+ "id": "B",
106
+ "type_id": 1
107
+ }
108
+ },
109
+ {
110
+ "SpecialToken": {
111
+ "id": "[SEP]",
112
+ "type_id": 1
113
+ }
114
+ }
115
+ ],
116
+ "special_tokens": {
117
+ "[CLS]": {
118
+ "id": "[CLS]",
119
+ "ids": [
120
+ 2
121
+ ],
122
+ "tokens": [
123
+ "[CLS]"
124
+ ]
125
+ },
126
+ "[SEP]": {
127
+ "id": "[SEP]",
128
+ "ids": [
129
+ 3
130
+ ],
131
+ "tokens": [
132
+ "[SEP]"
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "decoder": {
138
+ "type": "WordPiece",
139
+ "prefix": "##",
140
+ "cleanup": true
141
+ },
142
+ "model": {
143
+ "type": "WordLevel",
144
+ "vocab": {
145
+ "[UNK]": 0,
146
+ "[PAD]": 1,
147
+ "[CLS]": 2,
148
+ "[SEP]": 3,
149
+ "[MASK]": 4,
150
+ "fio2=h": 5,
151
+ "hr=n": 6,
152
+ "rr=n": 7,
153
+ "spo2=n": 8,
154
+ "sbp=n": 9,
155
+ "map=n": 10,
156
+ "dbp=n": 11,
157
+ "temp=n": 12,
158
+ "rr=h": 13,
159
+ "of=n": 14,
160
+ "fio2=vh": 15,
161
+ "hr=l": 16,
162
+ "hr=h": 17,
163
+ "spo2=vh": 18,
164
+ "dbp=h": 19,
165
+ "sbp=l": 20,
166
+ "sbp=h": 21,
167
+ "map=l": 22,
168
+ "map=h": 23,
169
+ "dbp=l": 24,
170
+ "temp=h": 25,
171
+ "rr=vh": 26,
172
+ "hr=vh": 27,
173
+ "spo2=l": 28,
174
+ "hr=vl": 29,
175
+ "temp=l": 30,
176
+ "map=vl": 31,
177
+ "sbp=vl": 32,
178
+ "spo2=h": 33,
179
+ "dbp=vl": 34,
180
+ "rr=vl": 35,
181
+ "sbp=vh": 36,
182
+ "dbp=vh": 37,
183
+ "map=vh": 38,
184
+ "plt=n": 39,
185
+ "bun=n": 40,
186
+ "cr=n": 41,
187
+ "wbc=n": 42,
188
+ "temp=vh": 43,
189
+ "of=h": 44,
190
+ "spo2=vl": 45,
191
+ "of=vh": 46,
192
+ "temp=vl": 47,
193
+ "cr=l": 48,
194
+ "cr=h": 49,
195
+ "bun=l": 50,
196
+ "wbc=h": 51,
197
+ "bun=h": 52,
198
+ "plt=l": 53,
199
+ "plt=h": 54,
200
+ "wbc=l": 55,
201
+ "wbc=vh": 56,
202
+ "br=n": 57,
203
+ "cr=vh": 58,
204
+ "bun=vh": 59,
205
+ "plt=vl": 60,
206
+ "lt=n": 61,
207
+ "plt=vh": 62,
208
+ "wbc=vl": 63,
209
+ "bun=vl": 64,
210
+ "rr=n>h": 65,
211
+ "rr=h>n": 66,
212
+ "of=vl": 67,
213
+ "cr=vl": 68,
214
+ "hr=h>n": 69,
215
+ "hr=n>l": 70,
216
+ "hr=n>h": 71,
217
+ "hr=l>n": 72,
218
+ "spo2=vh>n": 73,
219
+ "spo2=l>n": 74,
220
+ "bd=n": 75,
221
+ "spo2=n>l": 76,
222
+ "br=l": 77,
223
+ "rr=n>vl": 78,
224
+ "br=h": 79,
225
+ "spo2=n>vh": 80,
226
+ "dbp=h>n": 81,
227
+ "rr=vl>n": 82,
228
+ "spo2=n>h": 83,
229
+ "rr=h>vh": 84,
230
+ "spo2=h>n": 85,
231
+ "rr=vh>h": 86,
232
+ "dbp=l>n": 87,
233
+ "sbp=h>n": 88,
234
+ "dbp=n>l": 89,
235
+ "map=h>n": 90,
236
+ "sbp=n>l": 91,
237
+ "dbp=n>h": 92,
238
+ "hr=vh>h": 93,
239
+ "map=n>l": 94,
240
+ "sbp=l>n": 95,
241
+ "lt=h": 96,
242
+ "sbp=n>h": 97,
243
+ "map=l>n": 98,
244
+ "hr=h>vh": 99,
245
+ "sbp=vl>l": 100,
246
+ "spo2=vl>n": 101,
247
+ "map=n>h": 102,
248
+ "spo2=vh>h": 103,
249
+ "lt=l": 104,
250
+ "sbp=vh>h": 105,
251
+ "dbp=vh>h": 106,
252
+ "rr=vh>n": 107,
253
+ "map=vh>h": 108,
254
+ "rr=n>vh": 109,
255
+ "lt=vh": 110,
256
+ "map=vl>l": 111,
257
+ "hr=vl>l": 112,
258
+ "spo2=h>vh": 113,
259
+ "br=vh": 114,
260
+ "dbp=vl>l": 115,
261
+ "hr=l>vl": 116,
262
+ "fio2=h>vh": 117,
263
+ "dbp=l>vl": 118,
264
+ "spo2=vl>l": 119,
265
+ "sbp=l>vl": 120,
266
+ "map=l>vl": 121,
267
+ "dbp=vh>n": 122,
268
+ "dbp=n>n": 123,
269
+ "hr=n>n": 124,
270
+ "dbp=h>vh": 125,
271
+ "rr=h>vl": 126,
272
+ "spo2=n>vl": 127,
273
+ "dbp=vl>n": 128,
274
+ "sbp=n>n": 129,
275
+ "dbp=n>vl": 130,
276
+ "map=vh>n": 131,
277
+ "map=h>vh": 132,
278
+ "sbp=vh>n": 133,
279
+ "spo2=n>n": 134,
280
+ "sbp=h>vh": 135,
281
+ "rr=vl>h": 136,
282
+ "bd=vl": 137,
283
+ "spo2=l>vl": 138,
284
+ "hr=vh>n": 139,
285
+ "lt=vl": 140,
286
+ "bd=h": 141,
287
+ "fio2=vh>h": 142,
288
+ "map=vl>n": 143,
289
+ "map=n>n": 144,
290
+ "hr=n>vh": 145,
291
+ "hr=vl>n": 146,
292
+ "map=n>vl": 147,
293
+ "dbp=n>vh": 148,
294
+ "sbp=vl>n": 149,
295
+ "sbp=n>vl": 150,
296
+ "hr=n>vl": 151,
297
+ "temp=n>h": 152,
298
+ "br=vl": 153,
299
+ "fio2=vl": 154,
300
+ "temp=l>n": 155,
301
+ "of=h>n": 156,
302
+ "bd=vh": 157,
303
+ "of=n>h": 158,
304
+ "temp=n>l": 159,
305
+ "rr=n>n": 160,
306
+ "temp=h>n": 161,
307
+ "map=n>vh": 162,
308
+ "hr=h>h": 163,
309
+ "sbp=n>vh": 164,
310
+ "hr=l>l": 165,
311
+ "dbp=h>h": 166,
312
+ "rr=h>h": 167,
313
+ "rr=vl>vh": 168,
314
+ "spo2=vh>vh": 169,
315
+ "rr=vh>vl": 170,
316
+ "sbp=h>h": 171,
317
+ "dbp=l>l": 172,
318
+ "spo2=vh>l": 173,
319
+ "rr=vh>vh": 174,
320
+ "sbp=l>l": 175,
321
+ "map=h>h": 176,
322
+ "spo2=vl>vh": 177,
323
+ "map=l>l": 178,
324
+ "sr=n": 179,
325
+ "dbp=h>l": 180,
326
+ "spo2=l>vh": 181,
327
+ "dbp=vl>vl": 182,
328
+ "of=vh>h": 183,
329
+ "spo2=h>l": 184,
330
+ "crp=n": 185,
331
+ "of=h>vh": 186,
332
+ "map=vl>vl": 187,
333
+ "sbp=vl>vl": 188,
334
+ "temp=h>vh": 189,
335
+ "dbp=vh>vh": 190,
336
+ "hr=vh>vh": 191,
337
+ "spo2=l>h": 192,
338
+ "hr=vl>vl": 193,
339
+ "temp=vh>h": 194,
340
+ "dbp=l>h": 195,
341
+ "of=n>vh": 196,
342
+ "of=n>vl": 197,
343
+ "rr=vl>vl": 198,
344
+ "hr=h>l": 199,
345
+ "spo2=vl>h": 200,
346
+ "sbp=vh>vh": 201,
347
+ "temp=vh>n": 202,
348
+ "of=vh>n": 203,
349
+ "temp=n>vh": 204,
350
+ "spo2=l>l": 205,
351
+ "spo2=h>h": 206,
352
+ "hr=l>h": 207,
353
+ "temp=vl>n": 208,
354
+ "spo2=vh>vl": 209,
355
+ "temp=n>vl": 210,
356
+ "dbp=vh>l": 211,
357
+ "sbp=h>l": 212,
358
+ "map=vh>vh": 213,
359
+ "spo2=h>vl": 214,
360
+ "of=vl>n": 215,
361
+ "temp=vl>l": 216,
362
+ "dbp=vh>vl": 217,
363
+ "temp=l>vl": 218,
364
+ "sr=l": 219,
365
+ "map=h>l": 220,
366
+ "temp=l>h": 221,
367
+ "dbp=h>vl": 222,
368
+ "map=vh>l": 223,
369
+ "dbp=vl>h": 224,
370
+ "hr=vh>l": 225,
371
+ "map=h>vl": 226,
372
+ "hr=vl>h": 227,
373
+ "map=l>h": 228,
374
+ "dbp=vl>vh": 229,
375
+ "sr=h": 230,
376
+ "sr=vl": 231,
377
+ "temp=n>n": 232,
378
+ "dbp=l>vh": 233,
379
+ "crp=l": 234,
380
+ "hr=l>vh": 235,
381
+ "pct=n": 236,
382
+ "sbp=h>vl": 237,
383
+ "sbp=vh>l": 238,
384
+ "spo2=vl>vl": 239,
385
+ "hr=h>vl": 240,
386
+ "hr=vh>vl": 241,
387
+ "map=vh>vl": 242,
388
+ "sbp=vl>h": 243,
389
+ "crp=vl": 244,
390
+ "sbp=l>h": 245,
391
+ "temp=h>l": 246,
392
+ "hr=vl>vh": 247,
393
+ "sbp=l>vh": 248,
394
+ "crp=h": 249,
395
+ "map=l>vh": 250,
396
+ "map=vl>h": 251,
397
+ "map=vl>vh": 252,
398
+ "sr=vh": 253,
399
+ "temp=vh>l": 254,
400
+ "sbp=vh>vl": 255,
401
+ "temp=h>h": 256,
402
+ "temp=l>vh": 257,
403
+ "temp=vl>h": 258,
404
+ "pct=l": 259,
405
+ "crp=vh": 260,
406
+ "pct=vh": 261,
407
+ "sbp=vl>vh": 262,
408
+ "of=h>h": 263,
409
+ "fio2=h>vl": 264,
410
+ "fio2=vh>vh": 265,
411
+ "fio2=vl>h": 266,
412
+ "of=n>n": 267,
413
+ "pct=h": 268,
414
+ "fio2=h>h": 269,
415
+ "temp=h>vl": 270,
416
+ "temp=l>l": 271,
417
+ "fio2=vh>vl": 272,
418
+ "temp=vh>vh": 273,
419
+ "lt=h>n": 274,
420
+ "lt=l>n": 275,
421
+ "of=h>vl": 276,
422
+ "temp=vl>vh": 277,
423
+ "fio2=vl>vh": 278,
424
+ "lt=n>h": 279,
425
+ "lt=vl>l": 280,
426
+ "lt=h>vh": 281,
427
+ "lt=vh>h": 282,
428
+ "of=vl>h": 283,
429
+ "br=n>l": 284,
430
+ "bun=h>n": 285,
431
+ "lt=vl>n": 286,
432
+ "of=vh>vh": 287,
433
+ "plt=l>vl": 288,
434
+ "plt=n>h": 289,
435
+ "temp=vh>vl": 290,
436
+ "bd=n>vl": 291,
437
+ "cr=h>n": 292,
438
+ "cr=l>n": 293,
439
+ "cr=n>h": 294,
440
+ "cr=vl>l": 295,
441
+ "lt=h>l": 296,
442
+ "lt=n>l": 297,
443
+ "of=vh>vl": 298,
444
+ "of=vl>vh": 299,
445
+ "plt=n>l": 300,
446
+ "plt=vh>h": 301,
447
+ "plt=vh>n": 302,
448
+ "plt=vl>l": 303,
449
+ "temp=vl>vl": 304,
450
+ "wbc=h>n": 305,
451
+ "wbc=h>vh": 306,
452
+ "wbc=l>vl": 307,
453
+ "wbc=n>l": 308
454
+ },
455
+ "unk_token": "[UNK]"
456
+ }
457
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "name_or_path": "/content/drive/MyDrive/Experiments/tokenizers/cchs-bert-wordlevel-uncased",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "special_tokens_map_file": "/content/drive/MyDrive/Experiments/tokenizers/cchs-bert-wordlevel-uncased/special_tokens_map.json",
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "BertTokenizer",
12
+ "unk_token": "[UNK]"
13
+ }
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[UNK]":0,"[PAD]":1,"[CLS]":2,"[SEP]":3,"[MASK]":4,"fio2=h":5,"hr=n":6,"rr=n":7,"spo2=n":8,"sbp=n":9,"map=n":10,"dbp=n":11,"temp=n":12,"rr=h":13,"of=n":14,"fio2=vh":15,"hr=l":16,"hr=h":17,"spo2=vh":18,"dbp=h":19,"sbp=l":20,"sbp=h":21,"map=l":22,"map=h":23,"dbp=l":24,"temp=h":25,"rr=vh":26,"hr=vh":27,"spo2=l":28,"hr=vl":29,"temp=l":30,"map=vl":31,"sbp=vl":32,"spo2=h":33,"dbp=vl":34,"rr=vl":35,"sbp=vh":36,"dbp=vh":37,"map=vh":38,"plt=n":39,"bun=n":40,"cr=n":41,"wbc=n":42,"temp=vh":43,"of=h":44,"spo2=vl":45,"of=vh":46,"temp=vl":47,"cr=l":48,"cr=h":49,"bun=l":50,"wbc=h":51,"bun=h":52,"plt=l":53,"plt=h":54,"wbc=l":55,"wbc=vh":56,"br=n":57,"cr=vh":58,"bun=vh":59,"plt=vl":60,"lt=n":61,"plt=vh":62,"wbc=vl":63,"bun=vl":64,"rr=n>h":65,"rr=h>n":66,"of=vl":67,"cr=vl":68,"hr=h>n":69,"hr=n>l":70,"hr=n>h":71,"hr=l>n":72,"spo2=vh>n":73,"spo2=l>n":74,"bd=n":75,"spo2=n>l":76,"br=l":77,"rr=n>vl":78,"br=h":79,"spo2=n>vh":80,"dbp=h>n":81,"rr=vl>n":82,"spo2=n>h":83,"rr=h>vh":84,"spo2=h>n":85,"rr=vh>h":86,"dbp=l>n":87,"sbp=h>n":88,"dbp=n>l":89,"map=h>n":90,"sbp=n>l":91,"dbp=n>h":92,"hr=vh>h":93,"map=n>l":94,"sbp=l>n":95,"lt=h":96,"sbp=n>h":97,"map=l>n":98,"hr=h>vh":99,"sbp=vl>l":100,"spo2=vl>n":101,"map=n>h":102,"spo2=vh>h":103,"lt=l":104,"sbp=vh>h":105,"dbp=vh>h":106,"rr=vh>n":107,"map=vh>h":108,"rr=n>vh":109,"lt=vh":110,"map=vl>l":111,"hr=vl>l":112,"spo2=h>vh":113,"br=vh":114,"dbp=vl>l":115,"hr=l>vl":116,"fio2=h>vh":117,"dbp=l>vl":118,"spo2=vl>l":119,"sbp=l>vl":120,"map=l>vl":121,"dbp=vh>n":122,"dbp=n>n":123,"hr=n>n":124,"dbp=h>vh":125,"rr=h>vl":126,"spo2=n>vl":127,"dbp=vl>n":128,"sbp=n>n":129,"dbp=n>vl":130,"map=vh>n":131,"map=h>vh":132,"sbp=vh>n":133,"spo2=n>n":134,"sbp=h>vh":135,"rr=vl>h":136,"bd=vl":137,"spo2=l>vl":138,"hr=vh>n":139,"lt=vl":140,"bd=h":141,"fio2=vh>h":142,"map=vl>n":143,"map=n>n":144,"hr=n>vh":145,"hr=vl>n":146,"map=n>vl":147,"dbp=n>vh":148,"sbp=vl>n":149,"sbp=n>vl":150,"hr=n>vl":151,"temp=n>h":152,"br=vl":153,"fio2=vl":154,"temp=l>n":155,"of=h>n":156,"bd=vh":157,"of=n>h":158,"temp=n>l":159,"rr=n>n":160,"temp=h>n":161,"map=n>vh":162,"hr=h>h":163,"sbp=n>vh":164,"hr=l>l":165,"dbp=h>h":166,"rr=h>h":167,"rr=vl>vh":168,"spo2=vh>vh":169,"rr=vh>vl":170,"sbp=h>h":171,"dbp=l>l":172,"spo2=vh>l":173,"rr=vh>vh":174,"sbp=l>l":175,"map=h>h":176,"spo2=vl>vh":177,"map=l>l":178,"sr=n":179,"dbp=h>l":180,"spo2=l>vh":181,"dbp=vl>vl":182,"of=vh>h":183,"spo2=h>l":184,"crp=n":185,"of=h>vh":186,"map=vl>vl":187,"sbp=vl>vl":188,"temp=h>vh":189,"dbp=vh>vh":190,"hr=vh>vh":191,"spo2=l>h":192,"hr=vl>vl":193,"temp=vh>h":194,"dbp=l>h":195,"of=n>vh":196,"of=n>vl":197,"rr=vl>vl":198,"hr=h>l":199,"spo2=vl>h":200,"sbp=vh>vh":201,"temp=vh>n":202,"of=vh>n":203,"temp=n>vh":204,"spo2=l>l":205,"spo2=h>h":206,"hr=l>h":207,"temp=vl>n":208,"spo2=vh>vl":209,"temp=n>vl":210,"dbp=vh>l":211,"sbp=h>l":212,"map=vh>vh":213,"spo2=h>vl":214,"of=vl>n":215,"temp=vl>l":216,"dbp=vh>vl":217,"temp=l>vl":218,"sr=l":219,"map=h>l":220,"temp=l>h":221,"dbp=h>vl":222,"map=vh>l":223,"dbp=vl>h":224,"hr=vh>l":225,"map=h>vl":226,"hr=vl>h":227,"map=l>h":228,"dbp=vl>vh":229,"sr=h":230,"sr=vl":231,"temp=n>n":232,"dbp=l>vh":233,"crp=l":234,"hr=l>vh":235,"pct=n":236,"sbp=h>vl":237,"sbp=vh>l":238,"spo2=vl>vl":239,"hr=h>vl":240,"hr=vh>vl":241,"map=vh>vl":242,"sbp=vl>h":243,"crp=vl":244,"sbp=l>h":245,"temp=h>l":246,"hr=vl>vh":247,"sbp=l>vh":248,"crp=h":249,"map=l>vh":250,"map=vl>h":251,"map=vl>vh":252,"sr=vh":253,"temp=vh>l":254,"sbp=vh>vl":255,"temp=h>h":256,"temp=l>vh":257,"temp=vl>h":258,"pct=l":259,"crp=vh":260,"pct=vh":261,"sbp=vl>vh":262,"of=h>h":263,"fio2=h>vl":264,"fio2=vh>vh":265,"fio2=vl>h":266,"of=n>n":267,"pct=h":268,"fio2=h>h":269,"temp=h>vl":270,"temp=l>l":271,"fio2=vh>vl":272,"temp=vh>vh":273,"lt=h>n":274,"lt=l>n":275,"of=h>vl":276,"temp=vl>vh":277,"fio2=vl>vh":278,"lt=n>h":279,"lt=vl>l":280,"lt=h>vh":281,"lt=vh>h":282,"of=vl>h":283,"br=n>l":284,"bun=h>n":285,"lt=vl>n":286,"of=vh>vh":287,"plt=l>vl":288,"plt=n>h":289,"temp=vh>vl":290,"bd=n>vl":291,"cr=h>n":292,"cr=l>n":293,"cr=n>h":294,"cr=vl>l":295,"lt=h>l":296,"lt=n>l":297,"of=vh>vl":298,"of=vl>vh":299,"plt=n>l":300,"plt=vh>h":301,"plt=vh>n":302,"plt=vl>l":303,"temp=vl>vl":304,"wbc=h>n":305,"wbc=h>vh":306,"wbc=l>vl":307,"wbc=n>l":308}