File size: 72,892 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
{
    "paper_id": "2021",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T01:07:24.057054Z"
    },
    "title": "End-to-end Biomedical Entity Linking with Span-based Dictionary Matching",
    "authors": [
        {
            "first": "Shogo",
            "middle": [],
            "last": "Ujiie",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Nara Institute of Science and Technology \u2665 Megagon Labs",
                "location": {}
            },
            "email": "ujiie@is.naist.jp"
        },
        {
            "first": "Hayate",
            "middle": [],
            "last": "Iso",
            "suffix": "",
            "affiliation": {},
            "email": "hayate@magagon.ai"
        },
        {
            "first": "Shuntaro",
            "middle": [],
            "last": "Yada",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Nara Institute of Science and Technology \u2665 Megagon Labs",
                "location": {}
            },
            "email": "yada-s@is.naist.jp"
        },
        {
            "first": "Shoko",
            "middle": [],
            "last": "Wakamiya",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Nara Institute of Science and Technology \u2665 Megagon Labs",
                "location": {}
            },
            "email": "wakamiya@is.naist.jp"
        },
        {
            "first": "Eiji",
            "middle": [],
            "last": "Aramaki",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Nara Institute of Science and Technology \u2665 Megagon Labs",
                "location": {}
            },
            "email": "aramaki@is.naist.jp"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "Disease name recognition and normalization , which is generally called biomedical entity linking, is a fundamental process in biomedical text mining. Recently, neural joint learning of both tasks has been proposed to utilize the mutual benefits. While this approach achieves high performance, disease concepts that do not appear in the training dataset cannot be accurately predicted. This study introduces a novel end-to-end approach that combines span representations with dictionary-matching features to address this problem. Our model handles unseen concepts by referring to a dictionary while maintaining the performance of neural network-based models, in an end-to-end fashion. Experiments using two major datasets demonstrate that our model achieved competitive results with strong baselines, especially for unseen concepts during training.",
    "pdf_parse": {
        "paper_id": "2021",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "Disease name recognition and normalization , which is generally called biomedical entity linking, is a fundamental process in biomedical text mining. Recently, neural joint learning of both tasks has been proposed to utilize the mutual benefits. While this approach achieves high performance, disease concepts that do not appear in the training dataset cannot be accurately predicted. This study introduces a novel end-to-end approach that combines span representations with dictionary-matching features to address this problem. Our model handles unseen concepts by referring to a dictionary while maintaining the performance of neural network-based models, in an end-to-end fashion. Experiments using two major datasets demonstrate that our model achieved competitive results with strong baselines, especially for unseen concepts during training.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Identifying disease names , which is generally called biomedical entity linking, is the fundamental process of biomedical natural language processing, and it can be utilized in applications such as a literature search system ) and a biomedical relation extraction (Xu et al., 2016 ). The usual system to identify disease names consists of two modules: named entity recognition (NER) and named entity normalization (NEN). NER is the task that recognizes the span of a disease name, from the start position to the end position. NEN is the post-processing of NER, normalizing a disease name into a controlled vocabulary, such as a MeSH or Online Mendelian Inheritance in Man (OMIM).",
                "cite_spans": [
                    {
                        "start": 264,
                        "end": 280,
                        "text": "(Xu et al., 2016",
                        "ref_id": "BIBREF19"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Although most previous studies have developed pipeline systems, in which the NER model first recognizs disease mentions Weber et al., 2020) and the NEN model normalizes the recognized mention (Leaman et al., 2013; Ferr\u00e9 et al., 2020; Xu et al., 2020; Vashishth et al., 2020) , a few approaches employ a joint learning architecture for these tasks Lou et al., 2017) . These joint approaches simultaneously recognize and normalize disease names utilizing their mutual benefits. For example, Leaman et al. (2013) demonstrated that dictionary-matching features, which are commonly used for NEN, are also effective for NER. While these joint learning models achieve high performance for both NER and NEN, they predominately rely on hand-crafted features, which are difficult to construct because of the domain knowledge requirement.",
                "cite_spans": [
                    {
                        "start": 120,
                        "end": 139,
                        "text": "Weber et al., 2020)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 192,
                        "end": 213,
                        "text": "(Leaman et al., 2013;",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 214,
                        "end": 233,
                        "text": "Ferr\u00e9 et al., 2020;",
                        "ref_id": null
                    },
                    {
                        "start": 234,
                        "end": 250,
                        "text": "Xu et al., 2020;",
                        "ref_id": "BIBREF18"
                    },
                    {
                        "start": 251,
                        "end": 274,
                        "text": "Vashishth et al., 2020)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 347,
                        "end": 364,
                        "text": "Lou et al., 2017)",
                        "ref_id": "BIBREF12"
                    },
                    {
                        "start": 489,
                        "end": 509,
                        "text": "Leaman et al. (2013)",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Recently, a neural network (NN)-based model that does not require any hand-crafted features was applied to the joint learning of NER and NEN (Zhao et al., 2019) . NER and NEN were defined as two token-level classification tasks, i.e., their model classified each token into IOB2 tags and concepts, respectively. Although their model achieved the state-of-the-art performance for both NER and NEN, a concept that does not appear in training data (i.e., zero-shot situation) can not be predicted properly.",
                "cite_spans": [
                    {
                        "start": 141,
                        "end": 160,
                        "text": "(Zhao et al., 2019)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "One possible approach to handle this zero-shot situation is utilizing the dictionary-matching features. Suppose that an input sentence \"Classic polyarteritis nodosa is a systemic vasculitis\" is given, where \"polyarteritis nodosa\" is the target entity. Even if it does not appear in the training data, it can be recognized and normalized by referring to a controlled vocabulary that contains \"Polyarteritis Nodosa (MeSH: D010488).\" Combining such looking-up mechanisms with NN-based models, however, is not a trivial task; dictionary matching must be performed at the entity-level, whereas standard NN-based NER and NEN tasks are performed at the token-level (for example, Zhao et al., 2019) .",
                "cite_spans": [
                    {
                        "start": 672,
                        "end": 690,
                        "text": "Zhao et al., 2019)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "To overcome this problem, we propose a novel end-to-end approach for NER and NEN that com- The overview of our model. It combines the dictionary-matching scores with the context score obtained from PubMedBERT. The red boxes are the target span and \"ci\" in the figure is the \"i\"-th concept in the dictionary. bines dictionary-matching features with NN-based models. Based on the span-based model introduced by Lee et al. 2017, our model first computes span representations for all possible spans of the input sentence and then combines the dictionarymatching features with the span representations. Using the score obtained from both features, it directly classifies the disease concept. Thus, our model can handle the zero-shot problem by using dictionary-matching features while maintaining the performance of the NN-based models. Our model is also effective in situations other than the zero-shot condition. Consider the following input sentence: \"We report the case of a patient who developed acute hepatitis,\" where \"hepatitis\" is the target entity that should be normalized to \"drug-induced hepatitis.\" While the longer span \"acute hepatitis\" also appears plausible for standalone NER models, our end-to-end architecture assigns a higher score to the correct shorter span \"hepatitis\" due to the existence of the normalized term (\"drug-induced hepatitis\") in the dictionary.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Through the experiments using two major NER and NEN corpora, we demonstrate that our model achieves competitive results for both corpora. Further analysis illustrates that the dictionarymatching features improve the performance of NEN in the zero-shot and other situations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Our main contributions are twofold: (i) We propose a novel end-to-end model for disease name recognition and normalization that utilizes both NN-based features and dictionary-matching features; (ii) We demonstrate that combining dictionary-matching features with an NN-based model is highly effective for normalization, especially in the zero-shot situations.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Given an input sentence, which is a sequence of words",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Task Definition",
                "sec_num": "2.1"
            },
            {
                "text": "x = {x 1 , x 2 , \u2022 \u2022 \u2022 , x",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Task Definition",
                "sec_num": "2.1"
            },
            {
                "text": "|X| } in the biomedical literature, let us define S as a set of all possible spans, and L as a set of concepts that contains the special label Null for a non-disease span. Our goal is to predict a set of labeled spans",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Task Definition",
                "sec_num": "2.1"
            },
            {
                "text": "y = { i, j, d k } |Y | k=1",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Task Definition",
                "sec_num": "2.1"
            },
            {
                "text": ", where (i, j) \u2208 S is the word index in the sentence, and d \u2208 L is the concept of diseases.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Task Definition",
                "sec_num": "2.1"
            },
            {
                "text": "Our model predicts the concepts for each span based on the score, which is represented by the weighted sum of two factors: the context score score cont obtained from span representations and the dictionary-matching score score dict . Figure 1 illustrates the overall architecture of our model. We denote the score of the span s as follows:",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 234,
                        "end": 242,
                        "text": "Figure 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "score(s, c) = score cont (s, c) + \u03bbscore dict (s, c)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "where c \u2208 L is the candidate concept and \u03bb is the hyperparameter that balances the scores. For the concept prediction, the scores of all possible spans and concepts are calculated, and then the concept with the highest score is selected as the predicted concept for each span as follows:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "y = arg max c\u2208L score(s, c)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "Context score The context score is computed in a similar way to that of Lee et al. (2017) , which is based on the span representations. To compute the representations of each span, the input tokens are first encoded into the token embeddings. We used BioBERT as the encoder, which is a variation of bidirectional encoder representations from transformers (BERT) that is trained on a large amount of biomedical text. Given an input sentence containing T words, we can obtain the contextualized embeddings of each token using BioBERT as follows:",
                "cite_spans": [
                    {
                        "start": 72,
                        "end": 89,
                        "text": "Lee et al. (2017)",
                        "ref_id": "BIBREF9"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "h 1:T = BERT(x 1 , x 2 , \u2022 \u2022 \u2022 , x T )",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "where h 1:T is the input tokens embeddings.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "Span representations are obtained by concatenating several features from the token embeddings:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "g s = [h start(s) , h end(s) ,\u0125 s , \u03c6(s)] g s = GELU(FFNN(g s ))",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "where h start(s) and h end(s) are the start and end token embeddings of the span, respectively; and\u0125 s is the weighted sum of the token embeddings in the span, which is obtained using an attention mechanism (Bahdanau et al., 2015) . \u03c6(i) is the size of span s. These representations g s are then fed into a simple feed-forward NN, FFNN, and a nonlinear function, GELU (Hendrycks and Gimpel, 2016) .",
                "cite_spans": [
                    {
                        "start": 207,
                        "end": 230,
                        "text": "(Bahdanau et al., 2015)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 368,
                        "end": 396,
                        "text": "(Hendrycks and Gimpel, 2016)",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "Given a particular span representation and a candidate concept as the inputs, we formulate the context score as follows:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "score cont (s, c) = g s \u2022 W c",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "where W \u2208 R |L|\u00d7d g is the weight matrix associated with each concept c, and W c represents the weight vector for the concept c.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "Dictionary-matching score We used the cosine similarity of the TF-IDF vectors as the dictionarymatching features. Because there are several synonyms for a concept, we calculated the cosine similarity for all synonyms of the concept and used the maximum cosine similarity as the score for each concept. The TF-IDF is calculated using the character-level n-gram statistics computed for all diseases appearing in the training dataset and controlled vocabulary. For example, given the span \"breast cancer,\" synonyms with high cosine similarity are \"breast cancer (1.0)\" and \"male breast cancer (0.829).\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Model Architecture",
                "sec_num": "2.2"
            },
            {
                "text": "To evaluate our model, we chose two major datasets used in disease name recognition and normalization against a popular controlled vocabulary, MEDIC (Davis et al., 2012) . Both datasets, the National Center for Biotechnology Information Disease (NCBID) corpus (Dogan et al., 2014) and the BioCreative V Chemical Disease Relation (BC5CDR) task corpus (Li et al., 2016) , comprise of PubMed titles and abstracts annotated with disease names and their corresponding normalized term IDs (CUIs). NCBID provides 593 training, 100 development, and 100 test data splits, while BC5CDR evenly divides 1500 data into the three sets. We adopted the same version of MEDIC as TaggerOne used, and that we dismissed non-disease entity annotations contained in BC5CDR.",
                "cite_spans": [
                    {
                        "start": 149,
                        "end": 169,
                        "text": "(Davis et al., 2012)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 260,
                        "end": 280,
                        "text": "(Dogan et al., 2014)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 350,
                        "end": 367,
                        "text": "(Li et al., 2016)",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Experiment 3.1 Datasets",
                "sec_num": "3"
            },
            {
                "text": "We compared several baselines to evaluate our model. DNorm (Leaman et al., 2013) and NormCo (Wright et al., 2019) were used as pipeline models due to their high performance. In addition, we used the pipeline systems consisting of stateof-the-art models: BioBERT for NER and BioSyn (Sung et al., 2020) for NEN.",
                "cite_spans": [
                    {
                        "start": 59,
                        "end": 80,
                        "text": "(Leaman et al., 2013)",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 92,
                        "end": 113,
                        "text": "(Wright et al., 2019)",
                        "ref_id": "BIBREF17"
                    },
                    {
                        "start": 281,
                        "end": 300,
                        "text": "(Sung et al., 2020)",
                        "ref_id": "BIBREF14"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Baseline Models",
                "sec_num": "3.2"
            },
            {
                "text": "TaggerOne and Transition-based model (Lou et al., 2017) are used as joint-learning models. These models outperformed the pipeline models in NCBID and BC5CDR. For the model introduced by Zhao et al. (2019), we cannot reproduce the performance reported by them. Instead, we report the performance of the simple token-level joint learning model based on the BioBERT, which referred as \"joint (token)\".",
                "cite_spans": [
                    {
                        "start": 37,
                        "end": 55,
                        "text": "(Lou et al., 2017)",
                        "ref_id": "BIBREF12"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Baseline Models",
                "sec_num": "3.2"
            },
            {
                "text": "We performed several preprocessing steps: splitting the text into sentences using the NLTK toolkit (Bird et al., 2009) , removing punctuations, and resolving abbreviations using Ab3P (Sohn et al., 2008) , a common abbreviation resolution module. We also merged disease names in each training set into a controlled vocabulary, following the methods of Lou et al. (2017) .",
                "cite_spans": [
                    {
                        "start": 99,
                        "end": 118,
                        "text": "(Bird et al., 2009)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 183,
                        "end": 202,
                        "text": "(Sohn et al., 2008)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 351,
                        "end": 368,
                        "text": "Lou et al. (2017)",
                        "ref_id": "BIBREF12"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Implementation",
                "sec_num": "3.3"
            },
            {
                "text": "For training, we set the learning rate to 5e-5, and mini-batch size to 32. \u03bb was set to 0.9 using the development sets. For BC5CDR, we trained the model using both the training and development sets following . For computational efficiency, we only consider spans with up to 10 words.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Implementation",
                "sec_num": "3.3"
            },
            {
                "text": "We evaluated the recognition performance of our model using micro-F1 at the entity level. We consider the predicted spans as true positive when their spans are identical. Following the previous work (Wright et al., 2019; , the performance of NEN was evaluated using micro-F1 at the abstract level. If a predicted concept was found within the gold standard concepts in the abstract, regardless of its location, it was considered as a true positive. Table 1 illustrates that our model mostly achieved the highest F1-scores in both NER and NEN, except for the NEN in BC5CDR, in which the transition-based model displays its strength as a baseline. The proposed model outperformed the pipeline model of the state-of-the-art models for both tasks, which demonstrates that the improvement is attributed not to the strength of BioBERT but the model architecture, including the endto-end approach and combinations of dictionarymatching features.",
                "cite_spans": [
                    {
                        "start": 199,
                        "end": 220,
                        "text": "(Wright et al., 2019;",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 448,
                        "end": 455,
                        "text": "Table 1",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Evaluation Metrics",
                "sec_num": "3.4"
            },
            {
                "text": "Comparing the model variation results, adding dictionary-matching features improved the performance in NEN. The results clearly suggest that dictionary-matching features are effective for NNbased NEN models.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results & Discussions",
                "sec_num": "4"
            },
            {
                "text": "To analyze the behavior of our model in the zeroshot situation, we investigated the NEN performance on two subsets of both corpora: disease names with concepts that appear in the training data (i.e., standard situation), and disease names with concepts that do not appear in the training data (i.e., the zero-shot situation). Table 2 shows the number of mentions and concepts in each situation. Table 3 displays the results of the zero-shot and standard situation. The proposed model with dictionary-matching features can classify disease concepts in the zero-shot situation, whereas the NN-based classification model cannot normalize the disease names. The results of the standard situation demonstrate that combining dictionary-matching features also improves the performance even when target concepts appear in the training data. This finding implies that an NN-based model can benefit from dictionary-matching features, even if the models can learn from many training data.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 326,
                        "end": 333,
                        "text": "Table 2",
                        "ref_id": "TABREF3"
                    },
                    {
                        "start": 395,
                        "end": 402,
                        "text": "Table 3",
                        "ref_id": "TABREF4"
                    }
                ],
                "eq_spans": [],
                "section": "Contribution of Dictionary-Matching",
                "sec_num": "4.1"
            },
            {
                "text": "We examined 100 randomly sampled sentences to determine the contributions of dictionary-matching features. There are 32 samples in which the models predicted concepts correctly by adding dictionarymatching features. Most of these samples are disease concepts that do not appear in the training set but appear in the dictionary. For example, \"pure red cell aplasis (MeSH: D012010)\" is not in the BC5CDR training set while the MEDIC contains \"Pure Red-Cell Aplasias\" for \"D012010\". In this case, a high dictionary-matching score clearly leads to a correct prediction in the zero-shot situation.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Case study",
                "sec_num": "4.2"
            },
            {
                "text": "In contrast, there are 32 samples in which the dictionary-matching features cause errors. The sources of this error type are typically general disease names in the MEDIC. For example, \"Death (MeSH:D003643)\" is incorrectly predicted as a disease concept in NER. Because these words are also used in the general context, our model overestimated their dictionary-matching scores.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Case study",
                "sec_num": "4.2"
            },
            {
                "text": "Furthermore, in the remaining samples, our model predicted the code properly and the span incorrectly. For example, although \"thoracic hematomyelia\" is labeled as \"MeSH: D020758\" in the BC5CDR test set, our model recognized this as \"hematomyelia.\" In this case, our model mostly relied on the dictionary-matching features and misclassifies the span because 'hematomyelia\" is in the MEDIC but not in the training data.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Case study",
                "sec_num": "4.2"
            },
            {
                "text": "Our model is inferior to the transition-based model for BC5CDR. One possible reason is that the transition-based model utilizes normalized terms that co-occur within a sentence, whereas our model does not. Certain disease names that co-occur within a sentence are strongly useful for normalizing disease names. Although BERT implicitly considers the interaction between disease names via the attention mechanism, a more explicit method is preferable for normalizing diseases.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Limitations",
                "sec_num": "4.3"
            },
            {
                "text": "Another limitation is that our model treats the dictionary entries equally. Because certain terms in the dictionary may also be used for non-disease concepts, such as gene names, we must consider the relative importance of each concept.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Limitations",
                "sec_num": "4.3"
            },
            {
                "text": "We proposed a end-to-end model for disease name recognition and normalization that combines the NN-based model with the dictionary-matching features. Our model achieved highly competitive results for the NCBI disease corpus and BC5CDR corpus, demonstrating that incorporating dictionary-matching features into an NN-based model can improve its performance. Further experiments exhibited that dictionary-matching features enable our model to accurately predict the concepts in the zero-shot situation, and they are also beneficial in the other situation. While the results illustrate the effectiveness of our model, we found several areas for improvement, such as the general terms in the dictionary and the interaction between disease names within a sentence. A possible future direction to deal with general terms is to jointly train the parameters representing the importance of each synonyms.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "5"
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Neural machine translation by jointly learning to align and translate",
                "authors": [
                    {
                        "first": "Dzmitry",
                        "middle": [],
                        "last": "Bahdanau",
                        "suffix": ""
                    },
                    {
                        "first": "Kyunghyun",
                        "middle": [],
                        "last": "Cho",
                        "suffix": ""
                    },
                    {
                        "first": "Yoshua",
                        "middle": [],
                        "last": "Bengio",
                        "suffix": ""
                    }
                ],
                "year": 2015,
                "venue": "ICLR",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Ben- gio. 2015. Neural machine translation by jointly learning to align and translate. In ICLR.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Natural language processing with Python: Analyzing text with the natural language toolkit",
                "authors": [
                    {
                        "first": "Steven",
                        "middle": [],
                        "last": "Bird",
                        "suffix": ""
                    },
                    {
                        "first": "Ewan",
                        "middle": [],
                        "last": "Klein",
                        "suffix": ""
                    },
                    {
                        "first": "Edward",
                        "middle": [],
                        "last": "Loper",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Steven Bird, Ewan Klein, and Edward Loper. 2009. Natural language processing with Python: Analyz- ing text with the natural language toolkit. O'Reilly Media, Inc.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "MEDIC: A practical disease vocabulary used at the comparative toxicogenomics database",
                "authors": [
                    {
                        "first": "Allan",
                        "middle": [
                            "Peter"
                        ],
                        "last": "Davis",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Thomas",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Wiegers",
                        "suffix": ""
                    },
                    {
                        "first": "Carolyn",
                        "middle": [
                            "J"
                        ],
                        "last": "Michael C Rosenstein",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Mattingly",
                        "suffix": ""
                    }
                ],
                "year": 2012,
                "venue": "Database",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Allan Peter Davis, Thomas C Wiegers, Michael C Rosenstein, and Carolyn J Mattingly. 2012. MEDIC: A practical disease vocabulary used at the com- parative toxicogenomics database. Database, 2012:bar065.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "NCBI disease corpus: A resource for disease name recognition and concept normalization",
                "authors": [
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Rezarta Islamaj Dogan",
                        "suffix": ""
                    },
                    {
                        "first": "Zhiyong",
                        "middle": [],
                        "last": "Leaman",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Lu",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "J. Biomed. Inform",
                "volume": "47",
                "issue": "",
                "pages": "1--10",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Rezarta Islamaj Dogan, Robert Leaman, and Zhiyong Lu. 2014. NCBI disease corpus: A resource for dis- ease name recognition and concept normalization. J. Biomed. Inform., 47:1-10.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Pierre Zweigenbaum, and Claire N\u00e9dellec. 2020. C-Norm: a neural approach to few-shot entity normalization",
                "authors": [
                    {
                        "first": "Arnaud",
                        "middle": [],
                        "last": "Ferr\u00e9",
                        "suffix": ""
                    },
                    {
                        "first": "Louise",
                        "middle": [],
                        "last": "Del\u00e9ger",
                        "suffix": ""
                    },
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Bossy",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "BMC Bioinformatics",
                "volume": "21",
                "issue": "23",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Arnaud Ferr\u00e9, Louise Del\u00e9ger, Robert Bossy, Pierre Zweigenbaum, and Claire N\u00e9dellec. 2020. C-Norm: a neural approach to few-shot entity normalization. BMC Bioinformatics, 21(Suppl 23):579.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Gaussian error linear units (GELUs)",
                "authors": [
                    {
                        "first": "Dan",
                        "middle": [],
                        "last": "Hendrycks",
                        "suffix": ""
                    },
                    {
                        "first": "Kevin",
                        "middle": [],
                        "last": "Gimpel",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {
                    "arXiv": [
                        "arXiv:1606.08415"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Dan Hendrycks and Kevin Gimpel. 2016. Gaus- sian error linear units (GELUs). arXiv preprint arXiv:1606.08415.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "DNorm: Disease name normalization with pairwise learning to rank",
                "authors": [
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Leaman",
                        "suffix": ""
                    },
                    {
                        "first": "Zhiyong",
                        "middle": [],
                        "last": "Rezarta Islamaj Dogan",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Lu",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Bioinformatics",
                "volume": "29",
                "issue": "22",
                "pages": "2909--2917",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Robert Leaman, Rezarta Islamaj Dogan, and Zhiy- ong Lu. 2013. DNorm: Disease name normaliza- tion with pairwise learning to rank. Bioinformatics, 29(22):2909-2917.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "TaggerOne: Joint named entity recognition and normalization with semi-markov models",
                "authors": [
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Leaman",
                        "suffix": ""
                    },
                    {
                        "first": "Zhiyong",
                        "middle": [],
                        "last": "Lu",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Bioinformatics",
                "volume": "32",
                "issue": "18",
                "pages": "2839--2846",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Robert Leaman and Zhiyong Lu. 2016. TaggerOne: Joint named entity recognition and normaliza- tion with semi-markov models. Bioinformatics, 32(18):2839-2846.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "BioBERT: a pretrained biomedical language representation model for biomedical text mining",
                "authors": [
                    {
                        "first": "Jinhyuk",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Wonjin",
                        "middle": [],
                        "last": "Yoon",
                        "suffix": ""
                    },
                    {
                        "first": "Sungdong",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Donghyeon",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Sunkyu",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Chan",
                        "middle": [],
                        "last": "Ho So",
                        "suffix": ""
                    },
                    {
                        "first": "Jaewoo",
                        "middle": [],
                        "last": "Kang",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "Bioinformatics",
                "volume": "36",
                "issue": "4",
                "pages": "1234--1240",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Jinhyuk Lee, Wonjin Yoon, Sungdong Kim, Donghyeon Kim, Sunkyu Kim, Chan Ho So, and Jaewoo Kang. 2020. BioBERT: a pre- trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4):1234-1240.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "End-to-end neural coreference resolution",
                "authors": [
                    {
                        "first": "Kenton",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Luheng",
                        "middle": [],
                        "last": "He",
                        "suffix": ""
                    },
                    {
                        "first": "Mike",
                        "middle": [],
                        "last": "Lewis",
                        "suffix": ""
                    },
                    {
                        "first": "Luke",
                        "middle": [],
                        "last": "Zettlemoyer",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "EMNLP",
                "volume": "",
                "issue": "",
                "pages": "188--197",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kenton Lee, Luheng He, Mike Lewis, and Luke Zettle- moyer. 2017. End-to-end neural coreference resolu- tion. In EMNLP, pages 188-197.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "BEST: Next-Generation biomedical entity search tool for knowledge discovery from biomedical literature",
                "authors": [
                    {
                        "first": "Sunwon",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Donghyeon",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Kyubum",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Jaehoon",
                        "middle": [],
                        "last": "Choi",
                        "suffix": ""
                    },
                    {
                        "first": "Seongsoon",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Minji",
                        "middle": [],
                        "last": "Jeon",
                        "suffix": ""
                    },
                    {
                        "first": "Sangrak",
                        "middle": [],
                        "last": "Lim",
                        "suffix": ""
                    },
                    {
                        "first": "Donghee",
                        "middle": [],
                        "last": "Choi",
                        "suffix": ""
                    },
                    {
                        "first": "Sunkyu",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "Aik-Choon",
                        "middle": [],
                        "last": "Tan",
                        "suffix": ""
                    },
                    {
                        "first": "Jaewoo",
                        "middle": [],
                        "last": "Kang",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "PLoS One",
                "volume": "11",
                "issue": "10",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sunwon Lee, Donghyeon Kim, Kyubum Lee, Jae- hoon Choi, Seongsoon Kim, Minji Jeon, Sangrak Lim, Donghee Choi, Sunkyu Kim, Aik-Choon Tan, and Jaewoo Kang. 2016. BEST: Next-Generation biomedical entity search tool for knowledge dis- covery from biomedical literature. PLoS One, 11(10):e0164680.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "BioCreative V CDR task corpus: A resource for chemical disease relation extraction",
                "authors": [
                    {
                        "first": "Jiao",
                        "middle": [],
                        "last": "Li",
                        "suffix": ""
                    },
                    {
                        "first": "Yueping",
                        "middle": [],
                        "last": "Sun",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Robin",
                        "suffix": ""
                    },
                    {
                        "first": "Daniela",
                        "middle": [],
                        "last": "Johnson",
                        "suffix": ""
                    },
                    {
                        "first": "Chih-Hsuan",
                        "middle": [],
                        "last": "Sciaky",
                        "suffix": ""
                    },
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Wei",
                        "suffix": ""
                    },
                    {
                        "first": "Allan",
                        "middle": [
                            "Peter"
                        ],
                        "last": "Leaman",
                        "suffix": ""
                    },
                    {
                        "first": "Carolyn",
                        "middle": [
                            "J"
                        ],
                        "last": "Davis",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Mattingly",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Thomas",
                        "suffix": ""
                    },
                    {
                        "first": "Zhiyong",
                        "middle": [],
                        "last": "Wiegers",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Lu",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Database",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Jiao Li, Yueping Sun, Robin J Johnson, Daniela Sci- aky, Chih-Hsuan Wei, Robert Leaman, Allan Peter Davis, Carolyn J Mattingly, Thomas C Wiegers, and Zhiyong Lu. 2016. BioCreative V CDR task corpus: A resource for chemical disease relation extraction. Database, 2016:baw068.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "A transition-based joint model for disease named entity recognition and normalization",
                "authors": [
                    {
                        "first": "Yinxia",
                        "middle": [],
                        "last": "Lou",
                        "suffix": ""
                    },
                    {
                        "first": "Yue",
                        "middle": [],
                        "last": "Zhang",
                        "suffix": ""
                    },
                    {
                        "first": "Tao",
                        "middle": [],
                        "last": "Qian",
                        "suffix": ""
                    },
                    {
                        "first": "Fei",
                        "middle": [],
                        "last": "Li",
                        "suffix": ""
                    },
                    {
                        "first": "Shufeng",
                        "middle": [],
                        "last": "Xiong",
                        "suffix": ""
                    },
                    {
                        "first": "Donghong",
                        "middle": [],
                        "last": "Ji",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "Bioinformatics",
                "volume": "33",
                "issue": "15",
                "pages": "2363--2371",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Yinxia Lou, Yue Zhang, Tao Qian, Fei Li, Shufeng Xiong, and Donghong Ji. 2017. A transition-based joint model for disease named entity recognition and normalization. Bioinformatics, 33(15):2363-2371.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Abbreviation definition identification based on automatic precision estimates",
                "authors": [
                    {
                        "first": "Sunghwan",
                        "middle": [],
                        "last": "Sohn",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Donald",
                        "suffix": ""
                    },
                    {
                        "first": "Won",
                        "middle": [],
                        "last": "Comeau",
                        "suffix": ""
                    },
                    {
                        "first": "W John",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Wilbur",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "BMC Bioinformatics",
                "volume": "9",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sunghwan Sohn, Donald C Comeau, Won Kim, and W John Wilbur. 2008. Abbreviation definition iden- tification based on automatic precision estimates. BMC Bioinformatics, 9:402.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Biomedical entity representations with synonym marginalization",
                "authors": [
                    {
                        "first": "Mujeen",
                        "middle": [],
                        "last": "Sung",
                        "suffix": ""
                    },
                    {
                        "first": "Hwisang",
                        "middle": [],
                        "last": "Jeon",
                        "suffix": ""
                    },
                    {
                        "first": "Jinhyuk",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Jaewoo",
                        "middle": [],
                        "last": "Kang",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "ACL",
                "volume": "",
                "issue": "",
                "pages": "3641--3650",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mujeen Sung, Hwisang Jeon, Jinhyuk Lee, and Jae- woo Kang. 2020. Biomedical entity representations with synonym marginalization. In ACL, pages 3641- 3650.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "MedType: Improving Medical Entity Linking with Semantic Type Prediction",
                "authors": [
                    {
                        "first": "Shikhar",
                        "middle": [],
                        "last": "Vashishth",
                        "suffix": ""
                    },
                    {
                        "first": "Rishabh",
                        "middle": [],
                        "last": "Joshi",
                        "suffix": ""
                    },
                    {
                        "first": "Denis",
                        "middle": [],
                        "last": "Newman-Griffis",
                        "suffix": ""
                    },
                    {
                        "first": "Ritam",
                        "middle": [],
                        "last": "Dutt",
                        "suffix": ""
                    },
                    {
                        "first": "Carolyn",
                        "middle": [],
                        "last": "Rose",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {
                    "arXiv": [
                        "arXiv:2005.00460"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Shikhar Vashishth, Rishabh Joshi, Denis Newman- Griffis, Ritam Dutt, and Carolyn Rose. 2020. MedType: Improving Medical Entity Linking with Semantic Type Prediction. arXiv preprint arXiv:2005.00460.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "HUNER: improving biomedical NER with pretraining",
                "authors": [
                    {
                        "first": "Leon",
                        "middle": [],
                        "last": "Weber",
                        "suffix": ""
                    },
                    {
                        "first": "Jannes",
                        "middle": [],
                        "last": "M\u00fcnchmeyer",
                        "suffix": ""
                    },
                    {
                        "first": "Tim",
                        "middle": [],
                        "last": "Rockt\u00e4schel",
                        "suffix": ""
                    },
                    {
                        "first": "Maryam",
                        "middle": [],
                        "last": "Habibi",
                        "suffix": ""
                    },
                    {
                        "first": "Ulf",
                        "middle": [],
                        "last": "Leser",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "",
                "volume": "36",
                "issue": "",
                "pages": "295--302",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Leon Weber, Jannes M\u00fcnchmeyer, Tim Rockt\u00e4schel, Maryam Habibi, and Ulf Leser. 2020. HUNER: im- proving biomedical NER with pretraining. Bioinfor- matics, 36(1):295-302.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "NormCo: Deep disease normalization for biomedical knowledge base construction",
                "authors": [
                    {
                        "first": "Dustin",
                        "middle": [],
                        "last": "Wright",
                        "suffix": ""
                    },
                    {
                        "first": "Yannis",
                        "middle": [],
                        "last": "Katsis",
                        "suffix": ""
                    },
                    {
                        "first": "Raghav",
                        "middle": [],
                        "last": "Mehta",
                        "suffix": ""
                    },
                    {
                        "first": "Chun-Nan",
                        "middle": [],
                        "last": "Hsu",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "AKBC",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dustin Wright, Yannis Katsis, Raghav Mehta, and Chun-Nan Hsu. 2019. NormCo: Deep disease nor- malization for biomedical knowledge base construc- tion. In AKBC.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "A Generate-and-Rank framework with semantic type regularization for biomedical concept normalization",
                "authors": [
                    {
                        "first": "Dongfang",
                        "middle": [],
                        "last": "Xu",
                        "suffix": ""
                    },
                    {
                        "first": "Zeyu",
                        "middle": [],
                        "last": "Zhang",
                        "suffix": ""
                    },
                    {
                        "first": "Steven",
                        "middle": [],
                        "last": "Bethard",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "ACL",
                "volume": "",
                "issue": "",
                "pages": "8452--8464",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dongfang Xu, Zeyu Zhang, and Steven Bethard. 2020. A Generate-and-Rank framework with semantic type regularization for biomedical concept normal- ization. In ACL, pages 8452-8464.",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "CD-REST: A system for extracting chemical-induced disease relation in literature",
                "authors": [
                    {
                        "first": "Jun",
                        "middle": [],
                        "last": "Xu",
                        "suffix": ""
                    },
                    {
                        "first": "Yonghui",
                        "middle": [],
                        "last": "Wu",
                        "suffix": ""
                    },
                    {
                        "first": "Yaoyun",
                        "middle": [],
                        "last": "Zhang",
                        "suffix": ""
                    },
                    {
                        "first": "Jingqi",
                        "middle": [],
                        "last": "Wang",
                        "suffix": ""
                    },
                    {
                        "first": "Hee-Jin",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    },
                    {
                        "first": "Hua",
                        "middle": [],
                        "last": "Xu",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Database",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Jun Xu, Yonghui Wu, Yaoyun Zhang, Jingqi Wang, Hee-Jin Lee, and Hua Xu. 2016. CD-REST: A sys- tem for extracting chemical-induced disease relation in literature. Database, 2016:baw036.",
                "links": null
            },
            "BIBREF20": {
                "ref_id": "b20",
                "title": "A neural multi-task learning framework to jointly model medical named entity recognition and normalization",
                "authors": [
                    {
                        "first": "Sendong",
                        "middle": [],
                        "last": "Zhao",
                        "suffix": ""
                    },
                    {
                        "first": "Ting",
                        "middle": [],
                        "last": "Liu",
                        "suffix": ""
                    },
                    {
                        "first": "Sicheng",
                        "middle": [],
                        "last": "Zhao",
                        "suffix": ""
                    },
                    {
                        "first": "Fei",
                        "middle": [],
                        "last": "Wang",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "AAAI",
                "volume": "",
                "issue": "",
                "pages": "817--824",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sendong Zhao, Ting Liu, Sicheng Zhao, and Fei Wang. 2019. A neural multi-task learning framework to jointly model medical named entity recognition and normalization. In AAAI, pages 817-824.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "num": null,
                "text": "Figure 1: The overview of our model. It combines the dictionary-matching scores with the context score obtained from PubMedBERT. The red boxes are the target span and \"ci\" in the figure is the \"i\"-th concept in the dictionary.",
                "type_str": "figure",
                "uris": null
            },
            "TABREF1": {
                "content": "<table/>",
                "html": null,
                "text": "F1 scores of NER and NEN in NCBID and BC5CDR. Bold font represents the highest score.",
                "num": null,
                "type_str": "table"
            },
            "TABREF3": {
                "content": "<table><tr><td/><td>Methods</td><td colspan=\"2\">NCBID BC5CDR</td></tr><tr><td>zero-shot</td><td>Ours without dictionary Ours</td><td>0 0.704</td><td>0 0.597</td></tr><tr><td>standard</td><td>Ours without dictionary Ours</td><td>0.854 0.905</td><td>0.846 0.877</td></tr></table>",
                "html": null,
                "text": "Number of mentions and concepts in standard and zero-shot situations.",
                "num": null,
                "type_str": "table"
            },
            "TABREF4": {
                "content": "<table><tr><td>: F1 scores for NEN of NCBID and BC5CDR</td></tr><tr><td>subsets for zero-shot situation where disease concepts</td></tr><tr><td>do not appear in training data and the standard situation</td></tr><tr><td>where they do appear in training data.</td></tr></table>",
                "html": null,
                "text": "",
                "num": null,
                "type_str": "table"
            }
        }
    }
}