KEPTlongfomer is a medical knowledge enhanced version of Longformer that was further pre-trained using contrastive learning. The model achieves SOTA performance on auto ICD coding on MIMIC-III as of 11/12/2022. A sister model for better performance is available here.
Pre-training
We initialized this model from clinical longformer.
And then pretrained with Hierarchical Self-Alignment Pretrain (HSAP) using Knowledge Graph UMLS. This includes (a) Hierarchy, (b) Synonym, (c) Abbreviation. For more info, see section 3.3 in paper. The learning rate was 5e-5, weight decay was 0.01, adam epsilon was 1e-5.
Usage
See our github for how to use this with prompts on auto ICD coding.
With the following result:
Metric | Score |
---|---|
rec_micro | =0.5729403619819988 |
rec_macro | =0.11342156911120573 |
rec_at_8 | =0.4094837705486378 |
rec_at_75 | =0.8470734920535119 |
rec_at_50 | =0.8005338782352 |
rec_at_5 | =0.2891628170355805 |
rec_at_15 | =0.5768778119750537 |
prec_micro | =0.6411968713105065 |
prec_macro | =0.12227610414493029 |
prec_at_8 | =0.7760972716488731 |
prec_at_75 | =0.197504942665085 |
prec_at_50 | =0.2768090154211151 |
prec_at_5 | =0.8483392645314354 |
prec_at_15 | =0.6178529062870699 |
f1_micro | =0.6051499904242899 |
f1_macro | =0.11768251595637802 |
f1_at_8 | =0.536107150495997 |
f1_at_75 | =0.32032290907137506 |
f1_at_50 | =0.411373195944102 |
f1_at_5 | =0.43131028155283435 |
f1_at_15 | =0.5966627077602488 |
auc_micro | =0.9651754312635265 |
auc_macro | =0.8566590059725866 |
acc_micro | =0.43384592341105344 |
acc_macro | =0.08639139221100567 |
- Downloads last month
- 58
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.