--- language: - en tags: - biomedical - bionlp - entity linking - embedding - bert --- The GEBERT model pre-trained with GAT graph encoder. The model was published at [CLEF 2023 conference](https://clef2023.clef-initiative.eu/). The source code is available at [github](https://github.com/Andoree/GEBERT). Pretraining data: biomedical concept graph and concept names from the UMLS (2020AB release). Base model: [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext). ```bibtex @inproceedings{sakhovskiy2023gebert, author="Sakhovskiy, Andrey and Semenova, Natalia and Kadurin, Artur and Tutubalina, Elena", title="Graph-Enriched Biomedical Entity Representation Transformer", booktitle="Experimental IR Meets Multilinguality, Multimodality, and Interaction", year="2023", publisher="Springer Nature Switzerland", address="Cham", pages="109--120", isbn="978-3-031-42448-9" } ```