|
--- |
|
language: |
|
- multilingual |
|
- en |
|
- zh |
|
license: apache-2.0 |
|
pipeline_tag: fill-mask |
|
tags: |
|
- medical |
|
--- |
|
|
|
# KBioXLM |
|
|
|
The aligned corpus constructed using the knowledge-anchored method is combined with a multi task training strategy to continue training XLM-R, thus obtaining KBioXLM. It is the first multilingual biomedical pre-trained language model we know that has cross-lingual understanding capabilities in medical domain. It was introduced in the paper [KBioXLM: A Knowledge-anchored Biomedical |
|
Multilingual Pretrained Language Model](http://arxiv.org/abs/2311.11564) and released in [this repository](https://github.com/ngwlh-gl/KBioXLM/tree/main). |
|
|
|
## Model description |
|
KBioXLM model can be fintuned on downstream tasks. The downstream tasks here refer to biomedical cross-lingual understanding tasks, such as biomedical entity recognition, biomedical relationship extraction and biomedical text classification. |
|
|
|
## Usage |
|
|
|
You can follow the prompts below to load our model parameters: |
|
|
|
```python |
|
from transformers import RobertaModel |
|
model=RobertaModel.from_pretrained('ngwlh/KBioXLM') |
|
``` |
|
|
|
### BibTeX entry and citation info |
|
|
|
Coming soon. |