luqh commited on
Commit
39f4cfc
1 Parent(s): d41cdf5

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ ---
4
+
5
+ # Example Uses
6
+
7
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
8
+ ```
9
+ from transformers import AutoTokenizer, T5ForConditionalGeneration
10
+ tokenizer = AutoTokenizer.from_pretrained("luqh/ClinicalT5-base")
11
+ model = T5ForConditionalGeneration.from_pretrained("luqh/ClinicalT5-base", from_flax=True)
12
+ ```
13
+
14
+ # Citation
15
+ If you find this resource useful, please consider citing our work: [ClinicalT5: A Generative Language Model for Clinical Text](https://aclanthology.org/2022.findings-emnlp.398/)
16
+ ```
17
+ @inproceedings{lu-etal-2022-clinicalt5,
18
+ title = "{C}linical{T}5: A Generative Language Model for Clinical Text",
19
+ author = "Lu, Qiuhao and
20
+ Dou, Dejing and
21
+ Nguyen, Thien",
22
+ booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022",
23
+ month = dec,
24
+ year = "2022",
25
+ address = "Abu Dhabi, United Arab Emirates",
26
+ publisher = "Association for Computational Linguistics",
27
+ url = "https://aclanthology.org/2022.findings-emnlp.398",
28
+ pages = "5436--5443",
29
+ abstract = "In the past few years, large pre-trained language models (PLMs) have been widely adopted in different areas and have made fundamental improvements over a variety of downstream tasks in natural language processing (NLP). Meanwhile, domain-specific variants of PLMs are being proposed to address the needs of domains that demonstrate a specific pattern of writing and vocabulary, e.g., BioBERT for the biomedical domain and ClinicalBERT for the clinical domain. Recently, generative language models like BART and T5 are gaining popularity with their competitive performance on text generation as well as on tasks cast as generative problems. However, in the clinical domain, such domain-specific generative variants are still underexplored. To address this need, our work introduces a T5-based text-to-text transformer model pre-trained on clinical text, i.e., ClinicalT5. We evaluate the proposed model both intrinsically and extrinsically over a diverse set of tasks across multiple datasets, and show that ClinicalT5 dramatically outperforms T5 in the domain-specific tasks and compares favorably with its close baselines.",
30
+ }
31
+ '''
32
+
33
+
34
+