xtie commited on
Commit
cd5ac83
1 Parent(s): 24b36c7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +105 -10
README.md CHANGED
@@ -1,13 +1,108 @@
1
  ---
2
- language:
3
- - en
4
- metrics:
5
- - rouge
6
- - bertscore
7
- - bleu
8
- - chrf
9
- library_name: transformers
10
- pipeline_tag: summarization
11
  tags:
 
12
  - medical
13
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: en
 
 
 
 
 
 
 
 
3
  tags:
4
+ - summarization
5
  - medical
6
+ library_name: transformers
7
+ pipeline_tag: summarization
8
+ ---
9
+
10
+ # Automatic Personalized Impression Generation for PET Reports Using Large Language Models 📄✍
11
+
12
+ **Authored by**: Xin Tie, Muheon Shin, Ali Pirasteh, Nevein Ibrahim, Zachary Huemann, Sharon M. Castellino, Kara Kelly, John Garrett, Junjie Hu, Steve Y. Cho, Tyler J. Bradshaw
13
+
14
+ ## 📑 Model Description
15
+
16
+ This is the fine-tuned T5 model for summarizing findings in PET reports.
17
+
18
+ To check our fine-tuned large language models (LLMs) for PET report summarization:
19
+ - [BERT2BERT-PET](https://huggingface.co/xtie/Clinicallongformer2roberta-PET-impression)
20
+ - [BART-PET](https://huggingface.co/xtie/BART-PET-impression)
21
+ - [BioBART-PET](https://huggingface.co/xtie/BioBART-PET-impression)
22
+ - [PEGASUS-PET](https://huggingface.co/xtie/PEGASUS-PET-impression)
23
+ - [T5v1.1-PET](https://huggingface.co/xtie/T5v1.1-PET-impression)
24
+ - [Clinical-T5-PET](https://huggingface.co/xtie/ClinicalT5-PET-impression)
25
+ - [Flan-T5-PET](https://huggingface.co/xtie/Flan-T5-PET-impression)
26
+ - [GPT2-XL-PET](https://huggingface.co/xtie/GPT2-PET-impression)
27
+ - [OPT-1.3B-PET](https://huggingface.co/xtie/OPT-PET-impression)
28
+ - [LLaMA-LoRA-PET](https://huggingface.co/xtie/LLaMA-LoRA-PET-impression)
29
+ - [Alpaca-LoRA-PET](https://huggingface.co/xtie/Alpaca-LoRA-PET-impression)
30
+
31
+ ## 📑 Abstract
32
+
33
+ Purpose: To determine if fine-tuned large language models (LLMs) can generate accurate, personalized impressions for whole-body PET reports.
34
+
35
+ Materials and Methods: Twelve language models were trained on a corpus of PET reports using the teacher-forcing algorithm, with the report findings as input and the clinical impressions as reference. An extra input token encodes the reading physician’s identity, allowing models to learn physician-specific reporting styles. Our corpus comprised 37,370 retrospective PET reports collected from our institution between 2010 and 2022. To identify the best LLM, 30 evaluation metrics were benchmarked against quality scores from two nuclear medicine (NM) physicians, with the most aligned metrics selecting the model for expert evaluation. In a subset of data, model-generated impressions and original clinical impressions were assessed by three NM physicians according to 6 quality dimensions and an overall utility score (5-point scale). Each physician reviewed 12 of their own reports and 12 reports from other physicians. Bootstrap resampling was used for statistical analysis.
36
+
37
+ Results: Of all evaluation metrics, domain-adapted BARTScore and PEGASUSScore showed the highest Spearman’s ρ correlations (0.568 and 0.563) with physician preferences. Based on these metrics, the fine-tuned PEGASUS model was selected as the top LLM. When physicians reviewed PEGASUS-generated impressions in their own style, 89% were considered clinically acceptable, with a mean utility score of 4.08/5. Physicians rated these personalized impressions as comparable in overall utility to the impressions dictated by other physicians (4.03, P=0.41).
38
+
39
+ Conclusion: Personalized impressions generated by PEGASUS were clinically useful, highlighting its potential to expedite PET reporting.
40
+
41
+ [Read the full paper](https://arxiv.org/abs/2309.10066)
42
+ <!-- Link to our Arxiv paper -->
43
+
44
+ ## 🚀 Usage
45
+
46
+ ```bash
47
+ finetuned_model = "xtie/T5v1.1-PET-impression"
48
+ tokenizer = AutoTokenizer.from_pretrained(finetuned_model)
49
+ model = AutoModelForSeq2SeqLM.from_pretrained(finetuned_model, ignore_mismatched_sizes=True).eval()
50
+
51
+ findings_info =
52
+ """
53
+ Description: PET CT WHOLE BODY
54
+ Radiologist: James
55
+ Findings:
56
+ Head/Neck: xxx Chest: xxx Abdomen/Pelvis: xxx Extremities/Musculoskeletal: xxx
57
+ Indication:
58
+ The patient is a 60-year old male with a history of xxx
59
+ """
60
+
61
+ inputs = tokenizer(findings_info.replace('\n', ' '),
62
+ padding="max_length",
63
+ truncation=True,
64
+ max_length=1024,
65
+ return_tensors="pt")
66
+ input_ids = inputs.input_ids.to("cuda")
67
+ attention_mask = inputs.attention_mask.to("cuda")
68
+ outputs = model.generate(input_ids,
69
+ attention_mask=attention_mask,
70
+ max_new_tokens=512,
71
+ num_beam_groups=1,
72
+ num_beams=4,
73
+ do_sample=False,
74
+ diversity_penalty=0.0,
75
+ num_return_sequences=1,
76
+ length_penalty=2.0,
77
+ no_repeat_ngram_size=3,
78
+ early_stopping=True
79
+ )
80
+ # get the generated impressions
81
+ output_str = tokenizer.decode(outputs,
82
+ skip_special_tokens=True)
83
+ ```
84
+
85
+
86
+ ### 📊 Performance Metrics
87
+
88
+ For detailed evaluation results, please refer to our paper.
89
+ - **ROUGE-1**: 53.7
90
+ - **ROUGE-2**: 30.7
91
+ - **ROUGE-L**: 40.3
92
+ - **BLEU**: 24.1
93
+ - **BERTScore**: 0.747
94
+
95
+ ### 💡 Highlights
96
+
97
+ - The fine-tuned large language model provides clinically useful, personalized impressions based on PET findings.
98
+ - To our knowledge, this is the first attempt to automate impression generation for whole-body PET reports.
99
+
100
+ ### 🖥️ Hardware
101
+
102
+ The models were trained on NVIDIA A100 GPUs.
103
+
104
+ ---
105
+
106
+ ## 📁 Additional Resources
107
+ - **Finetuned from model:** [T5-Large](https://huggingface.co/google/t5-v1_1-large)
108
+ - **Codebase for training and inference:** [GitHub Repository](https://github.com/xtie97/PET-Report-Summarization)