RichardErkhov commited on
Commit
3819bf6
1 Parent(s): f7d779e

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +188 -0
README.md ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Asclepius-Llama3-8B - GGUF
11
+ - Model creator: https://huggingface.co/starmpcc/
12
+ - Original model: https://huggingface.co/starmpcc/Asclepius-Llama3-8B/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [Asclepius-Llama3-8B.Q2_K.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q2_K.gguf) | Q2_K | 2.96GB |
18
+ | [Asclepius-Llama3-8B.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
19
+ | [Asclepius-Llama3-8B.IQ3_S.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.IQ3_S.gguf) | IQ3_S | 3.43GB |
20
+ | [Asclepius-Llama3-8B.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
21
+ | [Asclepius-Llama3-8B.IQ3_M.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.IQ3_M.gguf) | IQ3_M | 3.52GB |
22
+ | [Asclepius-Llama3-8B.Q3_K.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q3_K.gguf) | Q3_K | 3.74GB |
23
+ | [Asclepius-Llama3-8B.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
24
+ | [Asclepius-Llama3-8B.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
25
+ | [Asclepius-Llama3-8B.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
26
+ | [Asclepius-Llama3-8B.Q4_0.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q4_0.gguf) | Q4_0 | 4.34GB |
27
+ | [Asclepius-Llama3-8B.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
28
+ | [Asclepius-Llama3-8B.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
29
+ | [Asclepius-Llama3-8B.Q4_K.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q4_K.gguf) | Q4_K | 4.58GB |
30
+ | [Asclepius-Llama3-8B.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
31
+ | [Asclepius-Llama3-8B.Q4_1.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q4_1.gguf) | Q4_1 | 4.78GB |
32
+ | [Asclepius-Llama3-8B.Q5_0.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q5_0.gguf) | Q5_0 | 5.21GB |
33
+ | [Asclepius-Llama3-8B.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
34
+ | [Asclepius-Llama3-8B.Q5_K.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q5_K.gguf) | Q5_K | 5.34GB |
35
+ | [Asclepius-Llama3-8B.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
36
+ | [Asclepius-Llama3-8B.Q5_1.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q5_1.gguf) | Q5_1 | 5.65GB |
37
+ | [Asclepius-Llama3-8B.Q6_K.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q6_K.gguf) | Q6_K | 6.14GB |
38
+ | [Asclepius-Llama3-8B.Q8_0.gguf](https://huggingface.co/RichardErkhov/starmpcc_-_Asclepius-Llama3-8B-gguf/blob/main/Asclepius-Llama3-8B.Q8_0.gguf) | Q8_0 | 7.95GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: cc-by-nc-sa-4.0
46
+ datasets:
47
+ - starmpcc/Asclepius-Synthetic-Clinical-Notes
48
+ language:
49
+ - en
50
+ pipeline_tag: text2text-generation
51
+ tags:
52
+ - medical
53
+ ---
54
+ # Model Card for Model ID
55
+
56
+ <!-- Provide a quick summary of what the model is/does. -->
57
+
58
+ This is an official model checkpoint for Asclepius-Llama3-8B [(arxiv)](https://arxiv.org/abs/2309.00237).
59
+ This model is an enhanced version of Asclepius-7B, by replacing the base model with Llama-3 and increasing the max sequence length to 8192.
60
+
61
+ ## UPDATE
62
+ ### 2024.01.10
63
+ - Asclepius-R, the variant of Asclepius that trained on MIMIC-III discharge summaries, is now available on [Physionet](https://physionet.org/content/asclepius-r/1.0.0/)!
64
+
65
+ ## Model Details
66
+
67
+ ### Model Description
68
+
69
+ <!-- Provide a longer summary of what this model is. -->
70
+
71
+
72
+
73
+ - **Model type:** Clinical LLM (Large Language Model)
74
+ - **Language(s) (NLP):** English
75
+ - **License:** CC-BY-NC-SA 4.0
76
+ - **Finetuned from model [optional]:** Llama3-8B
77
+
78
+ ### Model Sources [optional]
79
+
80
+ <!-- Provide the basic links for the model. -->
81
+
82
+ - **Repository:** https://github.com/starmpcc/Asclepius
83
+ - **Paper:** https://arxiv.org/abs/2309.00237
84
+ - **Data:** https://huggingface.co/datasets/starmpcc/Asclepius-Synthetic-Clinical-Notes
85
+
86
+ ## Uses
87
+
88
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
89
+ This model can perform below 8 clinical NLP tasks, with clincal notes.
90
+ - Named Entity Recognition
91
+ - Abbreviation Expansion
92
+ - Relation Extraction
93
+ - Temporal Information Extraction
94
+ - Coreference Resolution
95
+ - Paraphrasing
96
+ - Summarization
97
+ - Question Answering
98
+
99
+ ### Direct Use
100
+
101
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
102
+
103
+ [More Information Needed]
104
+
105
+ ### Downstream Use [optional]
106
+
107
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
108
+
109
+ [More Information Needed]
110
+
111
+ ### Out-of-Scope Use
112
+
113
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
114
+
115
+ ONLY USE THIS MODEL FOR RESEARCH PURPOSE!!
116
+
117
+ ## How to Get Started with the Model
118
+
119
+ ```python
120
+ prompt = """You are an intelligent clinical languge model.
121
+ Below is a snippet of patient's discharge summary and a following instruction from healthcare professional.
122
+ Write a response that appropriately completes the instruction.
123
+ The response should provide the accurate answer to the instruction, while being concise.
124
+
125
+ [Discharge Summary Begin]
126
+ {note}
127
+ [Discharge Summary End]
128
+
129
+ [Instruction Begin]
130
+ {question}
131
+ [Instruction End]
132
+ """
133
+
134
+ from transformers import AutoTokenizer, AutoModelForCausalLM
135
+ tokenizer = AutoTokenizer.from_pretrained("starmpcc/Asclepius-Llama3-8B", use_fast=False)
136
+ model = AutoModelForCausalLM.from_pretrained("starmpcc/Asclepius-Llama3-8B")
137
+
138
+ note = "This is a sample note"
139
+ question = "What is the diagnosis?"
140
+
141
+ model_input = prompt.format(note=note, question=question)
142
+ input_ids = tokenizer(model_input, return_tensors="pt").input_ids
143
+ output = model.generate(input_ids)
144
+ print(tokenizer.decode(output[0]))
145
+ ```
146
+
147
+ ## Training Details
148
+
149
+ ### Training Data
150
+
151
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
152
+
153
+ https://huggingface.co/datasets/starmpcc/Asclepius-Synthetic-Clinical-Notes
154
+
155
+ ### Training Procedure
156
+
157
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
158
+ - Initial training was conducted using causal language modeling on synthetic clinical notes.
159
+ - It was then fine-tuned with clinical instruction-response pairs.
160
+ - For a comprehensive overview of our methods, our upcoming paper will serve as a resource.
161
+
162
+ #### Training Hyperparameters
163
+
164
+ - We followed config used in [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca)
165
+ -
166
+ #### Speeds, Sizes, Times
167
+
168
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
169
+ - Pre-Training (1 epoch): 2h 59m with 4x A100 80G
170
+ - Instruction Fine-Tuning (3 epoch): 30h 41m with 4x A100 80G
171
+
172
+
173
+
174
+ ## Citation
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ ```
181
+ @article{kweon2023publicly,
182
+ title={Publicly Shareable Clinical Large Language Model Built on Synthetic Clinical Notes},
183
+ author={Kweon, Sunjun and Kim, Junu and Kim, Jiyoun and Im, Sujeong and Cho, Eunbyeol and Bae, Seongsu and Oh, Jungwoo and Lee, Gyubok and Moon, Jong Hak and You, Seng Chan and others},
184
+ journal={arXiv preprint arXiv:2309.00237},
185
+ year={2023}
186
+ }
187
+ ```
188
+