RichardErkhov commited on
Commit
55b165e
1 Parent(s): e4f8d41

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +204 -0
README.md ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Collaiborator-MEDLLM-Llama-3-8B-v2 - GGUF
11
+ - Model creator: https://huggingface.co/collaiborateorg/
12
+ - Original model: https://huggingface.co/collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q2_K.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q2_K.gguf) | Q2_K | 2.96GB |
18
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_XS.gguf) | IQ3_XS | 3.28GB |
19
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_S.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_S.gguf) | IQ3_S | 3.43GB |
20
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_S.gguf) | Q3_K_S | 3.41GB |
21
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_M.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.IQ3_M.gguf) | IQ3_M | 3.52GB |
22
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K.gguf) | Q3_K | 3.74GB |
23
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_M.gguf) | Q3_K_M | 3.74GB |
24
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q3_K_L.gguf) | Q3_K_L | 4.03GB |
25
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.IQ4_XS.gguf) | IQ4_XS | 4.18GB |
26
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_0.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_0.gguf) | Q4_0 | 4.34GB |
27
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.IQ4_NL.gguf) | IQ4_NL | 4.38GB |
28
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K_S.gguf) | Q4_K_S | 4.37GB |
29
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K.gguf) | Q4_K | 4.58GB |
30
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_K_M.gguf) | Q4_K_M | 4.58GB |
31
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_1.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q4_1.gguf) | Q4_1 | 4.78GB |
32
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_0.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_0.gguf) | Q5_0 | 5.21GB |
33
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K_S.gguf) | Q5_K_S | 5.21GB |
34
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K.gguf) | Q5_K | 5.34GB |
35
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_K_M.gguf) | Q5_K_M | 5.34GB |
36
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_1.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q5_1.gguf) | Q5_1 | 5.65GB |
37
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q6_K.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q6_K.gguf) | Q6_K | 6.14GB |
38
+ | [Collaiborator-MEDLLM-Llama-3-8B-v2.Q8_0.gguf](https://huggingface.co/RichardErkhov/collaiborateorg_-_Collaiborator-MEDLLM-Llama-3-8B-v2-gguf/blob/main/Collaiborator-MEDLLM-Llama-3-8B-v2.Q8_0.gguf) | Q8_0 | 7.95GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ license: llama3
46
+ library_name: transformers
47
+ tags:
48
+ - generated_from_trainer
49
+ - Healthcare & Lifesciences
50
+ - BioMed
51
+ - Medical
52
+ - CollAIborate
53
+ base_model: meta-llama/Meta-Llama-3-8B-Instruct
54
+ thumbnail: https://collaiborate.com/logo/logo-blue-bg-1.png
55
+ model-index:
56
+ - name: Collaiborator-MEDLLM-Llama-3-8B-v2
57
+ results: []
58
+ datasets:
59
+ - collaiborateorg/BioMedData
60
+ ---
61
+
62
+ # Collaiborator-MEDLLM-Llama-3-8B-v2
63
+
64
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/653f5b93cd52f288490edc83/wIES_YhNPKn--AqcEmzRJ.png)
65
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on our custom "BioMedData" dataset.
66
+
67
+ ## Model details
68
+
69
+ Model Name: Collaiborator-MEDLLM-Llama-3-8b-v2
70
+
71
+ Base Model: Llama-3-8B-Instruct
72
+
73
+ Parameter Count: 8 billion
74
+
75
+ Training Data: Custom high-quality biomedical dataset
76
+
77
+ Number of Entries in Dataset: 500,000+
78
+
79
+ Dataset Composition: The dataset comprises both synthetic and manually curated samples, ensuring a diverse and comprehensive coverage of biomedical knowledge.
80
+
81
+ ## Model description
82
+
83
+ Collaiborator-MEDLLM-Llama-3-8b-v2 is a specialized large language model designed for biomedical applications. It is finetuned from the Llama-3-8B-Instruct model using a custom dataset containing over 500,000 diverse entries. These entries include a mix of synthetic and manually curated data, ensuring high quality and broad coverage of biomedical topics.
84
+
85
+ The model is trained to understand and generate text related to various biomedical fields, making it a valuable tool for researchers, clinicians, and other professionals in the biomedical domain.
86
+
87
+ ## Evaluation Metrics
88
+ Collaiborator-MEDLLM-Llama-3-8b-v2 outperforms many of the leading LLMs and find below its metrics evaluated using the Eleuther AI Language Model Evaluation Harness framework against the tasks medmcqa, medqa_4options, mmlu_anatomy, mmlu_clinical_knowledge, mmlu_college_biology, mmlu_college_medicine, mmlu_medical_genetics, mmlu_professional_medicine and pubmedqa
89
+
90
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a78a59838b3acc53d656de/LAL2thZUhxtg56u5SYIGb.png)
91
+
92
+ ## Benchmark Results
93
+
94
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/653f5b93cd52f288490edc83/CxDpvXRSIhuG45TShELKw.png)
95
+
96
+ ## Quick Demo
97
+
98
+ <video controls autoplay src="https://cdn-uploads.huggingface.co/production/uploads/653f5b93cd52f288490edc83/piGRPwvcBTLmcgExL89zp.mp4"></video>
99
+
100
+ ## Intended uses & limitations
101
+
102
+ Collaiborator-MEDLLM-Llama-3-8b-v2 is intended for a wide range of applications within the biomedical field, including:
103
+
104
+ 1. Research Support: Assisting researchers in literature review and data extraction from biomedical texts.
105
+ 2. Clinical Decision Support: Providing information to support clinical decision-making processes.
106
+ 3. Educational Tool: Serving as a resource for medical students and professionals seeking to expand their knowledge base.
107
+
108
+ ## Limitations and Ethical Considerations
109
+
110
+ While Collaiborator-MEDLLM-Llama-3-8b-v2 performs well in various biomedical NLP tasks, users should be aware of the following limitations:
111
+
112
+ Biases: The model may inherit biases present in the training data. Efforts have been made to curate a balanced dataset, but some biases may persist.
113
+
114
+ Accuracy: The model's responses are based on patterns in the data it has seen and may not always be accurate or up-to-date. Users should verify critical information from reliable sources.
115
+
116
+ Ethical Use: The model should be used responsibly, particularly in clinical settings where the stakes are high. It should complement, not replace, professional judgment and expertise.
117
+
118
+
119
+ ## Training and evaluation
120
+
121
+ Collaiborator-MEDLLM-Llama-3-8b-v2 was trained using NVIDIA A40 GPU's, which provides the computational power necessary for handling large-scale data and model parameters efficiently. Rigorous evaluation protocols have been implemented to benchmark its performance against similar models, ensuring its robustness and reliability in real-world applications.
122
+
123
+ ## How to use
124
+
125
+ import transformers
126
+ import torch
127
+
128
+ model_id = "collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2"
129
+
130
+ pipeline = transformers.pipeline(
131
+ "text-generation",
132
+ model=model_id,
133
+ model_kwargs={"torch_dtype": torch.bfloat16},
134
+ device_map="auto",
135
+ )
136
+
137
+ messages = [
138
+ {"role": "system", "content": "You are an expert trained on healthcare and biomedical domain!"},
139
+ {"role": "user", "content": "I'm a 35-year-old male and for the past few months, I've been experiencing fatigue, increased sensitivity to cold, and dry, itchy skin. What is the diagnosis here?"},
140
+ ]
141
+
142
+ prompt = pipeline.tokenizer.apply_chat_template(
143
+ messages,
144
+ tokenize=False,
145
+ add_generation_prompt=True
146
+ )
147
+
148
+ terminators = [
149
+ pipeline.tokenizer.eos_token_id,
150
+ pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
151
+ ]
152
+
153
+ outputs = pipeline(
154
+ prompt,
155
+ max_new_tokens=256,
156
+ eos_token_id=terminators,
157
+ do_sample=True,
158
+ temperature=0.6,
159
+ top_p=0.9,
160
+ )
161
+
162
+ print(outputs[0]["generated_text"][len(prompt):])
163
+
164
+ ### Contact Information
165
+
166
+ For further information, inquiries, or issues related to Biomed-LLM, please contact:
167
+
168
+ Email: info@collaiborate.com
169
+
170
+ Website: https://www.collaiborate.com
171
+
172
+ ### Training hyperparameters
173
+
174
+ The following hyperparameters were used during training:
175
+ - learning_rate: 0.0002
176
+ - train_batch_size: 12
177
+ - eval_batch_size: 8
178
+ - seed: 42
179
+ - gradient_accumulation_steps: 4
180
+ - total_train_batch_size: 32
181
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
182
+ - lr_scheduler_type: cosine
183
+ - lr_scheduler_warmup_ratio: 0.03
184
+ - mixed_precision_training: Native AMP
185
+
186
+ ### Framework versions
187
+
188
+ - PEFT 0.11.0
189
+ - Transformers 4.40.2
190
+ - Pytorch 2.1.2
191
+ - Datasets 2.19.1
192
+ - Tokenizers 0.19.1
193
+
194
+ ### Citation
195
+
196
+ If you use Collaiborator-MEDLLM-Llama-3-8b in your research or applications, please cite it as follows:
197
+
198
+ @misc{Collaiborator_MEDLLM,
199
+ author = Collaiborator,
200
+ title = {Collaiborator-MEDLLM-Llama-3-8b: A High-Performance Biomedical Language Model},
201
+ year = {2024},
202
+ howpublished = {https://huggingface.co/collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v2},
203
+ }
204
+