mav23 commited on
Commit
06418ab
1 Parent(s): f4a3dad

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +49 -0
  3. biomedgpt-lm-7b.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ biomedgpt-lm-7b.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - medical
5
+ datasets:
6
+ - biomed
7
+ ---
8
+ # BioMedGPT-LM-7B
9
+
10
+ **BioMedGPT-LM-7B** is the first large generative language model based on Llama2 in the biomedical domain.
11
+ It was fine-tuned from the Llama2-7B-Chat with millions of biomedical papers from the [S2ORC corpus](https://github.com/allenai/s2orc/blob/master/README.md). Through further fine-tuning, BioMedGPT-LM-7B outperforms or is on par with human and significantly larger general-purpose foundation models on several biomedical QA benchmarks.
12
+
13
+ ### Training Details
14
+
15
+ The model was trained with the following hyperparameters:
16
+
17
+ * Epochs: 5
18
+ * Batch size: 192
19
+ * Context length: 2048
20
+ * Learning rate: 2e-5
21
+
22
+ BioMedGPT-LM-7B is fine-tuned on over 26 billion tokens highly pertinent to the field of biomedicine. The fine-tuning data are extracted from millions of biomedical papers in S2ORC data using PubMed Central (PMC)-ID and PubMed ID as criteria.
23
+
24
+ ### Model Developers
25
+
26
+ PharMolix
27
+
28
+ ### How to Use
29
+
30
+ BioMedGPT-LM-7B is the generative language model of **[BioMedGPT-10B](https://github.com/PharMolix/OpenBioMed)**, an open-source version of BioMedGPT.
31
+ BioMedGPT is an open multimodal generative pre-trained transformer (GPT) for biomedicine, which bridges the natural language modality and diverse biomedical data modalities via large generative language models.
32
+
33
+ ![The architecture of BioMedGPT-10B](BioMedGPT-10B.jpg)
34
+
35
+ ### Technical Report
36
+
37
+ More technical details of BioMedGPT-LM-7B, BioMedGPT-10B, and BioMedGPT can be found in the technical reprot: ["BioMedGPT: Open Multimodal Generative Pre-trained Transformer for BioMedicine"](https://arxiv.org/abs/2308.09442v2).
38
+
39
+ ### GitHub
40
+
41
+ [https://github.com/PharMolix/OpenBioMed](https://github.com/PharMolix/OpenBioMed)
42
+
43
+ ### Limitations
44
+
45
+ This repository holds BioMedGPT-LM-7B, and we emphasize the responsible and ethical use of this model. BioMedGPT-LM-7B should NOT be used to provide services to the general public. Generating any content that violates applicable laws and regulations, such as inciting subversion of state power, endangering national security and interests, propagating terrorism, extremism, ethnic hatred and discrimination, violence, pornography, or false and harmful information, etc. is strictly prohibited. BioMedGPT-LM-7B is not liable for any consequences arising from any content, data, or information provided or published by users.
46
+
47
+ ### Licenses
48
+
49
+ This repository is licensed under the Apache-2.0. The use of BioMedGPT-LM-7B model is accompanied with [Acceptable Use Policy](USE_POLICY.md).
biomedgpt-lm-7b.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d38aed91fb831fd5df4c6c929a7130aed1ca65e21f75101f31bf4e7af992a17
3
+ size 3825807520