TravelLeraLone commited on
Commit
44e0761
1 Parent(s): 811dc4c

release ChemDFM-v1.5-8B

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -15,9 +15,10 @@ ChemDFM is the pioneering open-sourced dialogue foundation model for Chemistry a
15
 
16
  ## News
17
 
 
18
  * **2024-06-13**: The results on the comprehensive science benchmark [SciKnowEval](https://huggingface.co/datasets/hicai-zju/SciKnowEval) show that "ChemDFM emerged as one of the top open-source models by continuing pre-training and fine-tuning on a vast corpus of scientific literature".
19
  * **2024-04-17**: The evaluation data (including instructions) we used in our paper is released on [GitHub](https://github.com/OpenDFM/ChemDFM)
20
- * **2024-03-12**: The parameter of ChemDFM-13B is open-sourced!
21
  * **2024-01-26**: The paper of ChemDFM-13B is released on arXiv: [ChemDFM: Dialogue Foundation Model for Chemistry](https://arxiv.org/abs/2401.14818)
22
 
23
  ## Usage Details
@@ -32,7 +33,7 @@ To load and run ChemDFM locally, here is an example:
32
  import torch
33
  from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig
34
 
35
- model_name_or_id = "X-LANCE/ChemDFM-13B-v1.0"
36
  tokenizer = LlamaTokenizer.from_pretrained(model_name_or_id)
37
  model = LlamaForCausalLM.from_pretrained(model_name_or_id, torch_dtype=torch.float16, device_map="auto")
38
 
 
15
 
16
  ## News
17
 
18
+ * 2024-11-09: [ChemDFM-v1.5-8B](https://huggingface.co/OpenDFM/ChemDFM-v1.5-8B) is released! We implemented our domain pre-training and instruction tuning precedure on a stronger base model LLaMA-3-8B.
19
  * **2024-06-13**: The results on the comprehensive science benchmark [SciKnowEval](https://huggingface.co/datasets/hicai-zju/SciKnowEval) show that "ChemDFM emerged as one of the top open-source models by continuing pre-training and fine-tuning on a vast corpus of scientific literature".
20
  * **2024-04-17**: The evaluation data (including instructions) we used in our paper is released on [GitHub](https://github.com/OpenDFM/ChemDFM)
21
+ * **2024-03-12**: The parameter of [ChemDFM-13B](https://huggingface.co/OpenDFM/ChemDFM-13B-v1.0) is open-sourced!
22
  * **2024-01-26**: The paper of ChemDFM-13B is released on arXiv: [ChemDFM: Dialogue Foundation Model for Chemistry](https://arxiv.org/abs/2401.14818)
23
 
24
  ## Usage Details
 
33
  import torch
34
  from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig
35
 
36
+ model_name_or_id = "OpenDFM/ChemDFM-13B-v1.0"
37
  tokenizer = LlamaTokenizer.from_pretrained(model_name_or_id)
38
  model = LlamaForCausalLM.from_pretrained(model_name_or_id, torch_dtype=torch.float16, device_map="auto")
39