codefuse-admin commited on
Commit
b5959bb
1 Parent(s): 13b678f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -18
README.md CHANGED
@@ -1,21 +1,14 @@
1
  ---
2
- frameworks:
3
- - Pytorch
4
- license: apache-2.0
5
- tasks:
6
- - text-generation
7
  ---
8
  # Model Card for CodeFuse-DeepSeek-33B
9
- <p align="center">
10
- <img src="https://modelscope.cn/api/v1/models/codefuse-ai/CodeFuse-DeepSeek-33B/repo?Revision=master&FilePath=LOGO.jpg&View=true" width="800"/>
11
- <p>
12
 
13
  [[中文]](#chinese) [[English]](#english)
14
 
15
- #### Clone with HTTP
16
- ```bash
17
- git clone https://www.modelscope.cn/codefuse-ai/CodeFuse-DeepSeek-33B.git
18
- ```
19
 
20
  <a id="english"></a>
21
 
@@ -27,7 +20,7 @@ CodeFuse-DeepSeek-33B is a 33B Code-LLM finetuned by QLoRA on multiple code-rela
27
 
28
  ## News and Updates
29
 
30
- 🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B has been released, achiving a pass@1 (greedy decoding) score of 78.65% on HumanEval.
31
 
32
  🔥🔥 2023-11-10 CodeFuse-CodeGeeX2-6B has been released, achieving a pass@1 (greedy decoding) score of 45.12% on HumanEval, which is a 9.22% increase compared to CodeGeeX2 35.9%.
33
 
@@ -39,7 +32,7 @@ CodeFuse-DeepSeek-33B is a 33B Code-LLM finetuned by QLoRA on multiple code-rela
39
 
40
  🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B-4bits/summary) of [CodeFuse-CodeLlama-34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary). Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
41
 
42
- 🔥🔥🔥 2023-09-11 [CodeFuse-CodeLlama34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary) has achived 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present.
43
 
44
  <br>
45
 
@@ -133,10 +126,10 @@ In this format, the system section is optional and the conversation can be eithe
133
 
134
  ```python
135
  import torch
136
- from modelscope import AutoTokenizer, AutoModelForCausalLM, GenerationConfig, snapshot_download
137
 
138
  def load_model_tokenizer(model_path):
139
- tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True, use_fast=False, legacy=False)
140
  tokenizer.eos_token = "<|end▁of▁sentence|>"
141
  tokenizer.pad_token = "<|end▁of▁sentence|>"
142
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
@@ -295,10 +288,10 @@ User prompt...
295
 
296
  ```python
297
  import torch
298
- from modelscope import AutoTokenizer, AutoModelForCausalLM, GenerationConfig, snapshot_download
299
 
300
  def load_model_tokenizer(model_path):
301
- tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True, use_fast=False, legacy=False)
302
  tokenizer.eos_token = "<|end▁of▁sentence|>"
303
  tokenizer.pad_token = "<|end▁of▁sentence|>"
304
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
 
1
  ---
2
+ license: other
3
+ tasks:
4
+ - code-generation
 
 
5
  ---
6
  # Model Card for CodeFuse-DeepSeek-33B
7
+ ![logo](LOGO.jpg)
 
 
8
 
9
  [[中文]](#chinese) [[English]](#english)
10
 
11
+
 
 
 
12
 
13
  <a id="english"></a>
14
 
 
20
 
21
  ## News and Updates
22
 
23
+ 🔥🔥🔥 2024-01-12 CodeFuse-DeepSeek-33B has been released, achieving a pass@1 (greedy decoding) score of 78.65% on HumanEval.
24
 
25
  🔥🔥 2023-11-10 CodeFuse-CodeGeeX2-6B has been released, achieving a pass@1 (greedy decoding) score of 45.12% on HumanEval, which is a 9.22% increase compared to CodeGeeX2 35.9%.
26
 
 
32
 
33
  🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B-4bits/summary) of [CodeFuse-CodeLlama-34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary). Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
34
 
35
+ 🔥🔥🔥 2023-09-11 [CodeFuse-CodeLlama34B](https://modelscope.cn/models/codefuse-ai/CodeFuse-CodeLlama-34B/summary) has achieved 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present.
36
 
37
  <br>
38
 
 
126
 
127
  ```python
128
  import torch
129
+ from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
130
 
131
  def load_model_tokenizer(model_path):
132
+ tokenizer = AutoTokenizer.from_pretrained("codefuse-ai/CodeFuse-DeepSeek-33B", trust_remote_code=True, use_fast=False, legacy=False)
133
  tokenizer.eos_token = "<|end▁of▁sentence|>"
134
  tokenizer.pad_token = "<|end▁of▁sentence|>"
135
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)
 
288
 
289
  ```python
290
  import torch
291
+ from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
292
 
293
  def load_model_tokenizer(model_path):
294
+ tokenizer = AutoTokenizer.from_pretrained("codefuse-ai/CodeFuse-DeepSeek-33B", trust_remote_code=True, use_fast=False, legacy=False)
295
  tokenizer.eos_token = "<|end▁of▁sentence|>"
296
  tokenizer.pad_token = "<|end▁of▁sentence|>"
297
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids(tokenizer.eos_token)