suolyer commited on
Commit
d1bf718
1 Parent(s): d5c654e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -3,16 +3,16 @@ language:
3
  - zh
4
  license: apache-2.0
5
  ---
6
- # Wenzhong-3.5B model (chinese),one model of [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM).
7
- As we all know, the single direction language model based on decoder structure has strong generation ability, such as GPT model. **The 3.5 billion parameter Wenzhong-3.5B large model, using 100G chinese common data, 32 A100 training for 28 hours,** is the largest open source **GPT2 large model of chinese**. **Our model performs well in Chinese continuation generation.**
8
 
9
  ## Usage
10
 
11
  ### load model
12
  ```python
13
  from transformers import GPT2Tokenizer, GPT2Model
14
- tokenizer = GPT2Tokenizer.from_pretrained('IDEA-CCNL/Wenzhong-3.5B')
15
- model = GPT2Model.from_pretrained('IDEA-CCNL/Wenzhong-3.5B')
16
  text = "Replace me by any text you'd like."
17
  encoded_input = tokenizer(text, return_tensors='pt')
18
  output = model(**encoded_input)
@@ -21,7 +21,7 @@ output = model(**encoded_input)
21
  ```python
22
  from transformers import pipeline, set_seed
23
  set_seed(55)
24
- generator = pipeline('text-generation', model='IDEA-CCNL/Wenzhong-3.5B')
25
  generator("北京位于", max_length=30, num_return_sequences=1)
26
 
27
  ```
 
3
  - zh
4
  license: apache-2.0
5
  ---
6
+ # Wenzhong-GPT2-3.5B model (chinese),one model of [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM).
7
+ As we all know, the single direction language model based on decoder structure has strong generation ability, such as GPT model. **The 3.5 billion parameter Wenzhong-GPT2-3.5B large model, using 100G chinese common data, 32 A100 training for 28 hours,** is the largest open source **GPT2 large model of chinese**. **Our model performs well in Chinese continuation generation.**
8
 
9
  ## Usage
10
 
11
  ### load model
12
  ```python
13
  from transformers import GPT2Tokenizer, GPT2Model
14
+ tokenizer = GPT2Tokenizer.from_pretrained('IDEA-CCNL/Wenzhong-GPT2-3.5B')
15
+ model = GPT2Model.from_pretrained('IDEA-CCNL/Wenzhong-GPT2-3.5B')
16
  text = "Replace me by any text you'd like."
17
  encoded_input = tokenizer(text, return_tensors='pt')
18
  output = model(**encoded_input)
 
21
  ```python
22
  from transformers import pipeline, set_seed
23
  set_seed(55)
24
+ generator = pipeline('text-generation', model='IDEA-CCNL/Wenzhong-GPT2-3.5B')
25
  generator("北京位于", max_length=30, num_return_sequences=1)
26
 
27
  ```