Text Generation
PyTorch
causal-lm
rwkv
BlinkDL commited on
Commit
512186e
1 Parent(s): 0616e70

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -27,7 +27,7 @@ datasets:
27
 
28
  ## Model Description
29
 
30
- RWKV-4 trained on 100+ world languages.
31
 
32
  How to use:
33
  * use latest rwkv pip package (0.7.4+)
@@ -35,10 +35,10 @@ How to use:
35
 
36
  The difference between World & Raven:
37
  * set pipeline = PIPELINE(model, "rwkv_vocab_v20230424") instead of 20B_tokenizer.json (EXACTLY AS WRITTEN HERE. "rwkv_vocab_v20230424" is included in rwkv 0.7.4+)
38
- * use Question/Answer or User/AI or Human/Bot prompt. **DO NOT USE Bob/Alice or Q/A**
39
  * use **fp32** (will overflow in fp16 at this moment - fixable in future)
40
 
41
- NOTE: the new greedy tokenizer will tokenize '\n\n' as one single token instead of ['\n','\n']
42
 
43
  A good prompt example:
44
  ```
 
27
 
28
  ## Model Description
29
 
30
+ RWKV-4 trained on 100+ world languages (70% English, 15% multilang, 15% code).
31
 
32
  How to use:
33
  * use latest rwkv pip package (0.7.4+)
 
35
 
36
  The difference between World & Raven:
37
  * set pipeline = PIPELINE(model, "rwkv_vocab_v20230424") instead of 20B_tokenizer.json (EXACTLY AS WRITTEN HERE. "rwkv_vocab_v20230424" is included in rwkv 0.7.4+)
38
+ * use Question/Answer or User/AI or Human/Bot prompt for Q&A. **DO NOT USE Bob/Alice or Q/A**
39
  * use **fp32** (will overflow in fp16 at this moment - fixable in future)
40
 
41
+ NOTE: the new greedy tokenizer (https://github.com/BlinkDL/ChatRWKV/blob/main/tokenizer/rwkv_tokenizer.py) will tokenize '\n\n' as one single token instead of ['\n','\n']
42
 
43
  A good prompt example:
44
  ```