JustinLin610 commited on
Commit
6f9a97e
·
verified ·
1 Parent(s): 42c72d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: other
3
- license_name: tongyi-qianwen
4
  license_link: >-
5
- https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT
6
  language:
7
  - en
8
  pipeline_tag: text-generation
@@ -30,6 +30,7 @@ For more details, please refer to our blog post and github repo.
30
 
31
  Qwen2 is a language model series including decoder language models of different model sizes. For each size, we release the base language model and the aligned chat model. It is based on the Transformer architecture with SwiGLU activation, attention QKV bias, group query attention, mixture of sliding window attention and full attention, etc. Additionally, we have an improved tokenizer adaptive to multiple natural languages and codes. For the beta version, temporarily we did not include GQA and the mixture of SWA and full attention.
32
 
 
33
  ## Requirements
34
 
35
  The code of Qwen2 has been in the latest Hugging face transformers and we advise you to install `transformers>=4.37.0`, or you might encounter the following error:
@@ -41,6 +42,7 @@ KeyError: 'qwen2'
41
  ## Usage
42
 
43
  We do not advise you to use base language models for text generation. Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., on this model.
 
44
 
45
 
46
  ## Citation
 
1
  ---
2
  license: other
3
+ license_name: tongyi-qianwen-research
4
  license_link: >-
5
+ https://huggingface.co/Qwen/Qwen2-beta-0_5B/blob/main/LICENSE
6
  language:
7
  - en
8
  pipeline_tag: text-generation
 
30
 
31
  Qwen2 is a language model series including decoder language models of different model sizes. For each size, we release the base language model and the aligned chat model. It is based on the Transformer architecture with SwiGLU activation, attention QKV bias, group query attention, mixture of sliding window attention and full attention, etc. Additionally, we have an improved tokenizer adaptive to multiple natural languages and codes. For the beta version, temporarily we did not include GQA and the mixture of SWA and full attention.
32
 
33
+
34
  ## Requirements
35
 
36
  The code of Qwen2 has been in the latest Hugging face transformers and we advise you to install `transformers>=4.37.0`, or you might encounter the following error:
 
42
  ## Usage
43
 
44
  We do not advise you to use base language models for text generation. Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., on this model.
45
+ <br>
46
 
47
 
48
  ## Citation