shaowenchen commited on
Commit
2f571d8
1 Parent(s): cad12a3

fix meta json

Browse files
Files changed (3) hide show
  1. README.md +32 -30
  2. README_zh.md +0 -58
  3. configuration.json +3 -0
README.md CHANGED
@@ -2,13 +2,15 @@
2
  inference: false
3
  language:
4
  - zh
5
- license: apache-2.0
6
  model_creator: ziqingyang
7
  model_link: https://huggingface.co/ziqingyang/chinese-llama-2-7b-16k
8
  model_name: chinese-llama-2-7b-16k
9
  model_type: llama
10
  pipeline_tag: text-generation
11
  quantized_by: shaowenchen
 
 
12
  tags:
13
  - meta
14
  - gguf
@@ -17,42 +19,42 @@ tags:
17
  - chinese
18
  ---
19
 
20
- > English | [中文](README_zh.md)
21
-
22
  ## Provided files
23
 
24
- | Name | Quant method | Size |
25
- | ------------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------ |
26
- | [chinese-llama-2-7b-16k.Q2_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q2_K.gguf) | Q2_K | 2.7 GB |
27
- | [chinese-llama-2-7b-16k.Q3_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K.gguf) | Q3_K | 3.2 GB |
28
- | [chinese-llama-2-7b-16k.Q3_K_L.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
- | [chinese-llama-2-7b-16k.Q3_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
- | [chinese-llama-2-7b-16k.Q4_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
- | [chinese-llama-2-7b-16k.Q4_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
- | [chinese-llama-2-7b-16k.Q4_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K.gguf) | Q4_K | 3.9 GB |
33
- | [chinese-llama-2-7b-16k.Q4_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
- | [chinese-llama-2-7b-16k.Q5_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
- | [chinese-llama-2-7b-16k.Q5_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
- | [chinese-llama-2-7b-16k.Q5_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K.gguf) | Q5_K | 4.6 GB |
37
- | [chinese-llama-2-7b-16k.Q5_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
- | [chinese-llama-2-7b-16k.Q6_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q6_K.gguf) | Q6_K | 5.3 GB |
39
- | [chinese-llama-2-7b-16k.Q8_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
- | [chinese-llama-2-7b-16k.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.gguf) | full | 13 GB |
41
 
42
  ## Provided images
43
 
44
- | Name | Quant method | Size |
45
- | ---------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------- |
46
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q2_K | 3.68 GB |
47
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K | 4.16 GB |
48
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_L | 4.46 GB |
49
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_S | 3.81 GB |
50
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_0 | 4.7 GB |
51
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K | 4.95 GB |
52
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K_S | 4.73 GB |
 
 
53
 
54
  ```
55
  docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K
56
  ```
57
 
58
- and open http://localhost:8000/docs to view the API documentation.
 
2
  inference: false
3
  language:
4
  - zh
5
+ license: other
6
  model_creator: ziqingyang
7
  model_link: https://huggingface.co/ziqingyang/chinese-llama-2-7b-16k
8
  model_name: chinese-llama-2-7b-16k
9
  model_type: llama
10
  pipeline_tag: text-generation
11
  quantized_by: shaowenchen
12
+ tasks:
13
+ - text2text-generation
14
  tags:
15
  - meta
16
  - gguf
 
19
  - chinese
20
  ---
21
 
 
 
22
  ## Provided files
23
 
24
+ | Name | Quant method | Size |
25
+ | ---------------------------------- | ------------ | ------ |
26
+ | chinese-llama-2-7b-16k.Q2_K.gguf | Q2_K | 2.7 GB |
27
+ | chinese-llama-2-7b-16k.Q3_K.gguf | Q3_K | 3.2 GB |
28
+ | chinese-llama-2-7b-16k.Q3_K_L.gguf | Q3_K_L | 3.5 GB |
29
+ | chinese-llama-2-7b-16k.Q3_K_S.gguf | Q3_K_S | 2.9 GB |
30
+ | chinese-llama-2-7b-16k.Q4_0.gguf | Q4_0 | 3.7 GB |
31
+ | chinese-llama-2-7b-16k.Q4_1.gguf | Q4_1 | 4.1 GB |
32
+ | chinese-llama-2-7b-16k.Q4_K.gguf | Q4_K | 3.9 GB |
33
+ | chinese-llama-2-7b-16k.Q4_K_S.gguf | Q4_K_S | 3.7 GB |
34
+ | chinese-llama-2-7b-16k.Q5_0.gguf | Q5_0 | 4.5 GB |
35
+ | chinese-llama-2-7b-16k.Q5_1.gguf | Q5_1 | 4.9 GB |
36
+ | chinese-llama-2-7b-16k.Q5_K.gguf | Q5_K | 4.6 GB |
37
+ | chinese-llama-2-7b-16k.Q5_K_S.gguf | Q5_K_S | 4.5 GB |
38
+ | chinese-llama-2-7b-16k.Q6_K.gguf | Q6_K | 5.3 GB |
39
+ | chinese-llama-2-7b-16k.Q8_0.gguf | Q8_0 | 6.9 GB |
40
+ | chinese-llama-2-7b-16k.gguf | full | 13 GB |
41
 
42
  ## Provided images
43
 
44
+ | Name | Quant method | Size |
45
+ | ---------------------------------------------- | ------------ | ------- |
46
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K | Q2_K | 3.68 GB |
47
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K | Q3_K | 4.16 GB |
48
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_L | Q3_K_L | 4.46 GB |
49
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_S | Q3_K_S | 3.81 GB |
50
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_0 | Q4_0 | 4.7 GB |
51
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K | Q4_K | 4.95 GB |
52
+ | shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K_S | Q4_K_S | 4.73 GB |
53
+
54
+ Usage:
55
 
56
  ```
57
  docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K
58
  ```
59
 
60
+ and you can view http://localhost:8000/docs to see the swagger UI.
README_zh.md DELETED
@@ -1,58 +0,0 @@
1
- ---
2
- inference: false
3
- language:
4
- - zh
5
- license: apache-2.0
6
- model_creator: ziqingyang
7
- model_link: https://huggingface.co/ziqingyang/chinese-llama-2-7b-16k
8
- model_name: chinese-llama-2-7b-16k
9
- model_type: llama
10
- pipeline_tag: text-generation
11
- quantized_by: shaowenchen
12
- tags:
13
- - meta
14
- - gguf
15
- - llama
16
- - llama-2
17
- - chinese
18
- ---
19
-
20
- > [English](README.md) | 中文
21
-
22
- ## 提供的文件
23
-
24
- | 名称 | 量化方法 | 大小 |
25
- | ---------------------------------------------------------------------------------------------------------------------------------------- | -------- | ------ |
26
- | [chinese-llama-2-7b-16k.Q2_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q2_K.gguf) | Q2_K | 2.7 GB |
27
- | [chinese-llama-2-7b-16k.Q3_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K.gguf) | Q3_K | 3.2 GB |
28
- | [chinese-llama-2-7b-16k.Q3_K_L.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
- | [chinese-llama-2-7b-16k.Q3_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
- | [chinese-llama-2-7b-16k.Q4_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
- | [chinese-llama-2-7b-16k.Q4_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
- | [chinese-llama-2-7b-16k.Q4_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K.gguf) | Q4_K | 3.9 GB |
33
- | [chinese-llama-2-7b-16k.Q4_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
- | [chinese-llama-2-7b-16k.Q5_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
- | [chinese-llama-2-7b-16k.Q5_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
- | [chinese-llama-2-7b-16k.Q5_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K.gguf) | Q5_K | 4.6 GB |
37
- | [chinese-llama-2-7b-16k.Q5_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
- | [chinese-llama-2-7b-16k.Q6_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q6_K.gguf) | Q6_K | 5.3 GB |
39
- | [chinese-llama-2-7b-16k.Q8_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
- | [chinese-llama-2-7b-16k.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.gguf) | 完整 | 13 GB |
41
-
42
- ## 提供的镜像
43
-
44
- | 名称 | 量化方法 | 大小 |
45
- | ---------------------------------------------------------------------------------------------------------------------------------- | -------- | ------- |
46
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q2_K | 3.68 GB |
47
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K | 4.16 GB |
48
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_L | 4.46 GB |
49
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_S | 3.81 GB |
50
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_0 | 4.7 GB |
51
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K | 4.95 GB |
52
- | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K_S | 4.73 GB |
53
-
54
- ```
55
- docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K
56
- ```
57
-
58
- 并打开 http://localhost:8000/docs 查看 API 文档。
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
configuration.json CHANGED
@@ -3,5 +3,8 @@
3
  "task": "text-generation",
4
  "model": {
5
  "type": "llama2"
 
 
 
6
  }
7
  }
 
3
  "task": "text-generation",
4
  "model": {
5
  "type": "llama2"
6
+ },
7
+ "pipeline": {
8
+ "type": "chinese-alpaca-2-7b-16k-text-generation-pipe"
9
  }
10
  }