shaowenchen commited on
Commit
368d708
1 Parent(s): e9ad72e
README.md CHANGED
@@ -17,42 +17,40 @@ tags:
17
  - chinese
18
  ---
19
 
20
- > English | [中文](README_zh.md)
21
-
22
  ## Provided files
23
 
24
- | Name | Quant method | Size |
25
- | ------------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------ |
26
- | [chinese-llama-2-7b.Q2_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q2_K.gguf) | Q2_K | 2.7 GB |
27
- | [chinese-llama-2-7b.Q3_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K.gguf) | Q3_K | 3.2 GB |
28
- | [chinese-llama-2-7b.Q3_K_L.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
- | [chinese-llama-2-7b.Q3_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
- | [chinese-llama-2-7b.Q4_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
- | [chinese-llama-2-7b.Q4_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
- | [chinese-llama-2-7b.Q4_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_K.gguf) | Q4_K | 3.9 GB |
33
- | [chinese-llama-2-7b.Q4_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
- | [chinese-llama-2-7b.Q5_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
- | [chinese-llama-2-7b.Q5_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
- | [chinese-llama-2-7b.Q5_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_K.gguf) | Q5_K | 4.6 GB |
37
- | [chinese-llama-2-7b.Q5_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
- | [chinese-llama-2-7b.Q6_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q6_K.gguf) | Q6_K | 5.3 GB |
39
- | [chinese-llama-2-7b.Q8_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
- | [chinese-llama-2-7b.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.gguf) | full | 13 GB |
41
 
42
  ## Provided images
43
 
44
- | Name | Quant method | Size |
45
- | ---------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------- |
46
- | [shaowenchen/chinese-llama-2-7b-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q2_K | 3.68 GB |
47
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K | 4.16 GB |
48
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K_L | 4.46 GB |
49
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K_S | 3.81 GB |
50
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_0 | 4.7 GB |
51
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_K | 4.95 GB |
52
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_K_S | 4.73 GB |
53
 
54
  ```
55
  docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-gguf:Q2_K
56
  ```
57
 
58
- and open http://localhost:8000/docs to view the API documentation.
 
17
  - chinese
18
  ---
19
 
 
 
20
  ## Provided files
21
 
22
+ | Name | Quant method | Size |
23
+ | ------------------------------ | ------------ | ------ |
24
+ | chinese-llama-2-7b.Q2_K.gguf | Q2_K | 2.7 GB |
25
+ | chinese-llama-2-7b.Q3_K.gguf | Q3_K | 3.2 GB |
26
+ | chinese-llama-2-7b.Q3_K_L.gguf | Q3_K_L | 3.5 GB |
27
+ | chinese-llama-2-7b.Q3_K_S.gguf | Q3_K_S | 2.9 GB |
28
+ | chinese-llama-2-7b.Q4_0.gguf | Q4_0 | 3.7 GB |
29
+ | chinese-llama-2-7b.Q4_1.gguf | Q4_1 | 4.1 GB |
30
+ | chinese-llama-2-7b.Q4_K.gguf | Q4_K | 3.9 GB |
31
+ | chinese-llama-2-7b.Q4_K_S.gguf | Q4_K_S | 3.7 GB |
32
+ | chinese-llama-2-7b.Q5_0.gguf | Q5_0 | 4.5 GB |
33
+ | chinese-llama-2-7b.Q5_1.gguf | Q5_1 | 4.9 GB |
34
+ | chinese-llama-2-7b.Q5_K.gguf | Q5_K | 4.6 GB |
35
+ | chinese-llama-2-7b.Q5_K_S.gguf | Q5_K_S | 4.5 GB |
36
+ | chinese-llama-2-7b.Q6_K.gguf | Q6_K | 5.3 GB |
37
+ | chinese-llama-2-7b.Q8_0.gguf | Q8_0 | 6.9 GB |
38
+ | chinese-llama-2-7b.gguf | full | 13 GB |
39
 
40
  ## Provided images
41
 
42
+ | Name | Quant method | Size |
43
+ | -------------------------------------------- | ------------ | ------- |
44
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q2_K` | Q2_K | 3.68 GB |
45
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q3_K` | Q3_K | 4.16 GB |
46
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q3_K_L` | Q3_K_L | 4.46 GB |
47
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q3_K_S` | Q3_K_S | 3.81 GB |
48
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q4_0` | Q4_0 | 4.7 GB |
49
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q4_K` | Q4_K | 4.95 GB |
50
+ | `shaowenchen/chinese-llama-2-7b-gguf:Q4_K_S` | Q4_K_S | 4.73 GB |
51
 
52
  ```
53
  docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-gguf:Q2_K
54
  ```
55
 
56
+ and you can view http://localhost:8000/docs to see the swagger UI.
README_zh.md DELETED
@@ -1,58 +0,0 @@
1
- ---
2
- inference: false
3
- language:
4
- - zh
5
- license: apache-2.0
6
- model_creator: ziqingyang
7
- model_link: https://www.modelscope.cn/ziqingyang/chinese-llama-2-7b
8
- model_name: chinese-llama-2-7b
9
- model_type: llama
10
- pipeline_tag: text-generation
11
- quantized_by: shaowenchen
12
- tags:
13
- - meta
14
- - gguf
15
- - llama
16
- - llama-2
17
- - chinese
18
- ---
19
-
20
- > [English](README.md) | 中文
21
-
22
- ## 提供的文件
23
-
24
- | 名称 | 量化方法 | 大小 |
25
- | ---------------------------------------------------------------------------------------------------------------------------------------- | -------- | ------ |
26
- | [chinese-llama-2-7b.Q2_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q2_K.gguf) | Q2_K | 2.7 GB |
27
- | [chinese-llama-2-7b.Q3_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K.gguf) | Q3_K | 3.2 GB |
28
- | [chinese-llama-2-7b.Q3_K_L.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
- | [chinese-llama-2-7b.Q3_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
- | [chinese-llama-2-7b.Q4_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
- | [chinese-llama-2-7b.Q4_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
- | [chinese-llama-2-7b.Q4_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_K.gguf) | Q4_K | 3.9 GB |
33
- | [chinese-llama-2-7b.Q4_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
- | [chinese-llama-2-7b.Q5_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
- | [chinese-llama-2-7b.Q5_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
- | [chinese-llama-2-7b.Q5_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_K.gguf) | Q5_K | 4.6 GB |
37
- | [chinese-llama-2-7b.Q5_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
- | [chinese-llama-2-7b.Q6_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q6_K.gguf) | Q6_K | 5.3 GB |
39
- | [chinese-llama-2-7b.Q8_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
- | [chinese-llama-2-7b.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-GGUF/blob/main/chinese-llama-2-7b.gguf) | 完整 | 13 GB |
41
-
42
- ## 提供的镜像
43
-
44
- | 名称 | 量化方法 | 大小 |
45
- | ---------------------------------------------------------------------------------------------------------------------------------- | -------- | ------- |
46
- | [shaowenchen/chinese-llama-2-7b-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q2_K | 3.68 GB |
47
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K | 4.16 GB |
48
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K_L | 4.46 GB |
49
- | [shaowenchen/chinese-llama-2-7b-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q3_K_S | 3.81 GB |
50
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_0 | 4.7 GB |
51
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_K | 4.95 GB |
52
- | [shaowenchen/chinese-llama-2-7b-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-gguf/general) | Q4_K_S | 4.73 GB |
53
-
54
- ```
55
- docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-gguf:Q2_K
56
- ```
57
-
58
- 并打开 http://localhost:8000/docs 查看 API 文档。
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
chinese-llama-2-7b.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beb5c78a461819fae3d8c51c72293c70237ae0af3bd7e0a69826932825122c6b
3
+ size 2936032800
chinese-llama-2-7b.Q3_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf785014e4ff2c6f18aed9467939ba41457bf2899abf6e9b95f74ce53307dd3
3
+ size 3417787936
chinese-llama-2-7b.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24abe12dce1a3981772647686f668c218190217aeef800358c4c0e43ce9e51c
3
+ size 3716894240
chinese-llama-2-7b.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd9e7a77814ca9acca0a85de0fd41973fd8a6f59c8bbbf3529699f450d8a322a
3
+ size 3068087840
chinese-llama-2-7b.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b28b7c388426594e6cf7fb5b4629b3008d7f16154838aff6cb3a9f3be8bed5a5
3
+ size 3958263328
chinese-llama-2-7b.Q4_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01bacb9ed904eb89b48d8f85a7f201c85f40f91f50099df5c5fcbbe999d52d65
3
+ size 4377169440
chinese-llama-2-7b.Q4_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cafc3f11b81e3a3b2d6ccc90d02385c09cdc221a437edb57380939537f26ed85
3
+ size 4213460512
chinese-llama-2-7b.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63824393c46d4756d41c46751cdaf764523c3d10209695c880ee947789618448
3
+ size 3989196320
chinese-llama-2-7b.Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a93664a3ac9ec07abfc1cbbf6c239dac355869330780a92978a8ff1c0484930
3
+ size 4796075552
chinese-llama-2-7b.Q5_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75776288868e502f52110fdc251f93fcc7d53aaf1cd458db353c2603d4f9292a
3
+ size 5214981664
chinese-llama-2-7b.Q5_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67b365ba50f6d94e9aef066ffe090be1fedfeb0787c4f677e3f648b8531b2e19
3
+ size 4927540768
chinese-llama-2-7b.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32539dad763566a452a9efd50e1c98ecf8a5ff9d3ef03ed9bf80675e70b0ae4e
3
+ size 4796075552
chinese-llama-2-7b.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d6bc82e9a0a69e96d7865008ba98dc4a38adc1dceef3fd2c64e0719d2ce4fc7
3
+ size 5686251040
chinese-llama-2-7b.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0182d4188160f6c283a4b1e624f9b97e80b271bddc9ecbd00498ddb318d57ebf
3
+ size 7364365856
chinese-llama-2-7b.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acbd95212ffc6215b91308065ca49bcb030c8f8b87395864e9c06449324c4c39
3
+ size 13860294144
configuration.json CHANGED
@@ -3,5 +3,8 @@
3
  "task": "text-generation",
4
  "model": {
5
  "type": "llama2"
 
 
 
6
  }
7
  }
 
3
  "task": "text-generation",
4
  "model": {
5
  "type": "llama2"
6
+ },
7
+ "pipeline": {
8
+ "type": "chinese-llama-2-7b-text-generation-pipe"
9
  }
10
  }