shaowenchen commited on
Commit
cad12a3
1 Parent(s): 0ada7fb

add model files

Browse files
.gitattributes CHANGED
@@ -1,34 +1,36 @@
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
  *.bz2 filter=lfs diff=lfs merge=lfs -text
 
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
 
11
  *.model filter=lfs diff=lfs merge=lfs -text
12
  *.msgpack filter=lfs diff=lfs merge=lfs -text
 
 
13
  *.onnx filter=lfs diff=lfs merge=lfs -text
14
  *.ot filter=lfs diff=lfs merge=lfs -text
15
  *.parquet filter=lfs diff=lfs merge=lfs -text
16
  *.pb filter=lfs diff=lfs merge=lfs -text
 
 
17
  *.pt filter=lfs diff=lfs merge=lfs -text
18
  *.pth filter=lfs diff=lfs merge=lfs -text
19
  *.rar filter=lfs diff=lfs merge=lfs -text
 
20
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
  *.tar.* filter=lfs diff=lfs merge=lfs -text
 
22
  *.tflite filter=lfs diff=lfs merge=lfs -text
23
  *.tgz filter=lfs diff=lfs merge=lfs -text
 
24
  *.xz filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *.tfevents* filter=lfs diff=lfs merge=lfs -text
28
- *.db* filter=lfs diff=lfs merge=lfs -text
29
- *.ark* filter=lfs diff=lfs merge=lfs -text
30
- **/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
31
- **/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
32
- **/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
33
- *.safetensors filter=lfs diff=lfs merge=lfs -text
34
- *.ckpt filter=lfs diff=lfs merge=lfs -text
 
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
  *.model filter=lfs diff=lfs merge=lfs -text
13
  *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
  *.onnx filter=lfs diff=lfs merge=lfs -text
17
  *.ot filter=lfs diff=lfs merge=lfs -text
18
  *.parquet filter=lfs diff=lfs merge=lfs -text
19
  *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
  *.pt filter=lfs diff=lfs merge=lfs -text
23
  *.pth filter=lfs diff=lfs merge=lfs -text
24
  *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
  *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
  *.tflite filter=lfs diff=lfs merge=lfs -text
30
  *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
  *.xz filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
README.md CHANGED
@@ -1,12 +1,58 @@
1
  ---
2
- #以下为文本生成的 "tasks"示例,您可以从此网页中了解更多相关信息:https://modelscope.cn/docs/%E4%BB%BB%E5%8A%A1%E7%9A%84%E4%BB%8B%E7%BB%8D
3
- #tasks:
4
- #- text-generation
5
- license: Apache License 2.0
 
 
 
 
 
 
 
 
 
 
 
 
6
  ---
7
- ###### 该模型当前使用的是默认介绍模版,处于“预发布”阶段,页面仅限所有者可见。
8
- ###### 请根据[模型贡献文档说明](https://www.modelscope.cn/docs/%E5%A6%82%E4%BD%95%E6%92%B0%E5%86%99%E5%A5%BD%E7%94%A8%E7%9A%84%E6%A8%A1%E5%9E%8B%E5%8D%A1%E7%89%87),及时完善模型卡片内容。ModelScope平台将在模型卡片完善后展示。谢谢您的理解。
9
- #### Clone with HTTP
10
- ```bash
11
- git clone https://www.modelscope.cn/shaowenchen/chinese-alpaca-2-7b-16k-gguf.git
12
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ inference: false
3
+ language:
4
+ - zh
5
+ license: apache-2.0
6
+ model_creator: ziqingyang
7
+ model_link: https://huggingface.co/ziqingyang/chinese-llama-2-7b-16k
8
+ model_name: chinese-llama-2-7b-16k
9
+ model_type: llama
10
+ pipeline_tag: text-generation
11
+ quantized_by: shaowenchen
12
+ tags:
13
+ - meta
14
+ - gguf
15
+ - llama
16
+ - llama-2
17
+ - chinese
18
  ---
19
+
20
+ > English | [中文](README_zh.md)
21
+
22
+ ## Provided files
23
+
24
+ | Name | Quant method | Size |
25
+ | ------------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------ |
26
+ | [chinese-llama-2-7b-16k.Q2_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q2_K.gguf) | Q2_K | 2.7 GB |
27
+ | [chinese-llama-2-7b-16k.Q3_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K.gguf) | Q3_K | 3.2 GB |
28
+ | [chinese-llama-2-7b-16k.Q3_K_L.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
+ | [chinese-llama-2-7b-16k.Q3_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
+ | [chinese-llama-2-7b-16k.Q4_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
+ | [chinese-llama-2-7b-16k.Q4_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
+ | [chinese-llama-2-7b-16k.Q4_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K.gguf) | Q4_K | 3.9 GB |
33
+ | [chinese-llama-2-7b-16k.Q4_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
+ | [chinese-llama-2-7b-16k.Q5_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
+ | [chinese-llama-2-7b-16k.Q5_1.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
+ | [chinese-llama-2-7b-16k.Q5_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K.gguf) | Q5_K | 4.6 GB |
37
+ | [chinese-llama-2-7b-16k.Q5_K_S.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
+ | [chinese-llama-2-7b-16k.Q6_K.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q6_K.gguf) | Q6_K | 5.3 GB |
39
+ | [chinese-llama-2-7b-16k.Q8_0.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
+ | [chinese-llama-2-7b-16k.gguf](https://huggingface.co/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.gguf) | full | 13 GB |
41
+
42
+ ## Provided images
43
+
44
+ | Name | Quant method | Size |
45
+ | ---------------------------------------------------------------------------------------------------------------------------------- | ------------ | ------- |
46
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q2_K | 3.68 GB |
47
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K | 4.16 GB |
48
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_L | 4.46 GB |
49
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_S | 3.81 GB |
50
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_0 | 4.7 GB |
51
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K | 4.95 GB |
52
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K_S | 4.73 GB |
53
+
54
+ ```
55
+ docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K
56
+ ```
57
+
58
+ and open http://localhost:8000/docs to view the API documentation.
README_zh.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ inference: false
3
+ language:
4
+ - zh
5
+ license: apache-2.0
6
+ model_creator: ziqingyang
7
+ model_link: https://huggingface.co/ziqingyang/chinese-llama-2-7b-16k
8
+ model_name: chinese-llama-2-7b-16k
9
+ model_type: llama
10
+ pipeline_tag: text-generation
11
+ quantized_by: shaowenchen
12
+ tags:
13
+ - meta
14
+ - gguf
15
+ - llama
16
+ - llama-2
17
+ - chinese
18
+ ---
19
+
20
+ > [English](README.md) | 中文
21
+
22
+ ## 提供的文件
23
+
24
+ | 名称 | 量化方法 | 大小 |
25
+ | ---------------------------------------------------------------------------------------------------------------------------------------- | -------- | ------ |
26
+ | [chinese-llama-2-7b-16k.Q2_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q2_K.gguf) | Q2_K | 2.7 GB |
27
+ | [chinese-llama-2-7b-16k.Q3_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K.gguf) | Q3_K | 3.2 GB |
28
+ | [chinese-llama-2-7b-16k.Q3_K_L.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_L.gguf) | Q3_K_L | 3.5 GB |
29
+ | [chinese-llama-2-7b-16k.Q3_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q3_K_S.gguf) | Q3_K_S | 2.9 GB |
30
+ | [chinese-llama-2-7b-16k.Q4_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_0.gguf) | Q4_0 | 3.7 GB |
31
+ | [chinese-llama-2-7b-16k.Q4_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_1.gguf) | Q4_1 | 4.1 GB |
32
+ | [chinese-llama-2-7b-16k.Q4_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K.gguf) | Q4_K | 3.9 GB |
33
+ | [chinese-llama-2-7b-16k.Q4_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q4_K_S.gguf) | Q4_K_S | 3.7 GB |
34
+ | [chinese-llama-2-7b-16k.Q5_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_0.gguf) | Q5_0 | 4.5 GB |
35
+ | [chinese-llama-2-7b-16k.Q5_1.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_1.gguf) | Q5_1 | 4.9 GB |
36
+ | [chinese-llama-2-7b-16k.Q5_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K.gguf) | Q5_K | 4.6 GB |
37
+ | [chinese-llama-2-7b-16k.Q5_K_S.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q5_K_S.gguf) | Q5_K_S | 4.5 GB |
38
+ | [chinese-llama-2-7b-16k.Q6_K.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q6_K.gguf) | Q6_K | 5.3 GB |
39
+ | [chinese-llama-2-7b-16k.Q8_0.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.Q8_0.gguf) | Q8_0 | 6.9 GB |
40
+ | [chinese-llama-2-7b-16k.gguf](https://www.modelscope.cn/shaowenchen/chinese-llama-2-7b-16k-gguf/blob/main/chinese-llama-2-7b-16k.gguf) | 完整 | 13 GB |
41
+
42
+ ## 提供的镜像
43
+
44
+ | 名称 | 量化方法 | 大小 |
45
+ | ---------------------------------------------------------------------------------------------------------------------------------- | -------- | ------- |
46
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q2_K | 3.68 GB |
47
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K | 4.16 GB |
48
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_L](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_L | 4.46 GB |
49
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q3_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q3_K_S | 3.81 GB |
50
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_0](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_0 | 4.7 GB |
51
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K | 4.95 GB |
52
+ | [shaowenchen/chinese-llama-2-7b-16k-gguf:Q4_K_S](https://hub.docker.com/repository/docker/shaowenchen/chinese-llama-2-7b-16k-gguf/general) | Q4_K_S | 4.73 GB |
53
+
54
+ ```
55
+ docker run --rm -p 8000:8000 shaowenchen/chinese-llama-2-7b-16k-gguf:Q2_K
56
+ ```
57
+
58
+ 并打开 http://localhost:8000/docs 查看 API 文档。
chinese-alpaca-2-7b-16k.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d2f53e19377100339b744d09fcdcdfa7183d24b78b441d07b3880c6efdc399f
3
+ size 2936032800
chinese-alpaca-2-7b-16k.Q3_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a3ebb86bf93ef2ec540a429c525354c3e9c94c652d11effd20e35e778682f6
3
+ size 3417787936
chinese-alpaca-2-7b-16k.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc5c29b3720e8928b1016187ef40dd5f5bb9246f3d28d64dee3be4a627f5036a
3
+ size 3716894240
chinese-alpaca-2-7b-16k.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d93b6324ac9b39caf598ba9b57b3351b7588f413fcb684756915f3608aa55e4f
3
+ size 3068087840
chinese-alpaca-2-7b-16k.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03c8268e5ebb2179efc8f8a090b1ef61592e71a70d21041d76eebaa6056cee3f
3
+ size 3958263328
chinese-alpaca-2-7b-16k.Q4_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc6b58b7b11111c7f25c7681d6239a9b671c0bb443b296b2fa2e00ff5ea30368
3
+ size 4377169440
chinese-alpaca-2-7b-16k.Q4_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26261d868b9bdb9a9ce8cd8f696240b675b6311ba4235f8d3fe87ff2eb132f4e
3
+ size 4213460512
chinese-alpaca-2-7b-16k.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a096e1dc0e2f48a428ae540a355392daf2d4bf105c5ad900e83181a7b97cff32
3
+ size 3989196320
chinese-alpaca-2-7b-16k.Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8768fa4121619202017a79c563de44ff9a492e0086ffe2db43c9959da92ee3a2
3
+ size 4796075552
chinese-alpaca-2-7b-16k.Q5_1.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4245225ab47cc2efed40808645c5d119495e5f38d3b2334ce4a08615e845d284
3
+ size 5214981664
chinese-alpaca-2-7b-16k.Q5_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79a9e4d6511850298c50bfbcd2baf81333638d55cc2bfe498974d96357bee3c8
3
+ size 4927540768
chinese-alpaca-2-7b-16k.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6955405c00ade27929e25da328671f69eaca7e3c50bf562baf367cfe8b61ce6
3
+ size 4796075552
chinese-alpaca-2-7b-16k.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd073f81de2e9c74343c01507ac62d2916efb796a6c7ad98b29fc91e1ceb30b8
3
+ size 5686251040
chinese-alpaca-2-7b-16k.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c3e53e9ff168674849211aaa30260bb3f031fe7bd95a99b4c91503e3939fbcb
3
+ size 7364365856
chinese-alpaca-2-7b-16k.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5253f49af145765eadd2cc59690bd500a789eabfe3e74258fc9ea5dd6ce9ee30
3
+ size 13860294112
configuration.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "framework": "pytorch",
3
+ "task": "text-generation",
4
+ "model": {
5
+ "type": "llama2"
6
+ }
7
+ }