apepkuss79 commited on
Commit
081e986
1 Parent(s): 953d84c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -86
README.md CHANGED
@@ -1,86 +1,87 @@
1
- ---
2
- base_model: CompVis/stable-diffusion-v-1-4-original
3
- license: creativeml-openrail-m
4
- library_name: "stable-diffusion"
5
- inference: false
6
- model_creator: runwayml
7
- model_name: stable-diffusion-v-1-4-original
8
- quantized_by: Second State Inc.
9
- tags:
10
- - stable-diffusion
11
- - text-to-image
12
- ---
13
-
14
- <!-- header start -->
15
- <!-- 200823 -->
16
- <div style="width: auto; margin-left: auto; margin-right: auto">
17
- <img src="https://github.com/LlamaEdge/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
18
- </div>
19
- <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
20
- <!-- header end -->
21
-
22
- # stable-diffusion-v-1-4-GGUF
23
-
24
- ## Original Model
25
-
26
- [CompVis/stable-diffusion-v-1-4-original](https://huggingface.co/CompVis/stable-diffusion-v-1-4-original)
27
-
28
- ## Run with LlamaEdge
29
-
30
- - LlamaEdge version: coming soon
31
-
32
- <!-- - LlamaEdge version: [v0.12.2](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.12.2) and above
33
-
34
- - Prompt template
35
-
36
- - Prompt type: `chatml`
37
-
38
- - Prompt string
39
-
40
- ```text
41
- <|im_start|>system
42
- {system_message}<|im_end|>
43
- <|im_start|>user
44
- {prompt}<|im_end|>
45
- <|im_start|>assistant
46
- ```
47
-
48
- - Context size: `4096`
49
-
50
- - Run as LlamaEdge service
51
-
52
- ```bash
53
- wasmedge --dir .:. --nn-preload default:GGML:AUTO:stablelm-2-12b-chat-Q5_K_M.gguf \
54
- llama-api-server.wasm \
55
- --prompt-template chatml \
56
- --ctx-size 4096 \
57
- --model-name stablelm-2-12b-chat
58
- ```
59
-
60
- - Run as LlamaEdge command app
61
-
62
- ```bash
63
- wasmedge --dir .:. \
64
- --nn-preload default:GGML:AUTO:stablelm-2-12b-chat-Q5_K_M.gguf \
65
- llama-chat.wasm \
66
- --prompt-template chatml \
67
- --ctx-size 4096
68
- ``` -->
69
-
70
- ## Quantized GGUF Models
71
-
72
- Using formats of different precisions will yield results of varying quality.
73
-
74
- | f32 | f16 |q8_0 |q5_0 |q5_1 |q4_0 |q4_1 |
75
- | ---- |---- |---- |---- |---- |---- |---- |
76
- | ![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/f32.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/f16.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q8_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q5_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q5_1.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q4_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q4_1.png?raw=true) |
77
-
78
- | Name | Quant method | Bits | Size | Use case |
79
- | ---- | ---- | ---- | ---- | ----- |
80
- | [stable-diffusion-v1-5-pruned-emaonly-Q4_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q4_0.gguf) | Q4_0 | 2 | 1.57 GB | |
81
- | [stable-diffusion-v1-5-pruned-emaonly-Q4_1.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q4_1.gguf) | Q4_1 | 3 | 1.59 GB | |
82
- | [stable-diffusion-v1-5-pruned-emaonly-Q5_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q5_0.gguf) | Q5_0 | 3 | 1.62 GB | |
83
- | [stable-diffusion-v1-5-pruned-emaonly-Q5_1.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q5_1.gguf) | Q5_1 | 3 | 1.64 GB | |
84
- | [stable-diffusion-v1-5-pruned-emaonly-Q8_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q8_0.gguf) | Q8_0 | 4 | 1.76 GB | |
85
- | [stable-diffusion-v1-5-pruned-emaonly-f16.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-f16.gguf) | f16 | 4 | 2.13 GB | |
86
- | [stable-diffusion-v1-5-pruned-emaonly-f32.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-f32.gguf) | f32 | 4 | 4.27 GB | |
 
 
1
+ ---
2
+ base_model: CompVis/stable-diffusion-v-1-4-original
3
+ license: creativeml-openrail-m
4
+ library_name: "stable-diffusion"
5
+ inference: false
6
+ model_creator: runwayml
7
+ model_name: stable-diffusion-v-1-4-original
8
+ quantized_by: Second State Inc.
9
+ tags:
10
+ - stable-diffusion
11
+ - text-to-image
12
+ ---
13
+
14
+ <!-- header start -->
15
+ <!-- 200823 -->
16
+ <div style="width: auto; margin-left: auto; margin-right: auto">
17
+ <img src="https://github.com/LlamaEdge/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
18
+ </div>
19
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
20
+ <!-- header end -->
21
+
22
+ # stable-diffusion-v-1-4-GGUF
23
+
24
+ ## Original Model
25
+
26
+ [CompVis/stable-diffusion-v-1-4-original](https://huggingface.co/CompVis/stable-diffusion-v-1-4-original)
27
+
28
+ ## Run with `sd-api-server`
29
+
30
+ Go to the [sd-api-server](https://github.com/LlamaEdge/sd-api-server/blob/main/README.md) repository for more information.
31
+
32
+
33
+ <!-- - LlamaEdge version: [v0.12.2](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.12.2) and above
34
+
35
+ - Prompt template
36
+
37
+ - Prompt type: `chatml`
38
+
39
+ - Prompt string
40
+
41
+ ```text
42
+ <|im_start|>system
43
+ {system_message}<|im_end|>
44
+ <|im_start|>user
45
+ {prompt}<|im_end|>
46
+ <|im_start|>assistant
47
+ ```
48
+
49
+ - Context size: `4096`
50
+
51
+ - Run as LlamaEdge service
52
+
53
+ ```bash
54
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:stablelm-2-12b-chat-Q5_K_M.gguf \
55
+ llama-api-server.wasm \
56
+ --prompt-template chatml \
57
+ --ctx-size 4096 \
58
+ --model-name stablelm-2-12b-chat
59
+ ```
60
+
61
+ - Run as LlamaEdge command app
62
+
63
+ ```bash
64
+ wasmedge --dir .:. \
65
+ --nn-preload default:GGML:AUTO:stablelm-2-12b-chat-Q5_K_M.gguf \
66
+ llama-chat.wasm \
67
+ --prompt-template chatml \
68
+ --ctx-size 4096
69
+ ``` -->
70
+
71
+ ## Quantized GGUF Models
72
+
73
+ Using formats of different precisions will yield results of varying quality.
74
+
75
+ | f32 | f16 |q8_0 |q5_0 |q5_1 |q4_0 |q4_1 |
76
+ | ---- |---- |---- |---- |---- |---- |---- |
77
+ | ![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/f32.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/f16.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q8_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q5_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q5_1.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q4_0.png?raw=true) |![](https://github.com/leejet/stable-diffusion.cpp/blob/master/assets/q4_1.png?raw=true) |
78
+
79
+ | Name | Quant method | Bits | Size | Use case |
80
+ | ---- | ---- | ---- | ---- | ----- |
81
+ | [stable-diffusion-v1-5-pruned-emaonly-Q4_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q4_0.gguf) | Q4_0 | 2 | 1.57 GB | |
82
+ | [stable-diffusion-v1-5-pruned-emaonly-Q4_1.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q4_1.gguf) | Q4_1 | 3 | 1.59 GB | |
83
+ | [stable-diffusion-v1-5-pruned-emaonly-Q5_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q5_0.gguf) | Q5_0 | 3 | 1.62 GB | |
84
+ | [stable-diffusion-v1-5-pruned-emaonly-Q5_1.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q5_1.gguf) | Q5_1 | 3 | 1.64 GB | |
85
+ | [stable-diffusion-v1-5-pruned-emaonly-Q8_0.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-Q8_0.gguf) | Q8_0 | 4 | 1.76 GB | |
86
+ | [stable-diffusion-v1-5-pruned-emaonly-f16.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-f16.gguf) | f16 | 4 | 2.13 GB | |
87
+ | [stable-diffusion-v1-5-pruned-emaonly-f32.gguf](https://huggingface.co/second-state/stable-diffusion-v1-5-GGUF/blob/main/stable-diffusion-v1-5-pruned-emaonly-f32.gguf) | f32 | 4 | 4.27 GB | |