Edit model card

stable-diffusion-2-1-GGUF

Original Model

stabilityai/stable-diffusion-2-1

Run with LlamaEdge

  • LlamaEdge version: coming soon

Quantized GGUF Models

Using formats of different precisions will yield results of varying quality.

f32 f16 q8_0 q5_0 q5_1 q4_0 q4_1
Name Quant method Bits Size Use case
v2-1_768-nonema-pruned-Q4_0.gguf Q4_0 2 1.70 GB
v2-1_768-nonema-pruned-Q4_1.gguf Q4_1 3 1.74 GB
v2-1_768-nonema-pruned-Q5_0.gguf Q5_0 3 1.78 GB
v2-1_768-nonema-pruned-Q5_1.gguf Q5_1 3 1.82 GB
v2-1_768-nonema-pruned-Q8_0.gguf Q8_0 4 2.01 GB
v2-1_768-nonema-pruned-f16.gguf f16 4 2.61 GB
v2-1_768-nonema-pruned-f32.gguf f32 4 5.21 GB
Downloads last month
20
GGUF
Model size
1.3B params
Architecture
undefined

4-bit

5-bit

8-bit

16-bit

32-bit

Unable to determine this model's library. Check the docs .

Quantized from