ChrisGoringe commited on
Commit
1e8e3e4
1 Parent(s): 867b980

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -16,18 +16,20 @@ They were created using the [convert.py script](https://github.com/chrisgoringe/
16
  They can be loaded in ComfyUI using the [ComfyUI GGUF Nodes](https://github.com/city96/ComfyUI-GGUF). Just put the gguf files in your
17
  models/unet directory.
18
 
 
 
19
  ## Naming convention (mx for 'mixed')
20
 
21
  [original_model_name]_mxNN_N.gguf
22
 
23
- where NN_N is the approximate reduction in VRAM usage compared the full 16 bit version.
24
-
25
  - 9_0 might just fit on a 16GB card
26
  - 10_6 is a good balance for 16GB cards,
27
  - 12_0 is roughly the size of an 8 bit model,
28
  - 14_1 should work for 12 GB cards
29
  - 15_2 is fully quantised to Q4_1
30
-
31
  ## How is this optimised?
32
 
33
  The process for optimisation is as follows:
 
16
  They can be loaded in ComfyUI using the [ComfyUI GGUF Nodes](https://github.com/city96/ComfyUI-GGUF). Just put the gguf files in your
17
  models/unet directory.
18
 
19
+ ## Bigger numbers in the name = smaller model!
20
+
21
  ## Naming convention (mx for 'mixed')
22
 
23
  [original_model_name]_mxNN_N.gguf
24
 
25
+ where NN_N is the approximate *reduction* in VRAM usage compared the full 16 bit version.
26
+ ```
27
  - 9_0 might just fit on a 16GB card
28
  - 10_6 is a good balance for 16GB cards,
29
  - 12_0 is roughly the size of an 8 bit model,
30
  - 14_1 should work for 12 GB cards
31
  - 15_2 is fully quantised to Q4_1
32
+ ```
33
  ## How is this optimised?
34
 
35
  The process for optimisation is as follows: