TheBloke commited on
Commit
c2d4b19
1 Parent(s): 4fd7ab4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -1,6 +1,13 @@
1
  ---
2
  inference: false
3
- license: other
 
 
 
 
 
 
 
4
  ---
5
 
6
  <!-- header start -->
@@ -26,7 +33,7 @@ Please note that these GGMLs are **not compatible with llama.cpp, or currently w
26
  ## Repositories available
27
 
28
  * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/WizardCoder-Guanaco-15B-V1.0-GPTQ)
29
- * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/WizardCoder-Guanaco-15B-V1.0-GGML)
30
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/LoupGarou/WizardCoder-Guanaco-15B-V1.0)
31
 
32
  ## Prompt template: Alpaca
@@ -69,7 +76,6 @@ As other options become available I will endeavour to update them here (do let m
69
  | wizardcoder-guanaco-15b-v1.0.ggmlv1.q5_1.bin | q5_1 | 5 | 14.26 GB| 16.76 GB | 5-bit. Even higher accuracy, resource usage and slower inference. |
70
  | wizardcoder-guanaco-15b-v1.0.ggmlv1.q8_0.bin | q8_0 | 8 | 20.11 GB| 22.61 GB | 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. |
71
 
72
-
73
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
74
 
75
  <!-- footer start -->
 
1
  ---
2
  inference: false
3
+ language:
4
+ - en
5
+ datasets:
6
+ - guanaco
7
+ model_hub_library:
8
+ - transformers
9
+ license:
10
+ - apache-2.0
11
  ---
12
 
13
  <!-- header start -->
 
33
  ## Repositories available
34
 
35
  * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/WizardCoder-Guanaco-15B-V1.0-GPTQ)
36
+ * [4, 5, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/WizardCoder-Guanaco-15B-V1.0-GGML)
37
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/LoupGarou/WizardCoder-Guanaco-15B-V1.0)
38
 
39
  ## Prompt template: Alpaca
 
76
  | wizardcoder-guanaco-15b-v1.0.ggmlv1.q5_1.bin | q5_1 | 5 | 14.26 GB| 16.76 GB | 5-bit. Even higher accuracy, resource usage and slower inference. |
77
  | wizardcoder-guanaco-15b-v1.0.ggmlv1.q8_0.bin | q8_0 | 8 | 20.11 GB| 22.61 GB | 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. |
78
 
 
79
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
80
 
81
  <!-- footer start -->