Transformers
GGUF
llama
TheBloke commited on
Commit
6a15066
1 Parent(s): bc17c31

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -2
README.md CHANGED
@@ -36,8 +36,6 @@ quantized_by: TheBloke
36
 
37
  This repo contains GGUF format model files for [Jon Durbin's Airoboros L2 70B 2.2](https://huggingface.co/jondurbin/airoboros-l2-70b-2.2).
38
 
39
- Note: these GGUF models were re-created on 15th September, as Jon has re-uploaded the original source weights. The first source upload was based on a new method for merging qLoRA weights. This has proved to cause problems, and therefore Jon has re-uploaded the weights in the usual way, and I have re-done all my GGUF and GPTQ models.
40
-
41
  <!-- description end -->
42
  <!-- README_GGUF.md-about-gguf start -->
43
  ### About GGUF
@@ -72,10 +70,12 @@ Here is an incomplate list of clients and libraries that are known to support GG
72
  A chat.
73
  USER: {prompt}
74
  ASSISTANT:
 
75
  ```
76
 
77
  <!-- prompt-template end -->
78
 
 
79
  <!-- compatibility_gguf start -->
80
  ## Compatibility
81
 
@@ -154,6 +154,63 @@ del airoboros-l2-70b-2.2.Q8_0.gguf-split-a airoboros-l2-70b-2.2.Q8_0.gguf-split-
154
  </details>
155
  <!-- README_GGUF.md-provided-files end -->
156
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
  <!-- README_GGUF.md-how-to-run start -->
158
  ## Example `llama.cpp` command
159
 
 
36
 
37
  This repo contains GGUF format model files for [Jon Durbin's Airoboros L2 70B 2.2](https://huggingface.co/jondurbin/airoboros-l2-70b-2.2).
38
 
 
 
39
  <!-- description end -->
40
  <!-- README_GGUF.md-about-gguf start -->
41
  ### About GGUF
 
70
  A chat.
71
  USER: {prompt}
72
  ASSISTANT:
73
+
74
  ```
75
 
76
  <!-- prompt-template end -->
77
 
78
+
79
  <!-- compatibility_gguf start -->
80
  ## Compatibility
81
 
 
154
  </details>
155
  <!-- README_GGUF.md-provided-files end -->
156
 
157
+ <!-- README_GGUF.md-how-to-download start -->
158
+ ## How to download GGUF files
159
+
160
+ **Note for manual downloaders:** You almost never want to clone the entire repo! Multiple different quantisation formats are provided, and most users only want to pick and download a single file.
161
+
162
+ The following clients/libraries will automatically download models for you, providing a list of available models to choose from:
163
+ - LM Studio
164
+ - LoLLMS Web UI
165
+ - Faraday.dev
166
+
167
+ ### In `text-generation-webui`
168
+
169
+ Under Download Model, you can enter the model repo: TheBloke/Airoboros-L2-70b-2.2-GGUF and below it, a specific filename to download, such as: airoboros-l2-70b-2.2.q4_K_M.gguf.
170
+
171
+ Then click Download.
172
+
173
+ ### On the command line, including multiple files at once
174
+
175
+ I recommend using the `huggingface-hub` Python library:
176
+
177
+ ```shell
178
+ pip3 install huggingface-hub>=0.17.1
179
+ ```
180
+
181
+ Then you can download any individual model file to the current directory, at high speed, with a command like this:
182
+
183
+ ```shell
184
+ huggingface-cli download TheBloke/Airoboros-L2-70b-2.2-GGUF airoboros-l2-70b-2.2.q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
185
+ ```
186
+
187
+ <details>
188
+ <summary>More advanced huggingface-cli download usage</summary>
189
+
190
+ You can also download multiple files at once with a pattern:
191
+
192
+ ```shell
193
+ huggingface-cli download TheBloke/Airoboros-L2-70b-2.2-GGUF --local-dir . --local-dir-use-symlinks False --include='*Q4_K*gguf'
194
+ ```
195
+
196
+ For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli).
197
+
198
+ To accelerate downloads on fast connections (1Gbit/s or higher), install `hf_transfer`:
199
+
200
+ ```shell
201
+ pip3 install hf_transfer
202
+ ```
203
+
204
+ And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
205
+
206
+ ```shell
207
+ HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download TheBloke/Airoboros-L2-70b-2.2-GGUF airoboros-l2-70b-2.2.q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
208
+ ```
209
+
210
+ Windows CLI users: Use `set HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1` before running the download command.
211
+ </details>
212
+ <!-- README_GGUF.md-how-to-download end -->
213
+
214
  <!-- README_GGUF.md-how-to-run start -->
215
  ## Example `llama.cpp` command
216