TheBloke commited on
Commit
3028b42
1 Parent(s): cd6783a

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -0
README.md CHANGED
@@ -124,8 +124,40 @@ Refer to the Provided Files table below to see what files use which methods, and
124
 
125
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
126
 
 
127
 
 
128
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129
  <!-- README_GGUF.md-provided-files end -->
130
 
131
  <!-- README_GGUF.md-how-to-run start -->
 
124
 
125
  **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
126
 
127
+ ### Q6_K and Q8_0 files are split and require joining
128
 
129
+ **Note:** HF does not support uploading files larger than 50GB. Therefore I have uploaded the Q6_K and Q8_0 files as split files.
130
 
131
+ <details>
132
+ <summary>Click for instructions regarding Q6_K and Q8_0 files</summary>
133
+
134
+ ### q6_K
135
+ Please download:
136
+ * `llama-2-70b.Q6_K.gguf-split-a`
137
+ * `llama-2-70b.Q6_K.gguf-split-b`
138
+
139
+ ### q8_0
140
+ Please download:
141
+ * `llama-2-70b.Q8_0.gguf-split-a`
142
+ * `llama-2-70b.Q8_0.gguf-split-b`
143
+
144
+ To join the files, do the following:
145
+
146
+ Linux and macOS:
147
+ ```
148
+ cat llama-2-70b.Q6_K.gguf-split-* > llama-2-70b.Q6_K.gguf && rm llama-2-70b.Q6_K.gguf-split-*
149
+ cat llama-2-70b.Q8_0.gguf-split-* > llama-2-70b.Q8_0.gguf && rm llama-2-70b.Q8_0.gguf-split-*
150
+ ```
151
+ Windows command line:
152
+ ```
153
+ COPY /B llama-2-70b.Q6_K.gguf-split-a + llama-2-70b.Q6_K.gguf-split-b llama-2-70b.Q6_K.gguf
154
+ del llama-2-70b.Q6_K.gguf-split-a fiction.live-Kimiko-V2-70B.Q6_K.gguf-split-b
155
+
156
+ COPY /B llama-2-70b.Q8_0.gguf-split-a + llama-2-70b.Q8_0.gguf-split-b llama-2-70b.Q8_0.gguf
157
+ del llama-2-70b.Q8_0.gguf-split-a llama-2-70b.Q8_0.gguf-split-b
158
+ ```
159
+
160
+ </details>
161
  <!-- README_GGUF.md-provided-files end -->
162
 
163
  <!-- README_GGUF.md-how-to-run start -->