TheBloke commited on
Commit
365f219
1 Parent(s): 4e36268

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -2,6 +2,17 @@
2
  inference: false
3
  license: other
4
  ---
 
 
 
 
 
 
 
 
 
 
 
5
  # LLaMa 13B GGML
6
 
7
  This repo contains GGML format model files for the original LLaMa.
@@ -10,7 +21,7 @@ These files are for CPU (+ CUDA) inference using [llama.cpp](https://github.com/
10
 
11
  I've uploaded them mostly for my own convenience, allowing me to easily grab them if and when I need them for future testing and comparisons.
12
 
13
- ## Provided files
14
 
15
  The following formats are included:
16
  * float16
@@ -27,3 +38,14 @@ llama.cpp recently made a breaking change to its quantisation methods.
27
  I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
28
 
29
  I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
 
 
 
 
 
 
 
 
 
 
 
 
2
  inference: false
3
  license: other
4
  ---
5
+ <div style="width: 100%;">
6
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
7
+ </div>
8
+ <div style="display: flex; justify-content: space-between; width: 100%;">
9
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
10
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
11
+ </div>
12
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
13
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
14
+ </div>
15
+ </div>
16
  # LLaMa 13B GGML
17
 
18
  This repo contains GGML format model files for the original LLaMa.
 
21
 
22
  I've uploaded them mostly for my own convenience, allowing me to easily grab them if and when I need them for future testing and comparisons.
23
 
24
+ ## Provided files
25
 
26
  The following formats are included:
27
  * float16
 
38
  I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
39
 
40
  I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
41
+ ## Want to support my work?
42
+
43
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
44
+
45
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
46
+
47
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
48
+
49
+ * Patreon: coming soon! (just awaiting approval)
50
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
51
+ * Discord: https://discord.gg/UBgz4VXf