Update README.md
Browse files
README.md
CHANGED
@@ -2,6 +2,7 @@
|
|
2 |
inference: false
|
3 |
license: other
|
4 |
---
|
|
|
5 |
<div style="width: 100%;">
|
6 |
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
7 |
</div>
|
@@ -10,9 +11,10 @@ license: other
|
|
10 |
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
11 |
</div>
|
12 |
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
13 |
-
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon
|
14 |
</div>
|
15 |
</div>
|
|
|
16 |
|
17 |
# LLaMa 7B GGML
|
18 |
|
@@ -38,14 +40,26 @@ llama.cpp recently made a breaking change to its quantisation methods.
|
|
38 |
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
|
39 |
|
40 |
I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
|
41 |
-
## Want to support my work?
|
42 |
|
43 |
-
|
|
|
44 |
|
45 |
-
|
46 |
|
47 |
-
|
48 |
|
49 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
51 |
-
|
|
|
|
|
|
|
|
|
|
2 |
inference: false
|
3 |
license: other
|
4 |
---
|
5 |
+
<!-- header start -->
|
6 |
<div style="width: 100%;">
|
7 |
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
8 |
</div>
|
|
|
11 |
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
12 |
</div>
|
13 |
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
14 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
15 |
</div>
|
16 |
</div>
|
17 |
+
<!-- header end --
|
18 |
|
19 |
# LLaMa 7B GGML
|
20 |
|
|
|
40 |
I have quantised the GGML files in this repo with the latest version. Therefore you will require llama.cpp compiled on May 12th or later (commit `b9fd7ee` or later) to use them.
|
41 |
|
42 |
I will not be providing GGML formats for the older llama.cpp code. They're already uploaded all over HF if you really need them!
|
|
|
43 |
|
44 |
+
<!-- footer start -->
|
45 |
+
## Discord
|
46 |
|
47 |
+
For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/UBgz4VXf)
|
48 |
|
49 |
+
## Thanks, and how to contribute.
|
50 |
|
51 |
+
Thanks to the [chirper.ai](https://chirper.ai) team!
|
52 |
+
|
53 |
+
I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training.
|
54 |
+
|
55 |
+
If you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on new AI projects.
|
56 |
+
|
57 |
+
Donaters will get priority support on any and all AI/LLM/model questions, plus other benefits.
|
58 |
+
|
59 |
+
* Patreon: https://patreon.com/TheBlokeAI
|
60 |
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
61 |
+
|
62 |
+
**Patreon special mentions**: Aemon Algiz; Johann-Peter Hartmann; Talal Aujan; Jonathan Leane; Illia Dulskyi; Khalefa Al-Ahmad; senxiiz; Sebastain Graf; Eugene Pentland; Nikolai Manek; Luke Pendergrass.
|
63 |
+
|
64 |
+
Thank you to all my generous patrons and donaters.
|
65 |
+
<!-- footer end -->
|