TheBloke commited on
Commit
dadc8c8
1 Parent(s): 6b6a638

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -11,6 +11,17 @@ tags:
11
  - gptq
12
  inference: false
13
  ---
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  ## About this repository
16
  This is an attempt to create a GPTQ 4-bit version of [Galpaca 30B](https://huggingface.co/GeorgiaTechResearchInstitute/galpaca-30b).
@@ -30,7 +41,7 @@ You will need 18+ GB VRAM to load these models on a GPU.
30
  * If you can get it working, please let me know!
31
 
32
 
33
- ## GPTQ
34
 
35
  The GPTQ code used to create these models can be found at [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
36
 
@@ -54,6 +65,17 @@ The model card from the original Galactica repo can be found [here](https://gith
54
  The dataset card for Alpaca can be found [here](https://huggingface.co/datasets/tatsu-lab/alpaca/blob/main/README.md), and the project homepage [here](https://crfm.stanford.edu/2023/03/13/alpaca.html).
55
  The Alpaca dataset was collected with a modified version of the [Self-Instruct Framework](https://github.com/yizhongw/self-instruct), and was built using OpenAI's `text-davinci-003` model. As such it is subject to OpenAI's terms of service.
56
 
 
 
 
 
 
 
 
 
 
 
 
57
  ## Model Details
58
 
59
  The GALACTICA models are trained on a large-scale scientific corpus and are designed to perform scientific tasks.
 
11
  - gptq
12
  inference: false
13
  ---
14
+ <div style="width: 100%;">
15
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
16
+ </div>
17
+ <div style="display: flex; justify-content: space-between; width: 100%;">
18
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
19
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
20
+ </div>
21
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
22
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
23
+ </div>
24
+ </div>
25
 
26
  ## About this repository
27
  This is an attempt to create a GPTQ 4-bit version of [Galpaca 30B](https://huggingface.co/GeorgiaTechResearchInstitute/galpaca-30b).
 
41
  * If you can get it working, please let me know!
42
 
43
 
44
+ ## GPTQ
45
 
46
  The GPTQ code used to create these models can be found at [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
47
 
 
65
  The dataset card for Alpaca can be found [here](https://huggingface.co/datasets/tatsu-lab/alpaca/blob/main/README.md), and the project homepage [here](https://crfm.stanford.edu/2023/03/13/alpaca.html).
66
  The Alpaca dataset was collected with a modified version of the [Self-Instruct Framework](https://github.com/yizhongw/self-instruct), and was built using OpenAI's `text-davinci-003` model. As such it is subject to OpenAI's terms of service.
67
 
68
+ ## Want to support my work?
69
+
70
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
71
+
72
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
73
+
74
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
75
+
76
+ * Patreon: coming soon! (just awaiting approval)
77
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
78
+ * Discord: https://discord.gg/UBgz4VXf
79
  ## Model Details
80
 
81
  The GALACTICA models are trained on a large-scale scientific corpus and are designed to perform scientific tasks.