Updating model files
Browse files
README.md
CHANGED
@@ -11,6 +11,17 @@ tags:
|
|
11 |
- gptq
|
12 |
inference: false
|
13 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
## About this repository
|
16 |
This is an attempt to create a GPTQ 4-bit version of [Galpaca 30B](https://huggingface.co/GeorgiaTechResearchInstitute/galpaca-30b).
|
@@ -30,7 +41,7 @@ You will need 18+ GB VRAM to load these models on a GPU.
|
|
30 |
* If you can get it working, please let me know!
|
31 |
|
32 |
|
33 |
-
## GPTQ
|
34 |
|
35 |
The GPTQ code used to create these models can be found at [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
36 |
|
@@ -54,6 +65,17 @@ The model card from the original Galactica repo can be found [here](https://gith
|
|
54 |
The dataset card for Alpaca can be found [here](https://huggingface.co/datasets/tatsu-lab/alpaca/blob/main/README.md), and the project homepage [here](https://crfm.stanford.edu/2023/03/13/alpaca.html).
|
55 |
The Alpaca dataset was collected with a modified version of the [Self-Instruct Framework](https://github.com/yizhongw/self-instruct), and was built using OpenAI's `text-davinci-003` model. As such it is subject to OpenAI's terms of service.
|
56 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
## Model Details
|
58 |
|
59 |
The GALACTICA models are trained on a large-scale scientific corpus and are designed to perform scientific tasks.
|
|
|
11 |
- gptq
|
12 |
inference: false
|
13 |
---
|
14 |
+
<div style="width: 100%;">
|
15 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
16 |
+
</div>
|
17 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
18 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
19 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
20 |
+
</div>
|
21 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
22 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
23 |
+
</div>
|
24 |
+
</div>
|
25 |
|
26 |
## About this repository
|
27 |
This is an attempt to create a GPTQ 4-bit version of [Galpaca 30B](https://huggingface.co/GeorgiaTechResearchInstitute/galpaca-30b).
|
|
|
41 |
* If you can get it working, please let me know!
|
42 |
|
43 |
|
44 |
+
## GPTQ
|
45 |
|
46 |
The GPTQ code used to create these models can be found at [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
|
47 |
|
|
|
65 |
The dataset card for Alpaca can be found [here](https://huggingface.co/datasets/tatsu-lab/alpaca/blob/main/README.md), and the project homepage [here](https://crfm.stanford.edu/2023/03/13/alpaca.html).
|
66 |
The Alpaca dataset was collected with a modified version of the [Self-Instruct Framework](https://github.com/yizhongw/self-instruct), and was built using OpenAI's `text-davinci-003` model. As such it is subject to OpenAI's terms of service.
|
67 |
|
68 |
+
## Want to support my work?
|
69 |
+
|
70 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
71 |
+
|
72 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
73 |
+
|
74 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
75 |
+
|
76 |
+
* Patreon: coming soon! (just awaiting approval)
|
77 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
78 |
+
* Discord: https://discord.gg/UBgz4VXf
|
79 |
## Model Details
|
80 |
|
81 |
The GALACTICA models are trained on a large-scale scientific corpus and are designed to perform scientific tasks.
|