Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
inference: false
|
3 |
+
license: other
|
4 |
+
datasets:
|
5 |
+
- gozfarb/ShareGPT_Vicuna_unfiltered
|
6 |
+
---
|
7 |
+
|
8 |
+
<div style="width: 100%;">
|
9 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
10 |
+
</div>
|
11 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
12 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
13 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
14 |
+
</div>
|
15 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
16 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
|
17 |
+
</div>
|
18 |
+
</div>
|
19 |
+
|
20 |
+
# Aeala's VicUnlocked Alpaca 65B QLoRA GGML
|
21 |
+
|
22 |
+
These files are fp16 mergedmodel files for [Aeala's VicUnlocked Alpaca 65B QLoRA](https://huggingface.co/Aeala/VicUnlocked-alpaca-65b-QLoRA).
|
23 |
+
|
24 |
+
## Other repositories available
|
25 |
+
|
26 |
+
* [4-bit GPTQ models for GPU inference](https://huggingface.co/Aeala/VicUnlocked-alpaca-65b-4bit)
|
27 |
+
* [4-bit, 5-bit, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/VicUnlocked-alpaca-65B-QLoRA-GGML)
|
28 |
+
* [Original unquantised fp16 model in HF format](https://huggingface.co/TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16)
|
29 |
+
|
30 |
+
## Want to support my work?
|
31 |
+
|
32 |
+
Thanks to the [chirper.ai](https://chirper.ai) team!
|
33 |
+
|
34 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
35 |
+
|
36 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
37 |
+
|
38 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
39 |
+
|
40 |
+
* Patreon: https://patreon.com/TheBlokeAI
|
41 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
42 |
+
* Discord: https://discord.gg/UBgz4VXf
|
43 |
+
|
44 |
+
Patreon special mention: Jonathan Leane; Talal Aujan. Thank you both, and to all my other patrons and donaters.
|
45 |
+
|
46 |
+
# Original model card: Aeala's VicUnlocked Alpaca 65B QLoRA
|
47 |
+
|
48 |
+
|
49 |
+
## LoRA Info:
|
50 |
+
Please note that this is a highly experimental LoRA model. It may do some good stuff, it might do some undesirable stuff. Training is paused for now. Feel free to try it!~
|
51 |
+
|
52 |
+
**Important Note**: While this is trained on a cleaned ShareGPT dataset like Vicuna used, this was trained in the *Alpaca* format, so prompting should be something like:
|
53 |
+
|
54 |
+
```
|
55 |
+
### Instruction:
|
56 |
+
|
57 |
+
<prompt> (without the <>)
|
58 |
+
|
59 |
+
### Response:
|
60 |
+
```
|
61 |
+
|
62 |
+
Current upload: checkpoint of step 1200 in training.
|
63 |
+
|
64 |
+
|
65 |
+
## Benchmarks
|
66 |
+
**wikitext2:** Coming soon...
|
67 |
+
|
68 |
+
**ptb-new:** Coming soon...
|
69 |
+
|
70 |
+
**c4-new:** Coming soon...
|