TheBloke commited on
Commit
65303eb
1 Parent(s): 857822a

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +23 -1
README.md CHANGED
@@ -10,6 +10,17 @@ tags:
10
  - gpt4
11
  inference: false
12
  ---
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  This is a 4bit 128g GPTQ of [chansung's gpt4-alpaca-lora-13b](https://huggingface.co/chansung/gpt4-alpaca-lora-13b).
15
 
@@ -31,7 +42,7 @@ Open the text-generation-webui UI as normal.
31
 
32
  Command to create was:
33
  ```
34
- CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
35
  ```
36
 
37
  Command to clone the latest Triton GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
@@ -47,6 +58,17 @@ git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa
47
 
48
  There is also a `no-act-order.safetensors` file which will work with oobabooga's fork of GPTQ-for-LLaMa; it does not require the latest GPTQ code.
49
 
 
 
 
 
 
 
 
 
 
 
 
50
  # Original model card is below
51
 
52
  This repository comes with LoRA checkpoint to make LLaMA into a chatbot like language model. The checkpoint is the output of instruction following fine-tuning process with the following settings on 8xA100(40G) DGX system.
 
10
  - gpt4
11
  inference: false
12
  ---
13
+ <div style="width: 100%;">
14
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
15
+ </div>
16
+ <div style="display: flex; justify-content: space-between; width: 100%;">
17
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
18
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
19
+ </div>
20
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
21
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
22
+ </div>
23
+ </div>
24
 
25
  This is a 4bit 128g GPTQ of [chansung's gpt4-alpaca-lora-13b](https://huggingface.co/chansung/gpt4-alpaca-lora-13b).
26
 
 
42
 
43
  Command to create was:
44
  ```
45
+ CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
46
  ```
47
 
48
  Command to clone the latest Triton GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
 
58
 
59
  There is also a `no-act-order.safetensors` file which will work with oobabooga's fork of GPTQ-for-LLaMa; it does not require the latest GPTQ code.
60
 
61
+ ## Want to support my work?
62
+
63
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
64
+
65
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
66
+
67
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
68
+
69
+ * Patreon: coming soon! (just awaiting approval)
70
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
71
+ * Discord: https://discord.gg/UBgz4VXf
72
  # Original model card is below
73
 
74
  This repository comes with LoRA checkpoint to make LLaMA into a chatbot like language model. The checkpoint is the output of instruction following fine-tuning process with the following settings on 8xA100(40G) DGX system.