TheBloke commited on
Commit
1e0feb9
1 Parent(s): 065ee41

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +25 -3
README.md CHANGED
@@ -2,7 +2,7 @@
2
  license: other
3
  library_name: transformers
4
  pipeline_tag: text-generation
5
- datasets:
6
  - RyokoAI/ShareGPT52K
7
  - Hello-SimpleAI/HC3
8
  tags:
@@ -12,6 +12,17 @@ tags:
12
  - gptq
13
  inference: false
14
  ---
 
 
 
 
 
 
 
 
 
 
 
15
  # Koala: A Dialogue Model for Academic Research
16
  This repo contains the weights of the Koala 7B model produced at Berkeley. It is the result of combining the diffs from https://huggingface.co/young-geng/koala with the original Llama 7B model.
17
 
@@ -24,7 +35,7 @@ I have the following Koala model repositories available:
24
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
25
  * [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
26
  * [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GGML)
27
-
28
  **7B models:**
29
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
30
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
@@ -57,7 +68,7 @@ Details of the files provided:
57
  * The older GPTQ code does not support all the latest features, so the quality may be fractionally lower.
58
  * Command to create:
59
  * `python3 llama.py koala-7B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save koala-7B-4bit-128g.no-act-order.ooba.pt`
60
-
61
  ## How to run in `text-generation-webui`
62
 
63
  File `koala-7B-4bit-128g.no-act-order.ooba.pt` can be loaded the same as any other GPTQ file, without requiring any updates to [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
@@ -122,6 +133,17 @@ PYTHON_PATH="${PWD}:$PYTHONPATH" python \
122
  --tokenizer_path=/content/LLaMA-7B/tokenizer.model
123
  ```
124
 
 
 
 
 
 
 
 
 
 
 
 
125
  ## Further info
126
 
127
  Check out the following links to learn more about the Berkeley Koala model.
 
2
  license: other
3
  library_name: transformers
4
  pipeline_tag: text-generation
5
+ datasets:
6
  - RyokoAI/ShareGPT52K
7
  - Hello-SimpleAI/HC3
8
  tags:
 
12
  - gptq
13
  inference: false
14
  ---
15
+ <div style="width: 100%;">
16
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
17
+ </div>
18
+ <div style="display: flex; justify-content: space-between; width: 100%;">
19
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
20
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
21
+ </div>
22
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
23
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
24
+ </div>
25
+ </div>
26
  # Koala: A Dialogue Model for Academic Research
27
  This repo contains the weights of the Koala 7B model produced at Berkeley. It is the result of combining the diffs from https://huggingface.co/young-geng/koala with the original Llama 7B model.
28
 
 
35
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
36
  * [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
37
  * [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GGML)
38
+
39
  **7B models:**
40
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
41
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
 
68
  * The older GPTQ code does not support all the latest features, so the quality may be fractionally lower.
69
  * Command to create:
70
  * `python3 llama.py koala-7B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save koala-7B-4bit-128g.no-act-order.ooba.pt`
71
+
72
  ## How to run in `text-generation-webui`
73
 
74
  File `koala-7B-4bit-128g.no-act-order.ooba.pt` can be loaded the same as any other GPTQ file, without requiring any updates to [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
 
133
  --tokenizer_path=/content/LLaMA-7B/tokenizer.model
134
  ```
135
 
136
+ ## Want to support my work?
137
+
138
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
139
+
140
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
141
+
142
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
143
+
144
+ * Patreon: coming soon! (just awaiting approval)
145
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
146
+ * Discord: https://discord.gg/UBgz4VXf
147
  ## Further info
148
 
149
  Check out the following links to learn more about the Berkeley Koala model.