TheBloke commited on
Commit
9f51330
1 Parent(s): 748baf3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -9,7 +9,7 @@ This version has then been quantized to 4-bit using [GPTQ-for-LLaMa](https://git
9
  ## Other Koala repos
10
 
11
  I have also made these other Koala repose available:
12
- * [GPTQ quantized 4bit 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
13
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
14
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
15
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
@@ -19,9 +19,10 @@ I have also made these other Koala repose available:
19
  This GPTQ model was quantized using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) with the following command:
20
  ```
21
  python3 llama.py /content/koala-7B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save /content/koala-7B-4bit-128g.pt
 
22
  ```
23
 
24
- I created this model using the latest Triton branch of GPTQ-for-LLaMa but I believe it can be run with the CUDA branch also.
25
 
26
  ## Provided files
27
 
@@ -29,7 +30,7 @@ I have provided both a `pt` and `safetensors` file. Either should work.
29
 
30
  If both are present in the model directory for text-generation-webui I am not sure which it picks, so if you need one or the other specifically I'd recommend just downloading the one you need.
31
 
32
- The `olderFormat` file was created with the aim of then converting it to GGML for use with llama.cpp. At present this file does not work.
33
 
34
  ## How to run with text-generation-webui
35
 
@@ -48,12 +49,12 @@ ln -s GPTQ-for-LLaMa text-generation-webui/repositories/GPTQ-for-LLaMa
48
  Then install this model into `text-generation-webui/models` and run text-generation-webui as follows:
49
  ```
50
  cd text-generation-webui
51
- python server.py --model koala-7B-GPTQ-4bit-128g --wbits 4 --groupsize 128 --model_type Llama
52
  ```
53
 
54
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
55
 
56
- If you cannot use the Triton branch for any reason, I believe it should also work to use the CUDA branch instead:
57
  ```
58
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
59
  ```
 
9
  ## Other Koala repos
10
 
11
  I have also made these other Koala repose available:
12
+ * [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
13
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
14
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
15
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
 
19
  This GPTQ model was quantized using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) with the following command:
20
  ```
21
  python3 llama.py /content/koala-7B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save /content/koala-7B-4bit-128g.pt
22
+ python3 llama.py /content/koala-7B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/koala-7B-4bit-128g.safetensors
23
  ```
24
 
25
+ I created this model using the latest Triton branch of GPTQ-for-LLaMa but it can be loaded with the CUDA branch also.
26
 
27
  ## Provided files
28
 
 
30
 
31
  If both are present in the model directory for text-generation-webui I am not sure which it picks, so if you need one or the other specifically I'd recommend just downloading the one you need.
32
 
33
+ The `olderFormat` file was created with the aim of then converting it to GGML for use with [llama.cpp](https://github.com/ggerganov/llama.cpp). At present this file does not work.
34
 
35
  ## How to run with text-generation-webui
36
 
 
49
  Then install this model into `text-generation-webui/models` and run text-generation-webui as follows:
50
  ```
51
  cd text-generation-webui
52
+ python server.py --model koala-7B-GPTQ-4bit-128g --wbits 4 --groupsize 128 --model_type Llama # add any other command line args you want
53
  ```
54
 
55
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
56
 
57
+ If you cannot use the Triton branch of GPTQ for any reason, it should also work to use the CUDA branch instead:
58
  ```
59
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
60
  ```