kingbri commited on
Commit
b26e96f
1 Parent(s): 4fc86c3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -7,4 +7,6 @@ This is a GPTQ quantized version of [airo-llongma-2-13B-16k](https://huggingface
7
 
8
  To run this model, make sure `compress_pos_emb` is set to 4 to apply proper rope scaling parameters. The `max_ctx_len` is 16384.
9
 
10
- The main branch of this repository is a 4bit 128g model with act order set to false. Sequence length was 4096 when quantizing.
 
 
 
7
 
8
  To run this model, make sure `compress_pos_emb` is set to 4 to apply proper rope scaling parameters. The `max_ctx_len` is 16384.
9
 
10
+ Branches:
11
+ - main: 4 bits, groupsize 128, act order false
12
+ - 4bit-32g-actorder: 4 bits, groupsize 32, act order true