seonglae commited on
Commit
8127d01
1 Parent(s): c6f17d1

build: AutoGPTQ for meta-llama/Llama-2-13b-chat-hf: 4bits, gr128, desc_act=False

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "llama-2-13b-chat-hf-gptq",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
gptq_model-4bit-128g.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b0348502ed12fbb94f68b654aa73b4621c349eed96d7b97cd58942208f8d7f7
3
- size 7259404840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:475ae41986b3baaad008f8da3731d83f1ff13eecdcd95119c1ca13725f52dcc3
3
+ size 7259449424
quantize_config.json CHANGED
@@ -5,6 +5,6 @@
5
  "desc_act": false,
6
  "sym": true,
7
  "true_sequential": true,
8
- "model_name_or_path": "llama-2-13b-chat-hf-gptq",
9
- "model_file_base_name": "gptq_model-4bit-128g"
10
  }
 
5
  "desc_act": false,
6
  "sym": true,
7
  "true_sequential": true,
8
+ "model_name_or_path": null,
9
+ "model_file_base_name": null
10
  }