seonglae commited on
Commit
c8f4f03
1 Parent(s): aadaf43

build: AutoGPTQ for meta-llama/Llama-2-7b-chat-hf: 4bits, gr128, desc_act=False

Browse files
Files changed (3) hide show
  1. config.json +1 -1
  2. gptq_model-4bit-128g.bin +2 -2
  3. quantize_config.json +2 -2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "llama-2-7b-chat-hf-gptq",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
gptq_model-4bit-128g.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:981e70a261a93caeb561ab1ece8050b367a3dd07cfe65e052cf47fdb77439bd3
3
- size 3896904828
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e86a4da7602a401317c4a49747d63a60dd6e56d4d216dd33a8e54e4309f833a8
3
+ size 3897013783
quantize_config.json CHANGED
@@ -5,6 +5,6 @@
5
  "desc_act": false,
6
  "sym": true,
7
  "true_sequential": true,
8
- "model_name_or_path": "llama-2-7b-chat-hf-gptq",
9
- "model_file_base_name": "gptq_model-4bit-128g"
10
  }
 
5
  "desc_act": false,
6
  "sym": true,
7
  "true_sequential": true,
8
+ "model_name_or_path": null,
9
+ "model_file_base_name": null
10
  }