amirm commited on
Commit
23250de
1 Parent(s): a7fb35a

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +5 -16
  2. model-00001-of-00002.safetensors +1 -1
config.json CHANGED
@@ -17,28 +17,17 @@
17
  "num_key_value_heads": 8,
18
  "pretraining_tp": 1,
19
  "quantization_config": {
20
- "batch_size": 1,
21
  "bits": 4,
22
- "block_name_to_quantize": null,
23
- "cache_block_outputs": true,
24
  "damp_percent": 0.01,
25
- "dataset": null,
26
  "desc_act": true,
27
- "exllama_config": {
28
- "version": 1
29
- },
30
  "group_size": 128,
31
- "max_input_length": null,
32
- "model_seqlen": null,
33
- "module_name_preceding_first_block": null,
34
- "modules_in_block_to_quantize": null,
35
- "pad_token_id": null,
36
  "quant_method": "gptq",
 
37
  "sym": true,
38
- "tokenizer": null,
39
- "true_sequential": true,
40
- "use_cuda_fp16": false,
41
- "use_exllama": true
42
  },
43
  "rms_norm_eps": 1e-05,
44
  "rope_scaling": null,
 
17
  "num_key_value_heads": 8,
18
  "pretraining_tp": 1,
19
  "quantization_config": {
 
20
  "bits": 4,
21
+ "checkpoint_format": "gptq",
 
22
  "damp_percent": 0.01,
 
23
  "desc_act": true,
 
 
 
24
  "group_size": 128,
25
+ "model_file_base_name": null,
26
+ "model_name_or_path": null,
 
 
 
27
  "quant_method": "gptq",
28
+ "static_groups": false,
29
  "sym": true,
30
+ "true_sequential": true
 
 
 
31
  },
32
  "rms_norm_eps": 1e-05,
33
  "rope_scaling": null,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:901c88e4b3b27c2054b4ab0cbc2ba8d57462c70916f364457e617f136a80f326
3
  size 4685047152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ef5bb24009e13201e153c2cc21b6e9bf1288336d2ea1232c4b4c0814e4f7b6f
3
  size 4685047152