OPEA
/

Safetensors
weiweiz1 commited on
Commit
80a3f3d
1 Parent(s): f85cfb5

auto_gptq format

Browse files
config.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d045f30d7d09714fecd0e63423e8814b66abb164aa7df9583f62276c26b0b322
3
- size 2601
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09cbfe58c9b2c463fec785cb44ffdc67ebe53f29356b7d3045076d19fe3e2813
3
+ size 2581
model.safetensors.index.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ee10f5b8e52d2a0ee2e6795c9ea649b8101ad81b1a0108de772e54fccb41d76
3
- size 210876
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd3587e0f4a4dc04b1b1a10b307b5dda785716817956c7815436bac85e573862
3
+ size 237410
quantization_config.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4814a2340cd541089354239396e9262cb662339634594b62948f122c8d18d812
3
- size 1381
 
 
 
 
quantize_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "sym": true,
5
+ "data_type": "int",
6
+ "enable_quanted_input": true,
7
+ "enable_minmax_tuning": true,
8
+ "seqlen": 512,
9
+ "batch_size": 8,
10
+ "scale_dtype": "torch.float16",
11
+ "lr": 1.0,
12
+ "minmax_lr": 1.0,
13
+ "gradient_accumulate_steps": 1,
14
+ "iters": 1,
15
+ "amp": true,
16
+ "nsamples": 512,
17
+ "low_gpu_mem_usage": false,
18
+ "to_quant_block_names": [
19
+ [
20
+ "model.layers.0",
21
+ "model.layers.1",
22
+ "model.layers.2",
23
+ "model.layers.3",
24
+ "model.layers.4",
25
+ "model.layers.5",
26
+ "model.layers.6",
27
+ "model.layers.7",
28
+ "model.layers.8",
29
+ "model.layers.9",
30
+ "model.layers.10",
31
+ "model.layers.11",
32
+ "model.layers.12",
33
+ "model.layers.13",
34
+ "model.layers.14",
35
+ "model.layers.15",
36
+ "model.layers.16",
37
+ "model.layers.17",
38
+ "model.layers.18",
39
+ "model.layers.19",
40
+ "model.layers.20",
41
+ "model.layers.21",
42
+ "model.layers.22",
43
+ "model.layers.23",
44
+ "model.layers.24",
45
+ "model.layers.25",
46
+ "model.layers.26",
47
+ "model.layers.27",
48
+ "model.layers.28",
49
+ "model.layers.29",
50
+ "model.layers.30",
51
+ "model.layers.31"
52
+ ]
53
+ ],
54
+ "enable_norm_bias_tuning": false,
55
+ "autoround_version": "0.4.2.dev",
56
+ "quant_method": "gptq",
57
+ "desc_act": false,
58
+ "true_sequential": false,
59
+ "damp_percent": 0.01
60
+ }