Upload folder using huggingface_hub
Browse files- config.json +6 -0
- model.safetensors +1 -1
- recipe.yaml +6 -0
config.json
CHANGED
@@ -46,6 +46,12 @@
|
|
46 |
],
|
47 |
"quant_method": "sparseml",
|
48 |
"quantization_status": "frozen"
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
}
|
50 |
},
|
51 |
"eos_token_id": 2,
|
|
|
46 |
],
|
47 |
"quant_method": "sparseml",
|
48 |
"quantization_status": "frozen"
|
49 |
+
},
|
50 |
+
"sparsity_config": {
|
51 |
+
"format": "dense",
|
52 |
+
"global_sparsity": 7.700428565876607,
|
53 |
+
"registry_requires_subclass": false,
|
54 |
+
"sparsity_structure": "0:0"
|
55 |
}
|
56 |
},
|
57 |
"eos_token_id": 2,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1528201733
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1413b4a713469abe8b3261fa9aa36d54712fdaa94ef5be5f0ab71b435714735
|
3 |
size 1528201733
|
recipe.yaml
CHANGED
@@ -8,3 +8,9 @@ test_stage:
|
|
8 |
input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true}
|
9 |
output_activations: null
|
10 |
targets: [Linear]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true}
|
9 |
output_activations: null
|
10 |
targets: [Linear]
|
11 |
+
SparseGPTModifier:
|
12 |
+
sparsity: 0.0
|
13 |
+
block_size: 128
|
14 |
+
sequential_update: false
|
15 |
+
quantize: true
|
16 |
+
targets: ['re:model.layers.\d+$']
|