sadkins65 commited on
Commit
01d7326
1 Parent(s): bfe8310

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +6 -0
  2. model.safetensors +1 -1
  3. recipe.yaml +6 -0
config.json CHANGED
@@ -46,6 +46,12 @@
46
  ],
47
  "quant_method": "sparseml",
48
  "quantization_status": "frozen"
 
 
 
 
 
 
49
  }
50
  },
51
  "eos_token_id": 2,
 
46
  ],
47
  "quant_method": "sparseml",
48
  "quantization_status": "frozen"
49
+ },
50
+ "sparsity_config": {
51
+ "format": "dense",
52
+ "global_sparsity": 7.700428565876607,
53
+ "registry_requires_subclass": false,
54
+ "sparsity_structure": "0:0"
55
  }
56
  },
57
  "eos_token_id": 2,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0966eef7b6292b59aeedc6a484e8886cbeb064ab0c29037804200c1e51105f2c
3
  size 1528201733
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1413b4a713469abe8b3261fa9aa36d54712fdaa94ef5be5f0ab71b435714735
3
  size 1528201733
recipe.yaml CHANGED
@@ -8,3 +8,9 @@ test_stage:
8
  input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true}
9
  output_activations: null
10
  targets: [Linear]
 
 
 
 
 
 
 
8
  input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true}
9
  output_activations: null
10
  targets: [Linear]
11
+ SparseGPTModifier:
12
+ sparsity: 0.0
13
+ block_size: 128
14
+ sequential_update: false
15
+ quantize: true
16
+ targets: ['re:model.layers.\d+$']