test_stage: obcq_modifiers: QuantizationModifier: ignore: - MistralRotaryEmbedding - MistralRMSNorm - SiLUActivation - model.layers.1.mlp.down_proj - model.layers.31.mlp.down_proj - model.layers.30.mlp.down_proj - model.layers.30.mlp.gate_proj - model.layers.30.mlp.up_proj post_oneshot_calibration: True scheme_overrides: Embedding: input_activations: null weights: num_bits: 8 symmetric: False SparseGPTModifier: sparsity: 0.5 block_size: 128 sequential_update: False quantize: True percdamp: 0.01 prunen: 0 prunem: 0 targets: [ "model.layers.0", "model.layers.1", "model.layers.2", "model.layers.3", "model.layers.4", "model.layers.5", "model.layers.6", "model.layers.7", "model.layers.8", "model.layers.9", "model.layers.10", "model.layers.11", "model.layers.12", "model.layers.13", "model.layers.14", "model.layers.15", "model.layers.16", "model.layers.17", "model.layers.18", "model.layers.19", "model.layers.20", "model.layers.21", "model.layers.22", "model.layers.23", "model.layers.24", "model.layers.25", "model.layers.26", "model.layers.27", "model.layers.28", "model.layers.29", "model.layers.30", "model.layers.31", ] target_ids: ["attention_mask", "position_ids"]