test_stage: | |
quant_modifiers: | |
vLLMQuantizationModifier: | |
ignore: [lm_head, model.layers.0.mlp.down_proj] | |
config_groups: | |
group_0: | |
weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor} | |
input_activations: {num_bits: 8, type: int, symmetric: false, strategy: tensor} | |
output_activations: null | |
targets: [Linear] | |
group_1: | |
weights: {num_bits: 8, type: int, symmetric: true, strategy: tensor} | |
input_activations: null | |
output_activations: null | |
targets: [Embedding] | |
SparseGPTModifier: | |
sparsity: 0.0 | |
block_size: 128 | |
sequential_update: false | |
quantize: true | |
targets: ['re:model.layers.\d+$'] | |