My first quantization uses the quantization method provided by vllm: | |
https://docs.vllm.ai/en/latest/quantization/int8.html | |
NUM_CALIBRATION_SAMPLES = 2048 | |
MAX_SEQUENCE_LENGTH = 8192 | |
smoothing_strength=0.8 | |
I will verify the validity of the model and update the readme as soon as possible. | |
edit: The performance in my ERP test was comparable to Mistral-Nemo-Instruct-2407-GPTQ-INT8, which I consider a successful quantization. |