| # OMP_NUM_THREADS=14 please Check issue: https://github.com/AutoGPTQ/AutoGPTQ/issues/439 | |
| OMP_NUM_THREADS=14 \ | |
| CUDA_VISIBLE_DEVICES=0 \ | |
| swift export \ | |
| --model Qwen/Qwen2.5-1.5B-Instruct \ | |
| --dataset 'AI-ModelScope/alpaca-gpt4-data-zh#500' \ | |
| 'AI-ModelScope/alpaca-gpt4-data-en#500' \ | |
| --quant_n_samples 256 \ | |
| --quant_batch_size 1 \ | |
| --max_length 2048 \ | |
| --quant_method gptq \ | |
| --quant_bits 4 \ | |
| --output_dir Qwen2.5-1.5B-Instruct-GPTQ-Int4 | |