NicoNico's picture
Update README.md
1b28ce1
metadata
license: apache-2.0

GreenBit LLMs

This is GreenBitAI's pretrained low-bit LLMs with extreme compression yet still strong performance.

Please refer to our Github page for the code to run the model and more information.

Zero-shot Evaluation

We evaluate the zero-shot ability of low-bit quantized Qwen1.5 models using the llm_eval library and list the results below:

Repository (Qwen Family) Avg Acc. OpenBQ ARC-E Winogr. HellaS. ARC-C PIQA BoolQ RACE ANLI-R1 ANLI-R2 ANLI-R3 WiC
Qwen-1.5-0.5B-layer-mix-bpw-2.2 0.398 0.170 0.443 0.527 0.332 0.238 0.634 0.620 0.318 0.332 0.338 0.330 0.500
Qwen-1.5-0.5B-layer-mix-bpw-2.5 0.394 0.170 0.514 0.541 0.337 0.232 0.637 0.496 0.318 0.316 0.358 0.326 0.490
Qwen-1.5-0.5B-layer-mix-bpw-3.0 0.407 0.198 0.533 0.536 0.348 0.234 0.671 0.552 0.323 0.330 0.333 0.335 0.495
Qwen-1.5-1.8B-layer-mix-bpw-2.2 0.415 0.218 0.539 0.586 0.392 0.260 0.678 0.622 0.333 0.333 0.333 0.336 0.464
Qwen-1.5-1.8B-layer-mix-bpw-2.5 0.423 0.222 0.592 0.585 0.406 0.267 0.695 0.629 0.336 0.314 0.339 0.361 0.507
Qwen-1.5-1.8B-layer-mix-bpw-3.0 0.438 0.246 0.576 0.563 0.413 0.277 0.694 0.645 0.352 0.323 0.336 0.343 0.492
Qwen-1.5-4B-layer-mix-bpw-2.2 0.480 0.254 0.663 0.623 0.463 0.339 0.712 0.718 0.349 0.326 0.355 0.384 0.513
Qwen-1.5-4B-layer-mix-bpw-2.5 0.490 0.266 0.677 0.629 0.473 0.365 0.732 0.717 0.351 0.372 0.352 0.360 0.502
Qwen-1.5-4B-layer-mix-bpw-3.0 0.502 0.268 0.678 0.642 0.494 0.358 0.755 0.757 0.380 0.395 0.395 0.392 0.519
Qwen-1.5-7B-layer-mix-bpw-2.2 0.513 0.278 0.669 0.654 0.504 0.389 0.741 0.759 0.376 0.383 0.410 0.403 0.517
Qwen-1.5-7B-layer-mix-bpw-2.5 0.520 0.294 0.705 0.650 0.520 0.387 0.750 0.769 0.371 0.445 0.424 0.398 0.564
Qwen-1.5-7B-layer-mix-bpw-3.0 0.531 0.292 0.713 0.654 0.545 0.405 0.764 0.807 0.383 0.424 0.393 0.414 0.627
Qwen-1.5-14B-layer-mix-bpw-2.5 0.553 0.318 0.727 0.682 0.564 0.413 0.775 0.792 0.390 0.472 0.434 0.446 0.623
Qwen-1.5-32B-layer-mix-bpw-3.0 0.599 0.346 0.775 0.722 0.620 0.492 0.807 0.853 0.444 0.515 0.494 0.478 0.642