yi-34b-w4a16g32 / README.md
yanghaojin's picture
Update README.md
e8fff0b
|
raw
history blame
4.27 kB
---
license: apache-2.0
language:
- en
- zh
---
license: apache-2.0
---
# GreenBit Yi
This is GreenBitAI's pretrained **4-bit** Yi 34B model with extreme compression yet still strong performance.
Please refer to our [Github page](https://github.com/GreenBitAI/low_bit_llama/tree/low_bit_yi) for the code to run the model and more information.
## Model Description
- **Developed by:** [GreenBitAI](https://github.com/GreenBitAI)
- **Evaluated By:** 01-Yi official
- **Model type:** Causal (Llama 2/Yi 6B)
- **Language(s) (NLP):** English
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0), [Llama 2 license agreement](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)
## Few Shot Evaluation (officially evaluated by 01-Yi)
| Model | Yi-34B FP16| [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) |
|----------------|-----------|----------|----------|---------|
| GroupSize | - | 32 | - | 8 |
| Model Size (GB)| 68.79 | 19.89 | 12.12 | 4.04 |
| AVG | 70.64 | 69.7 | 60.11 | 59.14 |
| **Detailed Evaluation** | | | | |
| MMLU | 76.32 | 75.42 | 63.24 | 62.09 |
| CMMLU | 83.65 | 83.07 | 75.53 | 72.85 |
| ARC-e | 84.42 | 84.13 | 77.23 | 76.52 |
| ARC-c | 61.77 | 59.56 | 50.34 | 48.47 |
| GAOKAO | 82.8 | 81.37 | 72.2 | 72.87 |
| GSM8K | 67.24 | 63.61 | 32.52 | 28.05 |
| HumanEval | 25.6 | 25 | 15.85 | 15.85 |
| BBH | 54.3 | 52.3 | 42.8 | 41.47 |
| WinoGrande | 78.68 | 78.53 | 70.63 | 71.19 |
| PIQA | 82.86 | 82.75 | 78.56 | 79.05 |
| SIQA | 74.46 | 73.44 | 64.53 | 64.53 |
| HellaSwag | 83.64 | 83.02 | 74.91 | 73.27 |
| OBQA | 91.6 | 90.8 | 85.4 | 82.6 |
| CSQA | 83.37 | 83.05 | 76.9 | 75.43 |
| TriviaQA | 81.52 | 80.73 | 64.85 | 61.75 |
| SquAD | 92.46 | 91.12 | 88.95 | 88.39 |
| BoolQ | 88.25 | 88.17 | 76.23 | 77.1 |
| MBPP | 41 | 39.68 | 26.32 | 25.13 |
| QUAC | 48.61 | 47.43 | 40.92 | 40.16 |
| Lambda | 73.18 | 73.39 | 67.74 | 67.8 |
| NaturalQuestion| 27.67 | 27.21 | 16.69 | 17.42 |
# Zero Shot Evaluation
| Task | Metric | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) | [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) |
|---------------|--------|---------|-------------|--------------|
| Openbookqa | acc | 0.314 | 0.324 | 0.344 |
| | ac_norm| 0.408 | 0.42 | 0.474 |
| arc_challenge | acc | 0.462 | 0.4573 | 0.569 |
| | ac_norm| 0.504 | 0.483 | 0.5964 |
| hellawswag | acc | 0.553 | 0.5447 | 0.628 |
| | ac_norm| 0.749 | 0.7327 | 0.83 |
| piqa | acc | 0.777 | 0.7709 | 0.8079 |
| | ac_norm| 0.787 | 0.7894 | 0.828 |
| arc_easy | acc | 0.777 | 0.7697 | 0.835 |
| | ac_norm| 0.774 | 0.7659 | 0.84 |
| Winogrande | acc | 0.707 | 0.7095 | 0.7853 |
| boolq | acc | 0.755 | 0.7648 | 0.886 |
| truthfulqa_mc | mc1 | 0.29 | 0.2729 | 0.4026 |
| | mc2 | 0.419 | 0.4033 | 0.5528 |
| anli_r1 | acc | 0.423 | 0.416 | 0.554 |
| anli_r2 | acc | 0.409 | 0.409 | 0.518 |
| anli_r3 | acc | 0.411 | 0.393 | 0.4983 |
| wic | acc | 0.529 | 0.545 | 0.5376 |
| rte | acc | 0.685 | 0.7039 | 0.7617 |
| record | f1 | 0.904 | 0.9011 | 0.924 |
| | em | 0.8962 | 0.8927 | 0.916 |
| Average | | 0.596 | 0.5937 | 0.6708 |