File size: 4,274 Bytes
e8fff0b
 
 
 
 
 
71bd3d5
 
d6fa456
 
bf78d16
d6fa456
5e16bc4
d6fa456
 
 
 
341bdd5
 
d6fa456
 
 
a42dd8a
 
 
b55fe0b
a42dd8a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d6fa456
 
a42dd8a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e8fff0b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
---
license: apache-2.0
language:
- en
- zh
---
license: apache-2.0
---
# GreenBit Yi

This is GreenBitAI's pretrained **4-bit** Yi 34B model with extreme compression yet still strong performance.

Please refer to our [Github page](https://github.com/GreenBitAI/low_bit_llama/tree/low_bit_yi) for the code to run the model and more information.

## Model Description

- **Developed by:** [GreenBitAI](https://github.com/GreenBitAI)
- **Evaluated By:** 01-Yi official
- **Model type:** Causal (Llama 2/Yi 6B)
- **Language(s) (NLP):** English
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0), [Llama 2 license agreement](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)

## Few Shot Evaluation (officially evaluated by 01-Yi)
| Model          | Yi-34B FP16| [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) |
|----------------|-----------|----------|----------|---------|
| GroupSize      | -         | 32       | -        | 32      |
| Model Size (GB)| 68.79     | 19.89    | 12.12    | 4.04    |
| AVG            | 70.64     | 69.7     | 60.11    | 59.14   |
| **Detailed Evaluation** | | | | |
| MMLU           | 76.32     | 75.42    | 63.24    | 62.09   |
| CMMLU          | 83.65     | 83.07    | 75.53    | 72.85   |
| ARC-e          | 84.42     | 84.13    | 77.23    | 76.52   |
| ARC-c          | 61.77     | 59.56    | 50.34    | 48.47   |
| GAOKAO         | 82.8      | 81.37    | 72.2     | 72.87   |
| GSM8K          | 67.24     | 63.61    | 32.52    | 28.05   |
| HumanEval      | 25.6      | 25       | 15.85    | 15.85   |
| BBH            | 54.3      | 52.3     | 42.8     | 41.47   |
| WinoGrande     | 78.68     | 78.53    | 70.63    | 71.19   |
| PIQA           | 82.86     | 82.75    | 78.56    | 79.05   |
| SIQA           | 74.46     | 73.44    | 64.53    | 64.53   |
| HellaSwag      | 83.64     | 83.02    | 74.91    | 73.27   |
| OBQA           | 91.6      | 90.8     | 85.4     | 82.6    |
| CSQA           | 83.37     | 83.05    | 76.9     | 75.43   |
| TriviaQA       | 81.52     | 80.73    | 64.85    | 61.75   |
| SquAD          | 92.46     | 91.12    | 88.95    | 88.39   |
| BoolQ          | 88.25     | 88.17    | 76.23    | 77.1    |
| MBPP           | 41        | 39.68    | 26.32    | 25.13   |
| QUAC           | 48.61     | 47.43    | 40.92    | 40.16   |
| Lambda         | 73.18     | 73.39    | 67.74    | 67.8    |
| NaturalQuestion| 27.67     | 27.21    | 16.69    | 17.42   |


# Zero Shot Evaluation
| Task          | Metric | Yi-6B FP16 | [Yi-6B 4 bit](https://huggingface.co/GreenBitAI/yi-6b-w4a16g32) | [Yi-34B 4 bit](https://huggingface.co/GreenBitAI/yi-34b-w4a16g32) |
|---------------|--------|---------|-------------|--------------|
| Openbookqa    | acc    | 0.314   | 0.324       | 0.344        |
|               | ac_norm| 0.408   | 0.42        | 0.474        |
| arc_challenge | acc    | 0.462   | 0.4573      | 0.569        |
|               | ac_norm| 0.504   | 0.483       | 0.5964       |
| hellawswag    | acc    | 0.553   | 0.5447      | 0.628        |
|               | ac_norm| 0.749   | 0.7327      | 0.83         |
| piqa          | acc    | 0.777   | 0.7709      | 0.8079       |
|               | ac_norm| 0.787   | 0.7894      | 0.828        |
| arc_easy      | acc    | 0.777   | 0.7697      | 0.835        |
|               | ac_norm| 0.774   | 0.7659      | 0.84         |
| Winogrande    | acc    | 0.707   | 0.7095      | 0.7853       |
| boolq         | acc    | 0.755   | 0.7648      | 0.886        |
| truthfulqa_mc | mc1    | 0.29    | 0.2729      | 0.4026       |
|               | mc2    | 0.419   | 0.4033      | 0.5528       |
| anli_r1       | acc    | 0.423   | 0.416       | 0.554        |
| anli_r2       | acc    | 0.409   | 0.409       | 0.518        |
| anli_r3       | acc    | 0.411   | 0.393       | 0.4983       |
| wic           | acc    | 0.529   | 0.545       | 0.5376       |
| rte           | acc    | 0.685   | 0.7039      | 0.7617       |
| record        | f1     | 0.904   | 0.9011      | 0.924        |
|               | em     | 0.8962  | 0.8927      | 0.916        |
| Average       |        | 0.596   | 0.5937      | 0.6708       |