|
--- |
|
license: apache-2.0 |
|
library_name: transformers |
|
tags: |
|
- mergekit |
|
- merge |
|
base_model: |
|
- openchat/openchat-3.5-0106 |
|
model-index: |
|
- name: OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
results: |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: IFEval (0-Shot) |
|
type: HuggingFaceH4/ifeval |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: inst_level_strict_acc and prompt_level_strict_acc |
|
value: 59.76 |
|
name: strict accuracy |
|
- type: inst_level_strict_acc and prompt_level_strict_acc |
|
value: 59.76 |
|
name: strict accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: BBH (3-Shot) |
|
type: BBH |
|
args: |
|
num_few_shot: 3 |
|
metrics: |
|
- type: acc_norm |
|
value: 24.06 |
|
name: normalized accuracy |
|
- type: acc_norm |
|
value: 24.06 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MATH Lvl 5 (4-Shot) |
|
type: hendrycks/competition_math |
|
args: |
|
num_few_shot: 4 |
|
metrics: |
|
- type: exact_match |
|
value: 6.8 |
|
name: exact match |
|
- type: exact_match |
|
value: 6.8 |
|
name: exact match |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: GPQA (0-shot) |
|
type: Idavidrein/gpqa |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: acc_norm |
|
value: 7.61 |
|
name: acc_norm |
|
- type: acc_norm |
|
value: 7.61 |
|
name: acc_norm |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MuSR (0-shot) |
|
type: TAUR-Lab/MuSR |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: acc_norm |
|
value: 11.78 |
|
name: acc_norm |
|
- type: acc_norm |
|
value: 11.78 |
|
name: acc_norm |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MMLU-PRO (5-shot) |
|
type: TIGER-Lab/MMLU-Pro |
|
config: main |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 25.44 |
|
name: accuracy |
|
- type: acc |
|
value: 25.44 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Pretergeek/OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
name: Open LLM Leaderboard |
|
--- |
|
<p align="center"> |
|
<a href="https://ko-fi.com/pretergeek">Buy me a Ko-Fi</a> • |
|
<a href="https://patreon.com/Pretergeek">Support my work using Patreon</a> |
|
</p> |
|
|
|
# OpenChat-3.5-0106_BlockExpansion-36Layers-End |
|
|
|
This is NOT your usual frankenmerge created using [mergekit](https://github.com/cg123/mergekit). |
|
|
|
## Merge Details |
|
### Merge Method |
|
|
|
This model was merged using the passthrough merge method, but employing a variation of the Block Expansion method described in the paper [LLaMA Pro: Progressive LLaMA with Block Expansion](https://arxiv.org/abs/2401.02415). |
|
|
|
The authors of the paper added new layers interleaved in between the original layers of the model, setting the parameters of the o_proj and down_proj layers to zero. This effectively adds layers that will just output their input (as if they were "transparent") allowing the model to remain functional even without further training. These new layers can then be targeted during training or fine-tuning without risking catastrophic forgetting, if you follow the author's training method to freeze the original layers and only train the new layers. |
|
|
|
I used the same method but added the new layers to the end of the model. My rationale is that the level of abstraction increases with each layer of the model. So, while new layers spread along the original layers will help the model to learn new tasks, adding layers to the end of the model and then re-training/fine-tuning the model on tasks it already performs well could improve the models understanding of those task and perform them better by employing more complex reasoning. |
|
|
|
This model has not yet received additional training, so it should perform close to the original model. |
|
|
|
### Models Merged |
|
|
|
The following models were included in the merge: |
|
* [openchat/openchat-3.5-0106](https://huggingface.co/openchat/openchat-3.5-0106) |
|
|
|
### Configuration |
|
|
|
The following YAML configuration was used to produce this model: |
|
|
|
```yaml |
|
slices: |
|
- sources: |
|
- model: openchat/openchat-3.5-0106 |
|
layer_range: [0, 32] |
|
- sources: |
|
- model: openchat/openchat-3.5-0106 |
|
layer_range: [31, 32] |
|
parameters: |
|
scale: |
|
- filter: o_proj |
|
value: 0.0 |
|
- filter: down_proj |
|
value: 0.0 |
|
- value: 1.0 |
|
- sources: |
|
- model: openchat/openchat-3.5-0106 |
|
layer_range: [31, 32] |
|
parameters: |
|
scale: |
|
- filter: o_proj |
|
value: 0.0 |
|
- filter: down_proj |
|
value: 0.0 |
|
- value: 1.0 |
|
- sources: |
|
- model: openchat/openchat-3.5-0106 |
|
layer_range: [31, 32] |
|
parameters: |
|
scale: |
|
- filter: o_proj |
|
value: 0.0 |
|
- filter: down_proj |
|
value: 0.0 |
|
- value: 1.0 |
|
- sources: |
|
- model: openchat/openchat-3.5-0106 |
|
layer_range: [31, 32] |
|
parameters: |
|
scale: |
|
- filter: o_proj |
|
value: 0.0 |
|
- filter: down_proj |
|
value: 0.0 |
|
- value: 1.0 |
|
merge_method: passthrough |
|
dtype: bfloat16 |
|
|
|
``` |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Pretergeek__OpenChat-3.5-0106_BlockExpansion-36Layers-End) |
|
|
|
| Metric |Value| |
|
|-------------------|----:| |
|
|Avg. |22.57| |
|
|IFEval (0-Shot) |59.76| |
|
|BBH (3-Shot) |24.06| |
|
|MATH Lvl 5 (4-Shot)| 6.80| |
|
|GPQA (0-shot) | 7.61| |
|
|MuSR (0-shot) |11.78| |
|
|MMLU-PRO (5-shot) |25.44| |
|
|
|
## Citation |
|
``` |
|
@misc{wu2024llamaproprogressivellama, |
|
title={LLaMA Pro: Progressive LLaMA with Block Expansion}, |
|
author={Chengyue Wu and Yukang Gan and Yixiao Ge and Zeyu Lu and Jiahao Wang and Ye Feng and Ying Shan and Ping Luo}, |
|
year={2024}, |
|
eprint={2401.02415}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL}, |
|
url={https://arxiv.org/abs/2401.02415}, |
|
} |
|
``` |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Pretergeek__OpenChat-3.5-0106_8.11B-36Layers-Last) |
|
|
|
| Metric |Value| |
|
|-------------------|----:| |
|
|Avg. |22.57| |
|
|IFEval (0-Shot) |59.76| |
|
|BBH (3-Shot) |24.06| |
|
|MATH Lvl 5 (4-Shot)| 6.80| |
|
|GPQA (0-shot) | 7.61| |
|
|MuSR (0-shot) |11.78| |
|
|MMLU-PRO (5-shot) |25.44| |
|
|
|
|