license: apache-2.0 | |
tags: | |
- alignment-handbook | |
- generated_from_trainer | |
datasets: | |
- allenai/ultrafeedback_binarized_cleaned | |
base_model: one-man-army/una-neural-chat-v3-3-P1-OMA | |
model-index: | |
- name: una-neural-chat-v3-3-P2 | |
results: [] | |
# una-neural-chat-v3-3-phase2 | |
OMA, OneManArmy proudly presents, `una-neural-chat-v3-3` **PHASE 2**. Powered by UNA (Uniform Neural Alignment), using zephyr trainer, allenai/ultrafeedback cleaned.. and JUST THAT. | |
Outperforming its base model, not adding any data.. just UNA Algorythm on Transformers Lib. | |
UNA Settings: | |
* MLP : 0.05 | |
* ATT : 0.03 | |
* LNOR : 0.02 | |
### Framework versions | |
- Transformers 4.35.0-UNA | |
- Pytorch 2.1.0 | |
- Datasets 2.14.6 | |
- Tokenizers 0.14.1 | |
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) | |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_one-man-army__una-neural-chat-v3-3-P2-OMA) | |
| Metric |Value| | |
|---------------------------------|----:| | |
|Avg. |70.72| | |
|AI2 Reasoning Challenge (25-Shot)|67.32| | |
|HellaSwag (10-Shot) |86.33| | |
|MMLU (5-Shot) |63.14| | |
|TruthfulQA (0-shot) |65.49| | |
|Winogrande (5-shot) |79.79| | |
|GSM8k (5-shot) |62.24| | |