## Fresh Alpasta, done Al Dente! It's da *logical* choice! Now with a similar personality emulation quality to [GPT4-X-Alpasta-30b!](https://huggingface.co/MetaIX/GPT4-X-Alpasta-30b) ## Model Info: ChanSung's [Alpaca-LoRA-30B-elina](https://huggingface.co/LLMs/Alpaca-LoRA-30B-elina) merged with [Open Assistant's second Finetune](https://huggingface.co/OpenAssistant/oasst-sft-7-llama-30b-xor) ## Benchmarks: **Wikitext2:** 4.662261962890625 **PTB:** 24.547462463378906 **C4:** 7.05504846572876 [4bit](https://huggingface.co/Aeala/GPT4-x-AlpacaDente2-30b/blob/main/4bit.safetensors): **Wikitext2:** 5.016242980957031 **PTB:** 25.576189041137695 **C4:** 7.332120418548584 ~ Thanks to [askmyteapot](https://huggingface.co/askmyteapot) for performing these benchmarks! # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Aeala__GPT4-x-AlpacaDente2-30b) | Metric | Value | |-----------------------|---------------------------| | Avg. | 57.05 | | ARC (25-shot) | 60.58 | | HellaSwag (10-shot) | 81.81 | | MMLU (5-shot) | 56.63 | | TruthfulQA (0-shot) | 48.38 | | Winogrande (5-shot) | 78.14 | | GSM8K (5-shot) | 26.76 | | DROP (3-shot) | 47.06 |