Text Generation
Transformers
Safetensors
llama
Not-For-All-Audiences
nsfw
text-generation-inference
Inference Endpoints
File size: 2,197 Bytes
1717392 c0cbe0b 1717392 a60f30c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 |
---
license: cc-by-nc-4.0
tags:
- not-for-all-audiences
- nsfw
---
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/bex2BCxY5n30mSsxvM-WW.png)
Attempt to recreate Amethyst-13B but in 20B. The two model was glued together with a sparkle of [Unholy](Undi95/Unholy-v1-12L-13B).
In addition, [LimaRP v3](https://huggingface.co/lemonilia/LimaRP-Llama2-13B-v3-EXPERIMENT) was used, is it recommanded to read the documentation.
<!-- description start -->
## Description
This repo contains fp16 files of Amethyst-20B.
<!-- description end -->
<!-- description start -->
## Models and loras used
- Xwin-LM/Xwin-LM-13B-V0.1
- The-Face-Of-Goonery/Huginn-13b-FP16
- zattio770/120-Days-of-LORA-v2-13B
- lemonilia/LimaRP-Llama2-13B-v3-EXPERIMENT
- Undi95/Unholy-v1-12L-13B
<!-- description end -->
<!-- prompt-template start -->
## Prompt template: Alpaca
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:
```
## LimaRP v3 usage and suggested settings
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/ZC_iP2KkcEcRdgG_iyxYE.png)
You can follow these instruction format settings in SillyTavern. Replace tiny with your desired response length:
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/PIn8_HSPTJEMdSEpNVSdm.png)
Special thanks to Sushi.
If you want to support me, you can [here](https://ko-fi.com/undiai).
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Undi95__U-Amethyst-20B)
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | 50.38 |
| ARC (25-shot) | 62.2 |
| HellaSwag (10-shot) | 83.11 |
| MMLU (5-shot) | 55.88 |
| TruthfulQA (0-shot) | 53.2 |
| Winogrande (5-shot) | 74.19 |
| GSM8K (5-shot) | 5.31 |
| DROP (3-shot) | 18.75 |
|