MLewd-v2.4-13B / README.md
Undi95's picture
Adding Evaluation Results (#3)
8e4a625
---
license: cc-by-nc-4.0
tags:
- not-for-all-audiences
- nsfw
---
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/_fVY7xvQ9tdoZ0nVDu_WB.png)
THIS MODEL IS MADE FOR LEWD
SEXUAL, CRUDE AND KINKY CONTENT IN OUTPUT CAN AND WILL HAPPEN. YOU'RE WARNED
Added the "magic touch" of MythoMax/Huginn/You call it.
In addition, [LimaRP v3](https://huggingface.co/lemonilia/LimaRP-Llama2-13B-v3-EXPERIMENT) was used, is it recommanded to read the documentation.
<!-- description start -->
## Description
This repo contains fp16 files of MLewd-2.4-13B, very hot and lewd model based on ReMM (SLERP).
<!-- description end -->
<!-- description start -->
## Models and loras used
- Undi95/ReMM-S-Light (base/private)
- Undi95/CreativeEngine
- Brouz/Slerpeno
- The-Face-Of-Goonery/Huginn-v3-13b
- zattio770/120-Days-of-LORA-v2-13B
- PygmalionAI/pygmalion-2-13b
- Undi95/StoryTelling
- TokenBender/sakhi_13B_roleplayer_NSFW_chat_adapter
- nRuaif/Kimiko-v2-13B
- The-Face-Of-Goonery/Huginn-13b-FP16
- lemonilia/LimaRP-Llama2-13B-v3-EXPERIMENT
<!-- description end -->
<!-- prompt-template start -->
## Prompt template: Alpaca
```
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:
```
## LimaRP v3 usage and suggested settings
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/ZC_iP2KkcEcRdgG_iyxYE.png)
You can follow these instruction format settings in SillyTavern. Replace tiny with your desired response length:
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/PIn8_HSPTJEMdSEpNVSdm.png)
Special thanks to Sushi and Shena ♥ | I love U hh_aa.
If you want to support me, you can [here](https://ko-fi.com/undiai).
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Undi95__MLewd-v2.4-13B)
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | 54.65 |
| ARC (25-shot) | 61.69 |
| HellaSwag (10-shot) | 83.83 |
| MMLU (5-shot) | 55.1 |
| TruthfulQA (0-shot) | 53.34 |
| Winogrande (5-shot) | 74.51 |
| GSM8K (5-shot) | 9.78 |
| DROP (3-shot) | 44.33 |