File size: 2,984 Bytes
246b5f9
 
ee6d7d6
f682ed1
246b5f9
 
 
57fa47a
 
 
 
2b219c0
57fa47a
 
 
 
 
f682ed1
57fa47a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2b219c0
57fa47a
 
 
 
 
 
2b219c0
57fa47a
2b219c0
 
 
 
 
 
57fa47a
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
---
license: cc-by-nc-4.0
library_name: transformers
base_model: secretmoon/LoRA-Llama-3-MLP
language:
- en
pipeline_tag: text-generation
---

## Overview

GGUF merged with base model version of **[Secretmoon/LoRA-Llama-3-MLP](https://huggingface.co/secretmoon/LoRA-Llama-3-MLP)** LoRA adapter, LoRA Alpha=48. Secretmoon/LoRA-Llama-3-MLP is 8-bit LoRA adapter for the Llama-3-8B model, primarily designed to expand the model's knowledge of the MLP:FiM (My Little Pony: Friendship is Magic) universe. This adapter is ideal for generating fan fiction, role-playing scenarios, and other creative projects. The training data includes factual content from the Fandom wiki and canonical fan works that deeply explore the universe.

![Night alicorn](https://huggingface.co/secretmoon/LoRA-Llama-3-MLP/resolve/main/profile.png)

## Base Model

The base model for this adapter is **[Sao10K/L3-8B-Stheno-v3.1](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.1)**, an excellent fine-tuned version of the original Llama-3-8B. It excels in story writing and role-playing without suffering from degradation due to overfitting.

## Training Details

- **Dataset:**
  1. Cleaned copy of the MLP Fandom Wiki, excluding information about recent and side projects unrelated to MLP:FiM. (Alpaca)
  2. Approximately 100 specially selected fan stories from FiMFiction. (RAW text)
  3. Additional data to train the model as a personal assistant and enhance its sensitivity to user emotions. (Alpaca)
- **Training Duration:** 3 hours
- **Hardware:** 1 x NVIDIA RTX A6000 48GB
- **PEFT Type:** LoRA 8-bit
- **Sequence Length:** 6144
- **Batch Size:** 2
- **Num Epochs:** 3
- **Optimizer:** AdamW_BNB_8bit
- **Learning Rate Scheduler:** Cosine
- **Learning Rate:** 0.00033
- **LoRA R:** 256
- **Sample Packing:** True
- **LoRA Target Linear:** True

### Recommendations for LoRA Alpha (If you merge LoRA to the model yourself)

- **16:** Low influence
- **48:** Suggested optimal value (recommended)
- **64:** High influence, significantly impacting model behavior
- **128:** Very high influence, drastically changing language model behavior (not recommended)

## How to Use

- **[llama.cpp](https://github.com/ggerganov/llama.cpp)**
  The opensource framework for running GGUF LLM models on which all other interfaces are made.
- **[koboldcpp](https://github.com/LostRuins/koboldcpp)**
  Lightweight open source fork llama.cpp with a simple graphical interface and many additional features. Optimized for RP.
- **[LM studio](https://lmstudio.ai/)**
  Proprietary free fork llama.cpp with a graphical interface.      

## Other:
<br> You can contact me on telegram @monstor86 or discord @starlight2288
<br> Also you can try some RP with this adapter for free in my bot on telegram @Luna_Pony_bot
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)