secretmoon
commited on
Commit
•
2b219c0
1
Parent(s):
57fa47a
Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,7 @@ pipeline_tag: text-generation
|
|
9 |
|
10 |
## Overview
|
11 |
|
12 |
-
**Secretmoon/LoRA-Llama-3-MLP** is
|
13 |
|
14 |
![Night alicorn](https://huggingface.co/secretmoon/LoRA-Llama-3-MLP/resolve/main/profile.png)
|
15 |
|
@@ -36,74 +36,21 @@ The base model for this adapter is **Sao10K/L3-8B-Stheno-v3.1**, an excellent fi
|
|
36 |
- **Sample Packing:** True
|
37 |
- **LoRA Target Linear:** True
|
38 |
|
39 |
-
|
40 |
-
|
41 |
-
You can apply the adapter to the original Safetensors weights of the model and load it through Transformers, or you can merge this adapter with the base model weights and convert it to f16 .gguf for use in llama.cpp.
|
42 |
-
|
43 |
-
### Recommendations for LoRA Alpha
|
44 |
|
45 |
- **16:** Low influence
|
46 |
- **48:** Suggested optimal value (recommended)
|
47 |
- **64:** High influence, significantly impacting model behavior
|
48 |
- **128:** Very high influence, drastically changing language model behavior (not recommended)
|
49 |
|
50 |
-
|
51 |
-
|
52 |
-
```python
|
53 |
-
import torch
|
54 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
55 |
-
from peft import PeftModel
|
56 |
-
|
57 |
-
# Loading tokenizer
|
58 |
-
tokenizer = AutoTokenizer.from_pretrained("Sao10K/L3-8B-Stheno-v3.1")
|
59 |
-
|
60 |
-
# Load base model in fp16, if you have ~15gb VRAM at least
|
61 |
-
base_model = AutoModelForCausalLM.from_pretrained(
|
62 |
-
"Sao10K/L3-8B-Stheno-v3.1",
|
63 |
-
trust_remote_code=True,
|
64 |
-
device_map="auto",
|
65 |
-
torch_dtype=torch.float16, # optional if you have enough VRAM
|
66 |
-
)
|
67 |
-
|
68 |
-
# Loading LoRA
|
69 |
-
adapter_name = "secretmoon/LoRA-Llama-3-MLP"
|
70 |
-
model = PeftModel.from_pretrained(base_model, adapter_name)
|
71 |
-
model = model.eval()
|
72 |
-
|
73 |
-
# Text generation function
|
74 |
-
def generate_text(prompt, max_length=100, num_return_sequences=1):
|
75 |
-
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
76 |
-
outputs = model.generate(
|
77 |
-
**inputs,
|
78 |
-
max_length=max_length,
|
79 |
-
num_return_sequences=num_return_sequences,
|
80 |
-
no_repeat_ngram_size=2,
|
81 |
-
early_stopping=True
|
82 |
-
)
|
83 |
-
return [tokenizer.decode(output, skip_special_tokens=True) for output in outputs]
|
84 |
-
|
85 |
-
prompt = "Once upon a time"
|
86 |
-
generated_texts = generate_text(prompt)
|
87 |
-
for i, text in enumerate(generated_texts):
|
88 |
-
print(f"Generated text {i+1}:\n{text}\n")
|
89 |
-
```
|
90 |
-
Example output:
|
91 |
-
```plaintext
|
92 |
-
Generated text 1:
|
93 |
-
Once upon a time, there was a young filly named Luna. She was the younger sister of a powerful princess named Celestia. Luna lived in a beautiful castle with her sister and their parents, the king and queen. The castle was surrounded by a lush, green forest, and it was always filled with the sounds of birds singing and animals playing.
|
94 |
-
```
|
95 |
-
|
96 |
-
## Merge:
|
97 |
-
|
98 |
-
1. **Using Axolotl** (https://github.com/OpenAccess-AI-Collective/axolotl)
|
99 |
-
```bash
|
100 |
-
python3 -m axolotl.cli.merge_lora lora.yml --lora_model_dir="./completed-model"
|
101 |
-
```
|
102 |
|
103 |
-
|
104 |
-
|
105 |
-
|
106 |
-
|
|
|
|
|
107 |
|
108 |
## Other:
|
109 |
<br> You can contact me on telegram @monstor86 or discord @starlight2288
|
|
|
9 |
|
10 |
## Overview
|
11 |
|
12 |
+
GGUF merged with base model version of **[Secretmoon/LoRA-Llama-3-MLP](https://huggingface.co/secretmoon/LoRA-Llama-3-MLP)** LoRA adapter, LoRA Alpha=48. Secretmoon/LoRA-Llama-3-MLP is 8-bit LoRA adapter for the Llama-3-8B model, primarily designed to expand the model's knowledge of the MLP:FiM (My Little Pony: Friendship is Magic) universe. This adapter is ideal for generating fan fiction, role-playing scenarios, and other creative projects. The training data includes factual content from the Fandom wiki and canonical fan works that deeply explore the universe.
|
13 |
|
14 |
![Night alicorn](https://huggingface.co/secretmoon/LoRA-Llama-3-MLP/resolve/main/profile.png)
|
15 |
|
|
|
36 |
- **Sample Packing:** True
|
37 |
- **LoRA Target Linear:** True
|
38 |
|
39 |
+
### Recommendations for LoRA Alpha (If you merge LoRA to the model yourself)
|
|
|
|
|
|
|
|
|
40 |
|
41 |
- **16:** Low influence
|
42 |
- **48:** Suggested optimal value (recommended)
|
43 |
- **64:** High influence, significantly impacting model behavior
|
44 |
- **128:** Very high influence, drastically changing language model behavior (not recommended)
|
45 |
|
46 |
+
## How to Use
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
48 |
+
- **[llama.cpp](https://github.com/ggerganov/llama.cpp)**
|
49 |
+
The opensource framework for running GGUF LLM models on which all other interfaces are made.
|
50 |
+
- **[koboldcpp](https://github.com/LostRuins/koboldcpp)**
|
51 |
+
Lightweight open source fork llama.cpp with a simple graphical interface and many additional features. Optimized for RP.
|
52 |
+
- **[LM studio](https://lmstudio.ai/)**
|
53 |
+
Proprietary free fork llama.cpp with a graphical interface.
|
54 |
|
55 |
## Other:
|
56 |
<br> You can contact me on telegram @monstor86 or discord @starlight2288
|