Steelskull commited on
Commit
b764197
1 Parent(s): fc804ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -145
README.md CHANGED
@@ -5,7 +5,6 @@ tags:
5
  - frankenmoe
6
  - merge
7
  - mergekit
8
- - lazymergekit
9
  - Himitsui/Kaiju-11B
10
  - Sao10K/Fimbulvetr-11B-v2
11
  - decapoda-research/Antares-11b-v2
@@ -16,157 +15,111 @@ base_model:
16
  - decapoda-research/Antares-11b-v2
17
  - beberik/Nyxene-v3-11B
18
  ---
 
 
 
 
 
 
 
 
 
 
19
 
20
- # Umbra-v3-MoE-4x11b
 
 
 
 
 
 
 
 
 
 
21
 
22
- Umbra-v3-MoE-4x11b is a Mixure of Experts (MoE) made with the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
23
- * [Himitsui/Kaiju-11B](https://huggingface.co/Himitsui/Kaiju-11B)
24
- * [Sao10K/Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2)
25
- * [decapoda-research/Antares-11b-v2](https://huggingface.co/decapoda-research/Antares-11b-v2)
26
- * [beberik/Nyxene-v3-11B](https://huggingface.co/beberik/Nyxene-v3-11B)
 
27
 
28
- ## 🧩 Configuration
 
 
 
 
29
 
30
- ```yaml
31
- base_model: vicgalle/CarbonBeagle-11B-truthy
32
- gate_mode: hidden
33
- dtype: bfloat16
34
- experts_per_token: 4
35
- experts:
36
- - source_model: Himitsui/Kaiju-11B
37
- positive_prompts:
38
- - "Imagine"
39
- - "Create"
40
- - "Envision"
41
- - "Fantasize"
42
- - "Invent"
43
- - "Narrate"
44
- - "Plot"
45
- - "Portray"
46
- - "Storytell"
47
- - "Visualize"
48
- - "Describe"
49
- - "Develop"
50
- - "Forge"
51
- - "Craft"
52
- - "Conceptualize"
53
- - "Dream"
54
- - "Concoct"
55
- - "Characterize"
56
- negative_prompts:
57
- - "Recite"
58
- - "Report"
59
- - "Summarize"
60
- - "Enumerate"
61
- - "List"
62
- - "Cite"
63
 
64
- - source_model: Sao10K/Fimbulvetr-11B-v2
65
- positive_prompts:
66
- - "Dramatize"
67
- - "Embody"
68
- - "Illustrate"
69
- - "Perform"
70
- - "Roleplay"
71
- - "Simulate"
72
- - "Stage"
73
- - "Unfold"
74
- - "Weave"
75
- - "Design"
76
- - "Outline"
77
- - "Script"
78
- - "Sketch"
79
- - "Spin"
80
- - "Depict"
81
- - "Render"
82
- - "Fashion"
83
- - "Conceive"
84
- negative_prompts:
85
- - "Analyze"
86
- - "Critique"
87
- - "Dissect"
88
- - "Explain"
89
- - "Clarify"
90
- - "Interpret"
91
 
92
- - source_model: decapoda-research/Antares-11b-v2
93
- positive_prompts:
94
- - "Solve"
95
- - "Respond"
96
- - "Convey"
97
- - "Disclose"
98
- - "Expound"
99
- - "Narrate"
100
- - "Present"
101
- - "Reveal"
102
- - "Specify"
103
- - "Uncover"
104
- - "Decode"
105
- - "Examine"
106
- - "Report"
107
- - "Survey"
108
- - "Validate"
109
- - "Verify"
110
- - "Question"
111
- - "Query"
112
- negative_prompts:
113
- - "Divert"
114
- - "Obscure"
115
- - "Overstate"
116
- - "Undermine"
117
- - "Misinterpret"
118
- - "Skew"
119
 
120
- - source_model: beberik/Nyxene-v3-11B
121
- positive_prompts:
122
- - "Explain"
123
- - "Instruct"
124
- - "Clarify"
125
- - "Educate"
126
- - "Guide"
127
- - "Inform"
128
- - "Teach"
129
- - "Detail"
130
- - "Elaborate"
131
- - "Enlighten"
132
- - "Advise"
133
- - "Interpret"
134
- - "Analyze"
135
- - "Define"
136
- - "Demonstrate"
137
- - "Illustrate"
138
- - "Simplify"
139
- - "Summarize"
140
- negative_prompts:
141
- - "Speculate"
142
- - "Fabricate"
143
- - "Exaggerate"
144
- - "Mislead"
145
- - "Confuse"
146
- - "Distort"
147
 
148
- ```
 
 
 
 
 
 
 
 
149
 
150
- ## 💻 Usage
 
 
151
 
152
- ```python
153
- !pip install -qU transformers bitsandbytes accelerate
154
-
155
- from transformers import AutoTokenizer
156
- import transformers
157
- import torch
158
-
159
- model = "Steelskull/Umbra-v3-MoE-4x11b"
160
-
161
- tokenizer = AutoTokenizer.from_pretrained(model)
162
- pipeline = transformers.pipeline(
163
- "text-generation",
164
- model=model,
165
- model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
166
- )
167
-
168
- messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
169
- prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
170
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
171
- print(outputs[0]["generated_text"])
172
- ```
 
 
 
 
 
 
 
 
 
 
 
5
  - frankenmoe
6
  - merge
7
  - mergekit
 
8
  - Himitsui/Kaiju-11B
9
  - Sao10K/Fimbulvetr-11B-v2
10
  - decapoda-research/Antares-11b-v2
 
15
  - decapoda-research/Antares-11b-v2
16
  - beberik/Nyxene-v3-11B
17
  ---
18
+ <!DOCTYPE html>
19
+ <style>
20
+ body {
21
+ font-family: 'Quicksand', sans-serif;
22
+ background: linear-gradient(135deg, #2E3440 0%, #1A202C 100%);
23
+ color: #D8DEE9;
24
+ margin: 0;
25
+ padding: 0;
26
+ font-size: 16px;
27
+ }
28
 
29
+ .container {
30
+ width: 80%;
31
+ max-width: 800px;
32
+ margin: 20px auto;
33
+ background-color: rgba(255, 255, 255, 0.02);
34
+ padding: 20px;
35
+ border-radius: 12px;
36
+ box-shadow: 0 4px 10px rgba(0, 0, 0, 0.2);
37
+ backdrop-filter: blur(10px);
38
+ border: 1px solid rgba(255, 255, 255, 0.1);
39
+ }
40
 
41
+ .header h1 {
42
+ font-size: 28px;
43
+ color: #ECEFF4;
44
+ margin: 0 0 20px 0;
45
+ text-shadow: 2px 2px 4px rgba(0, 0, 0, 0.3);
46
+ }
47
 
48
+ .info img {
49
+ width: 100%;
50
+ border-radius: 10px;
51
+ margin-bottom: 15px;
52
+ }
53
 
54
+ a {
55
+ color: #88C0D0;
56
+ text-decoration: none;
57
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
 
59
+ a:hover {
60
+ color: #A3BE8C;
61
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62
 
63
+ a::before {
64
+ content: '';
65
+ position: absolute;
66
+ width: 100%;
67
+ height: 2px;
68
+ bottom: -5px;
69
+ left: 0;
70
+ background-color: #A3BE8C;
71
+ visibility: hidden;
72
+ transform: scaleX(0);
73
+ transition: all 0.3s ease-in-out;
74
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
 
76
+ a:hover::before {
77
+ visibility: visible;
78
+ transform: scaleX(1);
79
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
 
81
+ .button {
82
+ display: inline-block;
83
+ background-color: #5E81AC;
84
+ color: #E5E9F0;
85
+ padding: 10px 20px;
86
+ border-radius: 5px;
87
+ cursor: pointer;
88
+ text-decoration: none;
89
+ }
90
 
91
+ .button:hover {
92
+ background-color: #81A1C1;
93
+ }
94
 
95
+ </style>
96
+ <html lang="en">
97
+ <head>
98
+ <meta charset="UTF-8">
99
+ <meta name="viewport" content="width=device-width, initial-scale=1.0">
100
+ <title>Umbra-v3-MoE-4x11b Data Card</title>
101
+ <link href="https://fonts.googleapis.com/css2?family=Quicksand:wght@400;500;600&display=swap" rel="stylesheet">
102
+ </head>
103
+ <body>
104
+ <div class="container">
105
+ <div class="header">
106
+ <h1>Umbra-v3-MoE-4x11b</h1>
107
+ </div>
108
+ <div class="info">
109
+ <img src="https://cdn-uploads.huggingface.co/umbra-v3-moe-visual.webp">
110
+ <p><strong>Creator:</strong> <a href="https://huggingface.co/Steelskull" target="_blank">SteelSkull</a></p>
111
+ <p><strong>About Umbra-v3-MoE-4x11b:</strong> A Mixture of Experts model designed for general assistance with a special knack for storytelling and RP/ERP, built using LazyMergekit.</p>
112
+ <p>Integrates models from notable sources for enhanced performance in diverse tasks.</p>
113
+ <p><strong>Source Models:</strong></p>
114
+ <ul>
115
+ <li><a href="https://huggingface.co/Himitsui/Kaiju-11B">Himitsui/Kaiju-11B</a></li>
116
+ <li><a href="https://huggingface.co/Sao10K/Fimbulvetr-11B-v2">Sao10K/Fimbulvetr-11B-v2</a></li>
117
+ <li><a href="https://huggingface.co/decapoda-research/Antares-11b-v2">decapoda-research/Antares-11b-v2</a></li>
118
+ <li><a href="https://huggingface.co/beberik/Nyxene-v3-11B">beberik/Nyxene-v3-11B</a></li>
119
+ </ul>
120
+ <p><strong>Configuration Highlights:</strong> Features a carefully curated mix of positive and negative prompts tailored to leverage the unique strengths of each model, enhancing its general AI capabilities while maintaining prowess in creative storytelling.</p>
121
+ <p><strong>Usage Instructions:</strong> Provided in the form of Python code snippets, making it easy for developers to integrate Umbra-v3 into their projects for advanced text generation tasks.</p>
122
+ </div>
123
+ </div>
124
+ </body>
125
+ </html>