DavidAU commited on
Commit
e48bdd9
1 Parent(s): 5f51832

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -52
README.md CHANGED
@@ -118,58 +118,6 @@ of these quants will function better.
118
 
119
  Rep pen adjustments may also be required to get the most out of this model at this/these quant level(s).
120
 
121
- <B>Brainstorm 5x</B>
122
-
123
- The BRAINSTORM process was developed by David_AU.
124
-
125
- Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
126
- scientific paper : Progressive LLaMA with Block Expansion </a>.
127
-
128
- However I went in a completely different direction from what was outlined in this paper.
129
-
130
- What is "Brainstorm" ?
131
-
132
- The reasoning center of an LLM is taken apart, reassembled, and expanded.
133
-
134
- In this case for this model: 5 times
135
-
136
- Then these centers are individually calibrated. These "centers" also interact with each other.
137
- This introduces subtle changes into the reasoning process.
138
- The calibrations further adjust - dial up or down - these "changes" further.
139
- The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
140
-
141
- The core aim of this process is to increase the model's detail, concept and connection to the "world",
142
- general concept connections, prose quality and prose length without affecting instruction following.
143
-
144
- This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
145
-
146
- Here are some of the enhancements this process brings to the model's performance:
147
-
148
- - Prose generation seems more focused on the moment to moment.
149
- - Sometimes there will be "preamble" and/or foreshadowing present.
150
- - Fewer or no "cliches"
151
- - Better overall prose and/or more complex / nuanced prose.
152
- - A greater sense of nuance on all levels.
153
- - Coherence is stronger.
154
- - Description is more detailed, and connected closer to the content.
155
- - Simile and Metaphors are stronger and better connected to the prose, story, and character.
156
- - Sense of "there" / in the moment is enhanced.
157
- - Details are more vivid, and there are more of them.
158
- - Prose generation length can be long to extreme.
159
- - Emotional engagement is stronger.
160
- - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
161
- - The MORE instructions and/or details you provide the more strongly the model will respond.
162
- - Depending on the model "voice" may be more "human" vs original model's "voice".
163
-
164
- Other "lab" observations:
165
-
166
- - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
167
- - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
168
- - From lab testing it seems to ponder, and consider more carefully roughly speaking.
169
- - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
170
-
171
- The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
172
-
173
  <B>Models Used:</b>
174
 
175
  This is a high precision "DARE TIES" merge at the layer level (each layer per model adjusted - 168 points of adjustment over the 4 models) comprised of these models:
@@ -343,3 +291,59 @@ A low, guttural moan echoed from somewhere within the tower, vibrating through m
343
  Frozen on the precarious metal perch, trapped between the skeletal promise of the fire escape and the seething emerald abyss closing in behind me, I tasted ash and understood. Escape wasn't an option. Survival was a forgotten dream. This wasn't a fight. This was supplication. A silent plea whispered against the symphony of horrors beginning to unfold beyond the shattered windows of my world. The emerald dawn had broken, and with it, the feast. heridoeand I was the unwilling appetizer.
344
 
345
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
118
 
119
  Rep pen adjustments may also be required to get the most out of this model at this/these quant level(s).
120
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
121
  <B>Models Used:</b>
122
 
123
  This is a high precision "DARE TIES" merge at the layer level (each layer per model adjusted - 168 points of adjustment over the 4 models) comprised of these models:
 
291
  Frozen on the precarious metal perch, trapped between the skeletal promise of the fire escape and the seething emerald abyss closing in behind me, I tasted ash and understood. Escape wasn't an option. Survival was a forgotten dream. This wasn't a fight. This was supplication. A silent plea whispered against the symphony of horrors beginning to unfold beyond the shattered windows of my world. The emerald dawn had broken, and with it, the feast. heridoeand I was the unwilling appetizer.
292
 
293
 
294
+ ---
295
+
296
+ <h2>What is Brainstorm?</h2>
297
+
298
+ <B>Brainstorm 5x</B>
299
+
300
+ The BRAINSTORM process was developed by David_AU.
301
+
302
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
303
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
304
+
305
+ However I went in a completely different direction from what was outlined in this paper.
306
+
307
+ What is "Brainstorm" ?
308
+
309
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
310
+
311
+ In this case for this model: 5 times
312
+
313
+ Then these centers are individually calibrated. These "centers" also interact with each other.
314
+ This introduces subtle changes into the reasoning process.
315
+ The calibrations further adjust - dial up or down - these "changes" further.
316
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
317
+
318
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
319
+ general concept connections, prose quality and prose length without affecting instruction following.
320
+
321
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
322
+
323
+ Here are some of the enhancements this process brings to the model's performance:
324
+
325
+ - Prose generation seems more focused on the moment to moment.
326
+ - Sometimes there will be "preamble" and/or foreshadowing present.
327
+ - Fewer or no "cliches"
328
+ - Better overall prose and/or more complex / nuanced prose.
329
+ - A greater sense of nuance on all levels.
330
+ - Coherence is stronger.
331
+ - Description is more detailed, and connected closer to the content.
332
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
333
+ - Sense of "there" / in the moment is enhanced.
334
+ - Details are more vivid, and there are more of them.
335
+ - Prose generation length can be long to extreme.
336
+ - Emotional engagement is stronger.
337
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
338
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
339
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
340
+
341
+ Other "lab" observations:
342
+
343
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
344
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
345
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
346
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
347
+
348
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
349
+