DavidAU commited on
Commit
e97c0f9
1 Parent(s): 25c2802

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +126 -0
README.md ADDED
@@ -0,0 +1,126 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - creative
7
+ - story
8
+ - writing
9
+ - fiction
10
+ - float32
11
+ - roleplaying
12
+ - rp
13
+ - enhanced
14
+ - space whale
15
+ - 32 bit upscale
16
+ ---
17
+
18
+ <font color=red><h3> Ultra Quality High Remaster of the incredible: Psyonic-Cetacean-20b - Imatrix Plus 2. </h3></font>
19
+
20
+ This is a Floating Point 32 upscale, where all components and merges were remastered to floating point 32.
21
+ This includes all the merges (recreated with master files), and where possible subbing full FP32 models.
22
+
23
+ This repo contains the new Imatrix Plus 2 quants using a new inhouse dataset merged with a master dataset
24
+ to push performance of the Ultra Quality remaster even higher.
25
+
26
+ <img src="space-whale-thinking.jpg">
27
+
28
+ The goal: Carry forward maximum precision right up to the point where it is "GUFFed".
29
+
30
+ This includes F32 master file for GGUF too... at a whopping 78 GBs.
31
+
32
+ WHY?
33
+
34
+ Because the difference between F32 vs BF16 is... over 8 DECIMAL places.
35
+
36
+ And as each merge / model is modified there are "losses" along the way.
37
+
38
+ These losses are carried forward and in turn lead to more losses.
39
+
40
+ And decimal points are critical to model performance.
41
+
42
+ SMALL?
43
+
44
+ Yes... but multipled by each merge(s), and compression(s): 20 billion times.
45
+
46
+ <B>The result:</b>
47
+
48
+ At Q2K an impressive drop of 533 points in perplexity. (lower is better)
49
+ (VS: Q2K original base model: PPL = 9.8077 +/- 0.06821 )
50
+
51
+ At Q4KM a whopping drop of 976 points in perplexity.
52
+ (VS: Q4km original base model -> PPL = 8.7858 +/- 0.06074)
53
+
54
+ At Q6 an awesome drop of 234 points in perplexity.
55
+ (VS: Q6 original base model -> PPL = 8.6070 +/- 0.05907 )
56
+
57
+ To put this in perspective "Q6" now operates ABOVE the original full precision version of "Psyonic-Cetacean-20b"
58
+ and Q4KM operates at close to Q6 level quality.
59
+
60
+ This because at "Q6" the quant / compressed model is considered to be accurate within "+0.0008 ppl" of the full,
61
+ uncompressed / unquanted model and it exceeds this threshold by over 200 points.
62
+
63
+ <I> Imatrix quants take this even further in most cases DOUBLING the "drop" in perplexity realized in the reg quants. </i>
64
+
65
+ Q4km-imatrix :
66
+
67
+ Final estimate: PPL = 8.6095 +/- 0.05898
68
+
69
+ (Non imatrix: Final estimate: PPL = 8.6902 +/- 0.05985 )
70
+
71
+ (VS: Q4km base model -> PPL = 8.7858 +/- 0.06074)
72
+
73
+ (VS: Q6 BASE model -> Final estimate: PPL = 8.6070 +/- 0.05907 Q6)
74
+
75
+
76
+ But... what about Q8?
77
+
78
+ The mountain moved:
79
+
80
+ 150 points better: PPL = 8.5850 +/- 0.05881 VS: BASE/ORGINAL: PPL = 8.6012 +/- 0.05900
81
+
82
+ <B>THE RESULTS ARE IN: </b>
83
+
84
+ AS per Jeb Carter, orginal creator of the model:
85
+
86
+ - instruction following has improved dramatically.
87
+ - new abilities have emerged.
88
+ - he had to REDUCE the instructions sets used because the model no longer needed as specific instructions.
89
+ - prose, nuance and depth have all improved.
90
+ - known issues with the original model have disappeared.
91
+
92
+ This is not "something for nothing" ; it is method of ensuring maximum precision at every step just before "ggufing" the model.
93
+
94
+ The methods employed only ensure precision loss is minimized or eliminated.
95
+
96
+ It is mathematical and theory sound.
97
+
98
+ <B>The bottom line here is this:</b>
99
+
100
+ Higher quality instruction following and output.
101
+
102
+ Likewise you can use a smaller compression, with higher token per second and still get great quality.
103
+
104
+ Same great model... turbo charged.
105
+
106
+ This is the first group of remasters.
107
+
108
+ <B>The FOUR Horsemen:</B>
109
+
110
+ This repo will be followed by a "reg quant plus" repo, which added additional components into the GGUF (all levels) at floating point 32
111
+ precision to further increase the sheer creativity and raw AI horsepower.
112
+
113
+ This process shaves at extra 50-100 points off perplexity... again.
114
+
115
+ Following this group will be a full float 32 precision Imatrix (including reg quants "imatrixed").
116
+
117
+ Test results VS org and "ultra" regular quants will be posted when they come in.
118
+
119
+ Imatrix Plus repo (with the same floating 32 enhancement at "reg quant plus") that will push the limit even more.
120
+
121
+ Details of all methods (and pitfalls to avoid) employed to make this high precision remasters will be
122
+ posted shortly along with comparisions of original model and new ultra remaster.
123
+
124
+ Thanks again to Jeb Carter, the original creator of "Psyonic-Cetacean 20B"
125
+
126
+ [ https://huggingface.co/jebcarter/psyonic-cetacean-20B ]