Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
llama3
enhanced quants
max quants
maxcpu quants
horror
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -18,26 +18,29 @@ tags:
|
|
18 |
- all genres
|
19 |
- story
|
20 |
- writing
|
21 |
-
- vivid
|
22 |
- vivid writing
|
23 |
- fiction
|
24 |
- roleplaying
|
25 |
- bfloat16
|
26 |
- swearing
|
27 |
- rp
|
|
|
|
|
|
|
|
|
28 |
- horror
|
29 |
-
- mistral nemo
|
30 |
- mergekit
|
31 |
pipeline_tag: text-generation
|
32 |
---
|
33 |
|
34 |
-
Updates Dec 21 2024: (uploading quants ... refreshed, and new quants)
|
35 |
-
-
|
36 |
-
-
|
37 |
-
-
|
38 |
-
-
|
39 |
-
- "MAX"
|
40 |
-
- "MAX-CPU"
|
41 |
|
42 |
<h2>L3-Dark-Planet-8B-GGUF</h2>
|
43 |
|
|
|
18 |
- all genres
|
19 |
- story
|
20 |
- writing
|
21 |
+
- vivid prose
|
22 |
- vivid writing
|
23 |
- fiction
|
24 |
- roleplaying
|
25 |
- bfloat16
|
26 |
- swearing
|
27 |
- rp
|
28 |
+
- llama3
|
29 |
+
- enhanced quants
|
30 |
+
- max quants
|
31 |
+
- maxcpu quants
|
32 |
- horror
|
|
|
33 |
- mergekit
|
34 |
pipeline_tag: text-generation
|
35 |
---
|
36 |
|
37 |
+
<B>Updates Dec 21 2024: (uploading quants ... refreshed, and new quants):</B>
|
38 |
+
- All quants have been "refreshed", quanted with the lastest LLAMACPP improvements : Better instruction following, output generation across all quants.
|
39 |
+
- All quants have also been upgraded with "more bits" for output tensor and embed for better performance (this is in addition to the "refresh")
|
40 |
+
- New "ARM" quants have been added for machines than can run them. (format: ".../Q4_0_4_4.gguf")
|
41 |
+
- New specialized additional quants (in addition to standard): "max, max-cpu" (will include this in the file name) for quants "Q2K" (max cpu only), "IQ4_XS", "Q6_K" and "Q8_0"
|
42 |
+
- "MAX": output tensor / embed at float 16. (better instruction following/output generation than standard quants)
|
43 |
+
- "MAX-CPU": output tensor / embed at bfloat 16, which forces these on to the CPU (Nvidia cards / other will vary), this frees up vram at cost of token/second and you get better instruction following/output generation too.
|
44 |
|
45 |
<h2>L3-Dark-Planet-8B-GGUF</h2>
|
46 |
|