Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
llama3
enhanced quants
max quants
maxcpu quants
horror
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -47,6 +47,24 @@ It requires Llama3 template and/or "Command-R" template.
|
|
47 |
|
48 |
Example outputs below.
|
49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
51 |
However this can be extended using "rope" settings up to 32k.
|
52 |
|
|
|
47 |
|
48 |
Example outputs below.
|
49 |
|
50 |
+
<B>Model Notes:</B>
|
51 |
+
|
52 |
+
- Detail, prose and fiction writing abilities are significantly increased vs L3 Instruct.
|
53 |
+
- For more varied prose (sentence/paragraph/dialog) raise the temp and/or add more instructions in your prompt(s).
|
54 |
+
- Role-players: Careful raising temp too high as it may affect instruction following.
|
55 |
+
- This model works with rep pen of 1 or higher, 1.05+ recommended.
|
56 |
+
- If you want a specific type of prose (IE horror) add in "(vivid horror)" or "(graphic vivid horror)" (no quotes) in your prompt(s).
|
57 |
+
- A lot of GPTisms have been removed. There are still a few however - errrrr.
|
58 |
+
- This is not a "happy ever after" model. It has a negative bias.
|
59 |
+
- Output length will vary however this model prefers shortly outputs unless you state the size.
|
60 |
+
- For creative uses, different quants will produce slightly different output.
|
61 |
+
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
62 |
+
- Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
|
63 |
+
|
64 |
+
Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
|
65 |
+
|
66 |
+
The Imatrix versions of this model have even lower perplexity then both this model and Llama3 Instruct and enhanced output.
|
67 |
+
|
68 |
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
69 |
However this can be extended using "rope" settings up to 32k.
|
70 |
|