Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
llama3
enhanced quants
max quants
maxcpu quants
horror
mergekit
Inference Endpoints
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -62,7 +62,9 @@ Example outputs below.
|
|
62 |
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
63 |
- Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
|
64 |
|
65 |
-
Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
|
|
|
|
|
66 |
|
67 |
The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
|
68 |
lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
|
@@ -70,6 +72,8 @@ lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhance
|
|
70 |
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
71 |
However this can be extended using "rope" settings up to 32k.
|
72 |
|
|
|
|
|
73 |
Here is the standard LLAMA3 template:
|
74 |
|
75 |
<PRE>
|
|
|
62 |
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
63 |
- Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
|
64 |
|
65 |
+
Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first. Usually I would
|
66 |
+
use the "float32" version only, however the "character range" displayed by the Bfloat16 and Float32 versions of this model
|
67 |
+
dictate they have their own repos.
|
68 |
|
69 |
The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
|
70 |
lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
|
|
|
72 |
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
73 |
However this can be extended using "rope" settings up to 32k.
|
74 |
|
75 |
+
If you use "Command-R" template your output will be very different from using "Llama3" template.
|
76 |
+
|
77 |
Here is the standard LLAMA3 template:
|
78 |
|
79 |
<PRE>
|