DavidAU commited on
Commit
097d21b
·
verified ·
1 Parent(s): b0f371f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -0
README.md CHANGED
@@ -47,6 +47,24 @@ It requires Llama3 template and/or "Command-R" template.
47
 
48
  Example outputs below.
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
51
  However this can be extended using "rope" settings up to 32k.
52
 
 
47
 
48
  Example outputs below.
49
 
50
+ <B>Model Notes:</B>
51
+
52
+ - Detail, prose and fiction writing abilities are significantly increased vs L3 Instruct.
53
+ - For more varied prose (sentence/paragraph/dialog) raise the temp and/or add more instructions in your prompt(s).
54
+ - Role-players: Careful raising temp too high as it may affect instruction following.
55
+ - This model works with rep pen of 1 or higher, 1.05+ recommended.
56
+ - If you want a specific type of prose (IE horror) add in "(vivid horror)" or "(graphic vivid horror)" (no quotes) in your prompt(s).
57
+ - A lot of GPTisms have been removed. There are still a few however - errrrr.
58
+ - This is not a "happy ever after" model. It has a negative bias.
59
+ - Output length will vary however this model prefers shortly outputs unless you state the size.
60
+ - For creative uses, different quants will produce slightly different output.
61
+ - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
62
+ - Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
63
+
64
+ Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
65
+
66
+ The Imatrix versions of this model have even lower perplexity then both this model and Llama3 Instruct and enhanced output.
67
+
68
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
69
  However this can be extended using "rope" settings up to 32k.
70