DavidAU commited on
Commit
a7e41a7
1 Parent(s): 3424995

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -62,7 +62,9 @@ Example outputs below.
62
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
63
  - Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
64
 
65
- Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
 
 
66
 
67
  The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
68
  lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
@@ -70,6 +72,8 @@ lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhance
70
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
71
  However this can be extended using "rope" settings up to 32k.
72
 
 
 
73
  Here is the standard LLAMA3 template:
74
 
75
  <PRE>
 
62
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
63
  - Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
64
 
65
+ Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first. Usually I would
66
+ use the "float32" version only, however the "character range" displayed by the Bfloat16 and Float32 versions of this model
67
+ dictate they have their own repos.
68
 
69
  The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
70
  lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
 
72
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
73
  However this can be extended using "rope" settings up to 32k.
74
 
75
+ If you use "Command-R" template your output will be very different from using "Llama3" template.
76
+
77
  Here is the standard LLAMA3 template:
78
 
79
  <PRE>