InferenceIllusionist commited on
Commit
edbd6b4
1 Parent(s): 073b62f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -26,7 +26,7 @@ PROUDLY PRESENTS
26
 
27
  ## experiment_2_8b-iMat-GGUF
28
 
29
- <b>Quantization Notes: Quantized from 3500 checkpoint. Use repetition penalty (--repeat-penalty on llama.cpp) of ~1.15 with Q6_K and lower quants for best results </b>
30
 
31
  Quantized from fp16 with love.
32
  * Weighted quantizations were created using fp16 GGUF and [groups_merged-enhancedV2-TurboMini.txt](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-9432658) in 189 chunks and n_ctx=512
 
26
 
27
  ## experiment_2_8b-iMat-GGUF
28
 
29
+ <b>Quantization Notes: Quantized from 3500 checkpoint. Use repetition penalty (--repeat-penalty on llama.cpp) of ~1.15 with Q6_K and lower and ~1.18 with IQ3_M and lower for best results. </b>
30
 
31
  Quantized from fp16 with love.
32
  * Weighted quantizations were created using fp16 GGUF and [groups_merged-enhancedV2-TurboMini.txt](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-9432658) in 189 chunks and n_ctx=512