Tristan Druyen commited on
Commit
a9770df
·
unverified ·
1 Parent(s): c8497d8

Add info abt imatrix dataset

Browse files
Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -17,5 +17,8 @@ imatrix generated with:
17
 
18
  ./imatrix -ofreq 4 -b 512 -c 512 -t 14 --chunks 24 -m ../models/Cerebrum-1.0-8x7b-GGUF/cerebrum-1.0-8x7b-Q8_0.gguf -f ./groups_merged.txt
19
 
 
 
 
20
  Sadly this means the imatrix is generated from the Q8 instead of the unquantized f16, like it should be, sadly I can't get it to work with the f16 on my machine at the moment. It should still improve the performance of the quants though.
21
 
 
17
 
18
  ./imatrix -ofreq 4 -b 512 -c 512 -t 14 --chunks 24 -m ../models/Cerebrum-1.0-8x7b-GGUF/cerebrum-1.0-8x7b-Q8_0.gguf -f ./groups_merged.txt
19
 
20
+ with the dataset from here:
21
+ https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384
22
+
23
  Sadly this means the imatrix is generated from the Q8 instead of the unquantized f16, like it should be, sadly I can't get it to work with the f16 on my machine at the moment. It should still improve the performance of the quants though.
24