Nexesenex commited on
Commit
7f70caf
1 Parent(s): e41b9ee

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -0
README.md ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Custom GGUF Quants with iMatrix for :
2
+ https://huggingface.co/MarsupialAI/LaDameBlanche-v2-95b
3
+
4
+ - Q8_0 used as quant base : https://huggingface.co/mradermacher/LaDameBlanche-v2-95b-GGUF
5
+ - iMatrix here : https://huggingface.co/mradermacher/LaDameBlanche-v2-95b-i1-GGUF
6
+
7
+ (Yes, I'm lazy, but I can live with a 0.01ppl bump ^^)
8
+
9
+ The model is a great merge, sensical and creative, imho working better for lesser requirements than the 100b+ Miqu which are worthy only for those with 48GB VRAM or more.
10
+
11
+ In IQ2_LR(2.7BPW, for 8k context with 36GB VRAM and an IGP running the OS display), ARC Challenge at 57, ARC Easy at 77, PPL 512 at 4.5860.
12
+
13
+ Mesdames et messieurs, vous êtes servis!