cgus commited on
Commit
229a8ed
1 Parent(s): 85f3628

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -53,7 +53,7 @@ tags:
53
 
54
  Experimental exl2 quantization for CausalLM-14B for Exllamav2.
55
  I had some issues during quantization process, so I suspect it might have quality issues.
56
- 3.5bpw version barely fits 12GB VRAM but has unusually high perplexity for wikitext dataset.
57
  I couldn't measure perplexity for 4bpw version and to compare it with TheBloke's GPTQ, so I have no idea if my quantization has issues or it supposed to be like this.
58
 
59
  You could try this exl2 version but I'd recommend to use [TheBloke's GPTQ](https://huggingface.co/TheBloke/CausalLM-14B-GPTQ) version instead.
 
53
 
54
  Experimental exl2 quantization for CausalLM-14B for Exllamav2.
55
  I had some issues during quantization process, so I suspect it might have quality issues.
56
+ 3.5bpw version barely fits my 12GB VRAM but has unusually high perplexity for wikitext dataset.
57
  I couldn't measure perplexity for 4bpw version and to compare it with TheBloke's GPTQ, so I have no idea if my quantization has issues or it supposed to be like this.
58
 
59
  You could try this exl2 version but I'd recommend to use [TheBloke's GPTQ](https://huggingface.co/TheBloke/CausalLM-14B-GPTQ) version instead.