adamo1139 commited on
Commit
9536b40
1 Parent(s): ea1a3cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -23,6 +23,25 @@ I had to lower max_positional_embeddings in config.json and model_max_length for
23
  My first attempt had max_positional_embeddings set to 16384 and model_max_length set to 200000. This allowed fine-tuning to finish, but that model was broken after applying LoRA and merging it. \
24
  This attempt had both max_position_embeddings and model_max_length set to 4096, which worked perfectly fine.
25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ## Prompt Format
27
 
28
  I recommend using ChatML format, as this was used during fine-tune. \
 
23
  My first attempt had max_positional_embeddings set to 16384 and model_max_length set to 200000. This allowed fine-tuning to finish, but that model was broken after applying LoRA and merging it. \
24
  This attempt had both max_position_embeddings and model_max_length set to 4096, which worked perfectly fine.
25
 
26
+ ## Quants!
27
+
28
+ Huge thank you to LoneStriker and TheBloke for providing quantized versions.
29
+
30
+ EXL2 \
31
+ 3bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-3.0bpw-h6-exl2 \
32
+ 4bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-4.0bpw-h6-exl2 \
33
+ 4.65bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-4.65bpw-h6-exl2 \
34
+ 5bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-5.0bpw-h6-exl2 \
35
+ 6bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-6.0bpw-h6-exl2 \
36
+ 8bpw - https://huggingface.co/LoneStriker/Yi-34B-200K-AEZAKMI-v2-8.0bpw-h8-exl2
37
+
38
+ GGUF - https://huggingface.co/TheBloke/Yi-34B-200K-AEZAKMI-v2-GGUF
39
+
40
+ GPTQ - https://huggingface.co/TheBloke/Yi-34B-200K-AEZAKMI-v2-GPTQ
41
+
42
+ AWQ - https://huggingface.co/TheBloke/Yi-34B-200K-AEZAKMI-v2-AWQ
43
+
44
+
45
  ## Prompt Format
46
 
47
  I recommend using ChatML format, as this was used during fine-tune. \