Kooten commited on
Commit
2167c42
1 Parent(s): f900adc

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ language:
4
+ - en
5
+ ---
6
+
7
+ # WinterGoddess-1.4x-70B-L2 IQ2-GGUF
8
+
9
+ ## Description
10
+ IQ2-GGUF quants of [Sao10K/WinterGoddess-1.4x-70B-L2](https://huggingface.co/Sao10K/WinterGoddess-1.4x-70B-L2)
11
+
12
+ Unlike regular GGUF quants this uses important matrix similar to Quip# to keep the quant from degrading too much even at 2bpw allowing you to run larger models on less powerful machines.
13
+
14
+ ***NOTE:*** As of uploading these this llamacpp can run these quants but i am unsure what guis like oobabooga / koboldcpp can run them.
15
+
16
+ [More info](https://github.com/ggerganov/llama.cpp/pull/4897)
17
+
18
+
19
+ # Models
20
+
21
+ Models: [IQ2-XS](https://huggingface.co/Kooten/WinterGoddess-1.4x-70B-L2-IQ2-GGUF/blob/main/WinterGoddess-1.4x-70B-L2-IQ2_XS.gguf), [IQ2-XXS](https://huggingface.co/Kooten/WinterGoddess-1.4x-70B-L2-IQ2-GGUF/blob/main/WinterGoddess-1.4x-70B-L2-IQ2_XXS.gguf)
22
+
23
+ Regular GGUF Quants: [Here](https://huggingface.co/TheBloke/WinterGoddess-1.4x-70B-L2-GGUF)
24
+
25
+ ## Prompt Format
26
+
27
+ ### Alpaca:
28
+ ```
29
+ ### Instruction:
30
+ <Prompt>
31
+
32
+ ### Response:
33
+
34
+ ```
35
+
36
+ OR
37
+
38
+ ```
39
+ ### Instruction:
40
+ <Prompt>
41
+
42
+ ### Input:
43
+ <Insert Context Here>
44
+
45
+ ### Response:
46
+
47
+
48
+ ## Contact
49
+ Kooten on discord