grimjim commited on
Commit
28324b9
1 Parent(s): 071f583

Initial release

Browse files
.gitattributes CHANGED
@@ -4,6 +4,7 @@
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
 
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gguf filter=lfs diff=lfs merge=lfs -text
8
  *.gz filter=lfs diff=lfs merge=lfs -text
9
  *.h5 filter=lfs diff=lfs merge=lfs -text
10
  *.joblib filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,52 @@
1
  ---
 
 
 
 
 
 
 
2
  license: cc-by-nc-4.0
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - grimjim/zephyr-beta-wizardLM-2-merge-7B
4
+ - alpindale/Mistral-7B-v0.2-hf
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
  license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
  ---
12
+ # madwind-wizard-7B-GGUF
13
+
14
+ This is a merge of pre-trained 7B language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ The intended goal of this merge was to combine the 32K context window of Mistral v0.2 base with the richness and strength of the Zephyr Beta and WizardLM 2 models. This was a mixed-precision merge, promoting Mistral v0.2 base from fp16 to bf16.
17
+
18
+ The result can be used for text generation. Note that Zephyr Beta training removed in-built alignment from datasets, resulting in a model more likely to generate problematic text when prompted. This merge appears to have inherited that feature.
19
+
20
+ - Full weights: [grimjim/madwind-wizard-7B](https://huggingface.co/grimjim/madwind-wizard-7B)
21
+ - GGUF quants: [grimjim/madwind-wizard-7B-GGUF](https://huggingface.co/grimjim/madwind-wizard-7B-GGUF)
22
+
23
+ ## Merge Details
24
+ ### Merge Method
25
+
26
+ This model was merged using the SLERP merge method.
27
+
28
+ ### Models Merged
29
+
30
+ The following models were included in the merge:
31
+ * [grimjim/zephyr-beta-wizardLM-2-merge-7B](https://huggingface.co/grimjim/zephyr-beta-wizardLM-2-merge-7B)
32
+ * [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf)
33
+
34
+ ### Configuration
35
+
36
+ The following YAML configuration was used to produce this model:
37
+
38
+ ```yaml
39
+ slices:
40
+ - sources:
41
+ - model: alpindale/Mistral-7B-v0.2-hf
42
+ layer_range: [0,32]
43
+ - model: grimjim/zephyr-beta-wizardLM-2-merge-7B
44
+ layer_range: [0,32]
45
+ merge_method: slerp
46
+ base_model: alpindale/Mistral-7B-v0.2-hf
47
+ parameters:
48
+ t:
49
+ - value: 0.5
50
+ dtype: bfloat16
51
+
52
+ ```
madwind-wizard-7B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83fd2b27f1a0102d2f3d273a4a43d8473c6a9aeaf723c63ad14ba43951386ea2
3
+ size 4368439072
madwind-wizard-7B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62821f889fc0f8ff869c93fcbf292ed7ebb0c0876d62e23282adfc0381ae72a9
3
+ size 5131409184
madwind-wizard-7B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21c057bd1694cd8f6965507e9aba2651d9ff8542a9e47d64511af0dddb3b5c83
3
+ size 5942064928
madwind-wizard-7B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb1e4bd92a8c7f0001ec0ab6d6512cad4ae60b1bca4f054ff57c21876d50d59e
3
+ size 7695857440