rAIfle commited on
Commit
0938d40
·
verified ·
1 Parent(s): fd8b336

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mistralai/Mixtral-8x7B-v0.1
4
+ - mistralai/Mixtral-8x7B-Instruct-v0.1
5
+ - jondurbin/bagel-dpo-8x7b-v0.2
6
+ - cognitivecomputations/dolphin-2.7-mixtral-8x7b
7
+ - NeverSleep/Noromaid-v0.4-Mixtral-Instruct-8x7b-Zloss
8
+ - ycros/BagelMIsteryTour-v2-8x7B
9
+ - smelborp/MixtralOrochi8x7B
10
+ library_name: transformers
11
+ tags:
12
+ - mergekit
13
+ - merge
14
+
15
+ ---
16
+
17
+ Quantized using 200 samples of 8192 tokens from an RP-oriented [PIPPA](https://huggingface.co/datasets/royallab/PIPPA-cleaned) dataset.
18
+
19
+ Branches:
20
+ - `main` -- `measurement.json`
21
+ - `6b6h` -- 6bpw, 6bit lm_head
22
+
23
+
24
+ Requires ExllamaV2 version 0.0.12 and up.
25
+
26
+ Original model link: [rhplus0831/maid-yuzu-v8-alter](https://huggingface.co/rhplus0831/maid-yuzu-v8-alter)
27
+
28
+ Original model README below.
29
+
30
+ ***
31
+
32
+ # maid-yuzu-v8-alter
33
+
34
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
35
+
36
+ v7's approach worked better than I thought, so I tried something even weirder as a test. I don't think a proper model will come out, but I'm curious about the results.
37
+
38
+ ## Merge Details
39
+ ### Merge Method
40
+
41
+ This model was merged using the SLERP merge method.
42
+
43
+ This models were merged using the SLERP method in the following order:
44
+
45
+ maid-yuzu-v8-base: mistralai/Mixtral-8x7B-v0.1 + mistralai/Mixtral-8x7B-Instruct-v0.1 = 0.5
46
+ maid-yuzu-v8-step1: above + jondurbin/bagel-dpo-8x7b-v0.2 = 0.25
47
+ maid-yuzu-v8-step2: above + cognitivecomputations/dolphin-2.7-mixtral-8x7b = 0.25
48
+ maid-yuzu-v8-step3: above + NeverSleep/Noromaid-v0.4-Mixtral-Instruct-8x7b-Zloss = 0.25
49
+ maid-yuzu-v8-step4-alter: above + ycros/BagelMIsteryTour-v2-8x7B = 0.5
50
+ maid-yuzu-v8-alter: above + smelborp/MixtralOrochi8x7B = 0.5
51
+
52
+ ### Models Merged
53
+
54
+ The following models were included in the merge:
55
+ * [smelborp/MixtralOrochi8x7B](https://huggingface.co/smelborp/MixtralOrochi8x7B)
56
+ * ../maid-yuzu-v8-step4-alter
57
+
58
+ ### Configuration
59
+
60
+ The following YAML configuration was used to produce this model:
61
+
62
+ ```yaml
63
+ base_model:
64
+ model:
65
+ path: ../maid-yuzu-v8-step4-alter
66
+ dtype: bfloat16
67
+ merge_method: slerp
68
+ parameters:
69
+ t:
70
+ - value: 0.5
71
+ slices:
72
+ - sources:
73
+ - layer_range: [0, 32]
74
+ model:
75
+ model:
76
+ path: ../maid-yuzu-v8-step4-alter
77
+ - layer_range: [0, 32]
78
+ model:
79
+ model:
80
+ path: smelborp/MixtralOrochi8x7B
81
+ ```
82
+