rAIfle commited on
Commit
fb2005e
1 Parent(s): 63ebd95

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -0
README.md ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+
7
+ ---
8
+
9
+ Quantized using 200 samples of 8192 tokens from an RP-oriented [PIPPA](https://huggingface.co/datasets/royallab/PIPPA-cleaned) dataset.
10
+
11
+ Branches:
12
+ - `main` -- `measurement.json`
13
+ - `2.25b6h` -- 2.25bpw, 6bit lm_head
14
+ - `3.5b6h` -- 3.5bpw, 6bit lm_head
15
+ - `6b6h` -- 6bpw, 6bit lm_head
16
+
17
+
18
+ Requires ExllamaV2 version 0.0.12 and up.
19
+
20
+ Original model link: [rAIfle/0x01-7x8b-hf](https://huggingface.co/rAIfle/0x01-7x8b-hf)
21
+
22
+ Original model README below.
23
+
24
+ ***
25
+ # 0x01-7x8B-hf
26
+
27
+ ![grinning female android, cyberpunk, robotic, biomechanical, serial number "0x01"](https://files.catbox.moe/je2zar.png)
28
+
29
+ here we go again. multi-step merge, various models involved at various ratios with various methods.
30
+
31
+ this thing came to me in a fever dream when I was hung over, but after slightly tweaking the recipe it turned out surprisingly decent. using with the settings included.
32
+
33
+ ## Constituent parts
34
+ ```yaml
35
+ # primordial_slop_a:
36
+ - model: mistralai/Mixtral-8x7B-v0.1+retrieval-bar/Mixtral-8x7B-v0.1_case-briefs
37
+ - model: mistralai/Mixtral-8x7B-v0.1+SeanWu25/Mixtral_8x7b_Medicine
38
+ - model: mistralai/Mixtral-8x7B-v0.1+SeanWu25/Mixtral_8x7b_WuKurtz
39
+ - model: mistralai/Mixtral-8x7B-v0.1+Epiculous/crunchy-onion-lora
40
+ - model: mistralai/Mixtral-8x7B-v0.1+maxkretchmer/gc-mixtral
41
+ # primordial_slop_b:
42
+ - model: Envoid/Mixtral-Instruct-ITR-8x7B
43
+ - model: crestf411/daybreak-mixtral-8x7b-v1.0-hf
44
+ - model: NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
45
+ - model: orangetin/OpenHermes-Mixtral-8x7B
46
+ - model: mistralai/Mixtral-8x7B-Instruct-v0.1+idegroup/PhyAssistant
47
+ - model: ycros/crunchy-onion-nx
48
+ - model: jondurbin/bagel-dpo-8x7b-v0.2
49
+ - model: amoldwalunj/Mixtral-8x7B-Instruct-v0.1-legal_finetune_mixtral_32k
50
+ # primordial_slop_c: a+b
51
+ # primordial_slop_d:
52
+ - model: Sao10K/Sensualize-Mixtral-bf16
53
+ - model: Envoid/Mixtral-Instruct-ITR-DADA-8x7B
54
+ ```
55
+
56
+ # mergekit
57
+
58
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
59
+
60
+ ## Merge Details
61
+ ### Merge Method
62
+
63
+ This model was merged using the SLERP merge method.
64
+
65
+ ### Models Merged
66
+
67
+ The following models were included in the merge:
68
+ * ./primordial_slop_d
69
+ * ./primordial_slop_c
70
+
71
+ ### Configuration
72
+
73
+ The following YAML configuration was used to produce this model:
74
+
75
+ ```yaml
76
+ models:
77
+ - model: ./primordial_slop_c
78
+ - model: ./primordial_slop_d
79
+ merge_method: slerp
80
+ base_model: ./primordial_slop_c
81
+ parameters:
82
+ t:
83
+ - value: 0.33
84
+ dtype: float16
85
+
86
+ ```