llmixer commited on
Commit
d19b7ff
1 Parent(s): 4fe291c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: unknown
3
+ language:
4
+ - en
5
+ pipeline_tag: conversational
6
+ tags:
7
+ - frankenmerge
8
+ - 108b
9
+ ---
10
+ # BigWeave v18 108b
11
+
12
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65a6db055c58475cf9e6def1/4CbbAN-X7ZWj702JrcCGH.png" width=600>
13
+
14
+ The BigWeave models aim to experimentally identify merge settings for increasing model performance. The version number merely tracks various attempts and is not a quality indicator. Only results demonstrating good performance are retained and shared.
15
+
16
+ # Prompting Format
17
+ Mistral, Vicuna and Alpaca.
18
+
19
+ # Merge process
20
+ This is a self-merge of 152334H/miqu-1-70b-sf. By conducting exl2 measurements, we identify the most relevant layers. The most important layers are extended with layers in-between to create longer series of consecutive layers.
21
+
22
+ Merge configuration:
23
+ ```
24
+ slices:
25
+ - sources:
26
+ - model: 152334H/miqu-1-70b-sf
27
+ layer_range: [0,5]
28
+ - sources:
29
+ - model: 152334H/miqu-1-70b-sf
30
+ layer_range: [1,9]
31
+ - sources:
32
+ - model: 152334H/miqu-1-70b-sf
33
+ layer_range: [5,33]
34
+ - sources:
35
+ - model: 152334H/miqu-1-70b-sf
36
+ layer_range: [16,51]
37
+ - sources:
38
+ - model: 152334H/miqu-1-70b-sf
39
+ layer_range: [34,77]
40
+ - sources:
41
+ - model: 152334H/miqu-1-70b-sf
42
+ layer_range: [75,79]
43
+ - sources:
44
+ - model: 152334H/miqu-1-70b-sf
45
+ layer_range: [77,80]
46
+ merge_method: passthrough
47
+ dtype: float16
48
+ ```