beberik commited on
Commit
9077d89
1 Parent(s): 89d8821

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +113 -0
README.md ADDED
@@ -0,0 +1,113 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ ## Description
6
+
7
+ This repo contains bf16 files of Nyxene-v1-11B. Same as the [previous version](https://huggingface.co/beberik/Nyxene-11B) but I used newer models and tried to repeat what I experimented with when there were older models.
8
+
9
+ ## Model used
10
+ - [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha)
11
+ - [openaccess-ai-collective/DPOpenHermes-7B](https://huggingface.co/openaccess-ai-collective/DPOpenHermes-7B)
12
+ - [fblgit/juanako-7b-UNA](https://huggingface.co/fblgit/juanako-7b-UNA)
13
+ - [chargoddard/loyal-piano-m7](https://huggingface.co/chargoddard/loyal-piano-m7)
14
+ - [argilla/notus-7b-v1](https://huggingface.co/argilla/notus-7b-v1)
15
+
16
+ I added a new model because after the same action but using zephyr and dolphin the model turned out to be more creative.
17
+
18
+ ## Prompt template
19
+
20
+ The best one after further testing is this one:
21
+
22
+ ```
23
+ <|system|>
24
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
25
+ <|user|>
26
+ {prompt}
27
+ <|assistant|>
28
+ ```
29
+
30
+ ## The secret sauce
31
+
32
+ Nyxene-11B :
33
+ ```
34
+ slices:
35
+ - sources:
36
+ - model: chargoddard/loyal-piano-m7
37
+ layer_range: [0, 48]
38
+ - model: argilla/notus-7b-v1
39
+ layer_range: [0, 48]
40
+ merge_method: slerp
41
+ base_model: argilla/notus-7b-v1
42
+ parameters:
43
+ t:
44
+ - filter: lm_head
45
+ value: [0.75]
46
+ - filter: embed_tokens
47
+ value: [0.75]
48
+ - filter: self_attn
49
+ value: [0.75, 0.25]
50
+ - filter: mlp
51
+ value: [0.25, 0.75]
52
+ - filter: layernorm
53
+ value: [0.5, 0.5]
54
+ - filter: modelnorm
55
+ value: [0.75]
56
+ - value: 0.99 # fallback for rest of tensors
57
+ dtype: bfloat16
58
+ ```
59
+
60
+ loyal-piano-m7-juanako-11B :
61
+ ```
62
+ slices:
63
+ - sources:
64
+ - model: fblgit/juanako-7b-UNA
65
+ layer_range: [0, 24]
66
+ - sources:
67
+ - model: ehartford/dolphin-2.1-mistral-7b
68
+ layer_range: [8, 32]
69
+ merge_method: passthrough
70
+ dtype: bfloat16
71
+ ```
72
+
73
+ Starling-DPOHermes-11B :
74
+ ```
75
+ slices:
76
+ - sources:
77
+ - model: berkeley-nest/Starling-LM-7B-alpha
78
+ layer_range: [0, 24]
79
+ - sources:
80
+ - model: mlabonne/NeuralHermes-2.5-Mistral-7B
81
+ layer_range: [8, 32]
82
+ merge_method: passthrough
83
+ dtype: bfloat16
84
+ ```
85
+
86
+ Nyxene-11B :
87
+ ```
88
+ slices:
89
+ - sources:
90
+ - model: dolphin-juanako-11B
91
+ layer_range: [0, 48]
92
+ - model: Starling-NeuralHermes-11B
93
+ layer_range: [0, 48]
94
+ merge_method: slerp
95
+ base_model: dolphin-juanako-11B
96
+ parameters:
97
+ t:
98
+ - filter: lm_head
99
+ value: [0.75]
100
+ - filter: embed_tokens
101
+ value: [0.75]
102
+ - filter: self_attn
103
+ value: [0.75, 0.25]
104
+ - filter: mlp
105
+ value: [0.25, 0.75]
106
+ - filter: layernorm
107
+ value: [0.5, 0.5]
108
+ - filter: modelnorm
109
+ value: [0.75]
110
+ - value: 0.5 # fallback for rest of tensors
111
+ dtype: bfloat16
112
+ ```
113
+ I use [mergekit](https://github.com/cg123/mergekit) for all the manipulation told here.