beberik commited on
Commit
eb79d14
1 Parent(s): fd5c3ba

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ ## Description
6
+
7
+ This repo contains bf16 files of Nyxene-v1-11B. It feels like with the new models, 1% is no longer needed as in the [previous version](https://huggingface.co/beberik/Nyxene-v1-11B).
8
+
9
+ ## Model used
10
+ - [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha)
11
+ - [openaccess-ai-collective/DPOpenHermes-7B](https://huggingface.co/openaccess-ai-collective/DPOpenHermes-7B)
12
+ - [fblgit/fblgit/una-cybertron-7b-v2](https://huggingface.co/fblgit/una-cybertron-7b-v2-bf16)
13
+ - [chargoddard/loyal-piano-m7-cdpo](https://huggingface.co/chargoddard/loyal-piano-m7-cdpo)
14
+
15
+ ## Prompt template
16
+
17
+ The best one after further testing is this one:
18
+
19
+ ```
20
+ <|system|>
21
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
22
+ <|user|>
23
+ {prompt}
24
+ <|assistant|>
25
+ ```
26
+
27
+ ## The secret sauce
28
+
29
+ loyal-piano-cybertron-11B :
30
+ ```
31
+ slices:
32
+ - sources:
33
+ - model: fblgit/una-cybertron-7b-v2
34
+ layer_range: [0, 24]
35
+ - sources:
36
+ - model: chargoddard/loyal-piano-m7-cdpo
37
+ layer_range: [8, 32]
38
+ merge_method: passthrough
39
+ dtype: bfloat16
40
+ ```
41
+
42
+ Starling-DPOHermes-11B :
43
+ ```
44
+ slices:
45
+ - sources:
46
+ - model: berkeley-nest/Starling-LM-7B-alpha
47
+ layer_range: [0, 24]
48
+ - sources:
49
+ - model: openaccess-ai-collective/DPOpenHermes-7B
50
+ layer_range: [8, 32]
51
+ merge_method: passthrough
52
+ dtype: bfloat16
53
+ ```
54
+
55
+ Nyxene-11B :
56
+ ```
57
+ slices:
58
+ - sources:
59
+ - model: loyal-piano-cybertron-11B
60
+ layer_range: [0, 48]
61
+ - model: Starling-NeuralHermes-11B
62
+ layer_range: [0, 48]
63
+ merge_method: slerp
64
+ base_model: dolphin-juanako-11B
65
+ parameters:
66
+ t:
67
+ - filter: lm_head
68
+ value: [0.75]
69
+ - filter: embed_tokens
70
+ value: [0.75]
71
+ - filter: self_attn
72
+ value: [0.75, 0.25]
73
+ - filter: mlp
74
+ value: [0.25, 0.75]
75
+ - filter: layernorm
76
+ value: [0.5, 0.5]
77
+ - filter: modelnorm
78
+ value: [0.75]
79
+ - value: 0.5 # fallback for rest of tensors
80
+ dtype: bfloat16
81
+ ```
82
+ I use [mergekit](https://github.com/cg123/mergekit) for all the manipulation told here.