beberik commited on
Commit
fb21218
1 Parent(s): ef779ae

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ ## Description
6
+
7
+ This repo contains bf16 files of Nyxene-v1-11B. Just new version with some new things.
8
+
9
+ ## Model used
10
+ - [Intel/neural-chat-7b-v3-3-Slerp](https://huggingface.co/Intel/neural-chat-7b-v3-3-Slerp)
11
+ - [AIDC-ai-business/Marcoroni-7B-v3](https://huggingface.co/AIDC-ai-business/Marcoroni-7B-v3)
12
+ - [rwitz/go-bruins-v2](https://huggingface.co/rwitz/go-bruins-v2)
13
+ - [chargoddard/loyal-piano-m7-cdpo](https://huggingface.co/chargoddard/loyal-piano-m7-cdpo)
14
+
15
+ ## Prompt template
16
+
17
+ Just use chatml.
18
+
19
+ ## The secret sauce
20
+
21
+ go-bruins-loyal-piano-11B :
22
+ ```
23
+ slices:
24
+ - sources:
25
+ - model: rwitz/go-bruins-v2
26
+ layer_range: [0, 24]
27
+ - sources:
28
+ - model: chargoddard/loyal-piano-m7-cdpo
29
+ layer_range: [8, 32]
30
+ merge_method: passthrough
31
+ dtype: bfloat16
32
+ ```
33
+
34
+ neural-marcoroni-11B :
35
+ ```
36
+ slices:
37
+ - sources:
38
+ - model: AIDC-ai-business/Marcoroni-7B-v3
39
+ layer_range: [0, 24]
40
+ - sources:
41
+ - model: Intel/neural-chat-7b-v3-3-Slerp
42
+ layer_range: [8, 32]
43
+
44
+ merge_method: passthrough
45
+ dtype: bfloat16
46
+ ```
47
+
48
+ Nyxene-11B :
49
+ ```
50
+ slices:
51
+ - sources:
52
+ - model: "./go-bruins-loyal-piano-11B"
53
+ layer_range: [0, 48]
54
+ - model: "./neural-marcoroni-11B"
55
+ layer_range: [0, 48]
56
+ merge_method: slerp
57
+ base_model: "./go-bruins-loyal-piano-11B"
58
+ parameters:
59
+ t:
60
+ - filter: lm_head
61
+ value: [0.5]
62
+ - filter: embed_tokens
63
+ value: [0.75]
64
+ - filter: self_attn
65
+ value: [0.75, 0.25]
66
+ - filter: mlp
67
+ value: [0.25, 0.75]
68
+ - filter: layernorm
69
+ value: [0.5, 0.5]
70
+ - filter: modelnorm
71
+ value: [0.5]
72
+ - value: 0.5 # fallback for rest of tensors
73
+ dtype: bfloat16
74
+ ```
75
+ I use [mergekit](https://github.com/cg123/mergekit) for all the manipulation told here.
76
+
77
+ Thanks to the [Undi95](https://huggingface.co/Undi95) for the original [11B mistral merge](https://huggingface.co/Undi95/Mistral-11B-OmniMix) recipe.