viethq188 commited on
Commit
7dae800
1 Parent(s): 805754e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -1,3 +1,36 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ Merge AIDC-ai-business/Marcoroni-7B-v3, Q-bert/MetaMath-Cybertron-Starling, mistralai/Mistral-7B-Instruct-v0.2 using slerp merge from https://github.com/cg123/mergekit.
5
+
6
+ After that we trained DPO with HF data.
7
+
8
+ *config.yaml*
9
+ ```
10
+ slices:
11
+ - sources:
12
+ - model: AIDC-ai-business/Marcoroni-7B-v3
13
+ layer_range: [0, 32]
14
+ - model: Q-bert/MetaMath-Cybertron-Starling
15
+ layer_range: [0, 32]
16
+ merge_method: slerp
17
+ base_model: AIDC-ai-business/Marcoroni-7B-v3
18
+ parameters:
19
+ t:
20
+ - filter: self_attn
21
+ value: [0, 0.5, 0.3, 0.7, 1]
22
+ - filter: mlp
23
+ value: [1, 0.5, 0.7, 0.3, 0]
24
+ - value: 0.5
25
+ dtype: float16
26
+ ```
27
+
28
+ You can use alpaca template.
29
+ ```
30
+ template_format = """{system}
31
+ ### Instruction:
32
+ {prompt}
33
+
34
+ ### Response:
35
+ """
36
+ ```