Wauplin HF staff commited on
Commit
41cac13
1 Parent(s): aa921bb

remove failing configs

Browse files
examples/gradient-slerp.yml DELETED
@@ -1,20 +0,0 @@
1
- slices:
2
- - sources:
3
- - model: psmathur/orca_mini_v3_13b
4
- layer_range: [0, 40]
5
- - model: garage-bAInd/Platypus2-13B
6
- layer_range: [0, 40]
7
- # or, the equivalent models: syntax:
8
- # models:
9
- # - model: psmathur/orca_mini_v3_13b
10
- # - model: garage-bAInd/Platypus2-13B
11
- merge_method: slerp
12
- base_model: psmathur/orca_mini_v3_13b
13
- parameters:
14
- t:
15
- - filter: self_attn
16
- value: [0, 0.5, 0.3, 0.7, 1]
17
- - filter: mlp
18
- value: [1, 0.5, 0.7, 0.3, 0]
19
- - value: 0.5 # fallback for rest of tensors
20
- dtype: float16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
examples/linear.yml DELETED
@@ -1,12 +0,0 @@
1
- models:
2
- - model: psmathur/orca_mini_v3_13b
3
- parameters:
4
- weight: 1.0
5
- - model: WizardLM/WizardLM-13B-V1.2
6
- parameters:
7
- weight: 0.3
8
- - model: garage-bAInd/Platypus2-13B
9
- parameters:
10
- weight: 0.5
11
- merge_method: linear
12
- dtype: float16
 
 
 
 
 
 
 
 
 
 
 
 
 
examples/mega.yml DELETED
@@ -1,37 +0,0 @@
1
- slices:
2
- - sources:
3
- - model: psmathur/orca_mini_v3_13b
4
- layer_range: [0, 40]
5
- - model: garage-bAInd/Platypus2-13B
6
- layer_range: [0, 40]
7
- merge_method: slerp
8
- base_model: psmathur/orca_mini_v3_13b
9
- parameters:
10
- t:
11
- - filter: self_attn
12
- value: [0, 0.5, 0.3, 0.7, 1]
13
- - filter: mlp
14
- value: [1, 0.5, 0.7, 0.3, 0]
15
- - value: 0.5 # fallback for rest of tensors
16
- dtype: float16
17
- name: gradient-slerp
18
- ---
19
- models:
20
- - model: gradient-slerp
21
- parameters:
22
- density: [1, 0.7, 0.1] # density gradient
23
- weight: 1.0
24
- - model: WizardLM/WizardMath-13B-V1.0
25
- parameters:
26
- density: 0.33
27
- weight:
28
- - filter: mlp
29
- value: 0.5
30
- - value: 0
31
- merge_method: ties
32
- base_model: TheBloke/Llama-2-13B-fp16
33
- parameters:
34
- normalize: true
35
- int8_mask: true
36
- dtype: float16
37
- name: gradient-slerp-ties
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
examples/orcamini-platy-44layer.yml DELETED
@@ -1,9 +0,0 @@
1
- slices:
2
- - sources:
3
- - model: psmathur/orca_mini_v3_13b
4
- layer_range: [0, 24]
5
- - sources:
6
- - model: garage-bAInd/Platypus2-13B
7
- layer_range: [20, 40]
8
- merge_method: passthrough
9
- dtype: float16
 
 
 
 
 
 
 
 
 
 
examples/ties.yml DELETED
@@ -1,22 +0,0 @@
1
- models:
2
- - model: psmathur/orca_mini_v3_13b
3
- parameters:
4
- density: [1, 0.7, 0.1] # density gradient
5
- weight: 1.0
6
- - model: garage-bAInd/Platypus2-13B
7
- parameters:
8
- density: 0.5
9
- weight: [0, 0.3, 0.7, 1] # weight gradient
10
- - model: WizardLM/WizardMath-13B-V1.0
11
- parameters:
12
- density: 0.33
13
- weight:
14
- - filter: mlp
15
- value: 0.5
16
- - value: 0
17
- merge_method: ties
18
- base_model: TheBloke/Llama-2-13B-fp16
19
- parameters:
20
- normalize: true
21
- int8_mask: true
22
- dtype: float16