Azazelle commited on
Commit
1069f7a
1 Parent(s): 009a978

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +137 -19
README.md CHANGED
@@ -4,37 +4,155 @@ library_name: transformers
4
  tags:
5
  - mergekit
6
  - merge
7
-
 
 
8
  ---
9
- # 6Mk0jGS
 
 
 
 
10
 
11
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
12
 
13
- ## Merge Details
14
- ### Merge Method
15
 
16
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using output/virgin_rp as a base.
 
 
 
 
 
 
17
 
18
- ### Models Merged
19
 
20
- The following models were included in the merge:
21
- * output/virgin_smart
22
- * output/virgin_dumb
23
 
24
  ### Configuration
25
 
26
  The following YAML configuration was used to produce this model:
27
 
28
  ```yaml
29
- base_model: output/virgin_rp
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  dtype: float32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  merge_method: model_stock
32
- slices:
33
- - sources:
34
- - layer_range: [0, 32]
35
- model: output/virgin_smart
36
- - layer_range: [0, 32]
37
- model: output/virgin_dumb
38
- - layer_range: [0, 32]
39
- model: output/virgin_rp
40
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  tags:
5
  - mergekit
6
  - merge
7
+ - llama
8
+ - conversational
9
+ license: llama3
10
  ---
11
+ # L3-Persephone-8B-v1.0
12
+
13
+ ![Persephone]
14
+
15
+ ## About:
16
 
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
 
19
+ **Recommended Samplers:**
 
20
 
21
+ ```
22
+ Temperature - 1.0
23
+ TFS - 0.85
24
+ Smoothing Factor - 0.3
25
+ Smoothing Curve - 1.1
26
+ Repetition Penalty - 1.1
27
+ ```
28
 
29
+ ### Merge Method
30
 
31
+ This model was merged a series of model stock and lora merges, followed by ExPO. It uses a mix of smart and roleplay centered models to improve performance.
 
 
32
 
33
  ### Configuration
34
 
35
  The following YAML configuration was used to produce this model:
36
 
37
  ```yaml
38
+ # Smart model mixing
39
+ models:
40
+ - model: migtissera/Llama-3-8B-Synthia-v3.5
41
+ - model: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct
42
+ - model: openchat/openchat-3.6-8b-20240522
43
+ - model: NousResearch/Hermes-2-Pro-Llama-3-8B
44
+ - model: WhiteRabbitNeo/Llama-3-WhiteRabbitNeo-8B-v2.0
45
+ - model: chujiezheng/LLaMA3-iterative-DPO-final-ExPO
46
+ - model: chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
47
+ - model: NousResearch/Hermes-2-Theta-Llama-3-8B
48
+ - model: mlabonne/Daredevil-8B-abliterated
49
+ - model: mlabonne/NeuralDaredevil-8B-abliterated
50
+ - model: iRyanBell/ARC1
51
+ - model: iRyanBell/ARC1-II
52
+ - model: aaditya/Llama3-OpenBioLLM-8B
53
+ - model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_Alpha
54
+ - model: Locutusque/Llama-3-Hercules-5.0-8B
55
+ - model: OwenArli/Awanllm-Llama-3-8B-Cumulus-v1.0
56
+ - model: TIGER-Lab/MAmmoTH2-8B-Plus
57
+ - model: refuelai/Llama-3-Refueled
58
+ - model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3
59
+ - model: HPAI-BSC/Llama3-Aloe-8B-Alpha
60
+ - model: abacusai/Llama-3-Smaug-8B
61
+ - model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3
62
+ - model: turboderp/llama3-turbcat-instruct-8b
63
+ - model: nbeerbower/llama-3-gutenberg-8B
64
+ - model: chargoddard/prometheus-2-llama-3-8b
65
+ - model: Magpie-Align/Llama-3-8B-OpenHermes-2.5-1M
66
+ - model: Magpie-Align/Llama-3-8B-Magpie-Pro-MT-SFT-v0.1
67
+ merge_method: model_stock
68
+ base_model: NousResearch/Meta-Llama-3-8B-Instruct
69
  dtype: float32
70
+ vocab_type: bpe
71
+ name: stop_it_nerd
72
+ # RP Lora Mixing
73
+ models:
74
+ - model: stop_it_nerd+Azazelle/Llama-3-8B-Abomination-LORA
75
+ - model: stop_it_nerd+Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B
76
+ - model: stop_it_nerd+ToastyPigeon/Llama-3-8B-Instruct-SpringDragon-V2-QLoRA
77
+ - model: stop_it_nerd+Azazelle/Llama-3-LongStory-LORA
78
+ - model: stop_it_nerd+Azazelle/Llama3_RP_ORPO_LoRA
79
+ - model: stop_it_nerd+Azazelle/RP_Format_QuoteAsterisk_Llama3
80
+ - model: stop_it_nerd+Azazelle/Theory_of_Mind_Llama3
81
+ - model: stop_it_nerd+Azazelle/Aura_Llama3
82
+ - model: stop_it_nerd+Azazelle/Luna_Llama3
83
+ - model: stop_it_nerd+Azazelle/BlueMoon_Llama3
84
+ - model: stop_it_nerd+Azazelle/Smarts_Llama3
85
+ - model: stop_it_nerd+Azazelle/Nimue-8B
86
+ - model: stop_it_nerd+Azazelle/Llama-3-Instruct-LiPPA-LoRA-8B
87
+ - model: stop_it_nerd+Azazelle/go-bruins-v3-lora
88
+ - model: stop_it_nerd+Azazelle/L3-Daybreak-8b-lora
89
  merge_method: model_stock
90
+ base_model: stop_it_nerd
91
+ dtype: float32
92
+ vocab_type: bpe
93
+ name: nerdy_rp
94
+ # RP Model Mixing
95
+ models:
96
+ - model: ChaoticNeutrals/Hathor_RP-v.01-L3-8B
97
+ - model: TheDrummer/Llama-3SOME-8B-v2
98
+ - model: cgato/TheSalt-L3-8b-v0.3.2
99
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
100
+ - model: Sao10K/L3-8B-Stheno-v3.2
101
+ - model: ChaoticNeutrals/T-900-8B
102
+ - model: ResplendentAI/Nymph_8B
103
+ - model: vicgalle/Roleplay-Llama-3-8B
104
+ - model: maldv/badger-mu-llama-3-8b
105
+ - model: maldv/badger-iota-llama-3-8b
106
+ - model: ContextualAI/Llama-3-8B-Instruct-EPO-checkpoint5376
107
+ - model: hf-100/Llama-3-Spellbound-Instruct-8B-0.3
108
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
109
+ - model: lodrick-the-lafted/Limon-8B
110
+ - model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B
111
+ - model: turboderp/llama3-turbcat-instruct-8b
112
+ merge_method: model_stock
113
+ base_model: NousResearch/Meta-Llama-3-8B-Instruct
114
+ dtype: float32
115
+ vocab_type: bpe
116
+ name: true_rp
117
+ # Component Mixing
118
+ models:
119
+ - model: true_rp
120
+ - model: nerdy_rp
121
+ merge_method: model_stock
122
+ base_model: NousResearch/Meta-Llama-3-8B-Instruct
123
+ dtype: float32
124
+ vocab_type: bpe
125
+ name: virgin_rp
126
+ # Normal ExPO
127
+ models:
128
+ - model: virgin_rp
129
+ parameters:
130
+ weight: 1.28
131
+ merge_method: task_arithmetic
132
+ base_model: NousResearch/Meta-Llama-3-8B-Instruct
133
+ parameters:
134
+ normalize: false
135
+ dtype: float32
136
+ vocab_type: bpe
137
+ name: virgin_dumb
138
+ # Instruct ExPO
139
+ models:
140
+ - model: virgin_rp
141
+ parameters:
142
+ weight: 1.12
143
+ merge_method: task_arithmetic
144
+ base_model: NousResearch/Meta-Llama-3-8B
145
+ parameters:
146
+ normalize: false
147
+ dtype: float32
148
+ vocab_type: bpe
149
+ name: virgin_smart
150
+ # ExPO Mixing
151
+ models:
152
+ - model: virgin_smart
153
+ - model: virgin_dumb
154
+ merge_method: model_stock
155
+ base_model: virgin_rp
156
+ dtype: float32
157
+ vocab_type: bpe
158
+ ```