base_model: | |
- princeton-nlp/gemma-2-9b-it-SimPO | |
- google/gemma-2-9b-it | |
- UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3 | |
- wzhouad/gemma-2-9b-it-WPO-HB | |
- google/gemma-2-9b | |
library_name: transformers | |
tags: | |
- mergekit | |
- merge | |
# merged_model | |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). | |
## Merge Details | |
### Merge Method | |
This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [google/gemma-2-9b](https://huggingface.co/google/gemma-2-9b) as a base. | |
### Models Merged | |
The following models were included in the merge: | |
* [princeton-nlp/gemma-2-9b-it-SimPO](https://huggingface.co/princeton-nlp/gemma-2-9b-it-SimPO) | |
* [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it) | |
* [UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3](https://huggingface.co/UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3) | |
* [wzhouad/gemma-2-9b-it-WPO-HB](https://huggingface.co/wzhouad/gemma-2-9b-it-WPO-HB) | |
### Configuration | |
The following YAML configuration was used to produce this model: | |
```yaml | |
models: | |
- model: google/gemma-2-9b | |
- model: google/gemma-2-9b-it | |
parameters: | |
density: 1.0 | |
weight: 0.6 | |
- model: wzhouad/gemma-2-9b-it-WPO-HB | |
parameters: | |
density: 0.55 | |
weight: 0.6 | |
- model: princeton-nlp/gemma-2-9b-it-SimPO | |
parameters: | |
density: 0.35 | |
weight: 0.6 | |
- model: UCLA-AGI/Gemma-2-9B-It-SPPO-Iter3 | |
parameters: | |
density: 0.25 | |
weight: 0.4 | |
merge_method: della | |
base_model: google/gemma-2-9b | |
parameters: | |
normalize: true | |
int8_mask: true | |
lambda: 1.0 | |
epsilon: 0.1 | |
dtype: float16 | |
``` | |