--- language: - da - sv license: cc-by-4.0 library_name: transformers tags: - merge - mergekit base_model: - danish-foundation-models/munin-7b-alpha - timpal0l/Mistral-7B-v0.1-flashback-v2 --- # Danish-Swedish Merged Model This is a merge of the following models, all based on `mistralai/Mistral-7B-v0.1`: 1. `danish-foundation-models/munin-7b-alpha`, continued pretraining on Danish data; 2. `timpal0l/Mistral-7B-v0.1-flashback-v2`, continued pretraining on Swedish data. ## Model Details - **Merged by:** [Dan Saattrup Nielsen](https://www.saattrupdan.com/) - **Model type:** Decoder model, based on `mistralai/Mistral-7B-v0.1` - **Language(s):** Danish and Swedish - **License:** [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/) - **Merge configuration:** ```python dict( models=[ dict( model="danish-foundation-models/munin-7b-alpha", parameters=dict( weight=1.0, density=0.6, ), ), dict( model="timpal0l/Mistral-7B-v0.1-flashback-v2", parameters=dict( weight=1.0, density=0.6, ), ), ], merge_method="dare_ties", random_seed=4242 base_model="mistralai/Mistral-7B-v0.1", parameters=dict( int8_mask=True, normalize=True, ), dtype="bfloat16", ) ```