--- base_model: [] library_name: transformers tags: - mergekit - merge --- # ung-merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method. ### Models Merged The following models were included in the merge: * Miqu-PlayMaid-70B-v0.1 * Senku-70B-Full ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: linear parameters: weight: 1.0 slices: - sources: - model: Miqu-PlayMaid-70B-v0.1 layer_range: [0, 17] - sources: - model: Senku-70B-Full layer_range: [10, 24] - sources: - model: Miqu-PlayMaid-70B-v0.1 layer_range: [17, 32] - sources: - model: Senku-70B-Full layer_range: [24, 40] - sources: - model: Miqu-PlayMaid-70B-v0.1 layer_range: [32, 48] - sources: - model: Senku-70B-Full layer_range: [40, 56] - sources: - model: Miqu-PlayMaid-70B-v0.1 layer_range: [49, 63] - sources: - model: Senku-70B-Full layer_range: [56, 70] - sources: - model: Miqu-PlayMaid-70B-v0.1 layer_range: [64, 80] dtype: float16 tokenizer_source: model:Miqu-PlayMaid-70B-v0.1 ```