metadata
base_model:
- MaziyarPanahi/calme-3.2-instruct-78b
- dfurman/CalmeRys-78B-Orpo-v0.1
library_name: transformers
tags:
- mergekit
- merge
Calme-Ties-78B
Calme-Ties-78B is a 78-billion-parameter model merged using the TIES methodology, based on the Qwen2 architecture. It integrates two sub-base models: calme-3.2-instruct-78B by MaziyarPanahi and CalmeRys-78B-Orpo-v0.1 by dfurman, which serves as the base model. The merging process assigns equal weight and density to both models, with additional parameters enabling normalization and int8 masking. The model operates using the bfloat16 data type.
Model | Model Name | Model Link |
---|---|---|
Base Model | CalmeRys-78B-Orpo-v0.1 | CalmeRys-78B-Orpo-v0.1 |
Model 1 | calme-3.2-instruct-78B | calme-3.2-instruct-78B |
Model 2 | CalmeRys-78B-Orpo-v0.1 | CalmeRys-78B-Orpo-v0.1 |
Merged Models
This is a merge of pre-trained language models created using mergekit.
Merge Method
This model was merged using the TIES merge method using dfurman/CalmeRys-78B-Orpo-v0.1 as a base.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
models:
- model: MaziyarPanahi/calme-3.2-instruct-78b
parameters:
weight: 1
density: 1
merge_method: ties
base_model: dfurman/CalmeRys-78B-Orpo-v0.1
parameters:
weight: 1
density: 1
normalize: true
int8_mask: true
dtype: bfloat16