Edit model card

Mistral Instruct MoE experimental

This is a merge of pre-trained language models created using mergekit using the mixtral branch.

This is an experimental model and has nothing to do with Mixtral. Mixtral is not a merge of models per se, but a transformer with MoE layers learned during training

This uses a random gate, so I expect not great results. We'll see!

Merge Details

Merge Method

This model was merged using the MoE merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

base_model: mistralai/Mistral-7B-Instruct-v0.2
gate_mode: random
dtype: bfloat16
experts:
  - source_model: mistralai/Mistral-7B-Instruct-v0.2
    positive_prompts: [""]
  - source_model: mistralai/Mistral-7B-Instruct-v0.1
    positive_prompts: [""]

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 61.39
AI2 Reasoning Challenge (25-Shot) 61.01
HellaSwag (10-Shot) 81.55
MMLU (5-Shot) 58.22
TruthfulQA (0-shot) 60.40
Winogrande (5-shot) 76.09
GSM8k (5-shot) 31.08
Downloads last month
2,208
Safetensors
Model size
12.9B params
Tensor type
BF16
·

Merge of

Collection including osanseviero/mistral-instruct-moe-experimental

Evaluation results