Edit model card

OpenCerebrum-2.0-7B

OpenCerebrum-2.0-7B is an open-source language model fine-tuned from the alpindale/Mistral-7B-v0.2-hf base model on a diverse dataset aimed at replicating capabilities of Aether Research's proprietary Cerebrum model.

The model was fine-tuned with SFT and DPO on approximately 7,000 examples across 15 data sources spanning coding, math, science, multi-turn conversation, RAG, reasoning, and general instruction-following. The goal was to assemble public datasets that could help the model achieve strong performance on benchmarks where Cerebrum excels.

Model Details

  • Base Model: alpindale/Mistral-7B-v0.2-hf
  • Parameters: 7 billion
  • Fine-Tuning Dataset Size: ~7,000 examples
  • Fine-Tuning Data: Advanced in-house curation techniques at Cognitive Computations, with 15 different data sources for DPO and SFT.
  • Language: English
  • License: Apache 2.0

Quants

EXL2 @bartowski

GGUF @bartowski

Intended Use

OpenCerebrum-2.0-7B is intended to be a powerful open-source model for coding, math, science, and general question-answering and text generation tasks. Its diverse fine-tuning data aims to equip it with broad knowledge and reasoning capabilities.

However, as an open-source replica trained on a subset of data compared to the original Cerebrum, it may not match Cerebrum's full performance. Additionally, biases and limitations of the fine-tuning data may be reflected in the model's outputs.

Limitations and Biases

  • The model may have biases and limitations inherited from its fine-tuning datasets. Thorough testing is needed to characterize these.
  • As the model is based on a 7B parameter model, it has computational and memory constraints compared to larger models.

Evaluations

Tasks Version Filter n-shot Metric Value Stderr
truthfulqa_mc2 2 none 0 acc 0.5182 ± 0.0152
ai2_arc N/A none 0 acc 0.7060 ± 0.0073
none 0 acc_norm 0.7049 ± 0.0074
- arc_challenge 1 none 0 acc 0.5000 ± 0.0146
none 0 acc_norm 0.5299 ± 0.0146
- arc_easy 1 none 0 acc 0.8077 ± 0.0081
none 0 acc_norm 0.7912 ± 0.0083
agieval_nous N/A none 0 acc 0.3778 ± 0.0093
none 0 acc_norm 0.3574 ± 0.0093
- agieval_aqua_rat 1 none 0 acc 0.2402 ± 0.0269
none 0 acc_norm 0.2205 ± 0.0261
- agieval_logiqa_en 1 none 0 acc 0.3164 ± 0.0182
none 0 acc_norm 0.3656 ± 0.0189
- agieval_lsat_ar 1 none 0 acc 0.2130 ± 0.0271
none 0 acc_norm 0.1913 ± 0.0260
- agieval_lsat_lr 1 none 0 acc 0.4078 ± 0.0218
none 0 acc_norm 0.3647 ± 0.0213
- agieval_lsat_rc 1 none 0 acc 0.4981 ± 0.0305
none 0 acc_norm 0.4498 ± 0.0304
- agieval_sat_en 1 none 0 acc 0.6650 ± 0.0330
none 0 acc_norm 0.5922 ± 0.0343
- agieval_sat_en_without_passage 1 none 0 acc 0.4612 ± 0.0348
none 0 acc_norm 0.3932 ± 0.0341
- agieval_sat_math 1 none 0 acc 0.3273 ± 0.0317
none 0 acc_norm 0.2818 ± 0.0304
Downloads last month
1,394
Safetensors
Model size
7.24B params
Tensor type
BF16
·

Collection including Locutusque/OpenCerebrum-2.0-7B