Maixtchup-4x7b / README.md
StatsGary's picture
Update README.md
c9abe46 verified
|
raw
history blame
No virus
440 Bytes
metadata
license: mit
library_name: transformers
tags:
  - mergekit
  - moe
  - mixture of experts

Custom MOE with Mergekit

Base model: mistralai/Mistral-7B-Instruct-v0.2. Fused with experts:

  • HuggingFaceH4/zephyr-7b-beta
  • mistralai/Mistral-7B-Instruct-v0.2
  • teknium/OpenHermes-2.5-Mistral-7B
  • meta-math/MetaMath-Mistral-7B

This can be downloaded to improve the base mistrals model to reason across mathematical and other objectives.