--- license: mit library_name: transformers tags: - mergekit - moe - mixture of experts --- # Custom MOE with Mergekit Base model: `mistralai/Mistral-7B-Instruct-v0.2`. Fused with experts: - HuggingFaceH4/zephyr-7b-beta - mistralai/Mistral-7B-Instruct-v0.2 - teknium/OpenHermes-2.5-Mistral-7B - meta-math/MetaMath-Mistral-7B This can be downloaded to improve the base mistrals model to reason across mathematical and other objectives.