license: apache-2.0 | |
MoE model build with: | |
1. https://github.com/cg123/mergekit/tree/mixtral | |
2. Mistral models, latest merges and fine tunes. | |
3. Expert prompts heavily inspired by https://huggingface.co/Kquant03/Eukaryote-8x7B-bf16 | |
For details check model files, there is config yaml I used to create that model. | |
Come back later for more details. |