Text Generation
Transformers
PyTorch
Safetensors
mistral
Inference Endpoints
text-generation-inference
Edit model card

Wukong-0.1-Mistral-7B-v0.2

Join Our Discord! https://discord.gg/cognitivecomputations

image/jpeg

Wukong-0.1-Mistral-7B-v0.2 is a dealigned chat finetune of the original fantastic Mistral-7B-v0.2 model by the Mistral team.

This model was trained on the teknium OpenHeremes-2.5 dataset, code datasets from Multimodal Art Projection https://m-a-p.ai, and the Dolphin dataset from Cognitive Computations https://erichartford.com/dolphin 🐬

This model was trained for 3 epochs over 4 4090's.

Example Outputs

TBD

Built with Axolotl

Downloads last month
3
Safetensors
Model size
7.24B params
Tensor type
BF16
·

Datasets used to train hflog/RESMPDEV-Wukong-0.1-Mistral-7B-v0.2