Edit model card

Introduction

This model vistagi/Mixtral-8x7b-v0.1-sft is trained with Ultrachat-200K dataset through supervised finetuning using Mixtral-8x7b-v0.1 as the baseline model. The training is done with bfloat16 precision using LoRA.

Details

Used Librarys

  • torch
  • deepspeed
  • pytorch lightning
  • transformers
  • peft
Downloads last month
732
Safetensors
Model size
46.7B params
Tensor type
BF16
·
Inference API
Input a message to start chatting with vistagi/Mixtral-8x7b-v0.1-dpo.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Dataset used to train vistagi/Mixtral-8x7b-v0.1-dpo