Edit model card

Model Card for Model ID

MistralAI 7B model fine-tuned for 1 epoch on Dataricks instruction tuning dataset.

Model Details

Model Description

Mistral Repository

Training Details

Training Data

Training Procedure

  • Quantized Low-Rank Adaptation (QLoRA)
  • Transformers Trainer
  • DataCollatorForSeq2Seq
  • Distributed Data Parallel (DDP) across two GPUs

Preprocessing

Manually created tokenized 'labels' for the dataset. Prompt template utilized basic template for instruction-tuning

Hardware

Performed fine-tuning with 2 * A100 GPUs

  • Provided by Gnewsoft during work period Model and dataset are too large for free run sessions on Google Colab
Downloads last month
0
Safetensors
Model size
7.24B params
Tensor type
F32
·

Dataset used to train Chahnwoo/Mistral-7B-v0.1-1E-QLoRA-SFT-Test