Edit model card

Model Card for Model ID

The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. This model uses peft finetuning with NEFTune for robustness.

Model Details

Model Description

This model is a finetuned model of the mistralai/Mistral-7B-v0.1.

Training Details

Training Data

This model is finetuned with kaist-ai/CoT-Collection.

Training Procedure

This model trained with SFT trainer and NEFTune method. (According to the paper, NEFTune adds noise to the embedding vectors during training)

Training Hyperparameters

  • lora alpha: 16
  • lora r: 64
  • lora dropout: 0.05
  • max sequence length: 4096
  • learning rate: 2e-4
  • max_grad_norm: 0.3
  • weight_decay: 0.001
  • gradient checkpoint: True
  • optim: paged_adamw_32bit
  • use_bf16: True
  • use_4bit: True
  • use_nested_quant: False
  • bnb_4bit_compute_dtype: float16
  • bnb_4bit_quant_type: nf4
Downloads last month
19
Safetensors
Model size
3.86B params
Tensor type
F32
FP16
U8