Edit model card

mistral_7b_instruct_v2_constitutional_rf_v2

This model is a fine-tuned version of mistralai/Mistral-7B-Instruct-v0.2 on the generator dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1829

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0002
  • train_batch_size: 1
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 4
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 1
  • training_steps: 1000

Training results

Training Loss Epoch Step Validation Loss
0.9391 0.1479 25 0.6647
0.6132 0.2959 50 0.6131
0.6037 0.4438 75 0.6059
0.5924 0.5917 100 0.5997
0.5962 0.7396 125 0.5944
0.6017 0.8876 150 0.5940
0.546 1.0355 175 0.6261
0.4233 1.1834 200 0.6179
0.4295 1.3314 225 0.6140
0.4165 1.4793 250 0.6192
0.428 1.6272 275 0.6221
0.4256 1.7751 300 0.6257
0.4254 1.9231 325 0.6235
0.3561 2.0710 350 0.6811
0.274 2.2189 375 0.7050
0.2681 2.3669 400 0.7119
0.2792 2.5148 425 0.7170
0.2738 2.6627 450 0.7101
0.2787 2.8107 475 0.7120
0.2794 2.9586 500 0.7014
0.1912 3.1065 525 0.8123
0.1592 3.2544 550 0.8475
0.1614 3.4024 575 0.8444
0.1628 3.5503 600 0.8430
0.1599 3.6982 625 0.8433
0.1667 3.8462 650 0.8468
0.1643 3.9941 675 0.8587
0.0958 4.1420 700 0.9951
0.0872 4.2899 725 1.0223
0.0892 4.4379 750 1.0246
0.0892 4.5858 775 1.0379
0.086 4.7337 800 1.0277
0.0871 4.8817 825 1.0588
0.0808 5.0296 850 1.0811
0.0512 5.1775 875 1.1683
0.0512 5.3254 900 1.1615
0.051 5.4734 925 1.1937
0.0493 5.6213 950 1.1847
0.0496 5.7692 975 1.1814
0.0469 5.9172 1000 1.1829

Framework versions

  • PEFT 0.10.0
  • Transformers 4.40.0
  • Pytorch 2.2.2
  • Datasets 2.19.0
  • Tokenizers 0.19.1
Downloads last month
0
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for edpowers/mistral_7b_instruct_v2_constitutional_rf_v2

Adapter
this model