Edit model card

Mixtral_8x7b_WuKurtz

Model is fine-tuned from the nephrology 80k dataset that we curated, injected into mixtral 8x7b instruct. This model is a fine-tuned version of mistralai/Mixtral-8x7B-v0.1 on the generator dataset.

Model description

Mixtral 8x7b WuKurtz was created by Sean Wu, Michael Koo, Andy Black, Lesley Blum, Fabien Scalzo, and Ira Kurtz at Pepperdine and UCLA. Arxiv paper out soon!

Intended uses & limitations

More information needed

Training and evaluation data

Training data out soon!

Training procedure

Parameter efficient fine tuning.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2.5e-05
  • train_batch_size: 4
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 0.03
  • num_epochs: 1

Training results

Framework versions

  • PEFT 0.8.1
  • Transformers 4.37.2
  • Pytorch 2.2.0+cu121
  • Datasets 2.16.1
  • Tokenizers 0.15.1
Downloads last month
2
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for SeanWu25/Mixtral_8x7b_WuKurtz

Adapter
(89)
this model