Adapters
medical
Laurent1's picture
Update README.md
11ff4c3
|
raw
history blame
2.44 kB
metadata
license: apache-2.0
datasets:
  - Laurent1/MedQuad-MedicalQnADataset_128tokens_max
library_name: adapter-transformers
tags:
  - medical

Model Card for mpt-7b-instruct2-QLoRa-medical-QA

image/gif

This is a QA model for answering medical questions

Foundation Model : https://huggingface.co/ibm/mpt-7b-instruct2
Dataset : https://huggingface.co/datasets/Laurent1/MedQuad-MedicalQnADataset_128tokens_max
The model has been fine tuned with 2 x GPU T4 (RAM : 2 x 14.8GB) + CPU (RAM : 29GB).

Model Details

The model is based upon the foundation model : ibm/mpt-7b-instruct2 (Apache 2.0 License).
It has been tuned with Supervised Fine-tuning Trainer and PEFT LoRa.

Librairies

  • bitsandbytes
  • einops
  • peft
  • trl
  • datasets
  • transformers
  • torch

Notebook used for the training

https://colab.research.google.com/drive/14nxSP5UuJcnIJtEERyk5nehBL3W03FR3?hl=fr

=> Improvements can be achieved by increasing the number of steps and using the full dataset.

Direct Use

image/png

Bias, Risks, and Limitations

In order to reduce training duration, the model has been trained only with the first 5100 rows of the dataset.

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
Generation of plausible yet incorrect factual information, termed hallucination, is an unsolved issue in large language models.

Training Details

  • per_device_train_batch_size = 1
  • gradient_accumulation_steps = 16
  • epoch = 5
  • 2 x GPU T4 (RAM : 14.8GB) + CPU (RAM : 29GB)

Training Data

https://huggingface.co/datasets/Laurent1/MedQuad-MedicalQnADataset_128tokens_max

Training Hyperparameters

image/png

Times

Training duration : 6287.4s

image/png