Model Card for Mistral-7B-Instruct-v0.1-QLoRa-medical-QA
This is a QA model for answering medical questions
Foundation Model : https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1
Dataset : https://huggingface.co/datasets/Laurent1/MedQuad-MedicalQnADataset_128tokens_max
The model has been fine tuned with 2 x GPU T4 (RAM : 2 x 14.8GB) + CPU (RAM : 29GB).
Model Details
The model is based upon the foundation model : Mistral-7B-Instruct-v0.1.
It has been tuned with Supervised Fine-tuning Trainer and PEFT LoRa.
Librairies
- bitsandbytes
- einops
- peft
- trl
- datasets
- transformers
- torch
Bias, Risks, and Limitations
In order to reduce training duration, the model has been trained only with the first 5100 rows of the dataset.
Generation of plausible yet incorrect factual information, termed hallucination, is an unsolved issue in large language models.
Training Details
- per_device_train_batch_size = 1
- gradient_accumulation_steps = 16
- epoch = 5
- 2 x GPU T4 (RAM : 14.8GB) + CPU (RAM : 29GB)
Notebook used for the training
You can find it in the files and versions tab
Training Data
https://huggingface.co/datasets/Laurent1/MedQuad-MedicalQnADataset_128tokens_max
Training Hyperparameters
Times
Training duration : 6287.4s
- Downloads last month
- 2
Model tree for Laurent1/Mistral-7B-Instruct-v0.1-QLoRa-medical-QA
Base model
mistralai/Mistral-7B-v0.1
Finetuned
mistralai/Mistral-7B-Instruct-v0.1