randomani's picture
Create README.md
bce2ae3 verified
|
raw
history blame
No virus
2.32 kB
---
license: apache-2.0
datasets:
- randomani/MedicalQnA-llama2
language:
- en
library_name: adapter-transformers
pipeline_tag: question-answering
tags:
- medical
---
# Fine-Tuning LLaMA-2 Chat Model with Medical QnA Dataset using QLoRA
This repository contains the code and configuration for fine-tuning the LLaMA-2 chat model using the Medical QnA dataset with the QLoRA technique.Used only 2k data elements for training due to constrained gpu resources.
## Model and Dataset
- **Pre-trained Model**: `NousResearch/Llama-2-7b-chat-hf`
- **Dataset for Fine-Tuning**: `randomani/MedicalQnA-llama2`
- **Fine-Tuned Model Name**: `Llama-2-7b-Medchat-finetune`
## QLoRA Parameters
- **LoRA Attention Dimension** (`lora_r`): 64
- **LoRA Scaling Alpha** (`lora_alpha`): 16
- **LoRA Dropout Probability** (`lora_dropout`): 0.1
## bitsandbytes Parameters
- **Use 4-bit Precision** (`use_4bit`): True
- **4-bit Compute Dtype** (`bnb_4bit_compute_dtype`): float16
- **4-bit Quantization Type** (`bnb_4bit_quant_type`): nf4
- **Use Nested Quantization** (`use_nested_quant`): False
## Training Arguments
- **Number of Training Epochs** (`num_train_epochs`): 1
- **Use fp16** (`fp16`): False
- **Use bf16** (`bf16`): False
- **Training Batch Size per GPU** (`per_device_train_batch_size`): 4
- **Evaluation Batch Size per GPU** (`per_device_eval_batch_size`): 4
- **Gradient Accumulation Steps** (`gradient_accumulation_steps`): 1
- **Enable Gradient Checkpointing** (`gradient_checkpointing`): True
- **Maximum Gradient Norm** (`max_grad_norm`): 0.3
- **Initial Learning Rate** (`learning_rate`): 2e-4
- **Weight Decay** (`weight_decay`): 0.001
- **Optimizer** (`optim`): paged_adamw_32bit
- **Learning Rate Scheduler Type** (`lr_scheduler_type`): cosine
- **Maximum Training Steps** (`max_steps`): -1
- **Warmup Ratio** (`warmup_ratio`): 0.03
- **Group Sequences by Length** (`group_by_length`): True
- **Save Checkpoints Every X Steps** (`save_steps`): 0
- **Logging Steps** (`logging_steps`): 25
## Supervised Fine-Tuning (SFT) Parameters
- **Maximum Sequence Length** (`max_seq_length`): None
- **Packing Multiple Short Examples** (`packing`): False
## References
For more details and access to the dataset, visit the [Hugging Face Dataset Page](https://huggingface.co/datasets/randomani/MedicalQnA-llama2).