Edit model card

Fine-Tuning LLaMA-2 Chat Model with Medical QnA Dataset using QLoRA

This repository contains the code and configuration for fine-tuning the LLaMA-2 chat model using the Medical QnA dataset with the QLoRA technique.Used only 2k data elements for training due to constrained gpu resources.

Model and Dataset

  • Pre-trained Model: NousResearch/Llama-2-7b-chat-hf
  • Dataset for Fine-Tuning: randomani/MedicalQnA-llama2
  • Fine-Tuned Model Name: Llama-2-7b-Medchat-finetune

QLoRA Parameters

  • LoRA Attention Dimension (lora_r): 64
  • LoRA Scaling Alpha (lora_alpha): 16
  • LoRA Dropout Probability (lora_dropout): 0.1

bitsandbytes Parameters

  • Use 4-bit Precision (use_4bit): True
  • 4-bit Compute Dtype (bnb_4bit_compute_dtype): float16
  • 4-bit Quantization Type (bnb_4bit_quant_type): nf4
  • Use Nested Quantization (use_nested_quant): False

Training Arguments

  • Number of Training Epochs (num_train_epochs): 1
  • Use fp16 (fp16): False
  • Use bf16 (bf16): False
  • Training Batch Size per GPU (per_device_train_batch_size): 4
  • Evaluation Batch Size per GPU (per_device_eval_batch_size): 4
  • Gradient Accumulation Steps (gradient_accumulation_steps): 1
  • Enable Gradient Checkpointing (gradient_checkpointing): True
  • Maximum Gradient Norm (max_grad_norm): 0.3
  • Initial Learning Rate (learning_rate): 2e-4
  • Weight Decay (weight_decay): 0.001
  • Optimizer (optim): paged_adamw_32bit
  • Learning Rate Scheduler Type (lr_scheduler_type): cosine
  • Maximum Training Steps (max_steps): -1
  • Warmup Ratio (warmup_ratio): 0.03
  • Group Sequences by Length (group_by_length): True
  • Save Checkpoints Every X Steps (save_steps): 0
  • Logging Steps (logging_steps): 25

Supervised Fine-Tuning (SFT) Parameters

  • Maximum Sequence Length (max_seq_length): None
  • Packing Multiple Short Examples (packing): False

References

For more details and access to the dataset, visit the Hugging Face Dataset Page.

Downloads last month
0

Dataset used to train randomani/Llama-2-7b-chat-Medchat-finetune