# Merging LoRA weights into base model weights Purpose: By merging our selected LoRA weights into the base model weights, we can benefit from all base model optimisation such as quantisation (available in this repo), pruning, caching, etc. ## How to run? After you have finish finetuning using LoRA, select your weight and run the converter script: ```bash python scripts/convert_lora_weights.py --lora_path out/lora/your-folder/your-weight-name.pth ``` The converted base weight file will be saved into the same folder with the name `{your-weight-name}-lora-merged-weights.pth`. Now you can run `generate.py` with the merged weights and apply quantisation: ```bash python generate.py --checkpoint_path out/lora/your-folder/your-weight-name-lora-merged-weights.pth --quantize llm.int8 ```