File size: 797 Bytes
7d52396 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 |
# Merging LoRA weights into base model weights
Purpose: By merging our selected LoRA weights into the base model weights, we can benefit from all base model optimisation such as quantisation (available in this repo), pruning, caching, etc.
## How to run?
After you have finish finetuning using LoRA, select your weight and run the converter script:
```bash
python scripts/convert_lora_weights.py --lora_path out/lora/your-folder/your-weight-name.pth
```
The converted base weight file will be saved into the same folder with the name `{your-weight-name}-lora-merged-weights.pth`. Now you can run `generate.py` with the merged weights and apply quantisation:
```bash
python generate.py --checkpoint_path out/lora/your-folder/your-weight-name-lora-merged-weights.pth --quantize llm.int8
```
|