Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
smangrul 
posted an update Feb 28
Post
🚨 New Release of 🤗PEFT!

1. New methods for merging LoRA weights. Refer this HF Post for more details: https://huggingface.co/posts/smangrul/850816632583824

2. AWQ and AQLM support for LoRA. You can now:
- Train adapters on top of 2-bit quantized models with AQLM
- Train adapters on top of powerful AWQ quantized models
Note for inference you can't merge the LoRA weights into the base model!

3. DoRA support: Enabling DoRA is as easy as adding use_dora=True to your LoraConfig. Find out more about this method here: https://arxiv.org/abs/2402.09353

4. Improved documentation, particularly docs regarding PEFT LoRA+DeepSpeed and PEFT LoRA+FSDP! 📄 Check out the docs at https://huggingface.co/docs/peft/index.

5. Full Release Notes: https://github.com/huggingface/peft/releases/tag/v0.9.0

Are there any examples or notebooks showing how to use AWQ in LORA fine-tuning a LLM? Or just use AWQ model from huggingface directly? I'm asking as neither the docs nor the Release notes explain anything.

·

cc @ybelkada for this question.

Hi @smangrul , apparently i can't push the merged adapter to the hub ???
Cuz when i do so it create num_of_adapters_to_merge + 1 (the merged adapter) and when i want to load the the merged adapter with model = PeftModel.from_pretrained(model, adapter) i got the error in image 2 !

image 1:
image.png

image 2:
image.png

Your help is much appreciated, tnx 🤗