dahara1's picture
Update finetune_sample/README.md
e46a468
autoGPTQ QLora finetune sample data and script.
```
git lfs install
git clone https://huggingface.co/dahara1/weblab-10b-instruction-sft-GPTQ
cd weblab-10b-instruction-sft-GPTQ/finetune_sample
python3 -m venv gptq_finetune
source gptq_finetune/bin/activate
pip install transformers==4.34.1
pip install datasets
pip install peft==0.5.0
pip install trl
pip install auto-gptq
pip install optimum
pip install torch==2.0.1
# finetune qlora
python3 finetune.py
# use qlora sample
python3 lora_test.py
```
Version is very very important.
For example if you get something like
```
ValueError: Target module QuantLinear() is not supported. Currently, only `torch.nn.Linear` and `Conv1D` are supported.
```
It's because peft old version.
I don't know if it's required, but the version of my running environment.
* auto-gptq 0.4.2
* trl 0.7.2
* optimum 1.13.2
* datasets 2.14.6
The official documentation says to install from source, but sometimes that causes errors.
If you can't get it to work, it might be better to wait until the stable version comes out.
Good luck!
If you encounter ```RuntimeError: Unrecognized tensor type ID: AutocastCUDA```, check your torch version.
auto-gptq 0.4.2 with torch 2.1.0 can't work for me.
- finetune.py gptq finetune sample file.
- jawiki3.csv sample data.(Japanese)
- lora_test.py after finetune, you can use lora with this script.
- checkpoint-700 Created sample LoRA for test.
The model.safetensors is ../gptq_model-4bit-128g.safetensors.
It's samefile. I can't find how to change script defaults model name, So I copied it.