qlora merge and load requires that base model isn't loaded in 4 or 8 bit
Browse files
README.md
CHANGED
@@ -24,7 +24,7 @@
|
|
24 |
|
25 |
## Quickstart β‘
|
26 |
|
27 |
-
**Requirements**: Python 3.9.
|
28 |
|
29 |
```bash
|
30 |
git clone https://github.com/OpenAccess-AI-Collective/axolotl
|
@@ -45,7 +45,7 @@ accelerate launch scripts/finetune.py examples/4bit-lora-7b/config.yml \
|
|
45 |
|
46 |
### Environment
|
47 |
|
48 |
-
- Docker
|
49 |
```bash
|
50 |
docker run --gpus '"all"' --rm -it winglian/axolotl:main
|
51 |
```
|
@@ -332,7 +332,7 @@ seed:
|
|
332 |
|
333 |
### Accelerate
|
334 |
|
335 |
-
Configure accelerate
|
336 |
|
337 |
```bash
|
338 |
accelerate config
|
@@ -363,12 +363,18 @@ Pass the appropriate flag to the train command:
|
|
363 |
|
364 |
### Merge LORA to base
|
365 |
|
366 |
-
Add below flag to train command above
|
367 |
|
368 |
```bash
|
369 |
--merge_lora --lora_model_dir="./completed-model"
|
370 |
```
|
371 |
|
|
|
|
|
|
|
|
|
|
|
|
|
372 |
## Common Errors π§°
|
373 |
|
374 |
> Cuda out of memory
|
@@ -383,7 +389,7 @@ Please reduce any below
|
|
383 |
Try set `fp16: true`
|
384 |
|
385 |
## Need help? πββοΈ
|
386 |
-
|
387 |
Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we can help you
|
388 |
|
389 |
## Contributing π€
|
|
|
24 |
|
25 |
## Quickstart β‘
|
26 |
|
27 |
+
**Requirements**: Python 3.9.
|
28 |
|
29 |
```bash
|
30 |
git clone https://github.com/OpenAccess-AI-Collective/axolotl
|
|
|
45 |
|
46 |
### Environment
|
47 |
|
48 |
+
- Docker
|
49 |
```bash
|
50 |
docker run --gpus '"all"' --rm -it winglian/axolotl:main
|
51 |
```
|
|
|
332 |
|
333 |
### Accelerate
|
334 |
|
335 |
+
Configure accelerate
|
336 |
|
337 |
```bash
|
338 |
accelerate config
|
|
|
363 |
|
364 |
### Merge LORA to base
|
365 |
|
366 |
+
Add below flag to train command above (and using LoRA)
|
367 |
|
368 |
```bash
|
369 |
--merge_lora --lora_model_dir="./completed-model"
|
370 |
```
|
371 |
|
372 |
+
Add below flag to train command above (and using QLoRA)
|
373 |
+
|
374 |
+
```bash
|
375 |
+
--merge_lora --lora_model_dir="./completed-model" --load_in_8bit False --load_in_4bit False
|
376 |
+
```
|
377 |
+
|
378 |
## Common Errors π§°
|
379 |
|
380 |
> Cuda out of memory
|
|
|
389 |
Try set `fp16: true`
|
390 |
|
391 |
## Need help? πββοΈ
|
392 |
+
|
393 |
Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we can help you
|
394 |
|
395 |
## Contributing π€
|