Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,8 @@ Differences in the qlora scripts:
|
|
10 |
- uses `--num_train_epochs` instead of `--max_steps`
|
11 |
- uses airoboros prompt format (mostly 1:1 with vicuna) rather than alpaca, and expects an input file in JSONL format with "instruction" and "response"
|
12 |
|
|
|
|
|
13 |
Full example of tuning (used for airoboros-mpt-30b-gpt4-1.4):
|
14 |
|
15 |
```
|
|
|
10 |
- uses `--num_train_epochs` instead of `--max_steps`
|
11 |
- uses airoboros prompt format (mostly 1:1 with vicuna) rather than alpaca, and expects an input file in JSONL format with "instruction" and "response"
|
12 |
|
13 |
+
__I think there's a bug in gradient accumulation, so if you try this, maybe set gradient accumulation steps to 1__
|
14 |
+
|
15 |
Full example of tuning (used for airoboros-mpt-30b-gpt4-1.4):
|
16 |
|
17 |
```
|