Commit
•
8fc3926
1
Parent(s):
a042dde
Update README.md
Browse files
README.md
CHANGED
@@ -15,19 +15,19 @@ Training Dataset: A combined dataset of alpaca, dolly and bactrainx which is tra
|
|
15 |
|
16 |
Training Method: Fine-tuned with Unsloth, which uses QLoRA. Used ORPO
|
17 |
|
18 |
-
#TrainingArguments
|
19 |
-
PER_DEVICE_BATCH_SIZE: 2
|
20 |
-
GRADIENT_ACCUMULATION_STEPS: 4
|
21 |
-
WARMUP_RATIO: 0.03
|
22 |
-
NUM_EPOCHS: 2
|
23 |
-
LR: 0.000008
|
24 |
-
OPTIM: "adamw_8bit"
|
25 |
-
WEIGHT_DECAY: 0.01
|
26 |
-
LR_SCHEDULER_TYPE: "linear"
|
27 |
BETA: 0.1
|
28 |
|
29 |
-
#PEFT Arguments
|
30 |
-
RANK: 128
|
31 |
TARGET_MODULES:
|
32 |
- "q_proj"
|
33 |
- "k_proj"
|
@@ -37,11 +37,11 @@ TARGET_MODULES:
|
|
37 |
- "up_proj"
|
38 |
- "down_proj"
|
39 |
|
40 |
-
LORA_ALPHA: 256
|
41 |
-
LORA_DROPOUT: 0
|
42 |
-
BIAS: "none"
|
43 |
-
GRADIENT_CHECKPOINT: 'unsloth'
|
44 |
-
USE_RSLORA: false
|
45 |
|
46 |
## Usage
|
47 |
This model is trained used Unsloth and uses it for fast inference. For Unsloth installation please refer to: https://github.com/unslothai/unsloth
|
|
|
15 |
|
16 |
Training Method: Fine-tuned with Unsloth, which uses QLoRA. Used ORPO
|
17 |
|
18 |
+
#TrainingArguments\
|
19 |
+
PER_DEVICE_BATCH_SIZE: 2\
|
20 |
+
GRADIENT_ACCUMULATION_STEPS: 4\
|
21 |
+
WARMUP_RATIO: 0.03\
|
22 |
+
NUM_EPOCHS: 2\
|
23 |
+
LR: 0.000008\
|
24 |
+
OPTIM: "adamw_8bit"\
|
25 |
+
WEIGHT_DECAY: 0.01\
|
26 |
+
LR_SCHEDULER_TYPE: "linear"\
|
27 |
BETA: 0.1
|
28 |
|
29 |
+
#PEFT Arguments\
|
30 |
+
RANK: 128\
|
31 |
TARGET_MODULES:
|
32 |
- "q_proj"
|
33 |
- "k_proj"
|
|
|
37 |
- "up_proj"
|
38 |
- "down_proj"
|
39 |
|
40 |
+
LORA_ALPHA: 256\
|
41 |
+
LORA_DROPOUT: 0\
|
42 |
+
BIAS: "none"\
|
43 |
+
GRADIENT_CHECKPOINT: 'unsloth'\
|
44 |
+
USE_RSLORA: false\
|
45 |
|
46 |
## Usage
|
47 |
This model is trained used Unsloth and uses it for fast inference. For Unsloth installation please refer to: https://github.com/unslothai/unsloth
|