tahauzumcu commited on
Commit
8fc3926
1 Parent(s): a042dde

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -16
README.md CHANGED
@@ -15,19 +15,19 @@ Training Dataset: A combined dataset of alpaca, dolly and bactrainx which is tra
15
 
16
  Training Method: Fine-tuned with Unsloth, which uses QLoRA. Used ORPO
17
 
18
- #TrainingArguments
19
- PER_DEVICE_BATCH_SIZE: 2
20
- GRADIENT_ACCUMULATION_STEPS: 4
21
- WARMUP_RATIO: 0.03
22
- NUM_EPOCHS: 2
23
- LR: 0.000008
24
- OPTIM: "adamw_8bit"
25
- WEIGHT_DECAY: 0.01
26
- LR_SCHEDULER_TYPE: "linear"
27
  BETA: 0.1
28
 
29
- #PEFT Arguments
30
- RANK: 128
31
  TARGET_MODULES:
32
  - "q_proj"
33
  - "k_proj"
@@ -37,11 +37,11 @@ TARGET_MODULES:
37
  - "up_proj"
38
  - "down_proj"
39
 
40
- LORA_ALPHA: 256
41
- LORA_DROPOUT: 0
42
- BIAS: "none"
43
- GRADIENT_CHECKPOINT: 'unsloth'
44
- USE_RSLORA: false
45
 
46
  ## Usage
47
  This model is trained used Unsloth and uses it for fast inference. For Unsloth installation please refer to: https://github.com/unslothai/unsloth
 
15
 
16
  Training Method: Fine-tuned with Unsloth, which uses QLoRA. Used ORPO
17
 
18
+ #TrainingArguments\
19
+ PER_DEVICE_BATCH_SIZE: 2\
20
+ GRADIENT_ACCUMULATION_STEPS: 4\
21
+ WARMUP_RATIO: 0.03\
22
+ NUM_EPOCHS: 2\
23
+ LR: 0.000008\
24
+ OPTIM: "adamw_8bit"\
25
+ WEIGHT_DECAY: 0.01\
26
+ LR_SCHEDULER_TYPE: "linear"\
27
  BETA: 0.1
28
 
29
+ #PEFT Arguments\
30
+ RANK: 128\
31
  TARGET_MODULES:
32
  - "q_proj"
33
  - "k_proj"
 
37
  - "up_proj"
38
  - "down_proj"
39
 
40
+ LORA_ALPHA: 256\
41
+ LORA_DROPOUT: 0\
42
+ BIAS: "none"\
43
+ GRADIENT_CHECKPOINT: 'unsloth'\
44
+ USE_RSLORA: false\
45
 
46
  ## Usage
47
  This model is trained used Unsloth and uses it for fast inference. For Unsloth installation please refer to: https://github.com/unslothai/unsloth