pr comments addressed
Browse files
examples/lora-openllama-3b/config.yml
CHANGED
@@ -13,7 +13,7 @@ dataset_prepared_path: last_run_prepared
|
|
13 |
val_set_size: 0.02
|
14 |
adapter: lora
|
15 |
lora_model_dir:
|
16 |
-
sequence_len:
|
17 |
max_packed_sequence_len:
|
18 |
lora_r: 8
|
19 |
lora_alpha: 16
|
@@ -43,7 +43,7 @@ train_on_inputs: false
|
|
43 |
group_by_length: false
|
44 |
bf16: false
|
45 |
fp16: true
|
46 |
-
tf32:
|
47 |
gradient_checkpointing: true
|
48 |
early_stopping_patience:
|
49 |
resume_from_checkpoint:
|
|
|
13 |
val_set_size: 0.02
|
14 |
adapter: lora
|
15 |
lora_model_dir:
|
16 |
+
sequence_len: 256
|
17 |
max_packed_sequence_len:
|
18 |
lora_r: 8
|
19 |
lora_alpha: 16
|
|
|
43 |
group_by_length: false
|
44 |
bf16: false
|
45 |
fp16: true
|
46 |
+
tf32: false
|
47 |
gradient_checkpointing: true
|
48 |
early_stopping_patience:
|
49 |
resume_from_checkpoint:
|