alvarobartt HF staff commited on
Commit
8c6ee2b
1 Parent(s): 91140df

Upload train-config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. train-config.yaml +56 -0
train-config.yaml ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_args:
2
+ path: argilla/dpo-mix-7k
3
+
4
+ format_args:
5
+ prompt_format: zephyr-gemma
6
+ add_generation_prompt: false
7
+
8
+ model_args:
9
+ pretrained_model_name_or_path: HuggingFaceH4/zephyr-7b-gemma-sft-v0.1
10
+ torch_dtype: bfloat16
11
+
12
+ wandb_args:
13
+ entity: argilla-io
14
+ project: zephyr-gemma-dpo
15
+ name: 16bit-no-gen-prompt
16
+
17
+ training_args:
18
+ # DPOTrainer
19
+ beta: 0.05
20
+ loss_type: sigmoid
21
+ max_length: 1024
22
+ max_prompt_length: 512
23
+ # Trainer (train)
24
+ bf16: true
25
+ do_train: true
26
+ gradient_accumulation_steps: 8
27
+ gradient_checkpointing: true
28
+ gradient_checkpointing_kwargs:
29
+ use_reentrant: false
30
+ learning_rate: 5.0e-7
31
+ logging_steps: 10
32
+ lr_scheduler_type: cosine
33
+ num_train_epochs: 2
34
+ optim: adamw_torch
35
+ output_dir: data/gemma-7b-it-dpo
36
+ per_device_train_batch_size: 2
37
+ seed: 42
38
+ warmup_ratio: 0.1
39
+ warmup_steps: 100
40
+ report_to:
41
+ - wandb
42
+ - tensorboard
43
+ # Trainer (eval)
44
+ do_eval: true
45
+ evaluation_strategy: steps
46
+ eval_steps: 100
47
+ per_device_eval_batch_size: 4
48
+ # Trainer (save)
49
+ hub_model_id: alvarobartt/zephyr-gemma-dpo-faithful
50
+ hub_private_repo: true
51
+ push_to_hub: true
52
+ save_strategy: "no" # Quoted, otherwise is casted to `False`
53
+ save_total_limit: null
54
+
55
+ use_accelerate: true
56
+ use_unsloth: false