gsmyrnis commited on
Commit
0e6fe90
1 Parent(s): b8f3669

Upload configs.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. configs.yaml +45 -0
configs.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ bf16: true
4
+ cutoff_len: 2048
5
+ dataset: mlfoundations-dev/oh-dcft-v3.1-gemini-1.5-flash
6
+ dataset_dir: ONLINE
7
+ ddp_timeout: 180000000
8
+ deepspeed: dcft/train/zero3.json
9
+ do_train: true
10
+ enable_liger_kernel: false
11
+ eval_strategy: epoch
12
+ finetuning_type: full
13
+ formatting: sharegpt
14
+ global_batch_size: 512
15
+ gradient_accumulation_steps: 1
16
+ gradient_checkpointing: true
17
+ hub_model_id: mlfoundations-dev/mistral_7b_0-3_oh-dcft-v3.1-gemini-1.5-flash
18
+ include_hp: dcft/train/hp_settings/mistral_sft.yaml
19
+ learning_rate: 5.0e-06
20
+ logging_steps: 10
21
+ lr_scheduler_kwargs:
22
+ min_lr: 5.0e-07
23
+ lr_scheduler_type: cosine_with_min_lr
24
+ max_grad_norm: 1
25
+ messages: conversations
26
+ model_name_or_path: mistralai/Mistral-7B-v0.3
27
+ neat_packing: true
28
+ num_train_epochs: 3.0
29
+ output_dir: /tmp/experiments/train/checkpoints/oh-dcft-v3.1-gemini-1.5-flash
30
+ overwrite_cache: true
31
+ overwrite_output_dir: true
32
+ packing: true
33
+ per_device_train_batch_size: 16
34
+ plot_loss: true
35
+ preprocessing_num_workers: 16
36
+ push_to_db: true
37
+ push_to_hub: true
38
+ report_to: wandb
39
+ run_name: mistral_7b_0-3_oh-dcft-v3.1-gemini-1.5-flash
40
+ save_strategy: epoch
41
+ stage: sft
42
+ template: mistral
43
+ val_size: 0.05
44
+ warmup_ratio: 0.05
45
+ weight_decay: 0.1