lomahony commited on
Commit
6abd1e5
1 Parent(s): 0179491

First model version

Browse files
Files changed (4) hide show
  1. config.yaml +44 -0
  2. optimizer.pt +3 -0
  3. policy.pt +3 -0
  4. scheduler.pt +3 -0
config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ seed: 0
2
+ exp_name: pythia_SFT_ga4
3
+ batch_size: 64
4
+ eval_batch_size: 32
5
+ debug: false
6
+ fsdp_port: 12355
7
+ datasets:
8
+ - hh
9
+ wandb:
10
+ enabled: true
11
+ entity: pythia_dpo
12
+ project: Pythia_LOM
13
+ local_dirs:
14
+ - /scr-ssd
15
+ - /scr
16
+ - .cache
17
+ sample_during_eval: false
18
+ n_eval_model_samples: 16
19
+ do_first_eval: true
20
+ local_run_dir: .cache/laura/pythia_SFT_ga4_2023-07-16_16-50-13_244945
21
+ lr: 1.0e-06
22
+ gradient_accumulation_steps: 4
23
+ max_grad_norm: 10.0
24
+ max_length: 512
25
+ max_prompt_length: 256
26
+ n_epochs: 1
27
+ n_examples: null
28
+ n_eval_examples: 256
29
+ trainer: FSDPTrainer
30
+ optimizer: RMSprop
31
+ warmup_steps: 150
32
+ activation_checkpointing: false
33
+ eval_every: 19968
34
+ minimum_log_interval_secs: 1.0
35
+ model:
36
+ name_or_path: EleutherAI/pythia-70m
37
+ tokenizer_name_or_path: null
38
+ archive: null
39
+ block_name: GPTNeoXLayer
40
+ policy_dtype: float32
41
+ fsdp_policy_mp: bfloat16
42
+ reference_dtype: float16
43
+ loss:
44
+ name: sft
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f187231f5c320ae9f98ea4d62e8baf3ce53fd9827bc0d4d0d6a4a4807cdc0a1
3
+ size 281732253
policy.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e4ab58154f0ec2c6f4062254793e8c7ed48e2972a2fa6b490b096bdd84d22f
3
+ size 306905707
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef3bb892ec5f85ebcd911ec0358781fc843702c92bef54eec6bc38b127a23878
3
+ size 627