step15000 / bigram_2 /training_config.yaml
gerou161's picture
Add files using upload-large-folder tool
fa3df34 verified
adam_beta1: 0.9
adam_beta2: 0.95
adjust_step: 0
autoresume: false
batch_size: 6
clip_grad_norm: 1.0
comment: null
cycle_length: null
dtype: bfloat16
emb_freeze: true
eval_dataset_path: /work01/yanokazuki/fineweb/valid_data_gpt2/
eval_every: 1000
gradient_accumulation: 13
keep_checkpoints: null
layer_freeze: null
load_optimizer_state_on_resume: true
lr: 0.0004
max_length: 1024
max_train_tokens: null
min_lr_ratio: 0.1
model_config: model_config/478m.json
model_name_or_path: null
model_revision: null
num_training_steps: 15000
optimizer: Adam
restart_warmup_steps: null
resume_from: null
run_name: easy-pyramid-21
save_dir: checkpoints/easy-pyramid-21
save_every: 1000
scheduler: cosine
seed: 0
shuffle: true
skip_batches: !!set {}
tags:
- 396m-for-680m
total_batch_size: 624
train_dataset_path: /work01/yanokazuki/fineweb/train_data_gpt2/
training_config: training_config/two_stage/478m_resume.yaml
wandb_watch: true
warmed_up_model: /lee_embedding/checkpoints/tough-snowflake-18/final_model/
warmup_steps: 1500
weight_decay: 0.0
workers: 8