--- license: apache-2.0 --- Experimental pre-training on instruction datasets. https://wandb.ai/open-assistant/supervised-finetuning/runs/ys9rt5ue Checkpoint: 3500 steps Used oasst dataset config: ``` pretrain: use_custom_sampler: true sort_by_length: false datasets: - joke - webgpt: val_split: 0.1 - gpt4all: val_split: 0.01 - alpaca: val_split: 0.025 - code_alpaca: val_split: 0.05 - minimath - humaneval_mbpp_codegen_qa - humaneval_mbpp_testgen_qa - grade_school_math_instructions - recipes - cmu_wiki_qa #- youtube_subs_howto100m # uses incompatible column names #- ubuntu_dialogue_qa # fails to load - oa_wiki_qa_bart_10000row - prosocial_dialogue: fraction: 0.1 - explain_prosocial: fraction: 0.05 ``` pythia parameters: ``` pythia-12b: dtype: fp16 log_dir: "pythia_log_12b" learning_rate: 6e-6 model_name: EleutherAI/pythia-12b-deduped output_dir: pythia_model_12b weight_decay: 0.0 max_length: 2048 use_flash_attention: true #deepspeed_config: configs/zero3_config.json warmup_steps: 50 gradient_checkpointing: true gradient_accumulation_steps: 8 per_device_train_batch_size: 2 per_device_eval_batch_size: 5 eval_steps: 200 save_steps: 500 num_train_epochs: 2 save_total_limit: 2 ```