base_model: unsloth/Meta-Llama-3.1-8B-Instruct load_in_8bit: false load_in_4bit: false strict: false # Layer unfreezing configuration unfreezing: enabled: true mode: "selective" layer_selection: patterns: - "model.layers.1" - "model.layers.2" - "model.layers.3" - "model.layers.4" - "model.layers.5" - "model.layers.6" - "model.layers.7" - "model.layers.8" - "model.layers.9" - "model.layers.10" - "model.layers.11" component_types: - "input_layernorm" - "self_attn.q_proj" - "self_attn.k_proj" - "self_attn.v_proj" - "self_attn.o_proj" - "post_attention_layernorm" - "mlp.gate_proj" - "mlp.up_proj" - "mlp.down_proj" # Pretraining dataset configuration pretraining_dataset: - path: LemTenku/adasdxacac text_column: text type: pretrain trust_remote_code: true # Required for streaming datasets max_steps: 14932 batching_strategy: stream datasets: - path: LemTenku/adasdxacac type: pretrain text_column: text dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./outputs/out sequence_len: 8192 sample_packing: true pad_to_sequence_len: true wandb_project: wandb_entity: wandb_watch: wandb_name: wandb_log_model: gradient_accumulation_steps: 8 micro_batch_size: 1 num_epochs: 3 optimizer: paged_adamw_8bit lr_scheduler: constant learning_rate: 4e-5 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 0 evaluation_strategy: "no" saves_per_epoch: 1 debug: deepspeed: weight_decay: 0.0 fsdp: fsdp_config: special_tokens: bos_token: "<|begin_of_text|>" eos_token: "<|eot_id|>" pad_token: "<|end_of_text|>" tokens: - "<|begin_of_text|>" - "<|eot_id|>" - "<|end_of_text|>"