dataset: name: alpaca_clean dataset_config: name: alpaca path: yahma/alpaca-cleaned chunk_size: 1024 # sequence length for distilling concat_data: true cache_dir: 'data/alpaca' # Change this to where you want to save pretrained_model_config: pretrained_model_name_or_path: 'mistralai/Mistral-7B-v0.1' # will be updated based on model_config cache_dir: '/scr-ssd/mzhang/models/mistral-v0.1' preprocess_config: null dataloader: batch_size: 1 num_workers: 2 drop_last: false pin_memory: true optimizer: optim: adamw_torch_fused lr: 0.01 weight_decay: 0.0 lr_scheduler: lr_scheduler_type: none trainer: # HuggingFace Trainer-like arguments name: null