#!/bin/bash torchrun --nproc_per_node=1 --master_port=9292 train.py \ --tokenizer_name_or_path /src/weights/tokenizer \ --data_path ./alpaca_data.json \ --model_name_or_path /src/weights/llama-7b \ --output_dir alpaca_out \ --num_train_epochs 3 \ --per_device_train_batch_size 4 \ --per_device_eval_batch_size 4 \ --gradient_accumulation_steps 8 \ --evaluation_strategy "no" \ --save_strategy "steps" \ --save_steps 2000 \ --learning_rate 2e-5 \ --weight_decay 0. \ --warmup_ratio 0.03 \ --lr_scheduler_type "cosine" \ --logging_steps 1 \ --fsdp "full_shard auto_wrap" \ --fsdp_transformer_layer_cls_to_wrap 'LLaMADecoderLayer' \