TOT_CUDA="2,3" CUDAs=(${TOT_CUDA//,/ }) CUDA_NUM=${#CUDAs[@]} PORT="12345" DATA_PATH="sample/legislation30k.jsonl" #"../dataset/instruction/guanaco_non_chat_mini_52K-utf8.json" OUTPUT_PATH="../llama2-30kjudgement-21sept" MODEL_PATH="../llama2-30kjudgement-20sept" #lora_remote_checkpoint="Chinese-Vicuna/Chinese-Vicuna-lora-7b-belle-and-guanaco-11600" #lora_checkpoint="./lora-Vicuna/checkpoint-11600" from_data_beginning=False # True TEST_SIZE=1000 #CUDA_VISIBLE_DEVICES=${TOT_CUDA} torchrun --nproc_per_node=$CUDA_NUM --master_port=$PORT finetune.py \ CUDA_VISIBLE_DEVICES=0 python finetune.py --data_path $DATA_PATH \ --output_path $OUTPUT_PATH \ --model_path $MODEL_PATH \ --eval_steps 200 \ --save_steps 200 \ --test_size $TEST_SIZE \ --resume_from_checkpoint $lora_checkpoint \ --lora_remote_checkpoint $lora_remote_checkpoint \ --ignore_data_skip $from_data_beginning