|
export WANDB_MODE=disabled |
|
export CUDA_LAUNCH_BLOCKING=1 |
|
|
|
DATASET=Games |
|
BASE_MODEL=huggyllama/llama-7b |
|
DATA_PATH=./data |
|
OUTPUT_DIR=./ckpt/$DATASET/ |
|
|
|
torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ |
|
--base_model $BASE_MODEL \ |
|
--output_dir $OUTPUT_DIR \ |
|
--dataset $DATASET \ |
|
--data_path $DATA_PATH \ |
|
--per_device_batch_size 8 \ |
|
--gradient_accumulation_steps 2 \ |
|
--learning_rate 5e-5 \ |
|
--epochs 4 \ |
|
--weight_decay 0.01 \ |
|
--save_and_eval_strategy epoch \ |
|
--deepspeed ./config/ds_z3_bf16.json \ |
|
--bf16 \ |
|
--only_train_response \ |
|
--tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ |
|
--train_prompt_sample_num 1,1,1,1,1,1 \ |
|
--train_data_sample_num 0,0,0,100000,0,0 \ |
|
--index_file .index.json |
|
|
|
|
|
cd convert |
|
nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & |
|
cd .. |
|
|
|
|
|
|
|
|
|
|
|
|
|
DATASET=Arts |
|
BASE_MODEL=huggyllama/llama-7b |
|
DATA_PATH=./data |
|
OUTPUT_DIR=./ckpt/$DATASET/ |
|
|
|
torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ |
|
--base_model $BASE_MODEL \ |
|
--output_dir $OUTPUT_DIR \ |
|
--dataset $DATASET \ |
|
--data_path $DATA_PATH \ |
|
--per_device_batch_size 8 \ |
|
--gradient_accumulation_steps 2 \ |
|
--learning_rate 5e-5 \ |
|
--epochs 4 \ |
|
--weight_decay 0.01 \ |
|
--save_and_eval_strategy epoch \ |
|
--deepspeed ./config/ds_z3_bf16.json \ |
|
--bf16 \ |
|
--only_train_response \ |
|
--tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ |
|
--train_prompt_sample_num 1,1,1,1,1,1 \ |
|
--train_data_sample_num 0,0,0,30000,0,0 \ |
|
--index_file .index.json |
|
|
|
|
|
cd convert |
|
nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & |
|
cd .. |
|
|
|
|
|
|
|
|
|
|
|
DATASET=Instruments |
|
BASE_MODEL=huggyllama/llama-7b |
|
DATA_PATH=./data |
|
OUTPUT_DIR=./ckpt/$DATASET/ |
|
|
|
torchrun --nproc_per_node=8 --master_port=3324 finetune.py \ |
|
--base_model $BASE_MODEL \ |
|
--output_dir $OUTPUT_DIR \ |
|
--dataset $DATASET \ |
|
--data_path $DATA_PATH \ |
|
--per_device_batch_size 8 \ |
|
--gradient_accumulation_steps 2 \ |
|
--learning_rate 5e-5 \ |
|
--epochs 4 \ |
|
--weight_decay 0.01 \ |
|
--save_and_eval_strategy epoch \ |
|
--deepspeed ./config/ds_z3_bf16.json \ |
|
--bf16 \ |
|
--only_train_response \ |
|
--tasks seqrec,item2index,index2item,fusionseqrec,itemsearch,preferenceobtain \ |
|
--train_prompt_sample_num 1,1,1,1,1,1 \ |
|
--train_data_sample_num 0,0,0,20000,0,0 \ |
|
--index_file .index.json |
|
|
|
|
|
cd convert |
|
nohup ./convert.sh $OUTPUT_DIR >convert.log 2>&1 & |
|
cd .. |
|
|