SuperCS's picture
Add files using upload-large-folder tool
e31e7b4 verified
set -x
sudo apt-get update && sudo apt-get install -y libgl1-mesa-glx
bash ./config/shell_scripts/cogvideo_i2v/train_wan_prepare.sh
git --no-pager log --decorate=short --pretty=oneline -n5
export OMNISTORE_LOAD_STRICT_MODE=0
export OMNISTORE_LOGGING_LEVEL=ERROR
#################################################################
## Torch
#################################################################
export TOKENIZERS_PARALLELISM=false
export TORCH_LOGS="+dynamo,recompiles,graph_breaks"
export TORCHDYNAMO_VERBOSE=1
export TORCH_NCCL_ENABLE_MONITORING=1
export PYTORCH_CUDA_ALLOC_CONF="expandable_segments:True,garbage_collection_threshold:0.9"
#################################################################
#################################################################
## NCCL
#################################################################
export NCCL_IB_GID_INDEX=3
export NCCL_IB_HCA=$ARNOLD_RDMA_DEVICE
export NCCL_SOCKET_IFNAME=eth0
export NCCL_SOCKET_TIMEOUT=3600000
export NCCL_DEBUG=WARN # disable the verbose NCCL logs
export NCCL_P2P_DISABLE=0
export NCCL_IB_DISABLE=0 # was 1
export NCCL_SHM_DISABLE=0 # was 1
export NCCL_P2P_LEVEL=NVL
export NCCL_PXN_DISABLE=0
export NCCL_NET_GDR_LEVEL=2
export NCCL_IB_QPS_PER_CONNECTION=4
export NCCL_IB_TC=160
export NCCL_IB_TIMEOUT=22
#################################################################
#################################################################
## WANDB
#################################################################
export WANDB__SERVICE_WAIT=6000
export WANDB_MODE=online
export WANDB_DISABLE_SERVICE=True
#################################################################
#################################################################
## DIST
#################################################################
MASTER_ADDR=$ARNOLD_WORKER_0_HOST
ports=(`echo $METIS_WORKER_0_PORT | tr ',' ' '`)
MASTER_PORT=${ports[0]}
NNODES=$ARNOLD_WORKER_NUM
NODE_RANK=$ARNOLD_ID
GPUS_PER_NODE=$ARNOLD_WORKER_GPU
# GPUS_PER_NODE=1
# NNODES=1
# NODE_RANK=0
WORLD_SIZE=$(($GPUS_PER_NODE*$NNODES))
DISTRIBUTED_ARGS="--nproc_per_node $GPUS_PER_NODE --nnodes $NNODES --node_rank $NODE_RANK --master_addr $MASTER_ADDR --master_port $MASTER_PORT"
if [ ! -z $RDZV_BACKEND ]; then
DISTRIBUTED_ARGS="${DISTRIBUTED_ARGS} --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_id 9863 --rdzv_backend c10d"
export NCCL_SHM_DISABLE=1
fi
region=$RUNTIME_IDC_NAME
if [ $region == 'maliva' ]; then
hdfs_prefix=hdfs://harunava/home/byte_icaip_nebudata
export ARNOLD_BASE_DIR=hdfs://harunava
else
hdfs_prefix=hdfs://harunasg/home/byte_icaip_nebudata_sg
export RUNTIME_IDC_NAME=my2
export ARNOLD_BASE_DIR=hdfs://harunasg
fi
echo -e "\033[31mDISTRIBUTED_ARGS: ${DISTRIBUTED_ARGS}\033[0m"
echo -e "\033[31mPERSISTENCE_PATH: ${hdfs_prefix}\033[0m"
#################################################################
#################################################################
## Training
#################################################################
learning_rate="1e-5"
lr_schedule="cosine_with_restarts"
optimizer="adamw"
steps="2000000"
version="v0.4"
DATASET_CONFIG="config/dataset_config/512_collection_config_vae1011_aligned_full_dump.yaml"
CKPT="/mnt/bn/icvg/users/yangxiao.0/Wan-AI/Wan2.1-I2V-14B-720P-patchsize1"
# CKPT="./models/Wan2.1-I2V-14B-720P"
output_dir="hdfs://harunasg/home/byte_icvg_aigc_cp/user/video/dali/dit_ckpt/i2v_wan_imageonly_lime_official_rl_1e-5_rm_with_1st_frame_round_4_2fps_rm_0812_color_VQ_MQ_MPS_0_cc_0814"
#output_dir="hdfs://harunasg/home/byte_icaip_nebudata_sg/fuwen/results/wan"
logging_dir="/mnt/bn/icvg/users/xinwei.huang/video_refl_new/log"
#logging_dir="./results/wan"
#################################################################
#TODO: prefetching
export WANDB_PROJECT=dc_ae_dit
export EXP_NAME=refl_2e-5_no_flowmatching_overall_fps6_rm_with_1st_frame_round_3_2fps_0812_RM_color_VQ_MQ_MPS_0_cc_loss
python3 -m torch.distributed.launch $DISTRIBUTED_ARGS ./training/train_wan_i2v_dc_ae.py \
--dataset_config $DATASET_CONFIG \
--frame_buckets 49 \
--dataloader_num_workers 1 \
--prefetch_factor 2 \
--pin_memory \
--seed 42 \
--mixed_precision bf16 \
--output_dir $output_dir \
--train_batch_size 1 \
--max_train_steps $steps \
--checkpointing_steps 50 \
--gradient_accumulation_steps 1 \
--learning_rate $learning_rate \
--lr_scheduler $lr_schedule \
--lr_warmup_steps 1 \
--lr_num_cycles 1 \
--optimizer $optimizer \
--beta1 0.9 \
--beta2 0.95 \
--weight_decay 0.001 \
--max_grad_norm 1.0 \
--allow_tf32 \
--report_to wandb \
--nccl_timeout 1800 \
--resume_from_checkpoint latest \
--wandb_project ${WANDB_PROJECT} \
--wandb_name ${EXP_NAME} \
--pretrained_model_name_or_path $CKPT \
--use_robust_loss \
--drop_first_frame_condition_threshold 0.00 \
--drop_last_frame_condition_threshold 0.0 \
--logging_dir $logging_dir \
--video_logging_interval 1000000 \
--scalar_logging_interval 1 \
--tp_size 8 \
--gradient_checkpointing \
--ema \
--ema_decay 0.99 \
--ema_interval 1 \
--sampling_steps 30 \
--max_turn_step 29 \
--min_turn_step 6 \
--optimizing_objective "VQ, MQ" \
--selected_frames 0 12 24 36 48 60 \
--half_input \
--use_cfg \
--rm_model_path "/mnt/bn/icvg/users/xinwei.huang/VideoAlign/rm_output_0801_first_color" \
--transformer_model_path "/mnt/bn/icvg/users/xinwei.huang/video_models/rm0806_round3_mps0.13000.pth/model.pt" \
--frame_reward_loss_weight 0