Tevatron ``` bs=512 epoch=40 save_steps=4000 backbone=bert-base-multilingual-cased output_dir=mlm.bs-$bs.epoch-$epoch.$backbone WANDB_PROJECT=mlm-mrtydi-DDR \ python examples/dense-adapter/dense-adapter-train.py \ --output_dir $output_dir \ --model_name_or_path $backbone \ --tokenizer_name bert-base-multilingual-cased \ --save_steps $save_steps \ --dataset_name Tevatron/msmarco-passage \ --fp16 \ --per_device_train_batch_size $bs \ --train_n_passages 2 \ --learning_rate 1e-5 \ --q_max_len 32 \ --p_max_len 128 \ --num_train_epochs $epoch \ --logging_steps 100 \ --overwrite_output_dir \ --dataloader_num_workers 4 \ ```