step_size=670 model_type=llama2 ckpt_path=/mnt/data/shared/multilingual_llm/experiments_megatron/continue_pretraining_tinyllama_all_20B/mc4_parallel_synth_checkpoints/ name=flavio for i in {1..10}; do iter=`expr $i \* $step_size` echo $iter echo ${iter} > $ckpt_path/latest_checkpointed_iteration.txt bash deploy.sh \ -p $ckpt_path \ -v "32000" \ -m "tiny_llama_${name}_${i}b_tokens" \ -t $model_type \ #-f /mnt/data/bpop/multilinguality_tower/extended-models/llama-2-7b-hf-merged-multi-32k-meaninit done #echo "6350" > $ckpt_path/latest_checkpointed_iteration.txt