|
python3 -m pip install --upgrade pip; |
|
|
|
export http_proxy=http://sys-proxy-rd-relay.byted.org:8118; |
|
export https_proxy=http://sys-proxy-rd-relay.byted.org:8118; |
|
|
|
export HF_HOME=/mnt/bn/vl-research-boli01-cn/.cache/huggingface; |
|
export HF_TOKEN="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"; |
|
export HF_HUB_ENABLE_HF_TRANSFER="1"; |
|
|
|
cd /mnt/bn/vl-research-boli01-cn/projects/zzz/lmms-eval; |
|
pip install -e .; |
|
|
|
cd /mnt/bn/vl-research-boli01-cn/projects/zzz/LLaVA_Next; |
|
pip install -e .; |
|
|
|
python3 -m pip install ninja; |
|
python3 -m pip install flash-attn --no-build-isolation; |
|
|
|
bash /mnt/bn/vl-research-boli01-cn/projects/zzz/LLaVA_Next/cn_scripts/vicuna/internal0.6m_finetune_llava1.6mix_7b_v0.2_unfreeze.sh |
|
|
|
|
|
accelerate launch --num_processes 8 --main_process_port 12345 -m lmms_eval \ |
|
--model llava \ |
|
--model_args pretrained="/mnt/bn/vl-research-boli01-cn/projects/zzz/LLaVA_Next/internal_project_checkpoints/llavanext-lmsys_vicuna-7b-v1.5-clip-vit-large-patch14-336-mlp2x_gelu-pretrain_internal0.6m_vicuna_v1_finetune_llava1.6_datamix_unfreezeVIS_1e" \ |
|
--tasks ok_vqa,textcaps_val,mme_test,mmmu,cmmmu,coco2017_cap_val,vizwiz_vqa_val,ai2d,chartqa,pope \ |
|
--batch_size 1 \ |
|
--log_samples \ |
|
--log_samples_suffix debug \ |
|
--output_path ./logs/ \ |
|
--wandb_args 'project=llava-next-lmms-eval,job_type=eval'; |