Spaces:
Running
Running
export PATH=/usr/local/cuda/bin:$PATH | |
export HF_ENDPOINT=https://hf-mirror.com | |
export OMP_NUM_THREADS=1 | |
export timestamp=`date +"%Y%m%d%H%M%S"` | |
export OLD_VERSION='False' | |
export PYTHONPATH=$(dirname $SELF_DIR):$PYTHONPATH | |
# gpu consumed | |
# fp16 17-18G | |
# int4 7-8G | |
# model to be used | |
# Example: MODELNAME=MiniCPM_V_2_6 | |
MODELNAME=$1 | |
# datasets to be tested | |
# Example: DATALIST="MMMU_DEV_VAL MathVista_MINI MMVet MMBench_DEV_EN_V11 MMBench_DEV_CN_V11 MMStar HallusionBench AI2D_TEST" | |
DATALIST=$2 | |
# test mode, all or infer | |
MODE=$3 | |
echo "Starting inference with model $MODELNAME on datasets $DATALIST" | |
# run on multi gpus with torchrun command | |
# remember to run twice, the first run may fail | |
torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE | |
torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE | |
# run on single gpu with python command | |
# python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE | |
# python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE | |
ls | |