export PATH=/usr/local/cuda/bin:$PATH export HF_ENDPOINT=https://hf-mirror.com export OMP_NUM_THREADS=1 export timestamp=`date +"%Y%m%d%H%M%S"` export OLD_VERSION='False' export PYTHONPATH=$(dirname $SELF_DIR):$PYTHONPATH # gpu consumed # fp16 17-18G # int4 7-8G # model to be used # Example: MODELNAME=MiniCPM_V_2_6 MODELNAME=$1 # datasets to be tested # Example: DATALIST="MMMU_DEV_VAL MathVista_MINI MMVet MMBench_DEV_EN_V11 MMBench_DEV_CN_V11 MMStar HallusionBench AI2D_TEST" DATALIST=$2 # test mode, all or infer MODE=$3 echo "Starting inference with model $MODELNAME on datasets $DATALIST" # run on multi gpus with torchrun command # remember to run twice, the first run may fail torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE torchrun --nproc_per_node=8 run.py --data $DATALIST --model $MODELNAME --mode $MODE # run on single gpu with python command # python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE # python run.py --data $DATALIST --model $MODELNAME --verbose --mode $MODE ls