mmselfsup/tools/dist_test.sh

18 lines
289 B
Bash

#!/bin/bash
set -x
CFG=$1
GPUS=$2
CHECKPOINT=$3
PORT=${PORT:-29500}
WORK_DIR="$(dirname $CHECKPOINT)/"
# test
python -m torch.distributed.launch --nproc_per_node=$GPUS --master_port=$PORT \
tools/test.py \
$CFG \
$CHECKPOINT \
--work_dir $WORK_DIR --launcher="pytorch"