# encoding: utf-8 """ @author: l1aoxingyu @contact: sherlockliao01@gmail.com """ import argparse import os import sys from os import mkdir import torch from torch.backends import cudnn sys.path.append('.') from config import cfg from data import get_data_bunch from engine.inference import inference from modeling import build_model def main(): parser = argparse.ArgumentParser(description="ReID Baseline Inference") parser.add_argument( "--config_file", default="", help="path to config file", type=str ) parser.add_argument("opts", help="Modify config options using the command-line", default=None, nargs=argparse.REMAINDER) args = parser.parse_args() num_gpus = int(os.environ["WORLD_SIZE"]) if "WORLD_SIZE" in os.environ else 1 if args.config_file != "": cfg.merge_from_file(args.config_file) cfg.merge_from_list(args.opts) cfg.freeze() output_dir = cfg.OUTPUT_DIR if output_dir and not os.path.exists(output_dir): mkdir(output_dir) logger = setup_logger("reid_baseline", output_dir, 0) logger.info("Using {} GPUS".format(num_gpus)) logger.info(args) if args.config_file != "": logger.info("Loaded configuration file {}".format(args.config_file)) with open(args.config_file, 'r') as cf: config_str = "\n" + cf.read() logger.info(config_str) logger.info("Running with config:\n{}".format(cfg)) cudnn.benchmark = True train_databunch, test_databunch, num_query = get_data_bunch(cfg) model = build_model(cfg, train_databunch.c) model.load_state_dict(torch.load(cfg.TEST.WEIGHT)) inference(cfg, model, test_databunch, num_query) if __name__ == '__main__': main()