# Copyright (c) OpenMMLab. All rights reserved. import argparse import copy import logging import os import os.path as osp import platform import re import shutil import sys import tarfile from distutils.util import get_platform from glob import glob from subprocess import CalledProcessError, check_output, run from typing import Dict import yaml from packaging import version logger = logging.getLogger() logger.setLevel(logging.INFO) CUR_DIR = os.path.dirname(os.path.abspath(__file__)) PACKAGING_DIR = osp.join(CUR_DIR, 'packaging') PLATFORM_TAG = get_platform().replace('-', '_').replace('.', '_') def get_version(version_file): with open(version_file, 'r') as f: exec(compile(f.read(), version_file, 'exec')) return locals()['__version__'] def _merge_cfg(cfg0, cfg1): cfg = copy.deepcopy(cfg0) for k, v in cfg1.items(): if k in cfg: cfg[k] = _merge_cfg(cfg0[k], cfg1[k]) else: cfg[k] = v return cfg def _remove_if_exist(path): if osp.exists(path): logging.info(f'Remove path: {path}.') if osp.isdir(path): shutil.rmtree(path) else: os.remove(path) def _copy(src_path, dst_path): assert osp.exists(src_path), f'src path: {src_path} not exist.' logging.info(f'copy path: {src_path} to {dst_path}.') if osp.isdir(src_path): if osp.exists(dst_path): shutil.rmtree(dst_path) shutil.copytree(src_path, dst_path) else: shutil.copy(src_path, dst_path) def _call_command(cmd, cwd, stdout=None, stderr=None): if cmd == '': return logging.info(f'Process cmd: {cmd}') logging.info(f'work_path: {cwd}') try: ret = run(cmd, stdout=stdout, stderr=stderr, cwd=cwd, shell=True) if ret.returncode != 0: logging.error(f'Process cmd: "{cmd}"' f' failed with returncode: {ret.returncode}') exit(-1) except Exception: logging.error(f'Process cmd: {cmd} failed.') exit(-1) def _create_tar(path, tar_name): logging.info(f'create tar file: {tar_name}') with tarfile.open(tar_name, 'w:gz') as tar: tar.add(path, arcname=os.path.basename(path)) def _create_bdist_cmd(cfg, c_ext=False, dist_dir=None): bdist_tags = cfg.get('bdist_tags', {}) # base bdist_cmd = 'python setup.py bdist_wheel ' # platform bdist_cmd += f' --plat-name {PLATFORM_TAG} ' # python tag python_tag = f'cp{sys.version_info.major}{sys.version_info.minor}'\ if c_ext else 'py3' if 'python_tag' in bdist_tags: python_tag = bdist_tags['python_tag'] bdist_cmd += f' --python-tag {python_tag} ' # dist dir if dist_dir is not None: dist_dir = osp.abspath(dist_dir) bdist_cmd += f' --dist-dir {dist_dir} ' return bdist_cmd def clear_mmdeploy(mmdeploy_dir: str): logging.info(f'cleaning mmdeploy: {mmdeploy_dir}') def _remove_in_mmdeploy(path): remove_dir = osp.join(mmdeploy_dir, path) _remove_if_exist(remove_dir) # remove build file _remove_in_mmdeploy('build') # remove dist _remove_in_mmdeploy('dist') # remove installed library _remove_in_mmdeploy('mmdeploy/lib') # remove onnx2ncnn and ncnn ext _remove_in_mmdeploy('mmdeploy/backend/ncnn/onnx2ncnn') _remove_in_mmdeploy('mmdeploy/backend/ncnn/onnx2ncnn.exe') _remove_in_mmdeploy('mmdeploy/backend/ncnn/mmdeploy_onnx2ncnn') _remove_in_mmdeploy('mmdeploy/backend/ncnn/mmdeploy_onnx2ncnn.exe') ncnn_ext_paths = glob( osp.join(mmdeploy_dir, 'mmdeploy/backend/ncnn/ncnn_ext.*')) for ncnn_ext_path in ncnn_ext_paths: os.remove(ncnn_ext_path) # remove ts_optmizer ts_optimizer_paths = glob( osp.join(mmdeploy_dir, 'mmdeploy/backend/torchscript/ts_optimizer.*')) for ts_optimizer_path in ts_optimizer_paths: os.remove(ts_optimizer_path) def build_mmdeploy(cfg, mmdeploy_dir, dist_dir=None): cmake_flags = cfg.get('cmake_flags', []) cmake_envs = cfg.get('cmake_envs', dict()) args = [f'-D{k}={v}' for k, v in cmake_envs.items()] # clear mmdeploy clear_mmdeploy(mmdeploy_dir) build_dir = osp.join(mmdeploy_dir, 'build') if not osp.exists(build_dir): os.mkdir(build_dir) # cmake cmd cmake_cmd = ' '.join(['cmake ..'] + cmake_flags + args) _call_command(cmake_cmd, build_dir) if sys.platform == 'win32': # build cmd build_cmd = 'cmake --build . --config Release -- /m' _call_command(build_cmd, build_dir) install_cmd = 'cmake --install . --config Release' _call_command(install_cmd, build_dir) else: # build cmd build_cmd = 'cmake --build . -- -j$(nproc) && cmake --install .' _call_command(build_cmd, build_dir) # build wheel bdist_cmd = _create_bdist_cmd(cfg, c_ext=False, dist_dir=dist_dir) _call_command(bdist_cmd, mmdeploy_dir) def build_mmdeploy_python(python_executable, cfg, mmdeploy_dir): cmake_flags = cfg.get('cmake_flags', []) cmake_envs = cfg.get('cmake_envs', dict()) args = [f'-D{k}={v}' for k, v in cmake_envs.items()] args.append( f'-DMMDeploy_DIR={mmdeploy_dir}/build/install/lib/cmake/MMDeploy') args.append(f'-DPYTHON_EXECUTABLE={python_executable}') if sys.platform == 'win32': build_cmd = 'cmake --build . --config Release -- /m' pass else: build_cmd = 'cmake --build . -- -j$(nproc)' cmake_cmd = ' '.join(['cmake ../csrc/mmdeploy/apis/python'] + cmake_flags + args) build_dir = osp.join(mmdeploy_dir, 'build_python') _remove_if_exist(build_dir) os.mkdir(build_dir) _call_command(cmake_cmd, build_dir) _call_command(build_cmd, build_dir) python_api_lib_path = [] lib_patterns = ['*mmdeploy_python*.so', '*mmdeploy_python*.pyd'] for pattern in lib_patterns: python_api_lib_path.extend( glob( osp.join(mmdeploy_dir, 'build_python/**', pattern), recursive=True, )) return python_api_lib_path[0] def get_dir_name(cfg, tag, default_name): if tag not in cfg: logging.warning(f'{tag} not found, use `{default_name}` as default.') else: default_name = cfg[tag] return cfg, default_name def check_env(cfg: Dict): env_info = {} cmake_envs = cfg.get('cmake_envs', dict()) # system platform_system = platform.system().lower() platform_machine = platform.machine().lower() env_info['system'] = platform_system env_info['machine'] = platform_machine # CUDA version cuda_version = 'unknown' CUDA_TOOLKIT_ROOT_DIR = cmake_envs.get('CUDA_TOOLKIT_ROOT_DIR', '') CUDA_TOOLKIT_ROOT_DIR = osp.expandvars(CUDA_TOOLKIT_ROOT_DIR) nvcc_cmd = ('nvcc' if len(CUDA_TOOLKIT_ROOT_DIR) <= 0 else osp.join( CUDA_TOOLKIT_ROOT_DIR, 'bin', 'nvcc')) try: nvcc = check_output(f'"{nvcc_cmd}" -V', shell=True) nvcc = nvcc.decode('utf-8').strip() pattern = r'Cuda compilation tools, release (\d+.\d+)' match = re.search(pattern, nvcc) if match is not None: cuda_version = match.group(1) except Exception: pass env_info['cuda_v'] = cuda_version # ONNX Runtime version onnxruntime_version = 'unknown' ONNXRUNTIME_DIR = os.getenv('ONNXRUNTIME_DIR', '') ONNXRUNTIME_DIR = cmake_envs.get('ONNXRUNTIME_DIR', ONNXRUNTIME_DIR) ONNXRUNTIME_DIR = osp.expandvars(ONNXRUNTIME_DIR) if osp.exists(ONNXRUNTIME_DIR): with open(osp.join(ONNXRUNTIME_DIR, 'VERSION_NUMBER'), mode='r') as f: onnxruntime_version = f.readlines()[0].strip() env_info['ort_v'] = onnxruntime_version # TensorRT version tensorrt_version = 'unknown' TENSORRT_DIR = os.getenv('TENSORRT_DIR', '') TENSORRT_DIR = cmake_envs.get('TENSORRT_DIR', TENSORRT_DIR) TENSORRT_DIR = osp.expandvars(TENSORRT_DIR) if osp.exists(TENSORRT_DIR): with open( osp.join(TENSORRT_DIR, 'include', 'NvInferVersion.h'), mode='r') as f: data = f.read() major = re.search(r'#define NV_TENSORRT_MAJOR (\d+)', data) minor = re.search(r'#define NV_TENSORRT_MINOR (\d+)', data) patch = re.search(r'#define NV_TENSORRT_PATCH (\d+)', data) build = re.search(r'#define NV_TENSORRT_BUILD (\d+)', data) if major is not None and minor is not None and patch is not None: tensorrt_version = (f'{major.group(1)}.' + f'{minor.group(1)}.' + f'{patch.group(1)}.' + f'{build.group(1)}') env_info['trt_v'] = tensorrt_version return env_info def create_package(cfg: Dict, mmdeploy_dir: str): build_dir = 'build' sdk_tar_name = 'sdk' # load flags cfg, build_dir = get_dir_name(cfg, 'BUILD_NAME', build_dir) cmake_envs = cfg.get('cmake_envs', dict()) build_sdk_flag = cmake_envs.get('MMDEPLOY_BUILD_SDK', 'OFF') if 'TAR_NAME' in cfg: cfg, sdk_tar_name = get_dir_name(cfg, 'TAR_NAME', sdk_tar_name) # fill name env_info = check_env(cfg) version_file = osp.join(mmdeploy_dir, 'mmdeploy', 'version.py') mmdeploy_version = get_version(version_file) build_dir = build_dir.format(mmdeploy_v=mmdeploy_version, **env_info) # create package directory. if osp.exists(build_dir): logging.info(f'{build_dir} existed, deleting...') shutil.rmtree(build_dir) os.mkdir(build_dir) logging.info(f'build mmdeploy in {build_dir}:') logging.debug(f'with config: {cfg}') try: # build dist dist_dir = osp.join(build_dir, 'dist') build_mmdeploy(cfg, mmdeploy_dir, dist_dir=dist_dir) if build_sdk_flag == 'ON': sdk_tar_dir = osp.join(build_dir, sdk_tar_name) # copy lib and install into sdk dir install_dir = osp.join(mmdeploy_dir, 'build/install/') _copy(install_dir, sdk_tar_dir) _copy(f'{mmdeploy_dir}/demo/python', f'{sdk_tar_dir}/example/python') _remove_if_exist(osp.join(sdk_tar_dir, 'example', 'build')) # build SDK Python API according to different python version for python_version in ['3.6', '3.7', '3.8', '3.9']: _version = version.parse(python_version) python_major, python_minor = _version.major, _version.minor # create sdk python api wheel sdk_python_package_dir = osp.join(build_dir, '.mmdeploy_python') _copy(PACKAGING_DIR, sdk_python_package_dir) _copy( osp.join(mmdeploy_dir, 'mmdeploy', 'version.py'), osp.join(sdk_python_package_dir, 'mmdeploy_python', 'version.py'), ) # build mmdeploy sdk python api python_executable = shutil.which('python')\ .replace('mmdeploy-3.6', f'mmdeploy-{python_version}') python_api_lib_path = build_mmdeploy_python( python_executable, cfg, mmdeploy_dir) _copy( python_api_lib_path, osp.join(sdk_python_package_dir, 'mmdeploy_python'), ) _remove_if_exist(osp.join(mmdeploy_dir, 'build_python')) sdk_wheel_dir = osp.abspath(osp.join(sdk_tar_dir, 'python')) bdist_cmd = (f'{python_executable} ' f'setup.py bdist_wheel --plat-name ' f'{PLATFORM_TAG} --python-tag ' f'cp{python_major}{python_minor} ' f'--dist-dir {sdk_wheel_dir}') _call_command(bdist_cmd, sdk_python_package_dir) # remove temp package dir _remove_if_exist(sdk_python_package_dir) logging.info('build finish.') except CalledProcessError: logging.error('build failed') exit(-1) def parse_args(): parser = argparse.ArgumentParser(description='Build mmdeploy from yaml.') parser.add_argument('build_cfgs', help='The build config yaml file.') parser.add_argument('mmdeploy_dir', help='The source code of MMDeploy.') args = parser.parse_args() return args def parse_configs(cfg_path: str): with open(cfg_path, mode='r') as f: cfgs = yaml.load(f, yaml.Loader) global_cfg = cfgs.get('global_config', dict()) local_cfgs = cfgs.get('local_configs', []) merged_cfgs = [ _merge_cfg(global_cfg, local_cfg) for local_cfg in local_cfgs ] return merged_cfgs def main(): args = parse_args() cfgs = parse_configs(args.build_cfgs) mmdeploy_dir = osp.abspath(args.mmdeploy_dir) logging.info(f'Using mmdeploy_dir: {mmdeploy_dir}') logging.info(f'Using PACKAGING_DIR: {PACKAGING_DIR}') for cfg in cfgs: create_package(cfg, mmdeploy_dir) if __name__ == '__main__': main()