# Copyright (c) OpenMMLab. All rights reserved.
import argparse
import copy
import logging
import os
import os.path as osp
import platform
import re
import shutil
import sys
import tarfile
from distutils.util import get_platform
from glob import glob
from subprocess import CalledProcessError, check_output, run
from typing import Dict

import yaml
from packaging import version

logger = logging.getLogger()
logger.setLevel(logging.INFO)

CUR_DIR = os.path.dirname(os.path.abspath(__file__))
PACKAGING_DIR = osp.join(CUR_DIR, 'packaging')
PLATFORM_TAG = get_platform().replace('-', '_').replace('.', '_')


def get_version(version_file):
    with open(version_file, 'r') as f:
        exec(compile(f.read(), version_file, 'exec'))
    return locals()['__version__']


def _merge_cfg(cfg0, cfg1):
    cfg = copy.deepcopy(cfg0)
    for k, v in cfg1.items():
        if k in cfg:
            cfg[k] = _merge_cfg(cfg0[k], cfg1[k])
        else:
            cfg[k] = v
    return cfg


def _remove_if_exist(path):
    if osp.exists(path):
        logging.info(f'Remove path: {path}.')
        if osp.isdir(path):
            shutil.rmtree(path)
        else:
            os.remove(path)


def _copy(src_path, dst_path):
    assert osp.exists(src_path), f'src path: {src_path} not exist.'

    logging.info(f'copy path: {src_path} to {dst_path}.')
    if osp.isdir(src_path):
        if osp.exists(dst_path):
            shutil.rmtree(dst_path)
        shutil.copytree(src_path, dst_path)
    else:
        shutil.copy(src_path, dst_path)


def _call_command(cmd, cwd, stdout=None, stderr=None):
    if cmd == '':
        return
    logging.info(f'Process cmd: {cmd}')
    logging.info(f'work_path: {cwd}')
    try:
        ret = run(cmd, stdout=stdout, stderr=stderr, cwd=cwd, shell=True)
        if ret.returncode != 0:
            logging.error(f'Process cmd: "{cmd}"'
                          f' failed with returncode: {ret.returncode}')
            exit(-1)
    except Exception:
        logging.error(f'Process cmd: {cmd} failed.')
        exit(-1)


def _create_tar(path, tar_name):
    logging.info(f'create tar file: {tar_name}')
    with tarfile.open(tar_name, 'w:gz') as tar:
        tar.add(path, arcname=os.path.basename(path))


def _create_bdist_cmd(cfg, c_ext=False, dist_dir=None):

    bdist_tags = cfg.get('bdist_tags', {})

    # base
    bdist_cmd = 'python setup.py bdist_wheel '

    # platform
    bdist_cmd += f' --plat-name {PLATFORM_TAG} '

    # python tag
    python_tag = f'cp{sys.version_info.major}{sys.version_info.minor}'\
        if c_ext else 'py3'
    if 'python_tag' in bdist_tags:
        python_tag = bdist_tags['python_tag']
    bdist_cmd += f' --python-tag {python_tag} '

    # dist dir
    if dist_dir is not None:
        dist_dir = osp.abspath(dist_dir)
        bdist_cmd += f' --dist-dir {dist_dir} '
    return bdist_cmd


def clear_mmdeploy(mmdeploy_dir: str):
    logging.info(f'cleaning mmdeploy: {mmdeploy_dir}')

    def _remove_in_mmdeploy(path):
        remove_dir = osp.join(mmdeploy_dir, path)
        _remove_if_exist(remove_dir)

    # remove build file
    _remove_in_mmdeploy('build')

    # remove dist
    _remove_in_mmdeploy('dist')

    # remove installed library
    _remove_in_mmdeploy('mmdeploy/lib')

    # remove onnx2ncnn and ncnn ext
    _remove_in_mmdeploy('mmdeploy/backend/ncnn/onnx2ncnn')
    _remove_in_mmdeploy('mmdeploy/backend/ncnn/onnx2ncnn.exe')
    _remove_in_mmdeploy('mmdeploy/backend/ncnn/mmdeploy_onnx2ncnn')
    _remove_in_mmdeploy('mmdeploy/backend/ncnn/mmdeploy_onnx2ncnn.exe')
    ncnn_ext_paths = glob(
        osp.join(mmdeploy_dir, 'mmdeploy/backend/ncnn/ncnn_ext.*'))
    for ncnn_ext_path in ncnn_ext_paths:
        os.remove(ncnn_ext_path)

    # remove ts_optmizer
    ts_optimizer_paths = glob(
        osp.join(mmdeploy_dir, 'mmdeploy/backend/torchscript/ts_optimizer.*'))
    for ts_optimizer_path in ts_optimizer_paths:
        os.remove(ts_optimizer_path)


def build_mmdeploy(cfg, mmdeploy_dir, dist_dir=None):
    cmake_flags = cfg.get('cmake_flags', [])
    cmake_envs = cfg.get('cmake_envs', dict())

    args = [f'-D{k}={v}' for k, v in cmake_envs.items()]

    # clear mmdeploy
    clear_mmdeploy(mmdeploy_dir)

    build_dir = osp.join(mmdeploy_dir, 'build')
    if not osp.exists(build_dir):
        os.mkdir(build_dir)

        # cmake cmd
        cmake_cmd = ' '.join(['cmake ..'] + cmake_flags + args)
        _call_command(cmake_cmd, build_dir)

    if sys.platform == 'win32':
        # build cmd
        build_cmd = 'cmake --build . --config Release -- /m'
        _call_command(build_cmd, build_dir)
        install_cmd = 'cmake --install . --config Release'
        _call_command(install_cmd, build_dir)
        _remove_if_exist(osp.join(build_dir, 'lib', 'Release'))
    else:
        # build cmd
        build_cmd = 'cmake --build . -- -j$(nproc) && cmake --install .'
        _call_command(build_cmd, build_dir)

    # build wheel
    bdist_cmd = _create_bdist_cmd(cfg, c_ext=False, dist_dir=dist_dir)
    _call_command(bdist_cmd, mmdeploy_dir)


def build_mmdeploy_python(python_executable, cfg, mmdeploy_dir):
    cmake_flags = cfg.get('cmake_flags', [])
    cmake_envs = cfg.get('cmake_envs', dict())

    args = [f'-D{k}={v}' for k, v in cmake_envs.items()]
    args.append(
        f'-DMMDeploy_DIR={mmdeploy_dir}/build/install/lib/cmake/MMDeploy')
    args.append(f'-DPYTHON_EXECUTABLE={python_executable}')

    if sys.platform == 'win32':
        build_cmd = 'cmake --build . --config Release -- /m'
        pass
    else:
        build_cmd = 'cmake --build . -- -j$(nproc)'
    cmake_cmd = ' '.join(['cmake ../csrc/mmdeploy/apis/python'] + cmake_flags +
                         args)

    build_dir = osp.join(mmdeploy_dir, 'build_python')
    _remove_if_exist(build_dir)
    os.mkdir(build_dir)

    _call_command(cmake_cmd, build_dir)
    _call_command(build_cmd, build_dir)

    python_api_lib_path = []
    lib_patterns = ['*mmdeploy_python*.so', '*mmdeploy_python*.pyd']
    for pattern in lib_patterns:
        python_api_lib_path.extend(
            glob(
                osp.join(mmdeploy_dir, 'build_python/**', pattern),
                recursive=True,
            ))
    return python_api_lib_path[0]


def get_dir_name(cfg, tag, default_name):
    if tag not in cfg:
        logging.warning(f'{tag} not found, use `{default_name}` as default.')
    else:
        default_name = cfg[tag]
    return cfg, default_name


def check_env(cfg: Dict):
    env_info = {}

    cmake_envs = cfg.get('cmake_envs', dict())

    # system
    platform_system = platform.system().lower()
    platform_machine = platform.machine().lower()
    env_info['system'] = platform_system
    env_info['machine'] = platform_machine

    # CUDA version
    cuda_version = 'unknown'

    CUDA_TOOLKIT_ROOT_DIR = cmake_envs.get('CUDA_TOOLKIT_ROOT_DIR', '')
    CUDA_TOOLKIT_ROOT_DIR = osp.expandvars(CUDA_TOOLKIT_ROOT_DIR)
    nvcc_cmd = ('nvcc' if len(CUDA_TOOLKIT_ROOT_DIR) <= 0 else osp.join(
        CUDA_TOOLKIT_ROOT_DIR, 'bin', 'nvcc'))

    try:
        nvcc = check_output(f'"{nvcc_cmd}" -V', shell=True)
        nvcc = nvcc.decode('utf-8').strip()
        pattern = r'Cuda compilation tools, release (\d+.\d+)'
        match = re.search(pattern, nvcc)
        if match is not None:
            cuda_version = match.group(1)
    except Exception:
        pass

    env_info['cuda_v'] = cuda_version

    # ONNX Runtime version
    onnxruntime_version = 'unknown'

    ONNXRUNTIME_DIR = os.getenv('ONNXRUNTIME_DIR', '')
    ONNXRUNTIME_DIR = cmake_envs.get('ONNXRUNTIME_DIR', ONNXRUNTIME_DIR)
    ONNXRUNTIME_DIR = osp.expandvars(ONNXRUNTIME_DIR)

    if osp.exists(ONNXRUNTIME_DIR):
        with open(osp.join(ONNXRUNTIME_DIR, 'VERSION_NUMBER'), mode='r') as f:
            onnxruntime_version = f.readlines()[0].strip()

    env_info['ort_v'] = onnxruntime_version

    # TensorRT version
    tensorrt_version = 'unknown'

    TENSORRT_DIR = os.getenv('TENSORRT_DIR', '')
    TENSORRT_DIR = cmake_envs.get('TENSORRT_DIR', TENSORRT_DIR)
    TENSORRT_DIR = osp.expandvars(TENSORRT_DIR)

    if osp.exists(TENSORRT_DIR):
        with open(
                osp.join(TENSORRT_DIR, 'include', 'NvInferVersion.h'),
                mode='r') as f:
            data = f.read()
            major = re.search(r'#define NV_TENSORRT_MAJOR (\d+)', data)
            minor = re.search(r'#define NV_TENSORRT_MINOR (\d+)', data)
            patch = re.search(r'#define NV_TENSORRT_PATCH (\d+)', data)
            build = re.search(r'#define NV_TENSORRT_BUILD (\d+)', data)
            if major is not None and minor is not None and patch is not None:
                tensorrt_version = (f'{major.group(1)}.' +
                                    f'{minor.group(1)}.' +
                                    f'{patch.group(1)}.' + f'{build.group(1)}')

    env_info['trt_v'] = tensorrt_version

    return env_info


def create_package(cfg: Dict, mmdeploy_dir: str):
    build_dir = 'build'
    sdk_tar_name = 'sdk'

    # load flags
    cfg, build_dir = get_dir_name(cfg, 'BUILD_NAME', build_dir)
    cmake_envs = cfg.get('cmake_envs', dict())
    build_sdk_flag = cmake_envs.get('MMDEPLOY_BUILD_SDK', 'OFF')
    if 'TAR_NAME' in cfg:
        cfg, sdk_tar_name = get_dir_name(cfg, 'TAR_NAME', sdk_tar_name)

    # fill name
    env_info = check_env(cfg)
    version_file = osp.join(mmdeploy_dir, 'mmdeploy', 'version.py')
    mmdeploy_version = get_version(version_file)
    build_dir = build_dir.format(mmdeploy_v=mmdeploy_version, **env_info)

    # create package directory.
    if osp.exists(build_dir):
        logging.info(f'{build_dir} existed, deleting...')
        shutil.rmtree(build_dir)
    os.mkdir(build_dir)

    logging.info(f'build mmdeploy in {build_dir}:')
    logging.debug(f'with config: {cfg}')

    try:
        # build dist
        dist_dir = osp.join(build_dir, 'dist')
        build_mmdeploy(cfg, mmdeploy_dir, dist_dir=dist_dir)

        if build_sdk_flag == 'ON':

            sdk_tar_dir = osp.join(build_dir, sdk_tar_name)

            # copy lib and install into sdk dir
            install_dir = osp.join(mmdeploy_dir, 'build/install/')
            _copy(install_dir, sdk_tar_dir)
            _copy(f'{mmdeploy_dir}/demo/python',
                  f'{sdk_tar_dir}/example/python')
            _remove_if_exist(osp.join(sdk_tar_dir, 'example', 'build'))

            # build SDK Python API according to different python version
            for python_version in ['3.6', '3.7', '3.8', '3.9']:
                _version = version.parse(python_version)
                python_major, python_minor = _version.major, _version.minor

                # create sdk python api wheel
                sdk_python_package_dir = osp.join(build_dir,
                                                  '.mmdeploy_python')
                _copy(PACKAGING_DIR, sdk_python_package_dir)
                _copy(
                    osp.join(mmdeploy_dir, 'mmdeploy', 'version.py'),
                    osp.join(sdk_python_package_dir, 'mmdeploy_python',
                             'version.py'),
                )

                # build mmdeploy sdk python api
                python_executable = shutil.which('python')\
                    .replace('mmdeploy-3.6', f'mmdeploy-{python_version}')
                python_api_lib_path = build_mmdeploy_python(
                    python_executable, cfg, mmdeploy_dir)
                _copy(
                    python_api_lib_path,
                    osp.join(sdk_python_package_dir, 'mmdeploy_python'),
                )
                _remove_if_exist(osp.join(mmdeploy_dir, 'build_python'))

                sdk_wheel_dir = osp.abspath(osp.join(sdk_tar_dir, 'python'))

                bdist_cmd = (f'{python_executable} '
                             f'setup.py bdist_wheel --plat-name '
                             f'{PLATFORM_TAG} --python-tag '
                             f'cp{python_major}{python_minor} '
                             f'--dist-dir {sdk_wheel_dir}')
                _call_command(bdist_cmd, sdk_python_package_dir)

                # remove temp package dir
                _remove_if_exist(sdk_python_package_dir)

        logging.info('build finish.')

    except CalledProcessError:
        logging.error('build failed')
        exit(-1)


def parse_args():
    parser = argparse.ArgumentParser(description='Build mmdeploy from yaml.')
    parser.add_argument('build_cfgs', help='The build config yaml file.')
    parser.add_argument('mmdeploy_dir', help='The source code of MMDeploy.')
    args = parser.parse_args()

    return args


def parse_configs(cfg_path: str):
    with open(cfg_path, mode='r') as f:
        cfgs = yaml.load(f, yaml.Loader)

    global_cfg = cfgs.get('global_config', dict())
    local_cfgs = cfgs.get('local_configs', [])

    merged_cfgs = [
        _merge_cfg(global_cfg, local_cfg) for local_cfg in local_cfgs
    ]

    return merged_cfgs


def main():
    args = parse_args()
    cfgs = parse_configs(args.build_cfgs)
    mmdeploy_dir = osp.abspath(args.mmdeploy_dir)
    logging.info(f'Using mmdeploy_dir: {mmdeploy_dir}')

    logging.info(f'Using PACKAGING_DIR: {PACKAGING_DIR}')

    for cfg in cfgs:
        create_package(cfg, mmdeploy_dir)


if __name__ == '__main__':
    main()