mirror of https://github.com/alibaba/EasyCV.git
85 lines
3.0 KiB
Python
85 lines
3.0 KiB
Python
# ------------------------------------------------------------------------------------------------
|
|
# Deformable DETR
|
|
# Copyright (c) 2020 SenseTime. All Rights Reserved.
|
|
# Licensed under the Apache License, Version 2.0 [see LICENSE for details]
|
|
# ------------------------------------------------------------------------------------------------
|
|
# Modified from https://github.com/chengdazhi/Deformable-Convolution-V2-PyTorch/tree/pytorch_1.0.0
|
|
# ------------------------------------------------------------------------------------------------
|
|
|
|
# Copyright (c) Facebook, Inc. and its affiliates.
|
|
# Modified by Bowen Cheng from https://github.com/fundamentalvision/Deformable-DETR
|
|
|
|
import glob
|
|
import os
|
|
from setuptools import find_packages, setup
|
|
|
|
import torch
|
|
from torch.utils.cpp_extension import CUDA_HOME, CppExtension, CUDAExtension
|
|
|
|
requirements = ['torch', 'torchvision']
|
|
|
|
|
|
def get_extensions():
|
|
this_dir = os.path.dirname(os.path.abspath(__file__))
|
|
extensions_dir = os.path.join(this_dir, 'src')
|
|
|
|
main_file = glob.glob(os.path.join(extensions_dir, '*.cpp'))
|
|
source_cpu = glob.glob(os.path.join(extensions_dir, 'cpu', '*.cpp'))
|
|
source_cuda = glob.glob(os.path.join(extensions_dir, 'cuda', '*.cu'))
|
|
|
|
sources = main_file + source_cpu
|
|
extension = CppExtension
|
|
extra_compile_args = {'cxx': []}
|
|
define_macros = []
|
|
|
|
# Force cuda since torch ask for a device, not if cuda is in fact available.
|
|
if (os.environ.get('FORCE_CUDA')
|
|
or torch.cuda.is_available()) and CUDA_HOME is not None:
|
|
extension = CUDAExtension
|
|
sources += source_cuda
|
|
define_macros += [('WITH_CUDA', None)]
|
|
extra_compile_args['nvcc'] = [
|
|
'-DCUDA_HAS_FP16=1',
|
|
'-D__CUDA_NO_HALF_OPERATORS__',
|
|
'-D__CUDA_NO_HALF_CONVERSIONS__',
|
|
'-D__CUDA_NO_HALF2_OPERATORS__',
|
|
]
|
|
else:
|
|
if CUDA_HOME is None:
|
|
raise NotImplementedError(
|
|
'CUDA_HOME is None. Please set environment variable CUDA_HOME.'
|
|
)
|
|
else:
|
|
raise NotImplementedError(
|
|
'No CUDA runtime is found. Please set FORCE_CUDA=1 or test it by running torch.cuda.is_available().'
|
|
)
|
|
|
|
sources = [os.path.join(extensions_dir, s) for s in sources]
|
|
include_dirs = [extensions_dir]
|
|
ext_modules = [
|
|
extension(
|
|
'MultiScaleDeformableAttention',
|
|
sources,
|
|
include_dirs=include_dirs,
|
|
define_macros=define_macros,
|
|
extra_compile_args=extra_compile_args,
|
|
)
|
|
]
|
|
return ext_modules
|
|
|
|
|
|
setup(
|
|
name='MultiScaleDeformableAttention',
|
|
version='1.0',
|
|
author='Weijie Su',
|
|
url='https://github.com/fundamentalvision/Deformable-DETR',
|
|
description=
|
|
'PyTorch Wrapper for CUDA Functions of Multi-Scale Deformable Attention',
|
|
packages=find_packages(exclude=(
|
|
'configs',
|
|
'tests',
|
|
)),
|
|
ext_modules=get_extensions(),
|
|
cmdclass={'build_ext': torch.utils.cpp_extension.BuildExtension},
|
|
)
|