[Feature]: Update github CI (#100)

* align mmedit static cfg

* add for test

* update requirments

* add dependencies from mmlab

* change name

* lower thresh for interrogate at first

* update test

* update to skip

* Move import tensorrt

* Move import statement

Co-authored-by: SingleZombie <singlezombie@163.com>
pull/12/head
RunningLeon 2021-09-27 17:43:41 +08:00 committed by GitHub
parent 09be668a27
commit f9b2cab3fa
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
8 changed files with 70 additions and 8 deletions

View File

@ -23,3 +23,40 @@ jobs:
source: backend_ops
extensions: h,c,cpp,hpp,cu,cuh
style: google
- name: Check docstring coverage
run: |
pip install interrogate
interrogate -v --ignore-init-method --ignore-module --ignore-nested-functions --ignore-regex "__repr__" --fail-under 30 mmdeploy
build_cpu:
runs-on: ubuntu-18.04
strategy:
matrix:
python-version: [3.7]
torch: [1.8.0]
include:
- torch: 1.8.0
torchvision: 0.9.0
mmcv: "latest+torch1.8.0+cpu"
steps:
- uses: actions/checkout@v2
- name: Set up Python ${{ matrix.python-version }}
uses: actions/setup-python@v2
with:
python-version: ${{ matrix.python-version }}
- name: Install PyTorch
run: pip install torch==${{matrix.torch}}+cpu torchvision==${{matrix.torchvision}}+cpu -f https://download.pytorch.org/whl/torch_stable.html
- name: Install MMCV
run: |
pip install mmcv-full -f https://download.openmmlab.com/mmcv/dist/cpu/torch${{matrix.torch}}/index.html
python -c 'import mmcv; print(mmcv.__version__)'
- name: Install unittest dependencies
run: |
pip install -r requirements/tests.txt -r requirements/optional.txt
- name: Build and install
run: rm -rf .eggs && pip install -e .
- name: Run unittests and generate coverage report
run: |
coverage run --branch --source mmdeploy -m pytest tests/
coverage xml
coverage report -m

View File

@ -18,6 +18,8 @@ def is_available():
return importlib.util.find_spec('tensorrt') is not None
__all__ = ['is_available']
if is_available():
from .onnx2tensorrt import onnx2tensorrt
from .tensorrt_utils import (TRTWrapper, create_trt_engine,
@ -26,7 +28,7 @@ if is_available():
# load tensorrt plugin lib
load_tensorrt_plugin()
__all__ = [
__all__ += [
'create_trt_engine', 'save_trt_engine', 'load_trt_engine',
'TRTWrapper', 'onnx2tensorrt'
]

View File

@ -1,4 +1,9 @@
mmcls
mmdet
mmedit
mmocr>=0.3.0
mmocr
mmsegmentation
ncnn
onnxruntime>=1.8.0
pyppl

View File

@ -1,3 +1,7 @@
asynctest
coverage
flake8
interrogate
isort==4.3.21
pytest
yapf

View File

@ -2,9 +2,20 @@ import os.path as osp
import tempfile
import mmcv
import pytest
import torch
from mmdeploy.apis import create_calib_table
from mmdeploy.apis.tensorrt.calib_utils import HDF5Calibrator
try:
from mmdeploy.apis.tensorrt.calib_utils import HDF5Calibrator
except ImportError:
pytest.skip(
'TensorRT should be installed from source.', allow_module_level=True)
if not torch.cuda.is_available():
pytest.skip(
'CUDA is required for this test module', allow_module_level=True)
calib_file = tempfile.NamedTemporaryFile(suffix='.h5').name
data_prefix = 'tests/data/tiger'

View File

@ -3,11 +3,10 @@ import tempfile
import mmcv
import pytest
import tensorrt as trt
import torch
import torch.nn as nn
from mmdeploy.apis.tensorrt import is_available, load_trt_engine, onnx2tensorrt
from mmdeploy.apis.tensorrt import is_available
onnx_file = tempfile.NamedTemporaryFile(suffix='.onnx').name
engine_file = tempfile.NamedTemporaryFile(suffix='.engine').name
@ -30,7 +29,7 @@ test_model = TestModel().eval().cuda()
def get_deploy_cfg():
import tensorrt as trt
deploy_cfg = mmcv.Config(
dict(
backend_config=dict(
@ -79,6 +78,7 @@ def generate_onnx_file(model):
@pytest.mark.skipif(trt_skip, reason='TensorRT not avaiable')
@pytest.mark.skipif(cuda_skip, reason='Cuda not avaiable')
def test_onnx2tensorrt():
from mmdeploy.apis.tensorrt import load_trt_engine, onnx2tensorrt
model = test_model
generate_onnx_file(model)
deploy_cfg = get_deploy_cfg()

View File

@ -4,8 +4,6 @@ import pytest
import torch
import torch.nn as nn
from mmdeploy.apis.tensorrt import (TRTWrapper, create_trt_engine,
save_trt_engine)
from mmdeploy.utils.constants import Backend
onnx_file = tempfile.NamedTemporaryFile(suffix='.onnx').name
@ -61,8 +59,8 @@ def check_backend_avaiable(backend):
def onnx2backend(backend, onnx_file):
if backend == Backend.TENSORRT:
from mmdeploy.apis.tensorrt import create_trt_engine, save_trt_engine
backend_file = tempfile.NamedTemporaryFile(suffix='.engine').name
engine = create_trt_engine(
onnx_file, {
@ -78,6 +76,7 @@ def onnx2backend(backend, onnx_file):
def create_wrapper(backend, engine_file):
if backend == Backend.TENSORRT:
from mmdeploy.apis.tensorrt import TRTWrapper
trt_model = TRTWrapper(engine_file)
return trt_model

View File

@ -10,6 +10,10 @@ import mmdeploy.apis.onnxruntime as ort_apis
import mmdeploy.apis.tensorrt as trt_apis
from mmdeploy.utils.test import assert_allclose
# PytestCollectionWarning: cannot collect test class 'TestOnnxRTExporter'
# because it has a __init__ constructor
pytest.skip('Skip test this file.', allow_module_level=True)
class TestOnnxRTExporter: