* DCN cpu version
* add modulated dcn cpu version
* move deform_conv_shape_check to deform conv utils
* add inline to deform_conv_shape_check
* add tests
* run linter
* add newline at file end
* run pre-commit against modulated deform conv cpp
* update saconv test
* run clang-format
* remove cuda device inline
* refactor dcn cuda/cpu functions
* remove DCN util
* remove DCN util hpp from all included files
* Addressing PR comment by refactoring modulated-DCN
* fix lint in cpp files
* add windows CI
* clean versions
* only allow pt1.7 on windows
* fix windows install issue
* add win cpu
* fix win command
* clean unnecessary command
* resolve turbojpeg & tempfile on win
* replace os.readlink with os.path.realpath
* fix windows ci
* close file before removing it
* fix windows ci
* fix symlink on windows
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* fix windows ci
* modify according to comment
Co-authored-by: zhouzaida <zhouzaida@163.com>
* add modulated_deform_conv in onnxruntime support
* Add docs descriptions
* Add gpu test in test_onnx.py
* code format
* remove new usage and move if outside for loop
* use memset when bias is nullptr
* Add gradient cumulative optimizer
fixes#190
* Update optimizer.py
* Update optimizer.py
* fix loss scale improperly in last equivalent_iter
* Add `GradientCumulativeOptimizerHook` in `__init__.py`.
* Add docstring of `GradientCumulativeOptimizerHook`.
* Add type check, BN warning and resume warning. And fix typo, lint the
code.
* Add unit test
* Update docstring example.
* Change GradientCumulativeOptimizerHook `__init__` arguments.
* Add GradientCumulativeOptimzierHook unit tests with IterBasedRunner.
* Add GradientCumulativeFp16OptimizerHook.
* Add unit tests of GradientCumulativeFp16OptimizerHook
* Use '!=' instead of '>' to determine resume
Co-authored-by: Zhiyuan Chen <this@zyc.ai>
* Support deprecation checking in Config
* add unittest for config deprecation
* support reference link in deprecation info
* add doc
* Update config.md
* [Fix] Fix the bug that training log and evaluating log are mixed
* [Fix] Fix the bug that training log and evaluating log are mixed
* fix comment
* fix import error
* refactor
* refactor
* refactor
* clear log_buffer before evaluation
* fix error
* add unittest
* [Refactor] Refactor the csrc directory
* update MANIFEST.in
* fix hip
* add csrc readme
* trailing whitespace
* fix syntax error in setup.py
* add compatibility docs
* move parrots_cudawarpfunction.cuh to common/cuda
* fix grammar, update directory tree
* fix MANIFEST.in
* Add new structre of csrc in compatibility.md
* Add original structre of csrc in compatibility.md
* fix typo
* remove TODO
* modify according to comment
* format
Co-authored-by: grimoire <yaoqian@sensetime.com>
* [Docs] Add PR document
* format
* idx modify
* revise according to comments
* add blank line
* grammar
* insert blank line after title
* revise according to comments
* revise according to comments
* typo
* resize image3
* lint
* del print
* fix lint
* del unused
Co-authored-by: zhouzaida <zhouzaida@163.com>
* improve digit_version & use it for version_checking
* more testing for digit_version
* setuptools >= 50 is needed
* fix CI
* add debuging log
* >= to ==
* fix lint
* remove
* add failure case
* replace
* fix
* consider TORCH_VERSION == 'parrots'
* add unittest
* digit_version do not deal with the case if 'parrots' in version name.
* [Fix] Fix example
* rename Net to Model
* support training without gpu in example
* modify according to comment
* replace data path with relative path
* add flat cosine lr updater
* add test
* add doc
* update doc
* reformat
* update unittest
* update test flat cos
* remove momentum hook test
* update test
* change assert to ValueError
* fix unittest
* add by_epoch=True unittest
* change to start_percent
* change to start_percent in test
* add torch.roll to onnx
* remove skip test
* add support to torch<170
* add dim=0 for torch==1.9.0
* update fixture function name, add comment in roll symbolic