* add initializers and BaseModule for unified parameter initialization
* fix circle import
* bug fix
* add is_init flag in BaseModule
* fix docstring
* sort import and fix doc format
* fix bug
* fix docformat and double quote string
* fix import sort
* import sort
* sort import
* revise according to comments
* fix doc format
* revise according to comments
* revise import and fix typo
* polish code
* revise minors
* revice minors
* revise apply function
* revise bias initialization with probability
* add type test for bias_prob
* revise minors
* Refactor _load_checkpoint fn
* Update _load_checkpoint fn
* Update docs str and add unit test
* Fix unit test
* Fix lint
* Add comment and Optimize function
* Fix docs str
* Update load_ckpt and fix doc str
* Update doc str and add sort unit test
* Update and fix unit test
* Fix unit test
* Update and add unit test
* Fix openmmlab prefix error
* Update lr_updater.py
since epoch/iteration in runner starts with 0, we shouldn't leave the latter iteration to former (12th epoch for example, with first period equal to 12) period.
* Update lr_updater.py
* Update test_hooks.py
* Support to specify LR of DCN's conv_offset
* Resolve comments & add unit test
* Resolve formats
* Fix CI for DCN
* Mock DCN when cpu only
* Use mock for cpu testing
* Fix docstring and support ModulatedDCN
* set offset_lr_mult as dcn's arguments, link CU-49u01p
* fix lr bug
* fall back to set LR in constructor
* resolve comments
* Add build_runner
* Parametrize test_runner
* Add imports to runner __init__
* Refactor max_iters and max_epochs from run to init
* Add assertion error messages
* Add test_builder
* Make change retro-compatible
* Raise ValueError if max_epochs and max_iters
* add ema hook
* add ema hook resume
* add ema hook test
* fix typo
* fix according to comment
* delete logger
* fix according to comment
* fix unitest
* fix typo
* fix according to comment
* change to resume_from
* typo
* fix isort
* fix: remove all module wrapper when saving checkpoint
* refactor: move position of if
* docs: add docstring
* refactor: add _save_to_state_dict from official torch
* docs: modify docstring of _save_to_state_dict
* docs: modify docstring
* feat: add unittest
* feat: add DataParallel to unittest
* fix: a bug when model has batchnorm
* docs: update docstring
* feat: support for os.environ port for slurm training
* fix: port data type
* feat: add flawed unittest
* feat: add flawed unittest
* docs: add comments
* fix: unittest
* fix: unittest
* feat: add CosineRestartLrUpdaterHook
* style: rename period to periods
* fix: bug in period 0
* feat: rename eta_min to min_lr and add min_lr_ratio
* docs: fix docstring of restart lr updater
* refactor: use annealing_cos
* docs: add docstring to annealing_cos
* feat: cosine restart lr update hook
* refactor: modify code order for unittest
* add a BaseRunner and rename Runner to EpochBasedRunner
* fix the train/val step
* bug fix
* update unit tests
* fix unit tests
* raise an error if both batch_processor and train_step are set
* add a unit test