* minor changes
* change to modulist
* change to Sequential
* replace dropout with attn_drop and proj_drop in MultiheadAttention
* add operation_name for attn
* add drop path and move all ffn args to ffncfgs
* fix typo
* fix a bug when use default value of ffn_cfgs
* fix ffns
* add deprecate warning
* fix deprecate warning
* change to pop kwargs
* support register FFN of transformer
* support batch first
* fix batch first wapper
* fix forward wapper
* fix typo
* fix lint
* add unitest for transformer
* fix unitest
* fix equal
* use allclose
* fix comments
* fix comments
* change configdict to dict
* move drop to a file
* add comments for drop path
* add noqa 501
* move bnc wapper to MultiheadAttention
* move bnc wapper to MultiheadAttention
* use dep warning
* resolve comments
* add unitest:
* rename residual to identity
* revert runner
* msda residual to identity
* rename inp_identity to identity
* fix name
* fix transformer
* remove key in msda
* remove assert for key
Co-authored-by: HIT-cwh <2892770585@qq.com>
Co-authored-by: bkhuang <congee524@gmail.com>
Co-authored-by: Wenwei Zhang <40779233+ZwwWayne@users.noreply.github.com>
* add c++ ms_deform_atten
* fix cpp lint
* fix cpp lint
* clang format
* remove cmakefile
* google style
* clang-format precommit
* use clang-format-lint-action
* add transformer base class
* add merge
* add docstr
* add pyargs
* fix according to commments
* resiger module
* change to use basemodule
* add _ between build function
* split the name
* fix according to comments
* fix lint and fix unitest
* fix cpp lint
* fix bug of deformdetr_atten
* fix drop out
* fix residual
* use CUDA_1D_KERNEL_LOOP