51 lines
1.3 KiB
YAML
51 lines
1.3 KiB
YAML
Collections:
|
|
- Name: iTPN
|
|
Metadata:
|
|
Architecture:
|
|
- Dense Connections
|
|
- GELU
|
|
- Layer Normalization
|
|
- Multi-Head Attention
|
|
- Scaled Dot-Product Attention
|
|
Paper:
|
|
Title: 'Integrally Pre-Trained Transformer Pyramid Networks'
|
|
URL: https://arxiv.org/abs/2211.12735
|
|
README: configs/itpn/README.md
|
|
Code:
|
|
URL: null
|
|
Version: null
|
|
|
|
Models:
|
|
- Name: itpn-clip-b_hivit-base-p16_8xb256-amp-coslr-800e_in1k
|
|
Metadata:
|
|
FLOPs: 18474000000
|
|
Parameters: 233000000
|
|
Training Data:
|
|
- ImageNet-1k
|
|
In Collection: iTPN
|
|
Results: null
|
|
Weights:
|
|
Config: configs/itpn/itpn-clip-b_hivit-base-p16_8xb256-amp-coslr-800e_in1k.py
|
|
|
|
- Name: itpn-pixel_hivit-base-p16_8xb512-amp-coslr-800e_in1k
|
|
Metadata:
|
|
FLOPs: 18474000000
|
|
Parameters: 103000000
|
|
Training Data:
|
|
- ImageNet-1k
|
|
In Collection: iTPN
|
|
Results: null
|
|
Weights:
|
|
Config: configs/itpn/itpn-pixel_hivit-base-p16_8xb512-amp-coslr-800e_in1k.py
|
|
|
|
- Name: itpn-pixel_hivit-large-p16_8xb512-amp-coslr-800e_in1k
|
|
Metadata:
|
|
FLOPs: 63977000000
|
|
Parameters: 314000000
|
|
Training Data:
|
|
- ImageNet-1k
|
|
In Collection: iTPN
|
|
Results: null
|
|
Weights:
|
|
Config: configs/itpn/itpn-pixel_hivit-large-p16_8xb512-amp-coslr-800e_in1k.py
|