12 Commits

Author SHA1 Message Date
Ross Wightman
511a8e8c96 Add official ResMLP weights. 2021-06-14 17:03:16 -07:00
Ross Wightman
4d96165989 Merge branch 'master' into cleanup_xla_model_fixes 2021-06-12 23:19:25 -07:00
Ross Wightman
8880f696b6 Refactoring, cleanup, improved test coverage.
* Add eca_nfnet_l2 weights, 84.7 @ 384x384
* All 'non-std' (ie transformer / mlp) models have classifier / default_cfg test added
* Fix #694 reset_classifer / num_features / forward_features / num_classes=0 consistency for transformer / mlp models
* Add direct loading of npz to vision transformer (pure transformer so far, hybrid to come)
* Rename vit_deit* to deit_*
* Remove some deprecated vit hybrid model defs
* Clean up classifier flatten for conv classifiers and unusual cases (mobilenetv3/ghostnet)
* Remove explicit model fns for levit conv, just pass in arg
2021-06-12 16:40:02 -07:00
Ross Wightman
d413eef1bf Add ResMLP-24 model weights that I trained in PyTorch XLA on TPU-VM. 79.2 top-1. 2021-06-08 14:22:05 -07:00
Ross Wightman
2f5ed2dec1 Update init_values const for 24 and 36 layer ResMLP models 2021-06-07 17:15:04 -07:00
Ross Wightman
bfc72f75d3 Expand scope of testing for non-std vision transformer / mlp models. Some related cleanup and create fn cleanup for all vision transformer and mlp models. More CoaT weights. 2021-05-24 21:13:26 -07:00
talrid
dc1a4efd28 mixer_b16_224_miil, mixer_b16_224_miil_in21k models 2021-05-20 10:35:50 +03:00
Ross Wightman
d5af752117 Add preliminary gMLP and ResMLP impl to Mlp-Mixer 2021-05-19 09:55:05 -07:00
Ross Wightman
e7f0db8664 Fix drop/drop_path arg on MLP-Mixer model. Fix #641 2021-05-16 08:31:52 -07:00
Ross Wightman
b2c305c2aa Move Mlp and PatchEmbed modules into layers. Being used in lots of models now... 2021-05-06 14:03:23 -07:00
Ross Wightman
2d8b09fe8b Add official pretrained weights to MLP-Mixer, complete model cfgs. 2021-05-05 15:59:40 -07:00
Ross Wightman
12efffa6b1 Initial MLP-Mixer attempt... 2021-05-05 00:59:45 -07:00