Ross Wightman
43aa84e861
Add 'fast' layer norm that doesn't cast to float32, support APEX LN impl for slight speed gain, update norm and act factories, tweak SE for ability to disable bias (needed by GCVit)
2022-08-17 14:32:58 -07:00
Ross Wightman
c486aa71f8
Add GCViT
2022-08-17 14:29:18 -07:00
Ross Wightman
fba6ecd39b
Add EfficientFormer
2022-08-17 14:08:53 -07:00
Ross Wightman
ff4a38e2c3
Add PyramidVisionTransformerV2
2022-08-17 12:06:05 -07:00
Ross Wightman
1d8ada359a
Add timm ConvNeXt 'atto' weights, change test resolution for FB ConvNeXt 224x224 weights, add support for different dw kernel_size
2022-08-15 17:56:08 -07:00
Ross Wightman
2544d3b80f
ConvNeXt pico, femto, and nano, pico, femto ols (overlapping stem) weights and model defs
2022-08-05 17:05:50 -07:00
Ross Wightman
13565aad50
Add edgenext_base model def & weight link, update to improve ONNX export #1385
2022-08-05 16:58:34 -07:00
Ross Wightman
8ad4bdfa06
Allow ntuple to be used with string values
2022-07-28 16:18:18 -07:00
Christoph Reich
faae93e62d
Fix typo in PositionalEncodingFourier
2022-07-28 19:08:08 -04:00
Ross Wightman
ec6a28830f
Add DeiT-III 'medium' model defs and weights
2022-07-28 15:03:20 -07:00
Ross Wightman
6f103a442b
Add convnext_nano weights, 80.8 @ 224, 81.5 @ 288
2022-07-26 16:40:27 -07:00
Ross Wightman
4042a94f8f
Add weights for two 'Edge' block (3x3->1x1) variants of CS3 networks.
2022-07-26 16:40:27 -07:00
Ross Wightman
c8f69e04a9
Merge pull request #1365 from veritable-tech/fix-resize-pos-embed
...
Take `no_emb_class` into account when calling `resize_pos_embed`
2022-07-24 21:03:01 -07:00
Ceshine Lee
0b64117592
Take `no_emb_class` into account when calling `resize_pos_embed`
2022-07-24 19:11:45 +08:00
Jasha10
56c3a84db3
Update type hint for `register_notrace_module`
...
register_notrace_module is used to decorate types (i.e. subclasses of nn.Module).
It is not called on module instances.
2022-07-22 16:59:55 -05:00
Ross Wightman
1b278136c3
Change models with mean 0,0,0 std 1,1,1 from int to float for consistency as mentioned in #1355
2022-07-21 17:36:15 -07:00
Ross Wightman
909705e7ff
Remove some redundant requires_grad=True from nn.Parameter in third party code
2022-07-20 12:37:41 -07:00
Ross Wightman
c5e0d1c700
Add dilation support to convnext, allows output_stride=8 and 16 use. Fix #1341
2022-07-19 17:52:10 -07:00
Ross Wightman
dc376e3676
Ensure all model entrypoint fn default to `pretrained=False` (a few didn't)
2022-07-19 13:58:41 -07:00
Ross Wightman
23b102064a
Add cs3sedarknet_x weights w/ 82.65 @ 288 top1. Add 2 cs3 edgenet models (w/ 3x3-1x1 block), remove aa from cspnet blocks (not needed)
2022-07-19 13:56:44 -07:00
Ross Wightman
05313940e2
Add cs3darknet_x, cs3sedarknet_l, and darknetaa53 weights from TPU sessions. Move SE btwn conv1 & conv2 in DarkBlock. Improve SE/attn handling in Csp/DarkNet. Fix leaky_relu bug on older csp models.
2022-07-15 16:55:16 -07:00
nateraw
51cca82aa1
👽 use hf_hub_download instead of cached_download
2022-07-14 16:41:45 -04:00
Ross Wightman
a45b4bce9a
x and xx small edgenext models do benefit from larger test input size
2022-07-08 10:53:27 -07:00
Ross Wightman
a8e34051c1
Unbreak gamma remap impacting beit checkpoint load, version bump to 0.6.4
2022-07-07 23:07:43 -07:00
Ross Wightman
a1cb25066e
Add edgnext_small_rw weights trained with swin like recipe. Better than original 'small' but not the recent 'USI' distilled weights.
2022-07-07 22:02:57 -07:00
Ross Wightman
7c7ecd2492
Add --use-train-size flag to force use of train input_size (over test input size) for validation. Default test-time pooling to use train input size (fixes issues).
2022-07-07 22:01:24 -07:00
Ross Wightman
ce65a7b29f
Update vit_relpos w/ some additional weights, some cleanup to match recent vit updates, more MLP log coord experiments.
2022-07-07 21:33:25 -07:00
Ross Wightman
58621723bd
Add CrossStage3 DarkNet (cs3) weights
2022-07-07 17:43:38 -07:00
Ross Wightman
db0cee9910
Refactor cspnet configuration using dataclasses, update feature extraction for new cs3 variants.
2022-07-07 14:43:27 -07:00
Ross Wightman
eca09b8642
Add MobileVitV2 support. Fix #1332 . Move GroupNorm1 to common layers (used in poolformer + mobilevitv2). Keep ol custom ConvNeXt LayerNorm2d impl as LayerNormExp2d for reference.
2022-07-07 14:41:01 -07:00
Ross Wightman
06307b8b41
Remove experimental downsample in block support in ConvNeXt. Experiment further before keeping it in.
2022-07-07 14:37:58 -07:00
Ross Wightman
7d4b3807d5
Support DeiT-3 (Revenge of the ViT) checkpoints. Add non-overlapping (w/ class token) pos-embed support to vit.
2022-07-04 22:25:22 -07:00
Ross Wightman
d0c5bd5722
Rename cs2->cs3 for darknets. Fix features_only for cs3 darknets.
2022-07-03 08:32:41 -07:00
Ross Wightman
d765305821
Remove first_conv for resnetaa50 def
2022-07-02 15:56:17 -07:00
Ross Wightman
dd9b8f57c4
Add feature_info to edgenext for features_only support, hopefully fix some fx / test errors
2022-07-02 15:20:45 -07:00
Ross Wightman
377e9bfa21
Add TPU trained darknet53 weights. Add mising pretrain_cfg for some csp/darknet models.
2022-07-02 15:18:52 -07:00
Ross Wightman
c170ba3173
Add weights for resnet10t, resnet14t, and resnetaa50 models. Fix #1314
2022-07-02 15:18:06 -07:00
Ross Wightman
188c194b0f
Left some experiment stem code in convnext by mistake
2022-07-02 15:17:28 -07:00
Ross Wightman
6064d16a2d
Add initial EdgeNeXt import. Significant cleanup / reorg (like ConvNeXt). Fix #1320
...
* edgenext refactored for torchscript compat, stage base organization
* slight refactor of ConvNeXt to match some EdgeNeXt additions
* remove use of funky LayerNorm layer in ConvNeXt and just use nn.LayerNorm and LayerNorm2d (permute)
2022-07-01 15:18:42 -07:00
Ross Wightman
7a9c6811c9
Add eps arg to LayerNorm2d, add 'tf' (tensorflow) variant of trunc_normal_ that applies scale/shift after sampling (instead of needing to move a/b)
2022-07-01 15:15:39 -07:00
Ross Wightman
82c311d082
Add more experimental darknet and 'cs2' darknet variants (different cross stage setup, closer to newer YOLO backbones) for train trials.
2022-07-01 15:14:01 -07:00
Ross Wightman
a050fde5cd
Add resnet10t (basic block) and resnet14t (bottleneck) with 1,1,1,1 repeats
2022-07-01 15:03:28 -07:00
Ross Wightman
e6d7df40ec
no longer a point using kwargs for pretrain_cfg resolve, just pass explicit arg
2022-06-24 21:36:23 -07:00
Ross Wightman
07d0c4ae96
Improve repr for DropPath module
2022-06-24 14:58:15 -07:00
Ross Wightman
e27c16b8a0
Remove unecessary code for synbn guard
2022-06-24 14:57:42 -07:00
Ross Wightman
0da3c9ebbf
Remove SiLU layer in default args that breaks import on old old PyTorch
2022-06-24 14:56:58 -07:00
Ross Wightman
7d657d2ef4
Improve resolve_pretrained_cfg behaviour when no cfg exists, warn instead of crash. Improve usability ex #1311
2022-06-24 14:55:25 -07:00
Ross Wightman
879df47c0a
Support BatchNormAct2d for sync-bn use. Fix #1254
2022-06-24 14:51:26 -07:00
Ross Wightman
4b30bae67b
Add updated vit_relpos weights, and impl w/ support for official swin-v2 differences for relpos. Add bias control support for MLP layers
2022-05-13 13:53:57 -07:00
Ross Wightman
d4c0588012
Remove persistent buffers from Swin-V2. Change SwinV2Cr cos attn + tau/logit_scale to match official, add ckpt convert, init_value zeros resid LN weight by default
2022-05-13 10:50:59 -07:00
Ross Wightman
27c42f0830
Fix torchscript use for offician Swin-V2, add support for non-square window/shift to WindowAttn/Block
2022-05-13 09:29:33 -07:00
Ross Wightman
c0211b0bf7
Swin-V2 test fixes, typo
2022-05-12 22:31:55 -07:00
Ross Wightman
9a86b900fa
Official SwinV2 models
2022-05-12 15:05:10 -07:00
Ross Wightman
d07d015173
Merge pull request #1249 from okojoalg/sequencer
...
Add Sequencer
2022-05-09 20:42:43 -07:00
Ross Wightman
39b725e1c9
Fix tests for rank-4 output where feature channels dim is -1 (3) and not 1
2022-05-09 15:20:24 -07:00
Ross Wightman
78a32655fa
Fix poolformer group_matcher to merge proj downsample with previous block, support coarse
2022-05-09 12:20:04 -07:00
Ross Wightman
d79f3d9d1e
Fix torchscript use for sequencer, add group_matcher, forward_head support, minor formatting
2022-05-09 12:09:39 -07:00
Ross Wightman
37b6920df3
Fix group_matcher regex for regnet.py
2022-05-09 10:40:40 -07:00
okojoalg
93a79a3dd9
Fix num_features in Sequencer
2022-05-06 23:16:32 +09:00
okojoalg
578d52e752
Add Sequencer
2022-05-06 00:36:01 +09:00
Ross Wightman
f5ca4141f7
Adjust arg order for recent vit model args, add a few comments
2022-05-02 22:41:38 -07:00
Ross Wightman
41dc49a337
Vision Transformer refactoring and Rel Pos impl
2022-05-02 15:37:39 -07:00
Ross Wightman
b7cb8d0337
Add Swin-V2 Small-NS weights (83.5 @ 224). Add layer scale like 'init_values' via post-norm LN weight scaling
2022-04-26 17:32:49 -07:00
jjsjann123
f88c606fcf
fixing channels_last on cond_conv2d; update nvfuser debug env variable
2022-04-25 12:41:46 -07:00
Li Dong
09e9f3defb
migrate azure blob for beit checkpoints
...
## Motivation
We are going to use a new blob account to store the checkpoints.
## Modification
Modify the azure blob storage URLs for BEiT checkpoints.
2022-04-23 13:02:29 +08:00
Ross Wightman
52ac881402
Missed first_conv in latest seresnext 'D' default_cfgs
2022-04-22 20:55:52 -07:00
Ross Wightman
7629d8264d
Add two new SE-ResNeXt101-D 32x8d weights, one anti-aliased and one not. Reshuffle default_cfgs vs model entrypoints for resnet.py so they are better aligned.
2022-04-22 16:54:53 -07:00
SeeFun
8f0bc0591e
fix convnext args
2022-04-05 20:00:57 +08:00
Ross Wightman
c5a8e929fb
Add initial swinv2 tiny / small weights
2022-04-03 15:22:55 -07:00
Ross Wightman
f670d98cb8
Make a few more layers symbolically traceable (remove from FX leaf modules)
...
* remove dtype kwarg from .to() calls in EvoNorm as it messed up script + trace combo
* BatchNormAct2d always uses custom forward (cut & paste from original) instead of super().forward. Fixes #1176
* BlurPool groups==channels, no need to use input.dim[1]
2022-03-24 21:43:56 -07:00
SeeFun
ec4e9aa5a0
Add ConvNeXt tiny and small pretrain in22k
...
Add ConvNeXt tiny and small pretrain in22k from ConvNeXt repo:
06f7b05f92
2022-03-24 15:18:08 +08:00
Ross Wightman
575924ed60
Update test crop for new RegNet-V weights to match Y
2022-03-23 21:40:53 -07:00
Ross Wightman
1618527098
Add layer scale and parallel blocks to vision_transformer
2022-03-23 16:09:07 -07:00
Ross Wightman
c42be74621
Add attrib / comments about Swin-S3 (AutoFormerV2) weights
2022-03-23 16:07:09 -07:00
Ross Wightman
474ac906a2
Add 'head norm first' convnext_tiny_hnf weights
2022-03-23 16:06:00 -07:00
Ross Wightman
dc51334cdc
Fix pruned adapt for EfficientNet models that are now using BatchNormAct layers
2022-03-22 20:33:01 -07:00
Ross Wightman
024fc4d9ab
version 0.6.1 for master
2022-03-21 22:03:13 -07:00
Ross Wightman
e1e037ba52
Fix bad tuple typing fix that was on XLA branch bust missed on master merge
2022-03-21 22:00:33 -07:00
Ross Wightman
fe457c1996
Update SwinTransformerV2Cr post-merge, update with grad checkpointing / grad matcher
...
* weight compat break, activate norm3 for final block of final stage (equivalent to pre-head norm, but while still in BLC shape)
* remove fold/unfold for TPU compat, add commented out roll code for TPU
* add option for end of stage norm in all stages
* allow weight_init to be selected between pytorch default inits and xavier / moco style vit variant
2022-03-21 14:50:28 -07:00
Ross Wightman
b049a5c5c6
Merge remote-tracking branch 'origin/master' into norm_norm_norm
2022-03-21 13:41:43 -07:00
Ross Wightman
9440a50c95
Merge branch 'mrT23-master'
2022-03-21 12:30:02 -07:00
Ross Wightman
d98aa47d12
Revert ml-decoder changes to model factory and train script
2022-03-21 12:29:02 -07:00
Ross Wightman
b20665d379
Merge pull request #1007 from qwertyforce/patch-1
...
update arxiv link
2022-03-21 12:12:58 -07:00
Ross Wightman
61d3493f87
Fix hf-hub handling when hf-hub is config source
2022-03-21 11:12:55 -07:00
Ross Wightman
5f47518f27
Fix pit implementation to be clsoer to deit/levit re distillation head handling
2022-03-21 11:12:14 -07:00
Ross Wightman
0862e6ebae
Fix correctness of some group matching regex (no impact on result), some formatting, missed forward_head for resnet
2022-03-19 14:58:54 -07:00
Ross Wightman
94bcdebd73
Add latest weights trained on TPU-v3 VM instances
2022-03-18 21:35:41 -07:00
Ross Wightman
0557c8257d
Fix bug introduced in non layer_decay weight_decay application. Remove debug print, fix arg desc.
2022-02-28 17:06:32 -08:00
Ross Wightman
372ad5fa0d
Significant model refactor and additions:
...
* All models updated with revised foward_features / forward_head interface
* Vision transformer and MLP based models consistently output sequence from forward_features (pooling or token selection considered part of 'head')
* WIP param grouping interface to allow consistent grouping of parameters for layer-wise decay across all model types
* Add gradient checkpointing support to a significant % of models, especially popular architectures
* Formatting and interface consistency improvements across models
* layer-wise LR decay impl part of optimizer factory w/ scale support in scheduler
* Poolformer and Volo architectures added
2022-02-28 13:56:23 -08:00
Ross Wightman
1420c118df
Missed comitting outstanding changes to default_cfg keys and test exclusions for swin v2
2022-02-23 19:50:26 -08:00
Ross Wightman
c6e4b7895a
Swin V2 CR impl refactor.
...
* reformat and change some naming so closer to existing timm vision transformers
* remove typing that wasn't adding clarity (or causing torchscript issues)
* support non-square windows
* auto window size adjust from image size
* post-norm + main-branch no
2022-02-23 17:28:52 -08:00
Christoph Reich
67d140446b
Fix bug in classification head
2022-02-20 22:28:05 +01:00
Christoph Reich
29add820ac
Refactor (back to relative imports)
2022-02-20 00:46:48 +01:00
Christoph Reich
74a04e0016
Add parameter to change normalization type
2022-02-20 00:46:00 +01:00
Christoph Reich
2a4f6c13dd
Create model functions
2022-02-20 00:40:22 +01:00
Christoph Reich
87b4d7a29a
Add get and reset classifier method
2022-02-19 22:47:02 +01:00
Christoph Reich
ff5f6bcd6c
Check input resolution
2022-02-19 22:42:02 +01:00
Christoph Reich
81bf0b4033
Change parameter names to match Swin V1
2022-02-19 22:37:22 +01:00
Christoph Reich
f227b88831
Add initials (CR) to model and file
2022-02-19 22:14:38 +01:00
Christoph Reich
90dc74c450
Add code from https://github.com/ChristophReich1996/Swin-Transformer-V2 and change docstring style to match timm
2022-02-19 22:12:11 +01:00
Ross Wightman
2c3870e107
semobilevit_s for good measure
2022-01-31 22:36:09 -08:00
Ross Wightman
58ba49c8ef
Add MobileViT models (w/ ByobNet base). Close #1038 .
2022-01-31 15:39:34 -08:00
Ross Wightman
5f81d4de23
Move DeiT to own file, vit getting crowded. Working towards fixing #1029 , make pooling interface for transformers and mlp closer to convnets. Still working through some details...
2022-01-26 22:53:57 -08:00
Ross Wightman
95cfc9b3e8
Merge remote-tracking branch 'origin/master' into norm_norm_norm
2022-01-25 22:20:45 -08:00
Ross Wightman
abc9ba2544
Transitioning default_cfg -> pretrained_cfg. Improving handling of pretrained_cfg source (HF-Hub, files, timm config, etc). Checkpoint handling tweaks.
2022-01-25 21:54:13 -08:00
Ross Wightman
07379c6d5d
Add vit_base2_patch32_256 for a model between base_patch16 and patch32 with a slightly larger img size and width
2022-01-24 14:46:47 -08:00
Ross Wightman
83b40c5a58
Last batch of small model weights (for now). mobilenetv3_small 050/075/100 and updated mnasnet_small with lambc/lamb optimizer.
2022-01-19 10:02:02 -08:00
Ross Wightman
1aa617cb3b
Add AvgPool2d anti-aliasing support to ResNet arch (as per OpenAI CLIP models), add a few blur aa models as well
2022-01-18 21:57:24 -08:00
Ross Wightman
010b486590
Add Dino pretrained weights (no head) for vit models. Add support to tests and helpers for models w/ no classifier (num_classes=0 in pretrained cfg)
2022-01-17 12:20:02 -08:00
Ross Wightman
738a9cd635
unbiased=False for torch.var_mean path of ConvNeXt LN. Fix #1090
2022-01-17 09:25:06 -08:00
Ross Wightman
e0c4eec4b6
Default conv_mlp to False across the board for ConvNeXt, causing issues on more setups than it's improving right now...
2022-01-16 14:20:08 -08:00
Ross Wightman
b669f4a588
Add ConvNeXt 22k->1k fine-tuned and 384 22k-1k fine-tuned weights after testing
2022-01-15 15:44:36 -08:00
Ross Wightman
e967c72875
Update REAMDE.md. Sneak in g/G (giant / gigantic?) ViT defs from scaling paper
2022-01-14 16:28:27 -08:00
Ross Wightman
9ca3437178
Add some more small model weights lcnet, mnas, mnv2
2022-01-14 16:28:27 -08:00
Ross Wightman
fa81164378
Fix stem width for really small mobilenetv3 arch defs
2022-01-14 16:28:27 -08:00
Ross Wightman
edd3d73695
Add missing dropout for head reset in ConvNeXt default head
2022-01-14 16:28:27 -08:00
Ross Wightman
b093dcb46d
Some convnext cleanup, remove in place mul_ for gamma, breaking symbolic trace, cleanup head a bit...
2022-01-14 16:28:27 -08:00
Ross Wightman
18934debc5
Add initial ConvNeXt impl (mods of official code)
2022-01-14 16:28:27 -08:00
Ross Wightman
656757d26b
Fix MobileNetV2 head conv size for multiplier < 1.0. Add some missing modification copyrights, fix starting date of some old ones.
2022-01-14 16:28:27 -08:00
Ross Wightman
ccfeb06936
Fix out_indices handling breakage, should have left as per vgg approach.
2022-01-07 19:30:51 -08:00
Ross Wightman
a9f91483a6
Fix #1078 , DarkNet has 6 feature maps. Make vgg and darknet out_indices handling/comments equivalent
2022-01-07 15:08:32 -08:00
Ross Wightman
c21b21660d
visformer supports spatial feat map, update pool_size in pretrained cfg to match
2022-01-07 14:31:43 -08:00
Ross Wightman
9c11dfd9cb
Fix fbnetv3 pretrained cfg changes
2022-01-07 14:09:50 -08:00
Ross Wightman
1406cddc2e
FBNetV3 timm trained weights added for b/d/g variants. Update version to 0.5.2 for pypi release.
2022-01-07 12:05:08 -08:00
Ross Wightman
4df51f3932
Add lcnet_100 and mnasnet_small weights
2022-01-06 22:21:05 -08:00
Ross Wightman
5ccf682a8f
Remove deprecated bn-tf train arg and create_model handler. Add evos/evob models back into fx test filter until norm_norm_norm branch merged.
2022-01-06 18:08:39 -08:00
Ross Wightman
b9a715c86a
Add more small model defs for MobileNetV3/V2/LCNet
2022-01-06 16:06:43 -08:00
Ross Wightman
b27c21b09a
Update drop_path and drop_block (fast impl) to be symbolically traceable, slightly faster
2022-01-06 16:04:58 -08:00
Ross Wightman
214c84a235
Disable use of timm nn.Linear wrapper since AMP autocast + torchscript use appears fixed
2022-01-06 16:01:51 -08:00
Ross Wightman
72b57163d1
Merge branch 'master' of https://github.com/mrT23/pytorch-image-models into mrT23-master
2022-01-06 13:57:16 -08:00
Ross Wightman
de5fa791c6
Merge branch 'master' into norm_norm_norm
2022-01-03 11:37:00 -08:00
Ross Wightman
26ff57f953
Add more small model defs for MobileNetV3/V2/LCNet
2022-01-03 11:30:54 -08:00
Ross Wightman
450ac6a0f5
Post merge tinynet fixes for pool_size, feature extraction
2021-12-21 23:51:54 -08:00
Ross Wightman
a04164cd75
Merge branch 'tinynet' of https://github.com/rsomani95/pytorch-image-models into rsomani95-tinynet
2021-12-21 22:45:56 -08:00
Ross Wightman
8a93ce6ee3
Fix regnetv/w tests, refactor regnet generator code a bit
2021-12-16 17:08:25 -08:00
Ross Wightman
4dec8c8087
Fix skip path regression for updated EfficientNet and RegNet def. Add Pre-Act RegNet support (experimental). Remove BN-TF flag. Add efficientnet_b0_g8_gn model.
2021-12-16 14:53:57 -08:00
Ross Wightman
a52a614475
Remove layer experiment which should not have been added
2021-12-14 14:29:32 -08:00
Ross Wightman
ab49d275de
Significant norm update
...
* ConvBnAct layer renamed -> ConvNormAct and ConvNormActAa for anti-aliased
* Significant update to EfficientNet and MobileNetV3 arch to support NormAct layers and grouped conv (as alternative to depthwise)
* Update RegNet to add Z variant
* Add Pre variant of XceptionAligned that works with NormAct layers
* EvoNorm matches bits_and_tpu branch for merge
2021-12-14 13:48:30 -08:00
Rahul Somani
31bcd36e46
add tinynet models
2021-12-14 19:34:04 +05:30
KAI ZHAO
b4b8d1ec18
fix hard-coded strides
2021-12-14 17:22:54 +08:00
Ross Wightman
d04f2f1377
Update drop_path and drop_block (fast impl) to be symbolically traceable, slightly faster
2021-12-05 15:36:56 -08:00
Ross Wightman
834a9ec721
Disable use of timm nn.Linear wrapper since AMP autocast + torchscript use appears fixed
2021-12-01 14:58:09 -08:00
Ross Wightman
78912b6375
Updated EvoNorm implementations with some experimentation. Add FilterResponseNorm. Updated RegnetZ and ResNetV2 model defs for trials.
2021-12-01 12:09:01 -08:00
talrid
c11f4c3218
support CNNs
2021-11-30 08:48:08 +02:00
mrT23
d6701d8a81
Merge branch 'rwightman:master' into master
2021-11-30 08:07:44 +02:00
qwertyforce
ccb3815360
update arxiv link
2021-11-29 21:41:00 +03:00
Ross Wightman
3dc71695bf
Merge pull request #989 from martinsbruveris/feat/resmlp-dino
...
Added DINO pretrained ResMLP models.
2021-11-24 09:26:07 -08:00
Ross Wightman
480c676ffa
Fix FX breaking assert in evonorm
2021-11-24 09:24:47 -08:00
Martins Bruveris
85c5ff26d7
Added DINO pretrained ResMLP models.
2021-11-24 15:02:46 +02:00
Ross Wightman
d633a014e6
Post merge cleanup. Fix potential security issue passing kwargs directly through to serialized web data.
2021-11-23 16:54:01 -08:00
Nathan Raw
b18c9e323b
Update helpers.py
2021-11-22 23:43:44 -05:00
Nathan Raw
308d0b9554
Merge branch 'master' into hf-save-and-push
2021-11-22 23:39:27 -05:00
talrid
41559247e9
use_ml_decoder_head
2021-11-22 17:50:39 +02:00
Ross Wightman
1f53db2ece
Updated lamhalobotnet weights, 81.5 top-1
2021-11-21 19:49:51 -08:00
Ross Wightman
15ef108eb4
Add better halo2botnet50ts weights, 82 top-1 @ 256
2021-11-21 14:09:12 -08:00
Ross Wightman
734b2244fe
Add RegNetZ-D8 (83.5 @ 256, 84 @ 320) and RegNetZ-E8 (84.5 @ 256, 85 @ 320) weights. Update names of existing RegZ models to include group size.
2021-11-20 15:52:04 -08:00
Ross Wightman
93cc08fdc5
Make evonorm variables 1d to match other PyTorch norm layers, will break weight compat for any existing use (likely minimal, easy to fix).
2021-11-20 15:50:51 -08:00
Ross Wightman
af607b75cc
Prep a set of ResNetV2 models with GroupNorm, EvoNormB0, EvoNormS0 for BN free model experiments on TPU and IPU
2021-11-19 17:37:00 -08:00
Ross Wightman
c976a410d9
Add ResNet-50 w/ GN (resnet50_gn) and SEBotNet-33-TS (sebotnet33ts_256) model defs and weights. Update halonet50ts weights w/ slightly better variant in1k val, more robust to test sets.
2021-11-19 14:24:43 -08:00
Ross Wightman
f2006b2437
Cleanup qkv_bias cat in beit model so it can be traced
2021-11-18 21:25:00 -08:00
Ross Wightman
1076a65df1
Minor post FX merge cleanup
2021-11-18 19:47:07 -08:00
Ross Wightman
32c9937dec
Merge branch 'fx-feature-extract-new' of https://github.com/alexander-soare/pytorch-image-models into alexander-soare-fx-feature-extract-new
2021-11-18 16:31:29 -08:00
Alexander Soare
65d827c7a6
rename notrace registration and standardize trace_utils imports
2021-11-15 21:03:21 +00:00
Ross Wightman
9b2daf2a35
Add ResNeXt-50 weights 81.1 top-1 @ 224, 82 @ 288 with A1 'high aug' recipe
2021-11-14 13:17:27 -08:00
Martins Bruveris
5220711d87
Added B/8 models to ViT.
2021-11-14 11:01:48 +00:00
Alexander Soare
0262a0e8e1
fx ready for review
2021-11-13 00:06:33 +00:00
Alexander Soare
d2994016e9
Add try/except guards
2021-11-12 21:16:53 +00:00
Alexander Soare
b25ff96768
wip - pre-rebase
2021-11-12 20:45:05 +00:00
Alexander Soare
e051dce354
Make all models FX traceable
2021-11-12 20:45:05 +00:00
Alexander Soare
cf4561ca72
Add FX based FeatureGraphNet capability
2021-11-12 20:45:05 +00:00
Alexander Soare
0149ec30d7
wip - attempting to rebase
2021-11-12 20:45:05 +00:00
Alexander Soare
02c3a75a45
wip - make it possible to use fx graph in train and eval mode
2021-11-12 20:45:05 +00:00
Alexander Soare
bc3d4eb403
wip -rebase
2021-11-12 20:45:05 +00:00
Alexander Soare
ab3ac3f25b
Add FX based FeatureGraphNet capability
2021-11-12 20:45:05 +00:00
Ross Wightman
ddc29da974
Add ResNet101 and ResNet152 weights from higher aug RSB recipes. 81.93 and 82.82 top-1 at 224x224.
2021-11-02 17:59:16 -07:00
Ross Wightman
b328e56f49
Update eca_halonext26ts weights to a better set
2021-11-02 16:52:53 -07:00
Ross Wightman
2ddef942b9
Better fix for #954 that doesn't break torchscript, pull torch._assert into timm namespace when it exists
2021-11-02 11:22:33 -07:00
Ross Wightman
4f0f9cb348
Fix #954 by bringing traceable _assert into timm to allow compat w/ PyTorch < 1.8
2021-11-02 09:21:40 -07:00
Ross Wightman
ae72d009fa
Add weights for lambda_resnet50ts, halo2botnet50ts, lamhalobotnet50ts, updated halonet50ts
2021-10-27 22:08:54 -07:00
Ross Wightman
b745d30a3e
Fix formatting of last commit
2021-10-25 15:15:14 -07:00
Ross Wightman
3478f1d7f1
Traceability fix for vit models for some experiments
2021-10-25 15:13:08 -07:00
Ross Wightman
f658a72e72
Cleanup re-use of Dropout modules in Mlp modules after some twitter feedback :p
2021-10-25 00:40:59 -07:00
Thomas Viehmann
f805ba86d9
use .unbind instead of explicitly listing the indices
2021-10-24 21:08:47 +02:00
Ross Wightman
0fe4fd3f1f
add d8 and e8 regnetz models with group size 8
2021-10-23 20:34:21 -07:00
Ross Wightman
25e7c8c5e5
Update broken resnetv2_50 weight url, add resnetv1_101 a1h recipe weights for 224x224 train
2021-10-20 22:14:12 -07:00
Ross Wightman
b6caa356d2
Fixed eca_botnext26ts_256 weights added, 79.27
2021-10-19 12:44:28 -07:00
Ross Wightman
c02334d9fa
Add weights for regnetz_d and haloregnetz_c, update regnetz_c weights. Add commented PyTorch XLA code for halo attention
2021-10-19 12:32:09 -07:00
Ross Wightman
02daf2ab94
Add option to include relative pos embedding in the attention scaling as per references. See discussion #912
2021-10-12 15:37:01 -07:00
Ross Wightman
cd34913278
Remove some outdated comments, botnet networks working great now.
2021-10-11 22:43:41 -07:00
Ross Wightman
6ed4cdccca
Update lambda_resnet26t weights with better set
2021-10-10 16:32:54 -07:00
ICLR Author
44d6d51668
Add ConvMixer
2021-10-09 21:09:51 -04:00
Ross Wightman
a85df34993
Update lambda_resnet26rpt weights to 78.9, add better halonet26t weights at 79.1 with tweak to attention dim
2021-10-08 17:44:13 -07:00
Ross Wightman
b544ad4d3f
regnetz model default cfg tweaks
2021-10-06 21:14:59 -07:00
Ross Wightman
e2b8d44ff0
Halo, bottleneck attn, lambda layer additions and cleanup along w/ experimental model defs
...
* align interfaces of halo, bottleneck attn and lambda layer
* add qk_ratio to all of above, control q/k dim relative to output dim
* add experimental haloregnetz, and trionet (lambda + halo + bottle) models
2021-10-06 16:32:48 -07:00
Ross Wightman
fbf59c04ee
Change crop ratio on correct resnet50 variant.
2021-10-04 22:31:08 -07:00
Ross Wightman
ae1ff5792f
Clean a1/a2/3 rsb _0 checkpoints properly, fix v2 loading.
2021-10-04 16:46:00 -07:00
Ross Wightman
da0d39bedd
Update default crop_pct for byoanet
2021-10-03 17:33:16 -07:00
Ross Wightman
cc9bedf373
Add initial ResNet Strikes Back weights for ResNet50 and ResNetV2-50 models
2021-10-03 17:32:02 -07:00
Ross Wightman
64495505b7
Add updated lambda resnet26 and botnet26 checkpoints with fixes applied
2021-10-03 17:31:39 -07:00
Ross Wightman
b2094f4ee8
support bits checkpoints in avg/load
2021-10-03 17:31:22 -07:00
Ross Wightman
007bc39323
Some halo and bottleneck attn code cleanup, add halonet50ts weights, use optimal crop ratios
2021-10-02 15:51:42 -07:00
Ross Wightman
b1c2e3eb92
Match rel_pos_indices attr rename in conv branch
2021-09-30 23:19:05 -07:00
Ross Wightman
b49630a138
Add relative pos embed option to LambdaLayer, fix last transpose/reshape.
2021-09-30 22:45:09 -07:00
Ross Wightman
d657e2cc0b
Remove dead code line from efficientnet
2021-09-30 21:54:42 -07:00
Ross Wightman
0ca687f224
Make 'regnetz' model experiments closer to actual RegNetZ, bottleneck expansion, expand from in_chs, no shortcut on stride 2, tweak model sizes
2021-09-30 21:49:38 -07:00
Ross Wightman
b81e79aae9
Fix bottleneck attn transpose typo, hopefully these train better now..
2021-09-28 16:38:41 -07:00
Ross Wightman
6478bcd02c
Fix regnetz_d conv layer name, use inception mean/std
2021-09-26 14:54:17 -07:00
Ross Wightman
515121cca1
Use reshape instead of view in std_conv, causing issues in recent PyTorch in channels_last
2021-09-23 15:43:48 -07:00
Ross Wightman
da06cc61d4
ResNetV2 seems to work best without zero_init residual
2021-09-23 15:43:22 -07:00
Ross Wightman
8e11da0ce3
Add experimental RegNetZ(ish) models for training / perf trials.
2021-09-23 15:42:57 -07:00
Alexander Soare
6bbc50beb4
make it possible to provide norm_layer via create_model
2021-09-21 10:19:04 +01:00
nateraw
adcb74f87f
🎨 Import load_state_dict_from_url directly
2021-09-14 01:11:40 -04:00
nateraw
e65a2cba3d
🎨 cleanup and add a couple comments
2021-09-14 01:07:04 -04:00
nateraw
2b6ade24b3
🎨 write model card to enable inference
2021-09-13 23:31:28 -04:00
Ross Wightman
cf5ac2800c
BotNet models were still off, remove weights for bad configs. Add good SE-HaloNet33-TS weights.
2021-09-13 17:18:59 -07:00
Ross Wightman
24720abe3b
Merge branch 'master' into attn_update
2021-09-13 16:51:10 -07:00
Ross Wightman
1c9284c640
Add BeiT 'finetuned' 1k weights and pretrained 22k weights, pretraining specific (masked) model excluded for now
2021-09-13 16:38:23 -07:00
Ross Wightman
f8a215cfe6
A few more crossvit tweaks, fix training w/ no_weight_decay names, add crop option for scaling, adjust default crop_pct for large img size to 1.0 for better results
2021-09-13 14:17:34 -07:00
Ross Wightman
7ab2491ab7
Better handling of crossvit for tests / forward_features, fix torchscript regression in my changes
2021-09-13 13:01:05 -07:00
Ross Wightman
f1808e0970
Post crossvit merge cleanup, change model names to reflect input size, cleanup img size vs scale handling, fix tests
2021-09-13 11:49:54 -07:00
Ross Wightman
4027412757
Add resnet33ts weights, update resnext26ts baseline weights
2021-09-09 14:46:41 -07:00
Richard Chen
9fe5798bee
fix bug for reset classifier and fix for validating the dimension
2021-09-08 21:58:17 -04:00
Richard Chen
3718c5a5bd
fix loading pretrained model
2021-09-08 11:53:05 -04:00
Richard Chen
bb50b69a57
fix for torch script
2021-09-08 11:20:59 -04:00
nateraw
abf9d51bc3
🚧 wip
2021-09-07 18:39:26 -06:00
Ross Wightman
5bd04714e4
Cleanup weight init for byob/byoanet and related
2021-09-05 15:34:05 -07:00
Ross Wightman
8642401e88
Swap botnet 26/50 weights/models after realizing a mistake in arch def, now figuring out why they were so low...
2021-09-05 15:17:19 -07:00
Ross Wightman
5f12de4875
Add initial AttentionPool2d that's being trialed. Fix comment and still trying to improve reliability of sgd test.
2021-09-05 12:41:14 -07:00
Ross Wightman
76881d207b
Add baseline resnet26t @ 256x256 weights. Add 33ts variant of halonet with at least one halo in stage 2,3,4
2021-09-04 14:52:54 -07:00
Ross Wightman
484e61648d
Adding the attn series weights, tweaking model names, comments...
2021-09-03 18:09:42 -07:00
nateraw
28d2841acf
💄 apply isort
2021-09-01 18:15:08 -06:00
Ross Wightman
492c0a4e20
Update HaloAttn comment
2021-09-01 17:14:31 -07:00
nateraw
e72c989973
✨ add ability to push to hf hub
2021-09-01 18:14:28 -06:00
Richard Chen
7ab9d4555c
add crossvit
2021-09-01 17:13:12 -04:00
Ross Wightman
3b9032ea48
Use Tensor.unfold().unfold() for HaloAttn, fast like as_strided but more clarity
2021-08-27 12:45:53 -07:00
Ross Wightman
78933122c9
Fix silly typo
2021-08-27 09:22:20 -07:00
Ross Wightman
2568ffc5ef
Merge branch 'master' into attn_update
2021-08-27 09:21:22 -07:00
Ross Wightman
708d87a813
Fix ViT SAM weight compat as weights at URL changed to not use repr layer. Fix #825 . Tweak optim test.
2021-08-27 09:20:13 -07:00
Ross Wightman
8449ba210c
Improve performance of HaloAttn, change default dim calc. Some cleanup / fixes for byoanet. Rename resnet26ts to tfs to distinguish (extra fc).
2021-08-26 21:56:44 -07:00
Ross Wightman
a8b65695f1
Add resnet26ts and resnext26ts models for non-attn baselines
2021-08-21 12:42:10 -07:00
Ross Wightman
a5a542f17d
Fix typo
2021-08-20 17:47:23 -07:00
Ross Wightman
925e102982
Update attention / self-attn based models from a series of experiments:
...
* remove dud attention, involution + my swin attention adaptation don't seem worth keeping
* add or update several new 26/50 layer ResNe(X)t variants that were used in experiments
* remove models associated with dead-end or uninteresting experiment results
* weights coming soon...
2021-08-20 16:13:11 -07:00
Ross Wightman
01cb46a9a5
Add gc_efficientnetv2_rw_t weights (global context instead of SE attn). Add TF XL weights even though the fine-tuned ones don't validate that well. Change default arg for GlobalContext to use scal (mul) mode.
2021-08-07 16:45:29 -07:00
Ross Wightman
d3f7440650
Add EfficientNetV2 XL model defs
2021-07-22 13:15:24 -07:00
Ross Wightman
72b227dcf5
Merge pull request #750 from drjinying/master
...
Specify "interpolation" mode in vision_transformer's resize_pos_embed
2021-07-13 11:01:20 -07:00
Ross Wightman
748ab852ca
Allow act_layer switch for xcit, fix in_chans for some variants
2021-07-12 13:27:29 -07:00
Ying Jin
20b2d4b69d
Use bicubic interpolation in resize_pos_embed()
2021-07-12 10:38:31 -07:00
Ross Wightman
d3255adf8e
Merge branch 'xcit' of https://github.com/alexander-soare/pytorch-image-models into alexander-soare-xcit
2021-07-12 08:30:30 -07:00
Ross Wightman
f8039c7492
Fix gc effv2 model cfg name
2021-07-11 12:14:31 -07:00
Alexander Soare
3a55a30ed1
add notes from author
2021-07-11 14:25:58 +01:00
Alexander Soare
899cf84ccc
bug fix - missing _dist postfix for many of the 224_dist models
2021-07-11 12:41:51 +01:00
Alexander Soare
623e8b8eb8
wip xcit
2021-07-11 09:39:38 +01:00
Ross Wightman
392368e210
Add efficientnetv2_rw_t defs w/ weights, and gc variant, as well as gcresnet26ts for experiments. Version 0.4.13
2021-07-09 16:46:52 -07:00
Ross Wightman
6d8272e92c
Add SAM pretrained model defs/weights for ViT B16 and B32 models.
2021-07-08 11:51:12 -07:00
Ross Wightman
ee4d8fc69a
Remove unecessary line from nest post refactor
2021-07-05 21:22:46 -07:00
Ross Wightman
8165cacd82
Realized LayerNorm2d won't work in all cases as is, fixed.
2021-07-05 18:21:34 -07:00
Ross Wightman
81cd6863c8
Move aggregation (convpool) for nest into NestLevel, cleanup and enable features_only use. Finalize weight url.
2021-07-05 18:20:49 -07:00
Ross Wightman
6ae0ac6420
Merge branch 'nested_transformer' of https://github.com/alexander-soare/pytorch-image-models into alexander-soare-nested_transformer
2021-07-03 12:45:26 -07:00
Alexander Soare
7b8a0017f1
wip to review
2021-07-03 12:10:12 +01:00
Alexander Soare
b11d949a06
wip checkpoint with some feature extraction work
2021-07-03 11:45:19 +01:00
Alexander Soare
23bb72ce5e
nested_transformer wip
2021-07-02 20:12:29 +01:00
Ross Wightman
766b4d3262
Fix features for resnetv2_50t
2021-06-28 15:56:24 -07:00
Ross Wightman
e8045e712f
Fix BatchNorm for ResNetV2 non GN models, add more ResNetV2 model defs for future experimentation, fix zero_init of last residual for pre-act.
2021-06-28 10:52:45 -07:00
Ross Wightman
20a2be14c3
Add gMLP-S weights, 79.6 top-1
2021-06-23 10:40:30 -07:00
Ross Wightman
85f894e03d
Fix ViT in21k representation (pre_logits) layer handling across old and new npz checkpoints
2021-06-23 10:38:34 -07:00
Ross Wightman
b41cffaa93
Fix a few issues loading pretrained vit/bit npz weights w/ num_classes=0 __init__ arg. Missed a few other small classifier handling detail on Mlp, GhostNet, Levit. Should fix #713
2021-06-22 23:16:05 -07:00
Ross Wightman
9c9755a808
AugReg release
2021-06-20 17:46:06 -07:00
Ross Wightman
381b279785
Add hybrid model fwds back
2021-06-19 22:28:44 -07:00
Ross Wightman
26f04a8e3e
Fix a weight link
2021-06-19 16:39:36 -07:00
Ross Wightman
8f4a0222ed
Add GMixer-24 MLP model weights, trained w/ TPU + PyTorch XLA
2021-06-18 16:49:28 -07:00
Ross Wightman
b319eb5b5d
Update ViT weights, more details to be added before merge.
2021-06-18 16:16:49 -07:00
Ross Wightman
8257b86550
Fix up resnetv2 bit/bitm model default res
2021-06-18 16:16:06 -07:00
Ross Wightman
1228f5a3d8
Add BiT distilled 50x1 and teacher 152x2 models from 'A good teacher is patient and consistent' paper.
2021-06-18 11:40:33 -07:00
Ross Wightman
511a8e8c96
Add official ResMLP weights.
2021-06-14 17:03:16 -07:00
Ross Wightman
b9cfb64412
Support npz custom load for vision transformer hybrid models. Add posembed rescale for npz load.
2021-06-14 12:31:44 -07:00
Ross Wightman
8319e0c373
Add file docstring to std_conv.py
2021-06-13 12:31:06 -07:00
Ross Wightman
4d96165989
Merge branch 'master' into cleanup_xla_model_fixes
2021-06-12 23:19:25 -07:00
Ross Wightman
8880f696b6
Refactoring, cleanup, improved test coverage.
...
* Add eca_nfnet_l2 weights, 84.7 @ 384x384
* All 'non-std' (ie transformer / mlp) models have classifier / default_cfg test added
* Fix #694 reset_classifer / num_features / forward_features / num_classes=0 consistency for transformer / mlp models
* Add direct loading of npz to vision transformer (pure transformer so far, hybrid to come)
* Rename vit_deit* to deit_*
* Remove some deprecated vit hybrid model defs
* Clean up classifier flatten for conv classifiers and unusual cases (mobilenetv3/ghostnet)
* Remove explicit model fns for levit conv, just pass in arg
2021-06-12 16:40:02 -07:00
Ross Wightman
ba2ca4b464
One codepath for stdconv, switch layernorm to batchnorm so gain included. Tweak epsilon values for nfnet, resnetv2, vit hybrid.
2021-06-12 12:27:43 -07:00
Ross Wightman
b7a568f065
Fix torchscript issue in bat
2021-06-08 23:19:51 -07:00
Ross Wightman
d17b374f0f
Minimum input_size needed to be higher
2021-06-08 21:31:39 -07:00
Ross Wightman
b3b90d944d
Add min_input_size to bat_resnext to prevent test breakage.
2021-06-08 17:32:08 -07:00
Ross Wightman
d413eef1bf
Add ResMLP-24 model weights that I trained in PyTorch XLA on TPU-VM. 79.2 top-1.
2021-06-08 14:22:05 -07:00
Ross Wightman
10d8fa4620
Add gc and bat attention resnext26ts variants to byob for test.
2021-06-08 14:21:07 -07:00
Ross Wightman
2f5ed2dec1
Update `init_values` const for 24 and 36 layer ResMLP models
2021-06-07 17:15:04 -07:00
Ross Wightman
8e4ac3549f
All ScaledStdConv and StdConv uses default to using F.layernorm so that they work with PyTorch XLA. eps value tweaking is a WIP.
2021-06-07 17:14:19 -07:00
Ross Wightman
2a63d0246b
Post merge cleanup
2021-06-07 14:38:30 -07:00
Ross Wightman
45dec179e5
Merge pull request #681 from lmk123568/master
...
Update convit.py
2021-06-07 14:10:53 -07:00
Dongyoon Han
ded1671483
Fix stochastic depth working only with a shortcut
2021-06-07 23:08:55 +09:00
Mike
b87d98b238
Update convit.py
...
Cut out the duplicates
2021-06-06 17:58:31 +08:00
Ross Wightman
bda8ab015a
Remove min channels for SelectiveKernel, divisor should cover cases well enough.
2021-05-31 15:38:56 -07:00
Ross Wightman
a27f4aec4a
Missed args for skresnext w/ refactoring.
2021-05-31 14:06:34 -07:00
Ross Wightman
307a935b79
Add non-local and BAT attention. Merge attn and self-attn factories into one. Add attention references to README. Add mlp 'mode' to ECA.
2021-05-31 13:18:11 -07:00
Ross Wightman
8bf63b6c6c
Able to use other attn layer in EfficientNet now. Create test ECA + GC B0 configs. Make ECA more configurable.
2021-05-30 12:47:02 -07:00
Ross Wightman
bcec14d3b5
Bring EfficientNet SE layer in line with others, pull se_ratio outside of blocks. Allows swapping w/ other attn layers.
2021-05-29 23:41:38 -07:00
Ross Wightman
9611458e19
Throw in some FBNetV3 code I had lying around, some refactoring of SE reduction channel calcs for all EffNet archs.
2021-05-28 20:47:24 -07:00
Ross Wightman
01b9108619
Merge branch 'master' into more_attn
2021-05-28 11:09:37 -07:00
Ross Wightman
d7bab8a6c5
Fix strict flag change for checkpoint load.
2021-05-28 09:54:50 -07:00
Ross Wightman
02f9d4bc34
Add weights for resnet51q model, add 61q def.
2021-05-28 09:53:16 -07:00
Ross Wightman
f615474be3
Fix broken test, repvgg block doesn't have attn_last attr.
2021-05-27 18:12:22 -07:00
Ross Wightman
742c2d5247
Add Gather-Excite and Global Context attn modules. Refactor existing SE-like attn for consistency and refactor byob/byoanet for less redundancy.
2021-05-27 18:03:29 -07:00
Ross Wightman
9c78de8c02
Fix #661 , move hardswish out of default args for LeViT. Enable native torch support for hardswish, hardsigmoid, mish if present.
2021-05-26 15:28:42 -07:00
Ross Wightman
5db7452173
Fix visformer in_chans stem handling
2021-05-25 14:11:36 -07:00
Ross Wightman
11ae795e99
Redo LeViT attention bias caching in a way that works with both torchscript and DataParallel
2021-05-25 10:15:32 -07:00
Ross Wightman
d400f1dbdd
Filter test models before creation for backward/torchscript tests
2021-05-25 10:14:45 -07:00
Ross Wightman
c4572cc5aa
Add Visformer-small weighs, tweak torchscript jit test img size.
2021-05-24 22:50:12 -07:00
Ross Wightman
bfc72f75d3
Expand scope of testing for non-std vision transformer / mlp models. Some related cleanup and create fn cleanup for all vision transformer and mlp models. More CoaT weights.
2021-05-24 21:13:26 -07:00
Ross Wightman
18bf520ad1
Add eca_nfnet_l2/l3 defs for future training
2021-05-22 21:55:37 -07:00
Ross Wightman
f45de37690
Merge branch 'master' into levit_visformer_rednet
2021-05-22 16:34:31 -07:00
Ross Wightman
23c18a33e4
Add efficientnetv2_rw_m weights trained in PyTorch. 84.8 top-1 @ 416 test. 53M params.
2021-05-21 21:16:25 -07:00
Ross Wightman
c2ba229d99
Prep for effcientnetv2_rw_m model weights that started training before official release..
2021-05-21 17:47:49 -07:00
Ross Wightman
30b9880d06
Minor adjustment, mutable default arg, extra check of valid len...
2021-05-21 17:20:51 -07:00
Ross Wightman
be0abfbcce
Merge branch 'master' of https://github.com/alexander-soare/pytorch-image-models into alexander-soare-master
2021-05-21 17:10:11 -07:00
Ross Wightman
b7de82e835
ConViT cleanup, fix torchscript, bit of reformatting, reuse existing layers.
2021-05-21 17:04:23 -07:00
Ross Wightman
306c86b668
Merge branch 'convit' of https://github.com/amaarora/pytorch-image-models into amaarora-convit
2021-05-21 16:27:10 -07:00
Ross Wightman
a569635045
Update twin weights to a copy in GitHub releases for faster dl. Tweak model class comment.
2021-05-21 16:23:14 -07:00
Ross Wightman
be99eef9c1
Remove redundant code, cleanup, fix torchscript.
2021-05-20 23:38:35 -07:00
Ross Wightman
5ab372a3ec
Merge branch 'master' of https://github.com/abcdvzz/pytorch-image-models into abcdvzz-master
2021-05-20 23:37:50 -07:00
Aman Arora
5db1eb6ba5
Add defaults
2021-05-21 02:11:20 +00:00
Aman Arora
8b1f2e8e1f
remote unused matplotlib import
2021-05-20 23:42:42 +00:00
Aman Arora
40c506ba1e
Add ConViT
2021-05-20 23:17:28 +00:00
Alexander Soare
7976019864
extend positional embedding resizing functionality to tnt
2021-05-20 11:55:48 +01:00
Alexander Soare
8086943b6f
allow resize positional embeddings to non-square grid
2021-05-20 11:27:58 +01:00
talrid
dc1a4efd28
mixer_b16_224_miil, mixer_b16_224_miil_in21k models
2021-05-20 10:35:50 +03:00
李鑫杰
7b799c4e79
add latest code
2021-05-20 11:15:49 +08:00
Ross Wightman
d5af752117
Add preliminary gMLP and ResMLP impl to Mlp-Mixer
2021-05-19 09:55:05 -07:00
李鑫杰
00548b8427
Add Twins
2021-05-18 19:21:53 +08:00
Ross Wightman
e7f0db8664
Fix drop/drop_path arg on MLP-Mixer model. Fix #641
2021-05-16 08:31:52 -07:00
Ross Wightman
9a3ae97311
Another set of byoanet models w/ ECA channel + SA + groups
2021-05-15 22:56:51 -07:00
Ross Wightman
7077f16c6a
Change 21k model naming from _21k to _in21k for consistency with existing 21k models.
2021-05-15 12:42:26 -07:00
Ross Wightman
94d4b53352
Add temporary default_cfgs to visformer models so they pass tests
2021-05-15 08:41:31 -07:00
Ross Wightman
3bffc701f1
Merge branch 'master' into levit_visformer_rednet
2021-05-14 23:02:12 -07:00
Ross Wightman
ecc7552c5c
Add levit, levit_c, and visformer model defs. Largely untested and not finished cleanup.
2021-05-14 17:16:34 -07:00
Ross Wightman
165fb354b2
Add initial RedNet model / Involution layer impl for testing
2021-05-14 17:16:34 -07:00
Ross Wightman
328249f11a
Update README, tweak fine-tune effv2 model names.
2021-05-14 16:37:43 -07:00
Ross Wightman
c4f482a08b
EfficientNetV2 official impl w/ weights ported from TF. Cleanup/refactor of related EfficientNet classes and models.
2021-05-14 15:50:00 -07:00
Ross Wightman
4fbc32d3d0
Fix crop_pct for cait models.
2021-05-11 10:50:15 -07:00
Ross Wightman
715519a5ef
Rethink name of patch embed grid info
2021-05-06 14:08:20 -07:00
Ross Wightman
b2c305c2aa
Move Mlp and PatchEmbed modules into layers. Being used in lots of models now...
2021-05-06 14:03:23 -07:00
Ross Wightman
3ba6b55cb2
More adjustments to ByoaNet models for further experiments.
2021-05-06 13:59:44 -07:00
Ross Wightman
5fcddb96a8
Merge branch 'master' into cait
2021-05-05 17:29:38 -07:00
Ross Wightman
3db12b4b6a
Finish CaiT cleanup
2021-05-05 17:28:19 -07:00
Ross Wightman
2d8b09fe8b
Add official pretrained weights to MLP-Mixer, complete model cfgs.
2021-05-05 15:59:40 -07:00
Ross Wightman
12efffa6b1
Initial MLP-Mixer attempt...
2021-05-05 00:59:45 -07:00
Ross Wightman
0721559511
Improved (hopefully) init for SA/SA-like layers used in ByoaNets
2021-05-04 21:40:39 -07:00
Ross Wightman
d5473c17f7
Fix incorrect name of shortcut/identity paths in many residual nets. Inherited from naming in old old torchvision, long fixed there.
2021-05-04 21:27:15 -07:00
Ross Wightman
0d87650fea
Remove filter hack from BlurPool w/ non-persistent buffer. Use BlurPool2d instead of AntiAliasing.. for TResNet. Breaks PyTorch < 1.6.
2021-05-04 16:56:28 -07:00
Ross Wightman
ddc743fdf8
Update ResNet-RS models to EMA weights
2021-05-04 16:16:55 -07:00
Ross Wightman
08d60f4a9a
resnetrs50 pool sizing wrong
2021-05-04 12:41:09 -07:00
Ross Wightman
1daa15ecc3
Initial Cait commit. Still some cleanup to do.
2021-05-04 11:19:27 -07:00
Ross Wightman
67d0665b46
Post ResNet-RS merge cleanup. Add weight urls, adjust train/test/crop pct.
2021-05-04 11:04:23 -07:00
Aman Arora
560eae38f5
[WIP] Add ResNet-RS models ( #554 )
...
* Add ResNet-RS models
* Only include resnet-rs changes
* remove whitespace diff
* EOF newline
* Update time
* increase time
* Add first conv
* Try running only resnetv2_101x1_bitm on Linux runner
* Add to exclude filter
* Run test_model_forward_features for all
* Add to exclude ftrs
* back to defaults
* only run test_forward_features
* run all tests
* Run all tests
* Add bigger resnetrs to model filters to fix Github CLI
* Remove resnetv2_101x1_bitm from exclude feat features
* Remove hardcoded values
* Make sure reduction ratio in resnetrs is 0.25
* There is no bias in replaced maxpool so remove it
2021-05-04 10:59:44 -07:00
Ross Wightman
9cc7dda6e5
Fixup byoanet configs to pass unit tests. Add swin_attn and swinnet26t model for testing.
2021-04-29 21:08:37 -07:00
Ross Wightman
e15c3886ba
Defaul lambda r=7. Define '26t' stage 4/5 256x256 variants for all of bot/halo/lambda nets for experiment. Add resnet50t for exp. Fix a few comments.
2021-04-29 10:58:49 -07:00
Ross Wightman
e5e15754c9
Fix coat first conv ident
2021-04-28 18:09:23 -07:00
Ross Wightman
76739a7589
CoaT merge. Bit of formatting, fix torchscript (for non features), remove einops/einsum dep, add pretrained weight hub (url) support.
2021-04-28 16:31:35 -07:00
Ross Wightman
026430c083
Merge branch 'master' of https://github.com/morizin/pytorch-image-models-1 into morizin-master
2021-04-28 11:24:27 -07:00
Ross Wightman
a0492e3b48
A few miil weights naming tweaks to improve compat with model registry and filtering wildcards.
2021-04-28 10:33:02 -07:00
talrid
8c1f03e56c
comment
2021-04-28 09:04:10 +03:00
talrid
19e1b67a84
old spaces
2021-04-28 09:01:29 +03:00
talrid
a443865876
update naming and scores
2021-04-28 08:57:06 +03:00
talrid
cf0e371594
84_0
2021-04-27 22:33:55 +03:00
talrid
0968bdeca3
vit, tresnet and mobilenetV3 ImageNet-21K-P weights
2021-04-27 20:54:06 +03:00
morizin
1e3b6d4dfc
Update __init__.py
2021-04-24 18:22:59 +05:30
morizin
fd022fd6a2
Update __init__.py
2021-04-24 18:22:36 +05:30
morizin
c2d5087eae
Add files via upload
2021-04-24 17:47:57 +05:30
Ross Wightman
c1cf9712fc
Add updated EfficientNet-V2S weights, 83.8 @ 384x384 test. Add PyTorch trained EfficientNet-B4 weights, 83.4 @ 384x384 test. Tweak non TF EfficientNet B1-B4 train/test res scaling.
2021-04-19 10:42:56 -07:00
Ross Wightman
e8a64fb881
Test input size for efficientnet_v2s was wrong in last results run
2021-04-17 16:17:41 -07:00
Ross Wightman
2df77ee5cb
Fix torchscript compat and features_only behaviour in GhostNet PR. A few minor formatting changes. Reuse existing layers.
2021-04-15 10:20:26 -07:00
Ross Wightman
d793deb51a
Merge branch 'master' of https://github.com/iamhankai/pytorch-image-models into iamhankai-master
2021-04-15 09:30:25 -07:00
Ross Wightman
f606c45c38
Add Swin Transformer models from https://github.com/microsoft/Swin-Transformer
2021-04-13 12:17:21 -07:00
iamhankai
de445e7827
Add GhostNet
2021-04-13 23:19:51 +08:00
Ross Wightman
b3d7580df1
Update ByoaNet comments. Fix first Steam feat chs for ByobNet.
2021-04-12 12:11:35 -07:00
Ross Wightman
16f7aa9f54
Add default_cfg options for min_input_size / fixed_input_size, queries in model registry, and use for testing self-attn models
2021-04-12 11:54:22 -07:00
Ross Wightman
4e4b863b15
Missed norm.py
2021-04-12 09:57:56 -07:00
Ross Wightman
7c97e66f7c
Remove commented code, add more consistent seed fn
2021-04-12 09:51:36 -07:00
Ross Wightman
ce62f96d4d
ByoaNet with bottleneck transformer, lambda resnet, and halo net experiments
2021-04-12 09:38:02 -07:00
Ross Wightman
21812d33aa
Add prelim efficientnet_v2s weights from 224x224 train, eval 83.3 @ 288. Add eca_nfnet_l1 weights, train at 256, eval 84 @ 320.
2021-04-11 23:26:13 -07:00
Ross Wightman
fb896c0b26
Update some comments re preliminary EfficientNet-V2 assumptions
2021-04-03 12:00:25 -07:00
Ross Wightman
2b49ab7a36
Fix ResNetV2 pretrained classifier issue. Fixes #540
2021-04-03 11:18:12 -07:00
Ross Wightman
de9dff933a
EfficientNet-V2S preliminary model def (for experimentation)
2021-04-02 09:36:51 -07:00
Ross Wightman
2bb65bd875
Wrong default_cfg pool_size for L1
2021-04-01 20:00:41 -07:00
Ross Wightman
bf2ca6bdf4
Merge jax and original weight init
2021-04-01 18:11:51 -07:00
Ross Wightman
acbd698c83
Update README.md with updates. Small tweak to head_dist handling.
2021-04-01 17:49:05 -07:00
Ross Wightman
9071568f0e
Add weights for SE NFNet-L0 model, rename nfnet_l0b -> nfnet_l0. 82.75 top-1 @ 288. Add nfnet_l1 model def for training.
2021-04-01 17:22:27 -07:00
Ross Wightman
c468c47a9c
Add regnety_160 weights from DeiT teacher model, update that and my regnety_032 weights to use higher test size.
2021-04-01 16:41:04 -07:00
Ross Wightman
288682796f
Update benchmark script to add precision arg. Fix some downstream (DeiT) compat issues with latest changes. Bump version to 0.4.7
2021-04-01 16:40:12 -07:00
Ross Wightman
ea9c9550b2
Fully move ViT hybrids to their own file, including embedding module. Remove some extra DeiT models that were for benchmarking only.
2021-04-01 14:17:38 -07:00
Ross Wightman
a5310a3451
Merge remote-tracking branch 'origin/benchmark-fixes-vit_hybrids' into pit_and_vit_update
2021-04-01 12:15:34 -07:00
Ross Wightman
7953e5d11a
Fix pos_embed scaling for ViT and num_classes != 1000 for pretrained distilled deit and pit models. Fix #426 and fix #433
2021-03-31 23:11:28 -07:00
Ross Wightman
a760a4c3f4
Some ViT cleanup, merge distilled model with main, fixup torchscript support for distilled models
2021-03-31 18:21:02 -07:00
Ross Wightman
0dfc5a66bb
Add PiT model from https://github.com/naver-ai/pit
2021-03-31 18:20:14 -07:00
Ross Wightman
51febd869b
Small tweak to tests for tnt model, reorder model imports.
2021-03-29 11:33:08 -07:00
contrastive
de86314655
Update TNT
2021-03-29 08:23:34 +08:00
contrastive
cfc15283a4
Update TNT url
2021-03-28 23:19:15 +08:00
contrastive
4a09bc851e
Add TNT model
2021-03-28 19:53:42 +08:00
Ross Wightman
cf5fec5047
Cleanup experimental vit weight init a bit
2021-03-20 09:44:24 -07:00
Ross Wightman
cbcb76d72c
Should have included Conv2d layers in original weight init. Lets see what the impact is...
2021-03-18 23:15:48 -07:00
Ross Wightman
4de57ccf01
Add weight init scheme that's closer to JAX impl
2021-03-18 15:35:22 -07:00
Ross Wightman
14ac4abf74
Change huggingface hub revision delimiter to '@', add hf_hub reference for eca_nfnet_l0 model as an example.
2021-03-18 11:38:06 -07:00