Ross Wightman
5078b28f8a
More kwarg handling tweaks, maxvit_base_rw def added
2023-01-09 08:57:31 -08:00
Ross Wightman
c0d7388a1b
Improving kwarg merging in more models
2023-01-09 08:57:31 -08:00
Ross Wightman
ae9153052f
Update version.py
2023-01-06 17:17:35 -08:00
Ross Wightman
60ebb6cefa
Re-order vit pretrained entries for more sensible default weights (no .tag specified)
2023-01-06 16:12:33 -08:00
Ross Wightman
e861b74cf8
Pass through --model-kwargs (and --opt-kwargs for train) from command line through to model __init__. Update some models to improve arg overlay. Cleanup along the way.
2023-01-06 16:12:33 -08:00
Ross Wightman
add3fb864e
Working on improved model card template for push_to_hf_hub
2023-01-06 16:12:33 -08:00
Ross Wightman
dd0bb327e9
Update version.py
...
Ver 0.8.4dev0
2023-01-05 07:55:18 -08:00
Ross Wightman
6e5553da5f
Add ConvNeXt-V2 support (model additions and weights) ( #1614 )
...
* Add ConvNeXt-V2 support (model additions and weights)
* ConvNeXt-V2 weights on HF Hub, tweaking some tests
* Update README, fixing convnextv2 tests
2023-01-05 07:53:32 -08:00
Ross Wightman
6902c48a5f
Fix ResNet based models to work w/ norm layers w/o affine params. Reformat long arg lists into vertical form.
2022-12-29 16:32:26 -08:00
Ross Wightman
d5aa17e415
Remove print from auto_augment
2022-12-28 17:11:35 -08:00
Ross Wightman
7c846d9970
Better vmap compat across recent torch versions
2022-12-24 14:37:04 -08:00
Ross Wightman
4e24f75289
Merge pull request #1593 from rwightman/multi-weight_effnet_convnext
...
Update efficientnet.py and convnext.py to multi-weight, add new 12k pretrained weights
2022-12-23 10:09:08 -08:00
Ross Wightman
8ece53e194
Switch BEiT to HF hub weights
2022-12-22 21:43:04 -08:00
Ross Wightman
d1bfa9a000
Support HF datasets and TFSD w/ a sub-path by fixing split, fix #1598 ... add class mapping support to HF datasets in case class label isn't in info.
2022-12-22 21:34:13 -08:00
Ross Wightman
e2fc43bc63
Version 0.8.2dev0
2022-12-22 17:34:09 -08:00
Ross Wightman
9a51e4ea2e
Add FlexiViT models and weights, refactoring, push more weights
...
* push all vision_transformer*.py weights to HF hub
* finalize more pretrained tags for pushed weights
* refactor pos_embed files and module locations, move some pos embed modules to layers
* tweak hf hub helpers to aid bulk uploading and updating
2022-12-22 17:23:09 -08:00
Fredo Guan
10b3f696b4
Davit std ( #6 )
...
Separate patch_embed module
2022-12-16 21:50:28 -08:00
Ross Wightman
656e1776de
Convert mobilenetv3 to multi-weight, tweak PretrainedCfg metadata
2022-12-16 09:29:13 -08:00
Fredo Guan
546590c5f5
Merge branch 'rwightman:main' into main
2022-12-14 23:44:15 -08:00
Ross Wightman
6a01101905
Update efficientnet.py and convnext.py to multi-weight, add ImageNet-12k pretrained EfficientNet-B5 and ConvNeXt-Nano.
2022-12-14 20:33:23 -08:00
alec.tu
74d6afb4cd
Add Adan to __init__.py
2022-12-15 11:37:29 +08:00
Fredo Guan
84178fca60
Merge branch 'rwightman:main' into main
2022-12-12 23:13:58 -08:00
Fredo Guan
c43340ddd4
Davit std ( #5 )
...
* Update davit.py
* Update test_models.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* starting point
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update test_models.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Davit revised (#4 )
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
clean up
* Update test_models.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update davit.py
* Update test_models.py
* Update davit.py
2022-12-11 03:03:22 -08:00
Ross Wightman
e7da205345
Fix aa min_max level clamp
2022-12-10 16:43:28 -08:00
Ross Wightman
e3b2f5be0a
Add 3-Augment support to auto_augment.py, clean up weighted choice handling, and allow adjust per op prob via arg string
2022-12-10 16:25:50 -08:00
Ross Wightman
d5e7d6b27e
Merge remote-tracking branch 'origin/main' into refactor-imports
2022-12-09 14:49:44 -08:00
Ross Wightman
cda39b35bd
Add a deprecation phase to module re-org
2022-12-09 14:39:45 -08:00
Fredo Guan
edea013dd1
Davit std ( #3 )
...
Davit with all features working
2022-12-09 02:53:21 -08:00
Ross Wightman
7c4ed4d5a4
Add EVA-large models
2022-12-08 16:21:30 -08:00
Fredo Guan
434a03937d
Merge branch 'rwightman:main' into main
2022-12-08 08:05:16 -08:00
Ross Wightman
98047ef5e3
Add EVA FT results, hopefully fix BEiT test failures
2022-12-07 08:54:06 -08:00
Ross Wightman
3cc4d7a894
Fix missing register for 224 eva model
2022-12-07 08:54:06 -08:00
Ross Wightman
eba07b0de7
Add eva models to beit.py
2022-12-07 08:54:06 -08:00
Fredo Guan
3bd96609c8
Davit ( #1 )
...
Implement the davit model from https://arxiv.org/abs/2204.03645 and https://github.com/dingmyu/davit
2022-12-06 17:19:25 -08:00
Ross Wightman
927f031293
Major module / path restructure, timm.models.layers -> timm.layers, add _ prefix to all non model modules in timm.models
2022-12-06 15:00:06 -08:00
Ross Wightman
3785c234d7
Remove clip vit models that won't be ft and comment two that aren't uploaded yet
2022-12-05 10:21:34 -08:00
Ross Wightman
f82239b30e
multi-weight branch version -> 0.8.0dev
2022-12-05 10:21:34 -08:00
Ross Wightman
755570e2d6
Rename _pretrained.py -> pretrained.py, not feasible to change the other files to same scheme without breaking uses
2022-12-05 10:21:34 -08:00
Ross Wightman
72cfa57761
Add ported Tensorflow MaxVit weights. Add a few more CLIP ViT fine-tunes. Tweak some model tag names. Improve model tag name sorting. Update HF hub push config layout.
2022-12-05 10:21:34 -08:00
Ross Wightman
4d5c395160
MaxVit, ViT, ConvNeXt, and EfficientNet-v2 updates
...
* Add support for TF weights and modelling specifics to MaxVit (testing ported weights)
* More fine-tuned CLIP ViT configs
* ConvNeXt and MaxVit updated to new pretrained cfgs use
* EfficientNetV2, MaxVit and ConvNeXt high res models use squash crop/resize
2022-12-05 10:21:34 -08:00
Ross Wightman
3db4e346e0
Switch TFDS dataset to use INTEGER_ACCURATE jpeg decode by default
2022-12-05 10:21:34 -08:00
Ross Wightman
9da7e3a799
Add crop_mode for pretraind config / image transforms. Add support for dynamo compilation to benchmark/train/validate
2022-12-05 10:21:34 -08:00
Ross Wightman
b2b6285af7
Add two more FT clip weights
2022-12-05 10:21:34 -08:00
Ross Wightman
5895056dc4
Add openai b32 ft
2022-12-05 10:21:34 -08:00
Ross Wightman
9dea5143d5
Adding more clip ft variants
2022-12-05 10:21:34 -08:00
Ross Wightman
444dcba4ad
CLIP B16 12k weights added
2022-12-05 10:21:34 -08:00
Ross Wightman
dff4717cbf
Add clip b16 384x384 finetunes
2022-12-05 10:21:34 -08:00
Ross Wightman
883fa2eeaa
Add fine-tuned B/16 224x224 in1k clip models
2022-12-05 10:21:34 -08:00
Ross Wightman
9a3d2ac2d5
Add latest CLIP ViT fine-tune pretrained configs / model entrypt updates
2022-12-05 10:21:34 -08:00
Ross Wightman
42bbbddee9
Add missing model config
2022-12-05 10:21:34 -08:00
Ross Wightman
def68befa7
Updating vit model defs for mult-weight support trial (vit first). Prepping for CLIP (laion2b and openai) fine-tuned weights.
2022-12-05 10:21:34 -08:00
Ross Wightman
0dadb4a6e9
Initial multi-weight support, handled so old pretraing config handling co-exists with new tags.
2022-12-05 10:21:34 -08:00
hongxin xiang
653bdc7105
Fix comment: https://github.com/rwightman/pytorch-image-models/pull/1564#issuecomment-1326743424
2022-11-25 09:52:52 +08:00
hongxin xiang
bdc9fad638
Fix compatible BUG: QMNIST and ImageNet datasets do not exist in torchvision 0.10.1.
2022-11-24 14:37:44 +08:00
Wauplin
9b114754db
refactor push_to_hub helper
2022-11-16 12:03:34 +01:00
Wauplin
ae0a0db7de
Create repo before cloning with Repository.clone_from
2022-11-15 15:17:20 +01:00
Ross Wightman
803254bb40
Fix spacing misalignment for fast norm path in LayerNorm modules
2022-10-24 21:43:49 -07:00
Ross Wightman
475ecdfa3d
cast env var args for dataset readers to int
2022-10-17 14:40:11 -07:00
Hoan Nguyen
39190f5f44
Remove inplace operators when calculating the loss
...
Remove inplace operators to overcome the following error when using `asymmetric_loss`
```
RuntimeError: one of the variables needed for gradient computation has been modified by an inplace operation
```
2022-10-17 10:41:21 +02:00
Ross Wightman
6635bc3f7d
Merge pull request #1479 from rwightman/script_cleanup
...
Train / val script enhancements, non-GPU (ie CPU) device support, HF datasets support, TFDS/WDS dataloading improvements
2022-10-15 09:29:39 -07:00
Ross Wightman
0e6023f032
Merge pull request #1381 from ChristophReich1996/master
...
Fix typo in PositionalEncodingFourier
2022-10-14 18:34:33 -07:00
Ross Wightman
66f4af7090
Merge remote-tracking branch 'origin/master' into script_cleanup
2022-10-14 15:54:00 -07:00
Ross Wightman
d3961536c9
comment some debug logs for WDS dataset
2022-10-14 15:39:00 -07:00
Ross Wightman
e9dccc918c
Rename dataset/parsers -> dataset/readers, create_parser to create_reader, etc
2022-10-14 15:14:38 -07:00
Ross Wightman
8c28363dc9
Version 0.7.dev0 for master
2022-10-14 09:38:02 -07:00
nateraw
30bafd7347
🔖 add dev suffix to version tag
2022-10-13 17:08:33 -04:00
Ross Wightman
f67a7ee8bd
Set num_workers in Iterable WDS/TFDS datasets early so sample estimate is correct
2022-10-11 15:11:18 -07:00
Ross Wightman
cea8df3d0c
Version 0.6.12
2022-10-10 21:49:52 -07:00
Ross Wightman
9914f744dc
Add more maxxvit weights includ ConvNeXt conv block based experiments.
2022-10-10 21:49:18 -07:00
Ross Wightman
b1b024dfed
Scheduler update, add v2 factory method, support scheduling on updates instead of just epochs. Add LR to summary csv. Add lr_base scaling calculations to train script. Fix #1168
2022-10-07 10:43:04 -07:00
Ross Wightman
4f18d6dc5f
Fix logs in WDS parser
2022-10-07 10:06:17 -07:00
Mohamed Rashad
8fda68aff6
Fix repo id bug
...
This to fix this issue #1482
2022-10-05 16:26:06 +02:00
Ross Wightman
b8c8550841
Data improvements. Improve train support for in_chans != 3. Add wds dataset support from bits_and_tpu branch w/ fixes and tweaks. TFDS tweaks.
2022-09-29 16:42:58 -07:00
Alex Fafard
7327792f39
update to support pickle based dictionaries
2022-09-27 11:13:48 -04:00
Ross Wightman
1199c5a1a4
clip_laion2b models need 1e-5 eps for LayerNorm
2022-09-25 10:36:54 -07:00
Ross Wightman
87939e6fab
Refactor device handling in scripts, distributed init to be less 'cuda' centric. More device args passed through where needed.
2022-09-23 16:08:59 -07:00
Ross Wightman
c88947ad3d
Add initial Hugging Face Datasets parser impl.
2022-09-23 16:08:19 -07:00
Ross Wightman
e858912e0c
Add brute-force checkpoint remapping option
2022-09-23 16:07:03 -07:00
Ross Wightman
b293dfa595
Add CL SE module
2022-09-23 16:06:09 -07:00
Ross Wightman
2a296412be
Add Adan optimizer
2022-09-23 16:05:52 -07:00
Ross Wightman
5dc4343308
version 0.6.11
2022-09-23 13:54:56 -07:00
Ross Wightman
a383ef99f5
Make huggingface_hub necessary if it's the only source for a pretrained weight
2022-09-23 13:54:21 -07:00
Ross Wightman
33e30f8c8b
Remove layer-decay print
2022-09-18 21:33:03 -07:00
Ross Wightman
e069249a2d
Add hf hub entries for laion2b clip models, add huggingface_hub dependency, update some setup/reqs, torch >= 1.7
2022-09-16 21:39:05 -07:00
Ross Wightman
9d65557be3
Fix errant import
2022-09-15 17:47:23 -07:00
Ross Wightman
9709dbaaa9
Adding support for fine-tune CLIP LAION-2B image tower weights for B/32, L/14, H/14 and g/14. Still WIP
2022-09-15 17:25:59 -07:00
Ross Wightman
a520da9b49
Update tresnet features_info for v2
2022-09-13 20:54:54 -07:00
Ross Wightman
c8ab747bf4
BEiT-V2 checkpoints didn't remove 'module' from weights, adapt checkpoint filter
2022-09-13 17:56:49 -07:00
Ross Wightman
73049dc2aa
Fix type in dla weight update
2022-09-13 17:52:45 -07:00
Ross Wightman
3599c7e6a4
version 0.6.10
2022-09-13 16:37:02 -07:00
Ross Wightman
e11efa872d
Update a bunch of weights with external links to timm release assets. Fixes issue with *aliyuncs.com returning forbidden. Did pickle scan / verify and re-hash. Add TresNet-V2-L weights.
2022-09-13 16:35:26 -07:00
Ross Wightman
fa8c84eede
Update maxvit_tiny_256 weight to better iter, add coatnet / maxvit / maxxvit model defs for future runs
2022-09-07 12:37:37 -07:00
Ross Wightman
c1b3cea19d
Add maxvit_rmlp_tiny_rw_256 model def and weights w/ 84.2 top-1 @ 256, 84.8 @ 320
2022-09-07 10:27:11 -07:00
Ross Wightman
914544fc81
Add beitv2 224x224 checkpoints from https://github.com/microsoft/unilm/tree/master/beit2
2022-09-06 20:25:18 -07:00
Ross Wightman
dc90816f26
Add `maxvit_tiny_rw_224` weights 83.5 @ 224 and `maxvit_rmlp_pico_rw_256` relpos weights, 80.5 @ 256, 81.3 @ 320
2022-09-06 16:14:41 -07:00
Ross Wightman
f489f02ad1
Make gcvit window size ratio based to improve resolution changing support #1449 . Change default init to original.
2022-09-06 16:14:00 -07:00
Ross Wightman
7f1b223c02
Add maxvit_rmlp_nano_rw_256 model def & weights, make window/grid size dynamic wrt img_size by default
2022-08-29 15:49:32 -07:00
Ross Wightman
e6a4361306
pretrained_cfg entry for mvitv2_small_cls
2022-08-28 15:27:01 -07:00
Ross Wightman
f66e5f0e35
Fix class token support in MViT-V2, add small_class variant to ensure it's tested. Fix #1443
2022-08-28 15:24:04 -07:00
Ross Wightman
f1d2160d85
Update a few maxxvit comments, rename PartitionAttention -> PartitionAttenionCl for consistency with other blocks
2022-08-26 12:53:49 -07:00
Ross Wightman
eca6f0a25c
Fix syntax error (extra dataclass comma) in maxxvit.py
2022-08-26 11:29:09 -07:00
Ross Wightman
ff6a919cf5
Add --fast-norm arg to benchmark.py, train.py, validate.py
2022-08-25 17:20:46 -07:00
Ross Wightman
769ab4b98a
Clean up no_grad for trunc normal weight inits
2022-08-25 16:29:52 -07:00
Ross Wightman
48e1df8b37
Add norm/norm_act header comments
2022-08-25 16:29:34 -07:00
Ross Wightman
7c2660576d
Tweak init for convnext block using maxxvit/coatnext.
2022-08-25 15:30:59 -07:00
Ross Wightman
1d8d6f6072
Fix two default args in DenseNet blocks... fix #1427
2022-08-25 15:00:35 -07:00
Ross Wightman
527f9a4cb2
Updated to correct maxvit_nano weights...
2022-08-24 12:42:11 -07:00
Ross Wightman
b2e8426fca
Make k=stride=2 ('avg2') pooling default for coatnet/maxvit. Add weight links. Rename 'combined' partition to 'parallel'.
2022-08-24 11:01:20 -07:00
Ross Wightman
837c68263b
For ConvNeXt, use timm internal LayerNorm for fast_norm in non conv_mlp mode
2022-08-23 15:17:12 -07:00
Ross Wightman
cac0a4570a
More test fixes, pool size for 256x256 maxvit models
2022-08-23 13:38:26 -07:00
Ross Wightman
e939ed19b9
Rename internal creation fn for maxvit, has not been just coatnet for a while...
2022-08-22 17:44:51 -07:00
Ross Wightman
ffaf97f813
MaxxVit! A very configurable MaxVit and CoAtNet impl with lots of goodies..
2022-08-22 17:42:10 -07:00
Ross Wightman
8c9696c9df
More model and test fixes
2022-08-22 17:40:31 -07:00
Ross Wightman
ca52108c2b
Fix some model support functions
2022-08-19 10:20:51 -07:00
Ross Wightman
f332fc2db7
Fix some test failures, torchscript issues
2022-08-18 16:19:46 -07:00
Ross Wightman
6e559e9b5f
Add MViT (Multi-Scale) V2
2022-08-17 15:12:31 -07:00
Ross Wightman
43aa84e861
Add 'fast' layer norm that doesn't cast to float32, support APEX LN impl for slight speed gain, update norm and act factories, tweak SE for ability to disable bias (needed by GCVit)
2022-08-17 14:32:58 -07:00
Ross Wightman
c486aa71f8
Add GCViT
2022-08-17 14:29:18 -07:00
Ross Wightman
fba6ecd39b
Add EfficientFormer
2022-08-17 14:08:53 -07:00
Ross Wightman
ff4a38e2c3
Add PyramidVisionTransformerV2
2022-08-17 12:06:05 -07:00
Ross Wightman
1d8ada359a
Add timm ConvNeXt 'atto' weights, change test resolution for FB ConvNeXt 224x224 weights, add support for different dw kernel_size
2022-08-15 17:56:08 -07:00
Ross Wightman
2544d3b80f
ConvNeXt pico, femto, and nano, pico, femto ols (overlapping stem) weights and model defs
2022-08-05 17:05:50 -07:00
Ross Wightman
13565aad50
Add edgenext_base model def & weight link, update to improve ONNX export #1385
2022-08-05 16:58:34 -07:00
Ross Wightman
8ad4bdfa06
Allow ntuple to be used with string values
2022-07-28 16:18:18 -07:00
Christoph Reich
faae93e62d
Fix typo in PositionalEncodingFourier
2022-07-28 19:08:08 -04:00
Ross Wightman
7430a85d07
Update README, bump version to 0.6.8
2022-07-28 15:07:11 -07:00
Ross Wightman
ec6a28830f
Add DeiT-III 'medium' model defs and weights
2022-07-28 15:03:20 -07:00
Ross Wightman
d875a1d3f6
version 0.6.7
2022-07-27 12:41:06 -07:00
Ross Wightman
6f103a442b
Add convnext_nano weights, 80.8 @ 224, 81.5 @ 288
2022-07-26 16:40:27 -07:00
Ross Wightman
4042a94f8f
Add weights for two 'Edge' block (3x3->1x1) variants of CS3 networks.
2022-07-26 16:40:27 -07:00
Ross Wightman
c8f69e04a9
Merge pull request #1365 from veritable-tech/fix-resize-pos-embed
...
Take `no_emb_class` into account when calling `resize_pos_embed`
2022-07-24 21:03:01 -07:00
Ceshine Lee
0b64117592
Take `no_emb_class` into account when calling `resize_pos_embed`
2022-07-24 19:11:45 +08:00
Jasha10
56c3a84db3
Update type hint for `register_notrace_module`
...
register_notrace_module is used to decorate types (i.e. subclasses of nn.Module).
It is not called on module instances.
2022-07-22 16:59:55 -05:00
Ross Wightman
1b278136c3
Change models with mean 0,0,0 std 1,1,1 from int to float for consistency as mentioned in #1355
2022-07-21 17:36:15 -07:00
Ross Wightman
909705e7ff
Remove some redundant requires_grad=True from nn.Parameter in third party code
2022-07-20 12:37:41 -07:00
Ross Wightman
c5e0d1c700
Add dilation support to convnext, allows output_stride=8 and 16 use. Fix #1341
2022-07-19 17:52:10 -07:00
Ross Wightman
dc376e3676
Ensure all model entrypoint fn default to `pretrained=False` (a few didn't)
2022-07-19 13:58:41 -07:00
Ross Wightman
23b102064a
Add cs3sedarknet_x weights w/ 82.65 @ 288 top1. Add 2 cs3 edgenet models (w/ 3x3-1x1 block), remove aa from cspnet blocks (not needed)
2022-07-19 13:56:44 -07:00
Ross Wightman
0dbd9352ce
Add bulk_runner script and updates to benchmark.py and validate.py for better error handling in bulk runs (used for benchmark and validation result runs). Improved batch size decay stepping on retry...
2022-07-18 18:04:54 -07:00
Ross Wightman
92b91af3bb
version 0.6.6
2022-07-15 17:00:18 -07:00
Ross Wightman
05313940e2
Add cs3darknet_x, cs3sedarknet_l, and darknetaa53 weights from TPU sessions. Move SE btwn conv1 & conv2 in DarkBlock. Improve SE/attn handling in Csp/DarkNet. Fix leaky_relu bug on older csp models.
2022-07-15 16:55:16 -07:00
nateraw
51cca82aa1
👽 use hf_hub_download instead of cached_download
2022-07-14 16:41:45 -04:00
Ross Wightman
324a4e58b6
disable nvfuser for jit te/legacy modes (for PT 1.12+)
2022-07-13 10:34:34 -07:00
Ross Wightman
2898cf6e41
version 0.6.5 for pypi release
2022-07-10 16:43:23 -07:00
Ross Wightman
a45b4bce9a
x and xx small edgenext models do benefit from larger test input size
2022-07-08 10:53:27 -07:00
Ross Wightman
a8e34051c1
Unbreak gamma remap impacting beit checkpoint load, version bump to 0.6.4
2022-07-07 23:07:43 -07:00
Ross Wightman
1c5cb819f9
bump version to 0.6.3 before merge
2022-07-07 22:05:56 -07:00
Ross Wightman
a1cb25066e
Add edgnext_small_rw weights trained with swin like recipe. Better than original 'small' but not the recent 'USI' distilled weights.
2022-07-07 22:02:57 -07:00
Ross Wightman
7c7ecd2492
Add --use-train-size flag to force use of train input_size (over test input size) for validation. Default test-time pooling to use train input size (fixes issues).
2022-07-07 22:01:24 -07:00
Ross Wightman
ce65a7b29f
Update vit_relpos w/ some additional weights, some cleanup to match recent vit updates, more MLP log coord experiments.
2022-07-07 21:33:25 -07:00
Ross Wightman
58621723bd
Add CrossStage3 DarkNet (cs3) weights
2022-07-07 17:43:38 -07:00
Ross Wightman
9be0c84715
Change set -> dict w/ None keys for dataset split synonym search, so always consistent if more than 1 exists. Fix #1224
2022-07-07 15:33:53 -07:00
Ross Wightman
db0cee9910
Refactor cspnet configuration using dataclasses, update feature extraction for new cs3 variants.
2022-07-07 14:43:27 -07:00
Ross Wightman
eca09b8642
Add MobileVitV2 support. Fix #1332 . Move GroupNorm1 to common layers (used in poolformer + mobilevitv2). Keep ol custom ConvNeXt LayerNorm2d impl as LayerNormExp2d for reference.
2022-07-07 14:41:01 -07:00
Ross Wightman
06307b8b41
Remove experimental downsample in block support in ConvNeXt. Experiment further before keeping it in.
2022-07-07 14:37:58 -07:00
Ross Wightman
bfc0dccb0e
Improve image extension handling, add methods to modify / get defaults. Fix #1335 fix #1274 .
2022-07-07 14:23:20 -07:00
Ross Wightman
7d4b3807d5
Support DeiT-3 (Revenge of the ViT) checkpoints. Add non-overlapping (w/ class token) pos-embed support to vit.
2022-07-04 22:25:22 -07:00
Ross Wightman
d0c5bd5722
Rename cs2->cs3 for darknets. Fix features_only for cs3 darknets.
2022-07-03 08:32:41 -07:00
Ross Wightman
d765305821
Remove first_conv for resnetaa50 def
2022-07-02 15:56:17 -07:00
Ross Wightman
dd9b8f57c4
Add feature_info to edgenext for features_only support, hopefully fix some fx / test errors
2022-07-02 15:20:45 -07:00
Ross Wightman
377e9bfa21
Add TPU trained darknet53 weights. Add mising pretrain_cfg for some csp/darknet models.
2022-07-02 15:18:52 -07:00
Ross Wightman
c170ba3173
Add weights for resnet10t, resnet14t, and resnetaa50 models. Fix #1314
2022-07-02 15:18:06 -07:00
Ross Wightman
188c194b0f
Left some experiment stem code in convnext by mistake
2022-07-02 15:17:28 -07:00
Ross Wightman
70d6d2c484
support test_crop_size in data config resolve
2022-07-02 15:17:05 -07:00
Ross Wightman
6064d16a2d
Add initial EdgeNeXt import. Significant cleanup / reorg (like ConvNeXt). Fix #1320
...
* edgenext refactored for torchscript compat, stage base organization
* slight refactor of ConvNeXt to match some EdgeNeXt additions
* remove use of funky LayerNorm layer in ConvNeXt and just use nn.LayerNorm and LayerNorm2d (permute)
2022-07-01 15:18:42 -07:00
Ross Wightman
7a9c6811c9
Add eps arg to LayerNorm2d, add 'tf' (tensorflow) variant of trunc_normal_ that applies scale/shift after sampling (instead of needing to move a/b)
2022-07-01 15:15:39 -07:00
Ross Wightman
82c311d082
Add more experimental darknet and 'cs2' darknet variants (different cross stage setup, closer to newer YOLO backbones) for train trials.
2022-07-01 15:14:01 -07:00
Ross Wightman
a050fde5cd
Add resnet10t (basic block) and resnet14t (bottleneck) with 1,1,1,1 repeats
2022-07-01 15:03:28 -07:00
Ross Wightman
e6d7df40ec
no longer a point using kwargs for pretrain_cfg resolve, just pass explicit arg
2022-06-24 21:36:23 -07:00
Ross Wightman
07d0c4ae96
Improve repr for DropPath module
2022-06-24 14:58:15 -07:00
Ross Wightman
e27c16b8a0
Remove unecessary code for synbn guard
2022-06-24 14:57:42 -07:00
Ross Wightman
0da3c9ebbf
Remove SiLU layer in default args that breaks import on old old PyTorch
2022-06-24 14:56:58 -07:00
Ross Wightman
7d657d2ef4
Improve resolve_pretrained_cfg behaviour when no cfg exists, warn instead of crash. Improve usability ex #1311
2022-06-24 14:55:25 -07:00
Ross Wightman
879df47c0a
Support BatchNormAct2d for sync-bn use. Fix #1254
2022-06-24 14:51:26 -07:00
Ross Wightman
7cedc8d474
Follow up to #1256 , fix interpolation warning in auto_autoaugment as well
2022-06-21 14:56:53 -07:00
Jakub Kaczmarzyk
db64393c0d
use `Image.Resampling` namespace for PIL mapping ( #1256 )
...
* use `Image.Resampling` namespace for PIL mapping
PIL shows a deprecation warning when accessing resampling constants via the `Image` namespace. The suggested namespace is `Image.Resampling`. This commit updates `_pil_interpolation_to_str` to use the `Image.Resampling` namespace.
```
/tmp/ipykernel_11959/698124036.py:2: DeprecationWarning: NEAREST is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.NEAREST or Dither.NONE instead.
Image.NEAREST: 'nearest',
/tmp/ipykernel_11959/698124036.py:3: DeprecationWarning: BILINEAR is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.BILINEAR instead.
Image.BILINEAR: 'bilinear',
/tmp/ipykernel_11959/698124036.py:4: DeprecationWarning: BICUBIC is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.BICUBIC instead.
Image.BICUBIC: 'bicubic',
/tmp/ipykernel_11959/698124036.py:5: DeprecationWarning: BOX is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.BOX instead.
Image.BOX: 'box',
/tmp/ipykernel_11959/698124036.py:6: DeprecationWarning: HAMMING is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.HAMMING instead.
Image.HAMMING: 'hamming',
/tmp/ipykernel_11959/698124036.py:7: DeprecationWarning: LANCZOS is deprecated and will be removed in Pillow 10 (2023-07-01). Use Resampling.LANCZOS instead.
Image.LANCZOS: 'lanczos',
```
* use new pillow resampling enum only if it exists
2022-06-12 22:30:57 -07:00
Ross Wightman
20a1fa63f8
Make dev version 0.6.2.dev0 for pypi pre
2022-05-15 14:29:57 -07:00
Ross Wightman
347308faad
Update README.md, version to 0.6.2
2022-05-13 13:54:41 -07:00
Ross Wightman
4b30bae67b
Add updated vit_relpos weights, and impl w/ support for official swin-v2 differences for relpos. Add bias control support for MLP layers
2022-05-13 13:53:57 -07:00
Ross Wightman
d4c0588012
Remove persistent buffers from Swin-V2. Change SwinV2Cr cos attn + tau/logit_scale to match official, add ckpt convert, init_value zeros resid LN weight by default
2022-05-13 10:50:59 -07:00
Ross Wightman
27c42f0830
Fix torchscript use for offician Swin-V2, add support for non-square window/shift to WindowAttn/Block
2022-05-13 09:29:33 -07:00
Ross Wightman
2f2b22d8c7
Disable nvfuser fma / opt level overrides per #1244
2022-05-13 09:27:13 -07:00
Ross Wightman
c0211b0bf7
Swin-V2 test fixes, typo
2022-05-12 22:31:55 -07:00
Ross Wightman
9a86b900fa
Official SwinV2 models
2022-05-12 15:05:10 -07:00
Ross Wightman
d07d015173
Merge pull request #1249 from okojoalg/sequencer
...
Add Sequencer
2022-05-09 20:42:43 -07:00
Ross Wightman
d30685c283
Merge pull request #1251 from hankyul2/fix-multistep-scheduler
...
fix: multistep lr decay epoch bugs
2022-05-09 16:07:46 -07:00
han
a16171335b
fix: change milestones to decay-milestones
...
- change argparser option `milestone` to `decay-milestone`
2022-05-10 07:57:19 +09:00
Ross Wightman
39b725e1c9
Fix tests for rank-4 output where feature channels dim is -1 (3) and not 1
2022-05-09 15:20:24 -07:00
Ross Wightman
78a32655fa
Fix poolformer group_matcher to merge proj downsample with previous block, support coarse
2022-05-09 12:20:04 -07:00
Ross Wightman
d79f3d9d1e
Fix torchscript use for sequencer, add group_matcher, forward_head support, minor formatting
2022-05-09 12:09:39 -07:00
Ross Wightman
37b6920df3
Fix group_matcher regex for regnet.py
2022-05-09 10:40:40 -07:00
okojoalg
93a79a3dd9
Fix num_features in Sequencer
2022-05-06 23:16:32 +09:00
han
57a988df30
fix: multistep lr decay epoch bugs
...
- add milestones arguments
- change decay_epochs to milestones variable
2022-05-06 13:14:43 +09:00
okojoalg
578d52e752
Add Sequencer
2022-05-06 00:36:01 +09:00
Ross Wightman
f5ca4141f7
Adjust arg order for recent vit model args, add a few comments
2022-05-02 22:41:38 -07:00
Ross Wightman
41dc49a337
Vision Transformer refactoring and Rel Pos impl
2022-05-02 15:37:39 -07:00
Ross Wightman
b7cb8d0337
Add Swin-V2 Small-NS weights (83.5 @ 224). Add layer scale like 'init_values' via post-norm LN weight scaling
2022-04-26 17:32:49 -07:00
jjsjann123
f88c606fcf
fixing channels_last on cond_conv2d; update nvfuser debug env variable
2022-04-25 12:41:46 -07:00
Li Dong
09e9f3defb
migrate azure blob for beit checkpoints
...
## Motivation
We are going to use a new blob account to store the checkpoints.
## Modification
Modify the azure blob storage URLs for BEiT checkpoints.
2022-04-23 13:02:29 +08:00
Ross Wightman
52ac881402
Missed first_conv in latest seresnext 'D' default_cfgs
2022-04-22 20:55:52 -07:00