Ross Wightman
4203efa36d
Fix #387 so that checkpoint saver works with max history of 1. Add checkpoint-hist arg to train.py.
2021-01-31 20:14:51 -08:00
Ross Wightman
38d8f67570
Fix potential issue with change to num_classes arg in train/validate.py defaulting to None (rely on model def / default_cfg)
2021-01-25 11:53:34 -08:00
Ross Wightman
5d4c3d0af3
Add enhanced ParserImageInTar that can read images from tars within tars, folders with multiple tars, etc. Additional comment cleanup.
2021-01-22 10:52:04 -08:00
Ross Wightman
9d5d4b8df6
Fix silly train.py typo during dataset work
2021-01-16 16:32:21 -08:00
Ross Wightman
855d6cc217
More dataset work including factories and a tensorflow datasets (TFDS) wrapper
...
* Add parser/dataset factory methods for more flexible dataset & parser creation
* Add dataset parser that wraps TFDS image classification datasets
* Tweak num_classes handling bug for 21k models
* Add initial deit models so they can be benchmarked in next csv results runs
2021-01-15 17:26:20 -08:00
Ross Wightman
fd9061dbf7
Remove debug print from train.py
2021-01-04 12:16:06 -08:00
Ross Wightman
59ec7e6a53
Merge branch 'master' into imagenet21k_datasets_more
2021-01-04 12:11:05 -08:00
Csaba Kertesz
e42b140ade
Add --input-size option to scripts to specify full input dimensions from command-line
2021-01-04 00:25:29 +02:00
Ross Wightman
231d04e91a
ResNetV2 pre-act and non-preact model, w/ BiT pretrained weights and support for ViT R50 model. Tweaks for in21k num_classes passing. More to do... tests failing.
2020-12-28 16:59:15 -08:00
Ross Wightman
de6046e213
Initial commit for dataset / parser reorg to support additional datasets / types
2020-12-28 16:59:15 -08:00
Ross Wightman
2ed8f24715
A few more changes for 0.3.2 maint release. Linear layer change for mobilenetv3 and inception_v3, support no bias for linear wrapper.
2020-11-30 16:19:52 -08:00
Ross Wightman
460eba7f24
Work around casting issue with combination of native torch AMP and torchscript for Linear layers
2020-11-30 13:30:51 -08:00
Ross Wightman
27bbc70d71
Add back old ModelEma and rename new one to ModelEmaV2 to avoid compat breaks in dependant code. Shuffle train script, add a few comments, remove DataParallel support, support experimental torchscript training.
2020-11-29 16:22:19 -08:00
Ross Wightman
9214ca0716
Simplifying EMA...
2020-11-16 12:51:52 -08:00
Ross Wightman
80078c47bb
Add Adafactor and Adahessian optimizers, cleanup optimizer arg passing, add gradient clipping support.
2020-10-09 17:24:43 -07:00
Ross Wightman
47a7b3b5b1
More flexible mixup mode, add 'half' mode.
2020-09-07 20:03:06 -07:00
Ross Wightman
532e3b417d
Reorg of utils into separate modules
2020-09-07 13:58:09 -07:00
Ross Wightman
751b0bba98
Add global_pool (--gp) arg changes to allow passing 'fast' easily for train/validate to avoid channels_last issue with AdaptiveAvgPool
2020-09-02 16:13:47 -07:00
Ross Wightman
9c297ec67d
Cleanup Apex vs native AMP scaler state save/load. Cleanup CheckpointSaver a bit.
2020-09-02 15:12:59 -07:00
Ross Wightman
c2cd1a332e
Improve torch amp support and add channels_last support for train/validate scripts
2020-08-31 17:58:16 -07:00
datamining99
5f563ca4df
fix save_checkpoint bug with native amp
2020-08-22 11:31:50 +09:00
datamining99
d98967ed5d
add support for native torch AMP in torch 1.6
2020-08-22 09:44:23 +09:00
Ross Wightman
8c9814e3f5
Final cleanup of mixup/cutmix. Element/batch modes working with both collate (prefetcher active) and without prefetcher.
2020-08-12 17:01:32 -07:00
Ross Wightman
f471c17c9d
More cutmix/mixup overhaul, ready to kick-off some trials.
2020-08-11 00:10:33 -07:00
Ross Wightman
92f2d0d65d
Merge branch 'master' into cutmix. Fixup a few issues.
2020-08-07 15:59:52 -07:00
Ross Wightman
fa28067704
Add more augmentation arguments, including a no_aug disable flag. Fix #209
2020-08-05 13:16:44 -07:00
Ross Wightman
7995295968
Merge branch 'logger' into features. Change 'logger' to '_logger'.
2020-07-27 18:00:46 -07:00
Ross Wightman
1998bd3180
Merge branch 'feature/AB/logger' of https://github.com/antoinebrl/pytorch-image-models into logger
2020-07-27 16:06:01 -07:00
Ross Wightman
6c17d57a2c
Fix some attributions, add copyrights to some file docstrings
2020-07-27 13:44:56 -07:00
Antoine Broyelle
78fa0772cc
Leverage python hierachical logger
...
with this update one can tune the kind of logs generated by timm but
training and inference traces are unchanged
2020-06-09 18:28:48 +01:00
Ross Wightman
6441e9cc1b
Fix memory_efficient mode for DenseNets. Add AntiAliasing (Blur) support for DenseNets and create one test model. Add lr cycle/mul params to train args.
2020-05-22 16:16:45 -07:00
AFLALO, Jonathan Isaac
a7f570c9b7
added MultiEpochsDataLoader
2020-05-05 14:47:02 +03:00
Ross Wightman
13cf68850b
Remove poorly named metrics from torch imagenet example origins. Use top1/top5 in csv output for consistency with existing validation results files, acc elsewhere. Fixes #111
2020-04-10 14:41:08 -07:00
Ross Wightman
27b3680d49
Revamp LR noise, move logic to scheduler base. Fixup PlateauLRScheduler and add it as an option.
2020-02-22 16:23:15 -08:00
Ross Wightman
514b0938c4
Experimenting with per-epoch learning rate noise w/ step scheduler
2020-02-21 11:51:05 -08:00
Ross Wightman
43225d110c
Unify drop connect vs drop path under 'drop path' name, switch all EfficientNet/MobilenetV3 refs to 'drop_path'. Update factory to handle new drop args.
2020-02-18 14:00:26 -08:00
Ross Wightman
b3cb5f3275
Working on CutMix impl as per #8 , integrating with Mixup, currently experimenting...
2020-02-16 20:09:11 -08:00
Andrew Lavin
b72013def8
Added commandline argument validation-batch-size-multiplier with default set to 1.
2020-02-04 01:02:27 -08:00
Ross Wightman
5b7cc16ac9
Add warning about using sync-bn with zero initialized BN layers. Fixes #54
2020-01-31 11:44:24 -08:00
Ross Wightman
d9a6a9d0af
Merge pull request #74 from rwightman/augmix-jsd
...
AugMix, JSD loss, SplitBatchNorm (Auxiliary BN), and more
2020-01-11 12:04:29 -08:00
Ross Wightman
3eb4a96eda
Update AugMix, JSD, etc comments and references
2020-01-11 12:02:05 -08:00
Ross Wightman
7547119891
Add SplitBatchNorm. AugMix, Rand/AutoAugment, Split (Aux) BatchNorm, Jensen-Shannon Divergence, RandomErasing all working together
2020-01-05 19:58:59 -08:00
Ross Wightman
40fea63ebe
Add checkpoint averaging script. Add headers, shebangs, exec perms to all scripts
2020-01-03 14:57:46 -08:00
Ross Wightman
4666cc9aed
Add --pin-mem arg to enable dataloader pin_memory (showing more benefit in some scenarios now), also add --torchscript arg to validate.py for testing models with jit.script
2020-01-02 16:22:06 -08:00
Ross Wightman
232ab7fb12
Working on an implementation of AugMix with JensenShannonDivergence loss that's compatible with my AutoAugment and RandAugment impl
2019-12-20 23:04:11 -08:00
Ross Wightman
5719b493ad
Missed update dist-bn logic for EMA model
2019-12-19 23:03:04 -08:00
Ross Wightman
a435ea1327
Change reduce_bn to distribute_bn, add ability to choose between broadcast and reduce (mean). Add crop_pct arg to allow selecting validation crop while training.
2019-12-19 22:56:54 -08:00
Ross Wightman
3bff2b21dc
Add support for keeping running bn stats the same across distributed training nodes before eval/save
2019-12-05 22:35:40 -08:00
Ross Wightman
1f39d15f15
Allow float decay epochs arg for training, works out with step lr math
2019-11-22 13:28:58 -08:00
Ross Wightman
7b83e67f77
Pass drop connect arg through to EfficientNet models
2019-11-22 13:27:43 -08:00