2595 Commits

Author SHA1 Message Date
Ross Wightman
d1e9a8622a Rename inception_next_atto pretrained str 2024-12-06 10:36:47 -08:00
Weihao Yu
0576175d85 Add inception_next_atto 2024-12-06 10:36:47 -08:00
Ross Wightman
7ab2b938e5 More tweaks to docstrings for hub/builder 2024-12-06 10:25:06 -08:00
Ross Wightman
dc1bb05e8e Punch cache_dir through model factory / builder / pretrain helpers. Improve some annotations in related code. 2024-12-06 10:25:06 -08:00
Ross Wightman
553ded5c6b Version 1.0.12 v1.0.12 2024-12-03 10:34:52 -08:00
Ross Wightman
464885e135 See if we can avoid some model / layer pickle issues with the aa attr in ConvNormAct 2024-12-03 08:02:55 -08:00
Ross Wightman
5fe5f9d488 Add a different mnv4 conv-small weight 2024-12-02 16:14:37 -08:00
Ross Wightman
303f7691a1 Add cautious mars, improve test reliability by skipping grad diff for first step 2024-12-02 11:29:02 -08:00
Ross Wightman
82e8677690 Make LaProp weight decay match typical PyTorch 'decoupled' behaviour where it's scaled by LR 2024-11-29 16:44:43 -08:00
Ross Wightman
886eb77938 Update README, missed small discrep in adafactor min dim update 2024-11-29 10:57:47 -08:00
Ross Wightman
e3e434bbc4 To be technically correct, need to check the in-place _ ver of op 2024-11-28 15:11:58 -08:00
Ross Wightman
7c32d3bd82 Work around _foreach_maximum issue, need scalar other support 2024-11-28 15:11:58 -08:00
Ross Wightman
7cf683628f Cautious optimizer impl plus some typing cleanup. 2024-11-28 15:11:58 -08:00
Ross Wightman
aeb1ed7a15 Keep basic optim test LR range closer to before w/ updated code 2024-11-26 15:10:15 -08:00
Ross Wightman
7a165fcb62 Remove rogue import, thanks IDE :/ 2024-11-26 15:10:15 -08:00
Ross Wightman
73d10ab482 Update tests, need handling for radamw with older PyTorch, need to back-off basic test LR in mars? 2024-11-26 15:10:15 -08:00
Ross Wightman
09bc21774e Update optimizers.mdx 2024-11-26 15:10:15 -08:00
Ross Wightman
4f64ec4e14 Add guard around 'somewhat' newer torch RAdam / NAdam imports 2024-11-26 15:10:15 -08:00
Ross Wightman
0903d98162 Reduce tolerance on model inference 'owl' test, pillow output varies a lot, was failing locally 2024-11-26 15:10:15 -08:00
Ross Wightman
1ab02a11a1 Update Adan with newer impl (from original source) that includes multi-tensor fn 2024-11-26 15:10:15 -08:00
Ross Wightman
a024ab3170 Replace radam & nadam impl with torch.optim ver, rename legacy adamw, nadam, radam impl in timm. Update optim factory & tests. 2024-11-26 15:10:15 -08:00
Ross Wightman
7b54eab807 Add MARS and LaProp impl, simplified from originals 2024-11-26 15:10:15 -08:00
Ross Wightman
e5aea357b1 Update Adopt to include clipping for stability, separate wd so no param decay if update not taken on first step 2024-11-26 15:10:15 -08:00
Ross Wightman
444c506ce3
Merge pull request #2346 from JohannesTheo/patch-1
Update timm torchvision resnet weight urls to the updated urls in torchvision
2024-11-26 11:15:17 -08:00
Johannes
093a234d01
Update torchvision resnet legacy weight urls in resnet.py 2024-11-26 15:53:54 +01:00
Ross Wightman
2fcf73e580 Add mini imagenet info files 2024-11-25 10:53:28 -08:00
Ross Wightman
900d2b508d add mnv4 conv_medium in12k -> in1k ft 2024-11-22 16:31:45 -08:00
Ross Wightman
6bcbdbfe41 CS3-DarkNet Small (Focus) w/ RA4 recipe. Fix #2122 2024-11-22 16:31:45 -08:00
Sina Hajimiri
3a6cc4fb17 Improve wandb logging 2024-11-20 21:04:07 -08:00
Ross Wightman
620cb4f3cb Improve the parsable results dump at end of train, stop excessive output, only display top-10. 2024-11-20 16:47:06 -08:00
Ross Wightman
36b5d1adaa In dist training, update loss running avg every step, only sync on log updates / final. 2024-11-20 16:47:06 -08:00
Ross Wightman
ae0737f5d0 Typo 2024-11-17 13:54:50 -08:00
Ross Wightman
84049d7f1e Missed input_size pretraind_cfg metadata for v2 34d @ 384 2024-11-17 12:44:08 -08:00
Ross Wightman
b7a4b49ae6 Add some 384x384 small model weights, 3 variants of mnv4 conv medium on in12k pretrain, and resnetv2-34d on in1k 2024-11-17 12:14:39 -08:00
Alina
facae65947 Update CODE_OF_CONDUCT.md 2024-11-17 11:43:39 -08:00
Alina Imtiaz
165c3dea98 Add CODE_OF_CONDUCT.md and CITATION.cff files 2024-11-17 11:43:39 -08:00
Antoine Broyelle
74196aceda Add py.typed file as recommended by PEP 561 2024-11-14 11:26:00 -08:00
Ross Wightman
e35ea733ab Fix compiler check for adopt so it doesn't fail for torch >= 2 but less than recent with .is_compiling() 2024-11-13 11:24:01 -08:00
Ross Wightman
0b5264a108 Missing optimizers in __init__.py, add bind_defaults=False for unit tests 2024-11-13 10:50:46 -08:00
Ross Wightman
d0161f303a Small optim factory tweak. default bind_defaults=True for get_optimizer_class 2024-11-13 10:45:48 -08:00
Ross Wightman
ef062eefe3
Update README.md 2024-11-13 10:21:51 -08:00
Ross Wightman
3bef09f831 Tweak a few docstrings 2024-11-13 10:12:31 -08:00
Ross Wightman
015ac30a91
Update README.md 2024-11-13 08:20:20 -08:00
Ross Wightman
8b9b6824ae Minor changes, has_eps=False missing for bnb lion 2024-11-12 20:49:01 -08:00
Ross Wightman
61305cc26a Fix adopt descriptions 2024-11-12 20:49:01 -08:00
Ross Wightman
ce42cc4846 Another doc class typo 2024-11-12 20:49:01 -08:00
Ross Wightman
dde990785e More fixes for new factory & tests, add back adahessian 2024-11-12 20:49:01 -08:00
Ross Wightman
45490ac52f Post merge fix reference of old param groups helper fn locations 2024-11-12 20:49:01 -08:00
Ross Wightman
53657a31b7 Try to fix documentation build, add better docstrings to public optimizer api 2024-11-12 20:49:01 -08:00
Ross Wightman
ee5f6e76bb A bit of an optimizer overhaul, added an improved factory, list_optimizers, class helper and add info classes with descriptions, arg configs 2024-11-12 20:49:01 -08:00