1328 Commits

Author SHA1 Message Date
Ross Wightman
1bf84b35c3 Update tests for aimv2 filtering 2024-12-30 19:24:21 -08:00
Ross Wightman
b33418713a Add (almost) full set of aimv2 model instances. Switch back to unpacked SwiGLU. Verify correctness. Add DFN L/14 39B weight. 2024-12-30 19:24:21 -08:00
Ross Wightman
5f12a25114 Add bias arg to Vitamin GeGLU 2024-12-30 19:24:21 -08:00
Ross Wightman
5804d92e4b Switch aimv2 to used packed SwiGLU 2024-12-30 19:24:21 -08:00
Ross Wightman
15406a939e Fixing RmsNorm to fix #2380 and noticed with aimv2 when comparing outputs. Still some work to do, need to look at AMP / fast mode behaviour, dispatch to torch when possible. Add SimpleNorm for 'LayerNorm w/o centering and bias' 2024-12-30 19:24:21 -08:00
Ross Wightman
a648a04834 Supporting aimv2 encoders 2024-12-30 19:24:21 -08:00
Ross Wightman
790decc89b Add more pali(2) weights. Switch rest of models adapting open_clip weights to their own weight instances. 2024-12-27 14:00:41 -08:00
Ross Wightman
01cf0f72af Add support for tag, license customization through push_to_hub 2024-12-27 14:00:41 -08:00
Ross Wightman
b12ecbd614 Move siglip timm weights to own repos 2024-12-27 14:00:41 -08:00
Ross Wightman
6fb7aaf37d Switching to timm specific weight instances for open_clip image encoders to facilitate hf-hub: use in timm and new transformers TimmWrapper 2024-12-27 14:00:41 -08:00
Ryan
ab0a70dfff fix feature_info.reduction 2024-12-18 21:12:40 +08:00
Álvaro Justen (@turicas)
9383f2880d Add cache_dir example 2024-12-06 10:39:13 -08:00
Ross Wightman
d1e9a8622a Rename inception_next_atto pretrained str 2024-12-06 10:36:47 -08:00
Weihao Yu
0576175d85 Add inception_next_atto 2024-12-06 10:36:47 -08:00
Ross Wightman
7ab2b938e5 More tweaks to docstrings for hub/builder 2024-12-06 10:25:06 -08:00
Ross Wightman
dc1bb05e8e Punch cache_dir through model factory / builder / pretrain helpers. Improve some annotations in related code. 2024-12-06 10:25:06 -08:00
Ross Wightman
5fe5f9d488 Add a different mnv4 conv-small weight 2024-12-02 16:14:37 -08:00
Johannes
093a234d01
Update torchvision resnet legacy weight urls in resnet.py 2024-11-26 15:53:54 +01:00
Ross Wightman
900d2b508d add mnv4 conv_medium in12k -> in1k ft 2024-11-22 16:31:45 -08:00
Ross Wightman
6bcbdbfe41 CS3-DarkNet Small (Focus) w/ RA4 recipe. Fix #2122 2024-11-22 16:31:45 -08:00
Ross Wightman
ae0737f5d0 Typo 2024-11-17 13:54:50 -08:00
Ross Wightman
84049d7f1e Missed input_size pretraind_cfg metadata for v2 34d @ 384 2024-11-17 12:44:08 -08:00
Ross Wightman
b7a4b49ae6 Add some 384x384 small model weights, 3 variants of mnv4 conv medium on in12k pretrain, and resnetv2-34d on in1k 2024-11-17 12:14:39 -08:00
Wojtek Jasiński
eb94efb218 fix pos embed dynamic resampling for eva 2024-11-06 16:03:27 -08:00
Wojtek Jasiński
3c7822c621 fix pos embed dynamic resampling for deit 2024-11-06 16:03:27 -08:00
Wojtek Jasiński
3ae3f44288 Fix positional embedding resampling for non-square inputs in ViT 2024-11-06 16:03:27 -08:00
Ross Wightman
d4dde48dd5 Missed first_conv from resnet18d 2024-10-31 19:29:53 -07:00
Ross Wightman
e6263bf64d Add resnet and resnet-v2 18/34 weights trained with mnv4 small based recipe 2024-10-31 16:39:35 -07:00
Ross Wightman
f689c850b9 One more small c&p issue 2024-10-23 21:51:09 -07:00
Ross Wightman
baa7242dd3 Fix c&p error, slight reformat 2024-10-23 21:51:09 -07:00
Ross Wightman
1b5cae681c Update some clip pretrained weights to point to new hub locations, add a few missing weights 2024-10-23 21:51:09 -07:00
Feraidoon Mehri
ca20e102fe
mambaout.py: fixed bug 2024-10-17 01:03:28 +03:30
Ross Wightman
89dffc5ff0 Another small fix for original mambaout models, no classifier nn.Linear when num_classe=0 on init 2024-10-16 12:36:36 -07:00
Ross Wightman
fad4538801 Elevate import deprecation warnings from DeprecationWarning to FutureWarning so messages are now seen 2024-10-16 11:30:01 -07:00
Ross Wightman
a1f379e712 Add intern300m vit w/ converted timm weights. Fix #2300 2024-10-16 10:29:06 -07:00
Ross Wightman
60f517c883 Fix wrong name in _all_ for models._registry 2024-10-15 07:39:46 -07:00
Ross Wightman
c3052fa19e
Merge pull request #2298 from huggingface/preact_resnet18
Add resnet18/18d pre-act model configs for potential training.
2024-10-14 19:39:04 -07:00
Ross Wightman
abdf33145c Add 34/34d pre-act resnet variants 2024-10-14 13:23:50 -07:00
Ross Wightman
c82ce86f8f Add 384x384 mambaout_base_plus model weights 2024-10-14 12:28:57 -07:00
Ross Wightman
82ae247879 MambaOut weights on hub, configs finalized 2024-10-11 11:07:40 -07:00
Ross Wightman
7efb60c299 Add first_conv for mambaout 2024-10-09 14:11:40 -07:00
Ross Wightman
5dc5ee5b42 Add global_pool to mambaout __init__ and pass to heads 2024-10-09 14:11:40 -07:00
Ross Wightman
9d1dfe8dbe Incorrectly named head_hidden_size 2024-10-09 14:11:40 -07:00
Ross Wightman
91e743f2dd Mambaout tweaks 2024-10-09 14:11:40 -07:00
Ross Wightman
4542cf03f9 Add features_only, other bits to mambaout, define different base alternatives 2024-10-09 14:11:40 -07:00
Ross Wightman
c2da12c7e1 Update rw models, fix heads 2024-10-09 14:11:40 -07:00
Ross Wightman
f2086f51a0 Add mambaout builder support, pretrained weight remap 2024-10-09 14:11:40 -07:00
Ross Wightman
c6ef54eefa Initial mambaout work 2024-10-09 14:11:40 -07:00
Ross Wightman
d9321b0e10 Add weights for fine-tuned siglip so400m. Add webli_i18n pretrained tags for the multi-lingual model variants (incl older base) 2024-10-09 09:04:44 -07:00
Ross Wightman
01b62264af Add i18n variant of so400m model w/ weights. Add two in1k fine-tunes of original so400m 384x384 but at 378x378 (better matches patch14) 2024-10-08 23:40:24 -07:00