Commit Graph

194 Commits (7efb60c2996dcc7335737e5cf6553e81476f548b)

Author SHA1 Message Date
Ross Wightman d9321b0e10 Add weights for fine-tuned siglip so400m. Add webli_i18n pretrained tags for the multi-lingual model variants (incl older base) 2024-10-09 09:04:44 -07:00
Ross Wightman 01b62264af Add i18n variant of so400m model w/ weights. Add two in1k fine-tunes of original so400m 384x384 but at 378x378 (better matches patch14) 2024-10-08 23:40:24 -07:00
Ross Wightman 545bd4056c Tag along test_vit3 weights 2024-09-30 12:03:32 -07:00
Ross Wightman 9067be6a30 Add weights for new tiny test models 2024-09-22 07:59:23 -07:00
Ross Wightman a2f539f055 Add a few more test model defs in prep for weight upload 2024-09-21 11:38:38 -07:00
Ross Wightman a50713ce6e Fix #2272 2024-09-02 13:20:05 -07:00
Ross Wightman 9fcbf39cdc Add remaining sbb vit betwixt/mediumd fine-tunes 2024-08-21 10:09:38 -07:00
Ross Wightman 2f3fed43b8 Fix hiera init with num_classes=0, fix weight tag names for sbb2 hiera/vit weights, add LayerScale/LayerScale2d to layers 2024-08-15 11:14:38 -07:00
Ross Wightman fee91fdd41 Update Hiera model for abswin, more stable weight init, layer-scale. ImageNet-12k weights for hiera_small_abswin, and two of the sbb vits with improved reg4 init. 2024-08-14 12:22:40 -07:00
Ross Wightman cec70b6779
Merge pull request #2225 from huggingface/small_things
Small things
2024-07-25 20:29:13 -07:00
Ross Wightman d2240745d3 Fix issue where feature out_indices out of order after wrapping with FeatureGetterNet due to use of set() 2024-07-22 13:33:30 -07:00
Ross Wightman 9b2b8014e8 Add weights for test models 2024-07-22 10:08:57 -07:00
Ross Wightman 392b78aee7 set_input_size initial impl for vit & swin v1. Move HybridEmbed to own location in timm/layers 2024-07-17 15:25:48 -07:00
Promisery 417cf7f871 Initialize weights of reg_token for ViT 2024-07-13 11:11:42 +08:00
Ross Wightman d5afe106dc Merge remote-tracking branch 'origin/tiny_test_models' into small_things 2024-07-09 12:49:57 -07:00
Ross Wightman 55101028bb Rename test_tiny* -> test*. Fix ByobNet BasicBlock attn location and add test_byobnet model. 2024-07-09 11:53:11 -07:00
Sejik c33a001397
Fix typo 2024-06-24 21:54:38 +09:00
Ross Wightman 71101ebba0 Refactor vit pooling to add more reduction options, separately callable 2024-06-14 23:16:58 -07:00
Fernando Cossio 9567cf6d84
Feature: add option global_pool='max' to VisionTransformer
Most of the CNNs have a max global pooling option. I would like to extend ViT to have this option.
2024-06-14 15:24:54 +02:00
Ross Wightman 66a0eb4673 Experimenting with tiny test models, how small can they go and be useful for regression tests? 2024-06-07 16:09:25 -07:00
Ross Wightman a5a2ad2e48 Fix consistency, testing for forward_head w/ pre_logits, reset_classifier, models with pre_logits size != unpooled feature size
* add test that model supports forward_head(x, pre_logits=True)
* add head_hidden_size attr to all models and set differently from num_features attr when head has hidden layers
* test forward_features() feat dim == model.num_features and pre_logits feat dim == self.head_hidden_size
* more consistency in reset_classifier signature, add typing
* asserts in some heads where pooling cannot be disabled
Fix #2194
2024-06-07 13:53:00 -07:00
Ross Wightman cc8a03daac Add ConvStem and MobileCLIP hybrid model for B variant. Add full norm disable support to ConvNormAct layers 2024-06-06 09:15:27 -07:00
Ross Wightman 1b66ec7cf3 Fixup ViTamin, add hub weight reference 2024-06-03 17:14:03 -07:00
Ross Wightman 5dce710101 Add vit_little in12k + in12k-ft-in1k weights 2024-05-27 14:56:03 -07:00
Ross Wightman 84cb225ecb Add in12k + 12k_ft_in1k vit_medium weights 2024-05-20 15:52:46 -07:00
Ross Wightman 83aee5c28c Add explicit GAP (avg pool) variants of other SigLIP models. 2024-05-15 07:53:19 -07:00
Ross Wightman 7b3b11b63f Support loading of paligemma weights into GAP variants of SigLIP ViT. Minor tweak to npz loading for packed transformer weights. 2024-05-14 15:44:37 -07:00
Ross Wightman a69863ad61
Merge pull request #2156 from huggingface/hiera
WIP Hiera implementation.
2024-05-13 14:58:12 -07:00
Ross Wightman f7aa0a1a71 Add missing vit_wee weight 2024-05-13 12:05:47 -07:00
Ross Wightman c838c4233f Add typing to reset_classifier() on other models 2024-05-12 11:12:00 -07:00
Ross Wightman 1d3ab176bc Remove debug / staging code 2024-05-10 22:16:34 -07:00
Ross Wightman aa4d06a11c sbb vit weights on hub, testing 2024-05-10 17:15:01 -07:00
Ross Wightman 3582ca499e Prepping weight push, benchmarking. 2024-05-10 14:14:06 -07:00
Ross Wightman cb57a96862 Fix early stop for efficientnet/mobilenetv3 fwd inter. Fix indices typing for all fwd inter. 2024-05-04 10:21:58 -07:00
Ross Wightman f8979d4f50 Comment out time local files while testing new vit weights 2024-05-03 20:26:56 -07:00
Ross Wightman c719f7eb86 More forward_intermediates() updates
* add convnext, resnet, efficientformer, levit support
* remove kwargs only for fn so that torchscript isn't broken for all :(
* use reset_classifier() consistently in prune
2024-05-03 16:22:32 -07:00
Ross Wightman c22efb9765 Add wee & little vits for some experiments 2024-05-02 10:51:35 -07:00
Ross Wightman 67332fce24 Add features_intermediate() support to coatnet, maxvit, swin* models. Refine feature interface. Start prep of new vit weights. 2024-04-30 16:56:33 -07:00
Ross Wightman 4b2565e4cb More forward_intermediates() / FeatureGetterNet work
* include relpos vit
* refactor reduction / size calcs so hybrid vits work and dynamic_img_size works
* fix -ve feature indices when pruning
* fix mvitv2 w/ class token
* refine naming
* add tests
2024-04-10 15:11:34 -07:00
Ross Wightman 679daef76a More forward_intermediates() & features_only work
* forward_intermediates() added to beit, deit, eva, mvitv2, twins, vit, vit_sam
* add features_only to forward intermediates to allow just intermediate features
* fix #2060
* fix #1374
* fix #657
2024-04-09 21:29:16 -07:00
Ross Wightman 5fdc0b4e93 Exploring vit features_only using get_intermediate_layers() as per #2131 2024-04-07 11:24:45 -07:00
Ross Wightman c559c3911f Improve vit conversions. OpenAI convert pass through main convert for patch & pos resize. Fix #2120 2024-03-21 10:00:43 -07:00
Ross Wightman 256cf19148 Rename tinyclip models to fit existing 'clip' variants, use consistently mapped OpenCLIP compatible checkpoint on hf hub 2024-03-20 15:21:46 -07:00
Thien Tran 1a1d07d479 add other tinyclip 2024-03-19 07:27:09 +08:00
Thien Tran dfffffac55 add tinyclip 8m 2024-03-19 07:02:17 +08:00
Cheng-Ling Lai db06b56d34
Saved computational costs of get_intermediate_layers() from unused blocks 2024-03-17 21:34:06 +08:00
Cheng-Ling Lai 4731e4efc4
Modified ViT get_intermediate_layers() to support dynamic image size 2024-03-16 23:07:21 +08:00
Ross Wightman ac1b08deb6 fix_init on vit & relpos vit 2024-02-10 20:15:37 -08:00
Ross Wightman 87fec3dc14 Update experimental vit model configs 2024-02-10 16:05:58 -08:00
Ross Wightman ada145b016 Literal use w/ python < 3.8 requires typing_extension, cach instead of check sys ver 2023-11-21 09:48:03 -08:00