Commit Graph

1786 Commits (c8c4f256b8c279575ed87012f38620a4cd25df57)

Author SHA1 Message Date
Ross Wightman 4542cf03f9 Add features_only, other bits to mambaout, define different base alternatives 2024-10-09 14:11:40 -07:00
Ross Wightman c2da12c7e1 Update rw models, fix heads 2024-10-09 14:11:40 -07:00
Ross Wightman f2086f51a0 Add mambaout builder support, pretrained weight remap 2024-10-09 14:11:40 -07:00
Ross Wightman c6ef54eefa Initial mambaout work 2024-10-09 14:11:40 -07:00
Ross Wightman d9321b0e10 Add weights for fine-tuned siglip so400m. Add webli_i18n pretrained tags for the multi-lingual model variants (incl older base) 2024-10-09 09:04:44 -07:00
Ross Wightman 01b62264af Add i18n variant of so400m model w/ weights. Add two in1k fine-tunes of original so400m 384x384 but at 378x378 (better matches patch14) 2024-10-08 23:40:24 -07:00
Ross Wightman 72f0edb7e8 missed first_conv for rnv2 18d 2024-10-08 12:38:54 -07:00
Ross Wightman 3ed603a2ce Add resnet18/18d pre-act model configs for potential training. Fix #2289 2024-10-08 11:28:07 -07:00
Ross Wightman 41a79e0fcb Add overlapped stem convnext zepto weights 2024-10-08 11:26:34 -07:00
Ross Wightman 545bd4056c Tag along test_vit3 weights 2024-09-30 12:03:32 -07:00
Ross Wightman 69b687d4cc Add zepto weights 2024-09-30 11:43:23 -07:00
Ross Wightman c6e5557a5a Mismatch pretrained_cfg 2024-09-30 11:43:23 -07:00
Ross Wightman 5d7bd2973e convnext zepto, rmsnorm experiments 2024-09-30 11:43:23 -07:00
Ross Wightman e3242a5258
Merge pull request #2277 from huggingface/more_tiny_test_models
Adding some more tiny test models to train...
2024-09-22 10:28:29 -07:00
Ross Wightman c1cb5641c7 Add weight fore mobilenetv4 small 0.5, change 0.25 -> 0.35 2024-09-22 10:27:01 -07:00
Ross Wightman a22ce0a329 Merge branch 'patch-1' of https://github.com/baorepo/pytorch-image-models into baorepo-patch-1 2024-09-22 10:14:35 -07:00
Ross Wightman 9067be6a30 Add weights for new tiny test models 2024-09-22 07:59:23 -07:00
Ross Wightman 65564f7da5 Fix reversed H & W padding for swin patch merging 2024-09-21 16:51:02 -07:00
Ross Wightman a2f539f055 Add a few more test model defs in prep for weight upload 2024-09-21 11:38:38 -07:00
Ross Wightman 6ab2af610d Adding some more tiny test models to train 2024-09-06 15:35:57 -07:00
alias pillar1989 d6b8816eda MobilenetV4: add two more lightweight models
Mobilenetv4 is very fast and ideal for embedded devices. However, for many low-cost, low-power embedded MCU devices, smaller models are required. Hopefully this PR will merge.
2024-09-05 02:34:11 +00:00
Ross Wightman f81cbdcca9
Merge pull request #2274 from huggingface/bulk_runner_tweaks
Better all res resolution for bulk runner
2024-09-03 12:11:56 -07:00
Ross Wightman 6c42299b3e Back to dev version 2024-09-03 10:29:05 -07:00
Ross Wightman a50713ce6e Fix #2272 2024-09-02 13:20:05 -07:00
Ross Wightman ebbe530ee4 Add MobileNetV3 RA4 (mnv4 recipe) weights 2024-09-02 13:10:34 -07:00
Ross Wightman fa4a1e597f Better all res resolution for bulk runner 2024-08-26 22:28:01 -07:00
Ross Wightman 072783331f Version 1.0.9 2024-08-23 11:47:42 -07:00
Ross Wightman 76b0e9931a Placeholder for new mnv3 model 2024-08-23 10:11:20 -07:00
Ross Wightman 39e92f0c0d mobilenet_edgetpu can use group_size override, more consistency in arg wrap/sadface w/ extra group_size arg 2024-08-22 11:44:02 -07:00
Ross Wightman b9f020a509 Allow group_size override for more efficientnet and mobilenetv3 based models 2024-08-21 16:51:38 -07:00
Ross Wightman 17923a66bb Add layer scale to hieradet 2024-08-21 11:23:39 -07:00
Ross Wightman 47e6958263 Add hierdet_small (non sam) model def 2024-08-21 11:05:54 -07:00
Ross Wightman 9fcbf39cdc Add remaining sbb vit betwixt/mediumd fine-tunes 2024-08-21 10:09:38 -07:00
Ross Wightman dc94cca0e5 Remaining Hiera sbb weights uploaded 2024-08-21 10:06:27 -07:00
Ross Wightman a256e50457 Move padding back in front of windowing 2024-08-17 11:22:53 -07:00
Ross Wightman 7d83749207 pool size test fixes 2024-08-17 08:27:13 -07:00
Ross Wightman 1bd92bca0e Add fused_attn flag to HieraDet attn block 2024-08-16 22:57:49 -07:00
Ross Wightman 691bb54443 Larger min input size needed 2024-08-16 17:09:19 -07:00
Ross Wightman de3a91a7a0 Add min_input_size of 128 for hieradet/sam2 2024-08-16 15:13:56 -07:00
Ross Wightman 0b05122cda Fixing hieradet (sam2) tests 2024-08-16 14:33:40 -07:00
Ross Wightman e035381171 Move padding out of windowing code for hieradet, fix torchscript typing issues, make pooling MaxPool unique instances across two modules 2024-08-16 13:36:33 -07:00
Ross Wightman 146c2fbe34 Add resnet50d and efficientnet_b1 ra4 (mnv4) hparam weights 2024-08-16 12:10:00 -07:00
Ross Wightman 962958723c More Hiera updates. Add forward_intermediates to hieradat/sam2 impl. Make both use same classifier module. Add coarse bool to intermediates. 2024-08-16 11:10:04 -07:00
Ross Wightman f2cfb4c677 Add WIP HieraDet impl (SAM2 backbone support) 2024-08-15 17:58:15 -07:00
Ross Wightman a50e53d41f Rename global pos embed for Hiera abswin, factor out commonly used vit weight init fns to layers. Add a channels-last ver of normmlp head. 2024-08-15 17:46:36 -07:00
Ross Wightman 2f3fed43b8 Fix hiera init with num_classes=0, fix weight tag names for sbb2 hiera/vit weights, add LayerScale/LayerScale2d to layers 2024-08-15 11:14:38 -07:00
Ross Wightman fee91fdd41 Update Hiera model for abswin, more stable weight init, layer-scale. ImageNet-12k weights for hiera_small_abswin, and two of the sbb vits with improved reg4 init. 2024-08-14 12:22:40 -07:00
Ross Wightman ac3470188b Add mobilenetv1_125 weights from ra4 / mnv4 hparam series 2024-08-14 08:47:02 -07:00
Ross Wightman a7b0bfc376 exception for unknown arg should have been TypeError 2024-08-12 16:09:36 -07:00
Ross Wightman bd0f79153b Add weights only flag to avoid warning, try to keep bwd compat. Default to True for remote load of pretrained weights, keep False for local checkpoing load to avoid training checkpoint breaks.. fix #2249 2024-08-12 11:41:41 -07:00
Ross Wightman 531215eded Bring changelog up to current, back to dev version, 1.0.9.dev0 2024-08-08 09:17:23 -07:00
Ross Wightman bf781bc3c2
Merge pull request #2245 from dhkim0225/feature/rdnet
Feature: Add RDNet
2024-08-07 16:19:22 -07:00
Ross Wightman 49f6e1da96 Remove eval to avoid safety concerns 2024-08-07 14:54:13 -07:00
Ross Wightman 2d5c9bf60d
Merge pull request #2253 from brendanartley/hgnet-grad-checkpointing
Add gradient checkpointing to hgnets
2024-08-07 12:45:14 -07:00
Ross Wightman e9ef9424f0 Add a few missing __all__ entries. 2024-08-07 09:35:51 -07:00
brendanartley 232a459e36 Added gradient checkpointing to hgnet 2024-08-07 09:22:51 -07:00
dong-hyun ea33d6a74d bugfix pool_size 2024-08-03 09:50:20 +09:00
Ross Wightman 10344625be proj_drop wasn't being called in sam vit 2024-08-02 16:12:02 -07:00
dong-hyun 6fdc9d8243 update code for ci pass 2024-08-02 15:14:40 +09:00
dong-hyun 7a866b6521 update code for torchscript 2024-08-02 09:58:13 +09:00
dong-hyun 8248122f82 add rdnet 2024-08-01 14:54:29 +09:00
dong-hyun 025259024d add rdnet 2024-08-01 14:51:15 +09:00
dong-hyun 225f4f92b3 add rdnet 2024-08-01 14:49:21 +09:00
Ross Wightman 4a10302754 Add mobilenet_edgetpu_v2_m weights 2024-07-28 17:19:36 -07:00
Ross Wightman 70182ea0cd Version 1.0.8 2024-07-26 17:05:36 -07:00
Ross Wightman ab8cb070fc Add xavier_uniform init of MNVC hybrid attention modules. Small improvement in training stability. 2024-07-26 17:03:40 -07:00
Ross Wightman cec70b6779
Merge pull request #2225 from huggingface/small_things
Small things
2024-07-25 20:29:13 -07:00
Ross Wightman 61df3fde89 Wrong hybrid_medium in12k pool sizes 2024-07-25 15:39:21 -07:00
Ross Wightman 9aa2930760 Add latest mobilenetv4 and baseline updates for mobilenetv1 and efficientnet_b0 weights 2024-07-25 14:20:54 -07:00
Ross Wightman 7b6a406474 remove swin debug prints 2024-07-24 21:05:56 -07:00
Ross Wightman 4c531be479 set_input_size(), always_partition, strict_img_size, dynamic mask option for all swin models. More flexibility in resolution, window resizing. 2024-07-24 16:41:31 -07:00
Ross Wightman 8efdc38213 Fix #2242 add checks for out indices with intermediate getter mode 2024-07-23 08:19:09 -07:00
Ross Wightman d2240745d3 Fix issue where feature out_indices out of order after wrapping with FeatureGetterNet due to use of set() 2024-07-22 13:33:30 -07:00
Ross Wightman 2b3f1a4633 Make channels for classic resnet configurable 2024-07-22 10:47:40 -07:00
Ross Wightman 9b2b8014e8 Add weights for test models 2024-07-22 10:08:57 -07:00
Ross Wightman 7e0caa1ba3 Padding helpers work if tuples/lists passed 2024-07-19 14:28:03 -07:00
Ross Wightman 2180800646 MQA query_strides bugs fix #2237. No padding for avg_pool2d if not 'same', use scale_factor for Upsample. 2024-07-19 14:26:54 -07:00
Ross Wightman 1a05ed29a1 Add to 'abswin' hiera models for train trials 2024-07-19 11:05:31 -07:00
Ross Wightman 0cbf4fa586 _orig_mod still causing issues even though I thought it was fixed in pytorch, add unwrap / clean helpers 2024-07-19 11:03:45 -07:00
Feraidoon Mehri 4cca568bd8
eva.py: fixed bug in applying attention mask
The mask should be applied before the softmax.
2024-07-19 15:12:04 +03:30
Ross Wightman 3a8a965891 Implement absolute+window pos embed for hiera, resizable but needs new weights 2024-07-18 21:43:37 -07:00
Ross Wightman 7160af4a24
Merge pull request #2229 from Promisery/reg_token
Initialize weights of reg_token for ViT
2024-07-18 09:25:29 -07:00
Ross Wightman 392b78aee7 set_input_size initial impl for vit & swin v1. Move HybridEmbed to own location in timm/layers 2024-07-17 15:25:48 -07:00
Ross Wightman 34c9fee554 Fix pass through of input / target keys so ImageDataset readers so args work with hfds instead of just hfids (iterable) 2024-07-17 10:11:46 -07:00
Promisery 417cf7f871 Initialize weights of reg_token for ViT 2024-07-13 11:11:42 +08:00
Ross Wightman f920119f3b Fixing tests 2024-07-09 14:53:20 -07:00
Ross Wightman 644abf9588 Fix default_cfg test for mobilenet_100 2024-07-09 12:52:24 -07:00
Ross Wightman d5afe106dc Merge remote-tracking branch 'origin/tiny_test_models' into small_things 2024-07-09 12:49:57 -07:00
Ross Wightman 55101028bb Rename test_tiny* -> test*. Fix ByobNet BasicBlock attn location and add test_byobnet model. 2024-07-09 11:53:11 -07:00
Ross Wightman 1334598462 Add support back to EfficientNet to disable head_conv / bn2 so mobilnetv1 can be implemented properly 2024-07-08 13:51:26 -07:00
Ross Wightman 800405d941 Add conv_large mobilenetv3 aa/blur model defs 2024-07-08 13:50:05 -07:00
Ross Wightman f81b094aaa Add 'qkv_bias_separate' flag for EVA/beit/swinv2 attn modules to allow an override for easy quantization wrappers. Fix #2098 2024-07-08 13:48:38 -07:00
Ross Wightman 83c2c2f0c5 Add 'Maybe' PIL / image tensor conversions in case image alread in tensor format 2024-07-08 13:43:51 -07:00
Steffen Schneider c01a47c9e7
Fix typo in type annotations in timm.models.hrnet 2024-07-08 00:53:16 +02:00
Daniel Suess 197c10463b Fix jit.script breaking with features_fx 2024-06-28 03:58:51 +00:00
Ross Wightman b751da692d Add latest ix (xavier init for mqa) hybrid medium & large weights for MobileNetV4 2024-06-24 13:49:55 -07:00
Ross Wightman d4d4d84fda Dev version 1.0.8.dev0 2024-06-24 11:34:13 -07:00
Ross Wightman f8342a045a
Merge pull request #2213 from huggingface/florence2
Fix #2212 map florence2 image tower to davit with a few changes
2024-06-24 11:01:08 -07:00
Sejik c33a001397
Fix typo 2024-06-24 21:54:38 +09:00
Ross Wightman 02d0f27721 cleanup davit padding 2024-06-22 12:06:46 -07:00
Ross Wightman c715c724e7 Fix tracing by removing float cast, should end up float anyways 2024-06-22 08:35:30 -07:00
Ross Wightman fb58a73033 Fix #2212 map florence2 image tower to davit with a few changes 2024-06-21 15:31:29 -07:00
Ross Wightman b28945ff05 Version 1.0.7, prep for release 2024-06-18 16:19:43 -07:00
Ross Wightman fb13e6385e
Merge pull request #2203 from huggingface/more_mobile
Add mobilenet edgetpu defs for exp, add ol mobilenet v1 back for comp…
2024-06-18 15:20:01 -07:00
Ross Wightman 16e082e1c2 Add mobilenetv4 hybrid-large weights 2024-06-17 11:08:31 -07:00
Ross Wightman e41125cc83
Merge pull request #2209 from huggingface/fcossio-vit-maxpool
ViT pooling refactor
2024-06-17 07:51:12 -07:00
Ross Wightman a22466852d Add 2400 epoch mobilenetv4 small weights, almost at paper, rounds to 73.8 2024-06-16 10:51:00 -07:00
Ross Wightman b1a6f4a946 Some missed reset_classifier() type annotations 2024-06-16 10:39:27 -07:00
Ross Wightman 71101ebba0 Refactor vit pooling to add more reduction options, separately callable 2024-06-14 23:16:58 -07:00
Ross Wightman a0bb5b4a44 Missing stem_kernel_size argument in EfficientNetFeatures 2024-06-14 13:39:31 -07:00
Fernando Cossio 9567cf6d84
Feature: add option global_pool='max' to VisionTransformer
Most of the CNNs have a max global pooling option. I would like to extend ViT to have this option.
2024-06-14 15:24:54 +02:00
Ross Wightman 9613c76844 Add mobilenet edgetpu defs for exp, add ol mobilenet v1 back for completeness / comparison 2024-06-13 17:33:04 -07:00
Ross Wightman 22de845add
Prepping for final MobileCLIP weight locations (#2199)
* Prepping for final MobileCLIP weight locations

* Update weight locations to coreml-projects

* Update mobileclip weight locations with final apple org location
2024-06-13 16:55:49 -07:00
Ross Wightman 575978ba55 Add mnv4_conv_large 384x384 weight location 2024-06-13 12:58:04 -07:00
Ross Wightman 7b5f17d1bd Update README.md, bump dev version 1.0.6 2024-06-12 12:35:44 -07:00
Ross Wightman e42e453128 Fix mmnv4 conv_large weight link, reorder mnv4 pretrained cfg for proper precedence 2024-06-12 11:16:49 -07:00
Ross Wightman 7b0a5321cb
Merge pull request #2198 from huggingface/openai_clip_resnet
Mapping OpenAI CLIP Modified ResNet weights -> ByobNet.
2024-06-12 09:33:30 -07:00
Ross Wightman 57adc1acc8 Fix rotary embed version of attn pool. Bit of cleanup/naming 2024-06-11 23:49:17 -07:00
Ross Wightman cdc7bcea69 Make 2d attention pool modules compatible with head interface. Use attention pool in CLIP ResNets as head. Make separate set of GAP models w/ avg pool instead of attn pool. 2024-06-11 21:32:07 -07:00
Ross Wightman c63da1405c Pretrained cfg name mismatch 2024-06-11 21:16:54 -07:00
Ross Wightman 88efca1be2 First set of MobileNetV4 weights trained in timm 2024-06-11 18:53:01 -07:00
Ross Wightman 30ffa152de Fix load of larger ResNet CLIP models, experimenting with making AttentionPool *the* head, seems to fine-tune better, one less layer. 2024-06-10 12:07:14 -07:00
Ross Wightman 5e9ff5798f Adding pos embed resize fns to FX autowrap exceptions 2024-06-10 12:06:47 -07:00
Ross Wightman f0fb471b26 Remove separate ConvNormActAa class, merge with ConvNormAct 2024-06-10 12:05:35 -07:00
Ross Wightman 5efa15b2a2 Mapping OpenAI CLIP Modified ResNet weights -> ByobNet. Improve AttentionPool2d layers. Fix #1731 2024-06-09 16:54:48 -07:00
Ross Wightman 7702d9afa1 ViTamin in_chans !=3 weight load fix 2024-06-07 20:39:23 -07:00
Ross Wightman 66a0eb4673 Experimenting with tiny test models, how small can they go and be useful for regression tests? 2024-06-07 16:09:25 -07:00
Ross Wightman 5ee06760dc Fix classifier input dim for mnv3 after last changes 2024-06-07 13:53:13 -07:00
Ross Wightman a5a2ad2e48 Fix consistency, testing for forward_head w/ pre_logits, reset_classifier, models with pre_logits size != unpooled feature size
* add test that model supports forward_head(x, pre_logits=True)
* add head_hidden_size attr to all models and set differently from num_features attr when head has hidden layers
* test forward_features() feat dim == model.num_features and pre_logits feat dim == self.head_hidden_size
* more consistency in reset_classifier signature, add typing
* asserts in some heads where pooling cannot be disabled
Fix #2194
2024-06-07 13:53:00 -07:00
Ross Wightman 4535a5412a Change default serialization for push_to_hf_hub to 'both' 2024-06-07 13:40:31 -07:00
Ross Wightman 5cce2185e1
Update version.py 2024-06-07 13:13:23 -07:00
Ross Wightman 7ccb10ebff Disable efficient_builder debug flag 2024-06-06 21:50:27 -07:00
Ross Wightman ad026e6e33 Fix in_chans switching on create 2024-06-06 17:56:14 -07:00
Ross Wightman fc1b66a51d Fix first conv name for mci vit-b 2024-06-06 13:42:26 -07:00
Ross Wightman 88a1006e02 checkpoint filter fns with consistent name, add mobileclip-b pretrained cfgs 2024-06-06 12:38:52 -07:00
Ross Wightman 7d4ada6d16 Update ViTamin model defs 2024-06-06 09:16:43 -07:00
Ross Wightman cc8a03daac Add ConvStem and MobileCLIP hybrid model for B variant. Add full norm disable support to ConvNormAct layers 2024-06-06 09:15:27 -07:00
Ross Wightman 3c9d8e5b33 Merge remote-tracking branch 'origin/efficientnet_x' into fastvit_mobileclip 2024-06-05 17:35:15 -07:00
Ross Wightman 5756a81c55 Merge remote-tracking branch 'origin/Beckschen-vitamin' into fastvit_mobileclip 2024-06-05 15:20:54 -07:00
Ross Wightman 58591a97f7 Enable features_only properly 2024-06-04 16:57:16 -07:00
Ross Wightman 1b66ec7cf3 Fixup ViTamin, add hub weight reference 2024-06-03 17:14:03 -07:00
Ross Wightman b2c0aeb0ec Merge branch 'main' of https://github.com/Beckschen/pytorch-image-models into Beckschen-vitamin 2024-06-02 14:16:30 -07:00
Ross Wightman 7f96538052 Add missing lkc act for mobileclip fastvits 2024-05-31 11:59:51 -07:00
Ross Wightman a503639bcc Add mobileclip fastvit model defs, support extra SE. Add forward_intermediates API to fastvit 2024-05-30 10:17:38 -07:00
Ross Wightman 5fa6efa158 Add anti-aliasing support to mobilenetv3 and efficientnet family models. Update MobileNetV4 model defs, resolutions. Fix #599
* create_aa helper function centralized for all timm uses (resnet, convbnact helper)
* allow BlurPool w/ pre-defined channels (expand)
* mobilenetv4 UIB block using ConvNormAct layers for improved clarity, esp with AA added
* improve more mobilenetv3 and efficientnet related type annotations
2024-05-27 22:06:22 -07:00
Ross Wightman 5dce710101 Add vit_little in12k + in12k-ft-in1k weights 2024-05-27 14:56:03 -07:00
Ross Wightman 3c0283f9ef Fix reparameterize for NextViT. Fix #2187 2024-05-27 14:48:58 -07:00
Ross Wightman 4ff7c25766 Pass layer_scale_init_value to Mnv3Features module 2024-05-24 16:44:50 -07:00
Ross Wightman a12b72b5c4 Fix missing head_norm arg pop for feature model 2024-05-24 15:50:34 -07:00
Ross Wightman 7fe96e7a92 More MobileNet-v4 fixes
* missed final norm after post pooling 1x1 PW head conv
* improve repr of model by flipping a few modules to None when not used, nn.Sequential for MultiQueryAttention query/key/value/output
* allow layer scaling to be enabled/disabled at model variant level, conv variants don't use it
2024-05-24 15:09:29 -07:00
Ross Wightman 28d76a97db Mixed up kernel size for last blocks in mnv4-conv-small 2024-05-24 11:50:42 -07:00
Ross Wightman 0c6a69e7ef Add comments to MNV4 model defs with block variants 2024-05-23 15:54:05 -07:00
Ross Wightman cb33956b20 Fix some mistakes in mnv4 model defs 2024-05-23 14:24:32 -07:00
Ross Wightman 70176a2dae torchscript typing fixes 2024-05-23 11:43:05 -07:00
Ross Wightman 2a1a6b1236 Adding missing attention2d.py 2024-05-23 11:06:32 -07:00
Ross Wightman cee79dada0 Merge remote-tracking branch 'origin/main' into efficientnet_x 2024-05-23 11:01:39 -07:00
Ross Wightman 6a8bb03330 Initial MobileNetV4 pass 2024-05-23 10:49:18 -07:00
Ross Wightman e748805be3 Add regex matching support to AttentionExtract. Add return_dict support to graph extractors and use returned output in AttentionExtractor 2024-05-22 14:33:39 -07:00
Ross Wightman 44f72c04b3 Change node/module name matching for AttentionExtract so it keeps outputs in order. #1232 2024-05-22 13:45:25 -07:00
Ross Wightman 84cb225ecb Add in12k + 12k_ft_in1k vit_medium weights 2024-05-20 15:52:46 -07:00
Ross Wightman 4634c3e134 Version 1.0.4.dev0 2024-05-20 15:52:27 -07:00
Beckschen 7a2ad6bce1 Add link to model weights on Hugging Face 2024-05-17 06:51:35 -04:00
Beckschen 530fb49e7e Add link to model weights on Hugging Face 2024-05-17 06:48:59 -04:00
Fernando Cossio 9b11801cb4
Credit earlier work with the same idea.
Hi, this earlier work has the same name and idea behind this layer. It could be useful for readers to keep both links here if they want to see the effects of introducing this layer on a very different domain. 😄
2024-05-16 22:50:34 +02:00
Ross Wightman cb0e4391be Release 1.0.3 2024-05-15 11:06:22 -07:00
Ross Wightman 27fd2f35d3
Merge pull request #2181 from huggingface/Delaunay-dist-backend
Delaunay dist backend flag
2024-05-15 10:00:59 -07:00
Ross Wightman e57625e814 Tweak dist_backend to use device_type (before possible :) 2024-05-15 08:49:25 -07:00
Ross Wightman 6ca92570f7 Merge branch 'patch-1' of https://github.com/Delaunay/pytorch-image-models into Delaunay-dist-backend 2024-05-15 08:40:58 -07:00
Ross Wightman cd0e7b11ff
Merge pull request #2180 from yvonwin/main
Remove a duplicate function in mobilenetv3.py
2024-05-15 07:54:17 -07:00
Ross Wightman 83aee5c28c Add explicit GAP (avg pool) variants of other SigLIP models. 2024-05-15 07:53:19 -07:00
yvonwin 58f2f79b04 Remove a duplicate function in mobilenetv3.py: `_gen_lcnet` is repeated in mobilenetv3.py.Remove the duplicate code. 2024-05-15 17:59:34 +08:00
Ross Wightman 7b3b11b63f Support loading of paligemma weights into GAP variants of SigLIP ViT. Minor tweak to npz loading for packed transformer weights. 2024-05-14 15:44:37 -07:00
Beckschen df304ffbf2 the dataclass init needs to use the default factory pattern, according to Ross 2024-05-14 15:10:05 -04:00
Ross Wightman cc5f2f6f70 version 1.0.2dev0 2024-05-13 15:25:15 -07:00
Ross Wightman 3bfd036b58 Add normalize flag to transforms factory, allow return of non-normalized native dtype torch.Tensors 2024-05-13 15:23:25 -07:00
Ross Wightman a69863ad61
Merge pull request #2156 from huggingface/hiera
WIP Hiera implementation.
2024-05-13 14:58:12 -07:00
Setepenre 8848dad362
Update distributed.py 2024-05-13 16:55:42 -04:00
Ross Wightman f7aa0a1a71 Add missing vit_wee weight 2024-05-13 12:05:47 -07:00
Ross Wightman 7a4e987b9f Hiera weights on hub 2024-05-13 11:43:22 -07:00
Ross Wightman 23f09af08e Merge branch 'main' into efficientnet_x 2024-05-12 21:31:08 -07:00
Ross Wightman c838c4233f Add typing to reset_classifier() on other models 2024-05-12 11:12:00 -07:00
Ross Wightman 3e03b2bf3f Fix a few more hiera API issues 2024-05-12 11:11:45 -07:00
Ross Wightman 211d18d8ac Move norm & pool into Hiera ClassifierHead. Misc fixes, update features_intermediate() naming 2024-05-11 23:37:35 -07:00
Ross Wightman 2ca45a4ff5 Merge remote-tracking branch 'upstream/main' into hiera 2024-05-11 15:43:05 -07:00
Ross Wightman 1d3ab176bc Remove debug / staging code 2024-05-10 22:16:34 -07:00
Ross Wightman aa4d06a11c sbb vit weights on hub, testing 2024-05-10 17:15:01 -07:00
Ross Wightman 3582ca499e Prepping weight push, benchmarking. 2024-05-10 14:14:06 -07:00
Ross Wightman 2bfa5e5d74 Remove JIT activations, take jit out of ME activations. Remove other instances of torch.jit.script. Breaks torch.compile and is much less performant. Remove SpaceToDepthModule 2024-05-06 16:32:49 -07:00
Beckschen 99d4c7d202 add ViTamin models 2024-05-05 02:50:14 -04:00
Ross Wightman 07535f408a Add AttentionExtract helper module 2024-05-04 14:10:00 -07:00
Ross Wightman 45b7ae8029 forward_intermediates() support for byob/byoanet models 2024-05-04 14:06:52 -07:00
Ross Wightman c4b8897e9e attention -> attn in davit for model consistency 2024-05-04 14:06:11 -07:00
Ross Wightman cb57a96862 Fix early stop for efficientnet/mobilenetv3 fwd inter. Fix indices typing for all fwd inter. 2024-05-04 10:21:58 -07:00
Ross Wightman 01dd01b70e forward_intermediates() for MlpMixer models and RegNet. 2024-05-04 10:21:03 -07:00
Ross Wightman f8979d4f50 Comment out time local files while testing new vit weights 2024-05-03 20:26:56 -07:00
Ross Wightman c719f7eb86 More forward_intermediates() updates
* add convnext, resnet, efficientformer, levit support
* remove kwargs only for fn so that torchscript isn't broken for all :(
* use reset_classifier() consistently in prune
2024-05-03 16:22:32 -07:00
Ross Wightman 301d0bb21f Stricter check on pool_type for adaptive pooling module. Fix #2159 2024-05-03 16:16:51 -07:00
Ross Wightman d6da4fb01e Add forward_intermediates() to efficientnet / mobilenetv3 based models as an exercise. 2024-05-02 14:19:16 -07:00
Ross Wightman c22efb9765 Add wee & little vits for some experiments 2024-05-02 10:51:35 -07:00
Ross Wightman 67332fce24 Add features_intermediate() support to coatnet, maxvit, swin* models. Refine feature interface. Start prep of new vit weights. 2024-04-30 16:56:33 -07:00