Commit Graph

2681 Commits (9ab5464e4d7fa6bd2350eda7d93d5ec3fe8b9927)
 

Author SHA1 Message Date
Ross Wightman 7fe96e7a92 More MobileNet-v4 fixes
* missed final norm after post pooling 1x1 PW head conv
* improve repr of model by flipping a few modules to None when not used, nn.Sequential for MultiQueryAttention query/key/value/output
* allow layer scaling to be enabled/disabled at model variant level, conv variants don't use it
2024-05-24 15:09:29 -07:00
Ross Wightman 28d76a97db Mixed up kernel size for last blocks in mnv4-conv-small 2024-05-24 11:50:42 -07:00
Ross Wightman 0c6a69e7ef Add comments to MNV4 model defs with block variants 2024-05-23 15:54:05 -07:00
Ross Wightman cb33956b20 Fix some mistakes in mnv4 model defs 2024-05-23 14:24:32 -07:00
Ross Wightman 70176a2dae torchscript typing fixes 2024-05-23 11:43:05 -07:00
Ross Wightman 2a1a6b1236 Adding missing attention2d.py 2024-05-23 11:06:32 -07:00
Ross Wightman cee79dada0 Merge remote-tracking branch 'origin/main' into efficientnet_x 2024-05-23 11:01:39 -07:00
Ross Wightman 6a8bb03330 Initial MobileNetV4 pass 2024-05-23 10:49:18 -07:00
Ross Wightman e748805be3 Add regex matching support to AttentionExtract. Add return_dict support to graph extractors and use returned output in AttentionExtractor 2024-05-22 14:33:39 -07:00
Ross Wightman 44f72c04b3 Change node/module name matching for AttentionExtract so it keeps outputs in order. #1232 2024-05-22 13:45:25 -07:00
Ross Wightman 84cb225ecb Add in12k + 12k_ft_in1k vit_medium weights 2024-05-20 15:52:46 -07:00
Ross Wightman 4634c3e134 Version 1.0.4.dev0 2024-05-20 15:52:27 -07:00
Ross Wightman 202199cac3
Merge pull request #2183 from lixit/patch-1
Fix a typo
2024-05-20 13:45:32 -07:00
xitong 14dc2c4df6
fix a typo 2024-05-20 14:34:39 -04:00
Beckschen 7a2ad6bce1 Add link to model weights on Hugging Face 2024-05-17 06:51:35 -04:00
Beckschen 530fb49e7e Add link to model weights on Hugging Face 2024-05-17 06:48:59 -04:00
Ross Wightman 6b9e38b589
Merge pull request #2182 from fcossio/patch-1
Credit earlier work with the same idea
2024-05-16 15:10:46 -07:00
Fernando Cossio 9b11801cb4
Credit earlier work with the same idea.
Hi, this earlier work has the same name and idea behind this layer. It could be useful for readers to keep both links here if they want to see the effects of introducing this layer on a very different domain. 😄
2024-05-16 22:50:34 +02:00
Ross Wightman cb0e4391be Release 1.0.3 2024-05-15 11:06:22 -07:00
Ross Wightman 27fd2f35d3
Merge pull request #2181 from huggingface/Delaunay-dist-backend
Delaunay dist backend flag
2024-05-15 10:00:59 -07:00
Ross Wightman e57625e814 Tweak dist_backend to use device_type (before possible :) 2024-05-15 08:49:25 -07:00
Ross Wightman 6ca92570f7 Merge branch 'patch-1' of https://github.com/Delaunay/pytorch-image-models into Delaunay-dist-backend 2024-05-15 08:40:58 -07:00
Ross Wightman cd0e7b11ff
Merge pull request #2180 from yvonwin/main
Remove a duplicate function in mobilenetv3.py
2024-05-15 07:54:17 -07:00
Ross Wightman 83aee5c28c Add explicit GAP (avg pool) variants of other SigLIP models. 2024-05-15 07:53:19 -07:00
yvonwin 58f2f79b04 Remove a duplicate function in mobilenetv3.py: `_gen_lcnet` is repeated in mobilenetv3.py.Remove the duplicate code. 2024-05-15 17:59:34 +08:00
Ross Wightman f04802fb8d
Update README.md 2024-05-14 17:19:34 -07:00
Ross Wightman 6653747d95
Merge pull request #2178 from huggingface/pali_siglip
Support loading of PaliGemma weights into GAP variants of SigLIP ViT.
2024-05-14 17:10:38 -07:00
Ross Wightman 7b3b11b63f Support loading of paligemma weights into GAP variants of SigLIP ViT. Minor tweak to npz loading for packed transformer weights. 2024-05-14 15:44:37 -07:00
Ross Wightman 04462f554f Fix typos in feat extract doc 2024-05-14 15:33:39 -07:00
Beckschen df304ffbf2 the dataclass init needs to use the default factory pattern, according to Ross 2024-05-14 15:10:05 -04:00
Ross Wightman cc5f2f6f70 version 1.0.2dev0 2024-05-13 15:25:15 -07:00
Ross Wightman 3bfd036b58 Add normalize flag to transforms factory, allow return of non-normalized native dtype torch.Tensors 2024-05-13 15:23:25 -07:00
Ross Wightman a69863ad61
Merge pull request #2156 from huggingface/hiera
WIP Hiera implementation.
2024-05-13 14:58:12 -07:00
Setepenre 8848dad362
Update distributed.py 2024-05-13 16:55:42 -04:00
Ross Wightman f7aa0a1a71 Add missing vit_wee weight 2024-05-13 12:05:47 -07:00
Ross Wightman 7a4e987b9f Hiera weights on hub 2024-05-13 11:43:22 -07:00
Ross Wightman 23f09af08e Merge branch 'main' into efficientnet_x 2024-05-12 21:31:08 -07:00
Ross Wightman c838c4233f Add typing to reset_classifier() on other models 2024-05-12 11:12:00 -07:00
Ross Wightman 3e03b2bf3f Fix a few more hiera API issues 2024-05-12 11:11:45 -07:00
Ross Wightman 211d18d8ac Move norm & pool into Hiera ClassifierHead. Misc fixes, update features_intermediate() naming 2024-05-11 23:37:35 -07:00
Ross Wightman 2ca45a4ff5 Merge remote-tracking branch 'upstream/main' into hiera 2024-05-11 15:43:05 -07:00
Ross Wightman 49de391470
Update feature_extraction.mdx 2024-05-11 12:14:02 -07:00
Ross Wightman a193d20b7b
Update README.md 2024-05-11 11:01:05 -07:00
Ross Wightman b2c10fec05
Merge pull request #2168 from huggingface/more_vit_better_getter_redux
A few more features_intermediate() models, AttentionExtract helper, related minor cleanup.
2024-05-11 08:13:50 -07:00
Ross Wightman 1d3ab176bc Remove debug / staging code 2024-05-10 22:16:34 -07:00
Ross Wightman aa4d06a11c sbb vit weights on hub, testing 2024-05-10 17:15:01 -07:00
Ross Wightman 3582ca499e Prepping weight push, benchmarking. 2024-05-10 14:14:06 -07:00
Ross Wightman 2bfa5e5d74 Remove JIT activations, take jit out of ME activations. Remove other instances of torch.jit.script. Breaks torch.compile and is much less performant. Remove SpaceToDepthModule 2024-05-06 16:32:49 -07:00
Beckschen 99d4c7d202 add ViTamin models 2024-05-05 02:50:14 -04:00
Ross Wightman 07535f408a Add AttentionExtract helper module 2024-05-04 14:10:00 -07:00