Ross Wightman
7fe96e7a92
More MobileNet-v4 fixes
...
* missed final norm after post pooling 1x1 PW head conv
* improve repr of model by flipping a few modules to None when not used, nn.Sequential for MultiQueryAttention query/key/value/output
* allow layer scaling to be enabled/disabled at model variant level, conv variants don't use it
2024-05-24 15:09:29 -07:00
Ross Wightman
28d76a97db
Mixed up kernel size for last blocks in mnv4-conv-small
2024-05-24 11:50:42 -07:00
Ross Wightman
0c6a69e7ef
Add comments to MNV4 model defs with block variants
2024-05-23 15:54:05 -07:00
Ross Wightman
cb33956b20
Fix some mistakes in mnv4 model defs
2024-05-23 14:24:32 -07:00
Ross Wightman
70176a2dae
torchscript typing fixes
2024-05-23 11:43:05 -07:00
Ross Wightman
2a1a6b1236
Adding missing attention2d.py
2024-05-23 11:06:32 -07:00
Ross Wightman
cee79dada0
Merge remote-tracking branch 'origin/main' into efficientnet_x
2024-05-23 11:01:39 -07:00
Ross Wightman
6a8bb03330
Initial MobileNetV4 pass
2024-05-23 10:49:18 -07:00
Ross Wightman
e748805be3
Add regex matching support to AttentionExtract. Add return_dict support to graph extractors and use returned output in AttentionExtractor
2024-05-22 14:33:39 -07:00
Ross Wightman
44f72c04b3
Change node/module name matching for AttentionExtract so it keeps outputs in order. #1232
2024-05-22 13:45:25 -07:00
Ross Wightman
84cb225ecb
Add in12k + 12k_ft_in1k vit_medium weights
2024-05-20 15:52:46 -07:00
Ross Wightman
4634c3e134
Version 1.0.4.dev0
2024-05-20 15:52:27 -07:00
Ross Wightman
202199cac3
Merge pull request #2183 from lixit/patch-1
...
Fix a typo
2024-05-20 13:45:32 -07:00
xitong
14dc2c4df6
fix a typo
2024-05-20 14:34:39 -04:00
Beckschen
7a2ad6bce1
Add link to model weights on Hugging Face
2024-05-17 06:51:35 -04:00
Beckschen
530fb49e7e
Add link to model weights on Hugging Face
2024-05-17 06:48:59 -04:00
Ross Wightman
6b9e38b589
Merge pull request #2182 from fcossio/patch-1
...
Credit earlier work with the same idea
2024-05-16 15:10:46 -07:00
Fernando Cossio
9b11801cb4
Credit earlier work with the same idea.
...
Hi, this earlier work has the same name and idea behind this layer. It could be useful for readers to keep both links here if they want to see the effects of introducing this layer on a very different domain. 😄
2024-05-16 22:50:34 +02:00
Ross Wightman
cb0e4391be
Release 1.0.3
2024-05-15 11:06:22 -07:00
Ross Wightman
27fd2f35d3
Merge pull request #2181 from huggingface/Delaunay-dist-backend
...
Delaunay dist backend flag
2024-05-15 10:00:59 -07:00
Ross Wightman
e57625e814
Tweak dist_backend to use device_type (before possible :)
2024-05-15 08:49:25 -07:00
Ross Wightman
6ca92570f7
Merge branch 'patch-1' of https://github.com/Delaunay/pytorch-image-models into Delaunay-dist-backend
2024-05-15 08:40:58 -07:00
Ross Wightman
cd0e7b11ff
Merge pull request #2180 from yvonwin/main
...
Remove a duplicate function in mobilenetv3.py
2024-05-15 07:54:17 -07:00
Ross Wightman
83aee5c28c
Add explicit GAP (avg pool) variants of other SigLIP models.
2024-05-15 07:53:19 -07:00
yvonwin
58f2f79b04
Remove a duplicate function in mobilenetv3.py: `_gen_lcnet` is repeated in mobilenetv3.py.Remove the duplicate code.
2024-05-15 17:59:34 +08:00
Ross Wightman
f04802fb8d
Update README.md
2024-05-14 17:19:34 -07:00
Ross Wightman
6653747d95
Merge pull request #2178 from huggingface/pali_siglip
...
Support loading of PaliGemma weights into GAP variants of SigLIP ViT.
2024-05-14 17:10:38 -07:00
Ross Wightman
7b3b11b63f
Support loading of paligemma weights into GAP variants of SigLIP ViT. Minor tweak to npz loading for packed transformer weights.
2024-05-14 15:44:37 -07:00
Ross Wightman
04462f554f
Fix typos in feat extract doc
2024-05-14 15:33:39 -07:00
Beckschen
df304ffbf2
the dataclass init needs to use the default factory pattern, according to Ross
2024-05-14 15:10:05 -04:00
Ross Wightman
cc5f2f6f70
version 1.0.2dev0
2024-05-13 15:25:15 -07:00
Ross Wightman
3bfd036b58
Add normalize flag to transforms factory, allow return of non-normalized native dtype torch.Tensors
2024-05-13 15:23:25 -07:00
Ross Wightman
a69863ad61
Merge pull request #2156 from huggingface/hiera
...
WIP Hiera implementation.
2024-05-13 14:58:12 -07:00
Setepenre
8848dad362
Update distributed.py
2024-05-13 16:55:42 -04:00
Ross Wightman
f7aa0a1a71
Add missing vit_wee weight
2024-05-13 12:05:47 -07:00
Ross Wightman
7a4e987b9f
Hiera weights on hub
2024-05-13 11:43:22 -07:00
Ross Wightman
23f09af08e
Merge branch 'main' into efficientnet_x
2024-05-12 21:31:08 -07:00
Ross Wightman
c838c4233f
Add typing to reset_classifier() on other models
2024-05-12 11:12:00 -07:00
Ross Wightman
3e03b2bf3f
Fix a few more hiera API issues
2024-05-12 11:11:45 -07:00
Ross Wightman
211d18d8ac
Move norm & pool into Hiera ClassifierHead. Misc fixes, update features_intermediate() naming
2024-05-11 23:37:35 -07:00
Ross Wightman
2ca45a4ff5
Merge remote-tracking branch 'upstream/main' into hiera
2024-05-11 15:43:05 -07:00
Ross Wightman
49de391470
Update feature_extraction.mdx
2024-05-11 12:14:02 -07:00
Ross Wightman
a193d20b7b
Update README.md
2024-05-11 11:01:05 -07:00
Ross Wightman
b2c10fec05
Merge pull request #2168 from huggingface/more_vit_better_getter_redux
...
A few more features_intermediate() models, AttentionExtract helper, related minor cleanup.
2024-05-11 08:13:50 -07:00
Ross Wightman
1d3ab176bc
Remove debug / staging code
2024-05-10 22:16:34 -07:00
Ross Wightman
aa4d06a11c
sbb vit weights on hub, testing
2024-05-10 17:15:01 -07:00
Ross Wightman
3582ca499e
Prepping weight push, benchmarking.
2024-05-10 14:14:06 -07:00
Ross Wightman
2bfa5e5d74
Remove JIT activations, take jit out of ME activations. Remove other instances of torch.jit.script. Breaks torch.compile and is much less performant. Remove SpaceToDepthModule
2024-05-06 16:32:49 -07:00
Beckschen
99d4c7d202
add ViTamin models
2024-05-05 02:50:14 -04:00
Ross Wightman
07535f408a
Add AttentionExtract helper module
2024-05-04 14:10:00 -07:00