Ross Wightman
4542cf03f9
Add features_only, other bits to mambaout, define different base alternatives
2024-10-09 14:11:40 -07:00
Ross Wightman
c2da12c7e1
Update rw models, fix heads
2024-10-09 14:11:40 -07:00
Ross Wightman
f2086f51a0
Add mambaout builder support, pretrained weight remap
2024-10-09 14:11:40 -07:00
Ross Wightman
c6ef54eefa
Initial mambaout work
2024-10-09 14:11:40 -07:00
Ross Wightman
d9321b0e10
Add weights for fine-tuned siglip so400m. Add webli_i18n pretrained tags for the multi-lingual model variants (incl older base)
2024-10-09 09:04:44 -07:00
Ross Wightman
01b62264af
Add i18n variant of so400m model w/ weights. Add two in1k fine-tunes of original so400m 384x384 but at 378x378 (better matches patch14)
2024-10-08 23:40:24 -07:00
Ross Wightman
72f0edb7e8
missed first_conv for rnv2 18d
2024-10-08 12:38:54 -07:00
Ross Wightman
3ed603a2ce
Add resnet18/18d pre-act model configs for potential training. Fix #2289
2024-10-08 11:28:07 -07:00
Ross Wightman
41a79e0fcb
Add overlapped stem convnext zepto weights
2024-10-08 11:26:34 -07:00
Ross Wightman
545bd4056c
Tag along test_vit3 weights
2024-09-30 12:03:32 -07:00
Ross Wightman
69b687d4cc
Add zepto weights
2024-09-30 11:43:23 -07:00
Ross Wightman
c6e5557a5a
Mismatch pretrained_cfg
2024-09-30 11:43:23 -07:00
Ross Wightman
5d7bd2973e
convnext zepto, rmsnorm experiments
2024-09-30 11:43:23 -07:00
Ross Wightman
e3242a5258
Merge pull request #2277 from huggingface/more_tiny_test_models
...
Adding some more tiny test models to train...
2024-09-22 10:28:29 -07:00
Ross Wightman
c1cb5641c7
Add weight fore mobilenetv4 small 0.5, change 0.25 -> 0.35
2024-09-22 10:27:01 -07:00
Ross Wightman
a22ce0a329
Merge branch 'patch-1' of https://github.com/baorepo/pytorch-image-models into baorepo-patch-1
2024-09-22 10:14:35 -07:00
Ross Wightman
9067be6a30
Add weights for new tiny test models
2024-09-22 07:59:23 -07:00
Ross Wightman
65564f7da5
Fix reversed H & W padding for swin patch merging
2024-09-21 16:51:02 -07:00
Ross Wightman
a2f539f055
Add a few more test model defs in prep for weight upload
2024-09-21 11:38:38 -07:00
Ross Wightman
6ab2af610d
Adding some more tiny test models to train
2024-09-06 15:35:57 -07:00
alias pillar1989
d6b8816eda
MobilenetV4: add two more lightweight models
...
Mobilenetv4 is very fast and ideal for embedded devices. However, for many low-cost, low-power embedded MCU devices, smaller models are required. Hopefully this PR will merge.
2024-09-05 02:34:11 +00:00
Ross Wightman
f81cbdcca9
Merge pull request #2274 from huggingface/bulk_runner_tweaks
...
Better all res resolution for bulk runner
2024-09-03 12:11:56 -07:00
Ross Wightman
6c42299b3e
Back to dev version
2024-09-03 10:29:05 -07:00
Ross Wightman
a50713ce6e
Fix #2272
2024-09-02 13:20:05 -07:00
Ross Wightman
ebbe530ee4
Add MobileNetV3 RA4 (mnv4 recipe) weights
2024-09-02 13:10:34 -07:00
Ross Wightman
fa4a1e597f
Better all res resolution for bulk runner
2024-08-26 22:28:01 -07:00
Ross Wightman
072783331f
Version 1.0.9
2024-08-23 11:47:42 -07:00
Ross Wightman
76b0e9931a
Placeholder for new mnv3 model
2024-08-23 10:11:20 -07:00
Ross Wightman
39e92f0c0d
mobilenet_edgetpu can use group_size override, more consistency in arg wrap/sadface w/ extra group_size arg
2024-08-22 11:44:02 -07:00
Ross Wightman
b9f020a509
Allow group_size override for more efficientnet and mobilenetv3 based models
2024-08-21 16:51:38 -07:00
Ross Wightman
17923a66bb
Add layer scale to hieradet
2024-08-21 11:23:39 -07:00
Ross Wightman
47e6958263
Add hierdet_small (non sam) model def
2024-08-21 11:05:54 -07:00
Ross Wightman
9fcbf39cdc
Add remaining sbb vit betwixt/mediumd fine-tunes
2024-08-21 10:09:38 -07:00
Ross Wightman
dc94cca0e5
Remaining Hiera sbb weights uploaded
2024-08-21 10:06:27 -07:00
Ross Wightman
a256e50457
Move padding back in front of windowing
2024-08-17 11:22:53 -07:00
Ross Wightman
7d83749207
pool size test fixes
2024-08-17 08:27:13 -07:00
Ross Wightman
1bd92bca0e
Add fused_attn flag to HieraDet attn block
2024-08-16 22:57:49 -07:00
Ross Wightman
691bb54443
Larger min input size needed
2024-08-16 17:09:19 -07:00
Ross Wightman
de3a91a7a0
Add min_input_size of 128 for hieradet/sam2
2024-08-16 15:13:56 -07:00
Ross Wightman
0b05122cda
Fixing hieradet (sam2) tests
2024-08-16 14:33:40 -07:00
Ross Wightman
e035381171
Move padding out of windowing code for hieradet, fix torchscript typing issues, make pooling MaxPool unique instances across two modules
2024-08-16 13:36:33 -07:00
Ross Wightman
146c2fbe34
Add resnet50d and efficientnet_b1 ra4 (mnv4) hparam weights
2024-08-16 12:10:00 -07:00
Ross Wightman
962958723c
More Hiera updates. Add forward_intermediates to hieradat/sam2 impl. Make both use same classifier module. Add coarse bool to intermediates.
2024-08-16 11:10:04 -07:00
Ross Wightman
f2cfb4c677
Add WIP HieraDet impl (SAM2 backbone support)
2024-08-15 17:58:15 -07:00
Ross Wightman
a50e53d41f
Rename global pos embed for Hiera abswin, factor out commonly used vit weight init fns to layers. Add a channels-last ver of normmlp head.
2024-08-15 17:46:36 -07:00
Ross Wightman
2f3fed43b8
Fix hiera init with num_classes=0, fix weight tag names for sbb2 hiera/vit weights, add LayerScale/LayerScale2d to layers
2024-08-15 11:14:38 -07:00
Ross Wightman
fee91fdd41
Update Hiera model for abswin, more stable weight init, layer-scale. ImageNet-12k weights for hiera_small_abswin, and two of the sbb vits with improved reg4 init.
2024-08-14 12:22:40 -07:00
Ross Wightman
ac3470188b
Add mobilenetv1_125 weights from ra4 / mnv4 hparam series
2024-08-14 08:47:02 -07:00
Ross Wightman
a7b0bfc376
exception for unknown arg should have been TypeError
2024-08-12 16:09:36 -07:00
Ross Wightman
bd0f79153b
Add weights only flag to avoid warning, try to keep bwd compat. Default to True for remote load of pretrained weights, keep False for local checkpoing load to avoid training checkpoint breaks.. fix #2249
2024-08-12 11:41:41 -07:00
Ross Wightman
531215eded
Bring changelog up to current, back to dev version, 1.0.9.dev0
2024-08-08 09:17:23 -07:00
Ross Wightman
bf781bc3c2
Merge pull request #2245 from dhkim0225/feature/rdnet
...
Feature: Add RDNet
2024-08-07 16:19:22 -07:00
Ross Wightman
49f6e1da96
Remove eval to avoid safety concerns
2024-08-07 14:54:13 -07:00
Ross Wightman
2d5c9bf60d
Merge pull request #2253 from brendanartley/hgnet-grad-checkpointing
...
Add gradient checkpointing to hgnets
2024-08-07 12:45:14 -07:00
Ross Wightman
e9ef9424f0
Add a few missing __all__ entries.
2024-08-07 09:35:51 -07:00
brendanartley
232a459e36
Added gradient checkpointing to hgnet
2024-08-07 09:22:51 -07:00
dong-hyun
ea33d6a74d
bugfix pool_size
2024-08-03 09:50:20 +09:00
Ross Wightman
10344625be
proj_drop wasn't being called in sam vit
2024-08-02 16:12:02 -07:00
dong-hyun
6fdc9d8243
update code for ci pass
2024-08-02 15:14:40 +09:00
dong-hyun
7a866b6521
update code for torchscript
2024-08-02 09:58:13 +09:00
dong-hyun
8248122f82
add rdnet
2024-08-01 14:54:29 +09:00
dong-hyun
025259024d
add rdnet
2024-08-01 14:51:15 +09:00
dong-hyun
225f4f92b3
add rdnet
2024-08-01 14:49:21 +09:00
Ross Wightman
4a10302754
Add mobilenet_edgetpu_v2_m weights
2024-07-28 17:19:36 -07:00
Ross Wightman
70182ea0cd
Version 1.0.8
2024-07-26 17:05:36 -07:00
Ross Wightman
ab8cb070fc
Add xavier_uniform init of MNVC hybrid attention modules. Small improvement in training stability.
2024-07-26 17:03:40 -07:00
Ross Wightman
cec70b6779
Merge pull request #2225 from huggingface/small_things
...
Small things
2024-07-25 20:29:13 -07:00
Ross Wightman
61df3fde89
Wrong hybrid_medium in12k pool sizes
2024-07-25 15:39:21 -07:00
Ross Wightman
9aa2930760
Add latest mobilenetv4 and baseline updates for mobilenetv1 and efficientnet_b0 weights
2024-07-25 14:20:54 -07:00
Ross Wightman
7b6a406474
remove swin debug prints
2024-07-24 21:05:56 -07:00
Ross Wightman
4c531be479
set_input_size(), always_partition, strict_img_size, dynamic mask option for all swin models. More flexibility in resolution, window resizing.
2024-07-24 16:41:31 -07:00
Ross Wightman
8efdc38213
Fix #2242 add checks for out indices with intermediate getter mode
2024-07-23 08:19:09 -07:00
Ross Wightman
d2240745d3
Fix issue where feature out_indices out of order after wrapping with FeatureGetterNet due to use of set()
2024-07-22 13:33:30 -07:00
Ross Wightman
2b3f1a4633
Make channels for classic resnet configurable
2024-07-22 10:47:40 -07:00
Ross Wightman
9b2b8014e8
Add weights for test models
2024-07-22 10:08:57 -07:00
Ross Wightman
7e0caa1ba3
Padding helpers work if tuples/lists passed
2024-07-19 14:28:03 -07:00
Ross Wightman
2180800646
MQA query_strides bugs fix #2237 . No padding for avg_pool2d if not 'same', use scale_factor for Upsample.
2024-07-19 14:26:54 -07:00
Ross Wightman
1a05ed29a1
Add to 'abswin' hiera models for train trials
2024-07-19 11:05:31 -07:00
Ross Wightman
0cbf4fa586
_orig_mod still causing issues even though I thought it was fixed in pytorch, add unwrap / clean helpers
2024-07-19 11:03:45 -07:00
Feraidoon Mehri
4cca568bd8
eva.py: fixed bug in applying attention mask
...
The mask should be applied before the softmax.
2024-07-19 15:12:04 +03:30
Ross Wightman
3a8a965891
Implement absolute+window pos embed for hiera, resizable but needs new weights
2024-07-18 21:43:37 -07:00
Ross Wightman
7160af4a24
Merge pull request #2229 from Promisery/reg_token
...
Initialize weights of reg_token for ViT
2024-07-18 09:25:29 -07:00
Ross Wightman
392b78aee7
set_input_size initial impl for vit & swin v1. Move HybridEmbed to own location in timm/layers
2024-07-17 15:25:48 -07:00
Ross Wightman
34c9fee554
Fix pass through of input / target keys so ImageDataset readers so args work with hfds instead of just hfids (iterable)
2024-07-17 10:11:46 -07:00
Promisery
417cf7f871
Initialize weights of reg_token for ViT
2024-07-13 11:11:42 +08:00
Ross Wightman
f920119f3b
Fixing tests
2024-07-09 14:53:20 -07:00
Ross Wightman
644abf9588
Fix default_cfg test for mobilenet_100
2024-07-09 12:52:24 -07:00
Ross Wightman
d5afe106dc
Merge remote-tracking branch 'origin/tiny_test_models' into small_things
2024-07-09 12:49:57 -07:00
Ross Wightman
55101028bb
Rename test_tiny* -> test*. Fix ByobNet BasicBlock attn location and add test_byobnet model.
2024-07-09 11:53:11 -07:00
Ross Wightman
1334598462
Add support back to EfficientNet to disable head_conv / bn2 so mobilnetv1 can be implemented properly
2024-07-08 13:51:26 -07:00
Ross Wightman
800405d941
Add conv_large mobilenetv3 aa/blur model defs
2024-07-08 13:50:05 -07:00
Ross Wightman
f81b094aaa
Add 'qkv_bias_separate' flag for EVA/beit/swinv2 attn modules to allow an override for easy quantization wrappers. Fix #2098
2024-07-08 13:48:38 -07:00
Ross Wightman
83c2c2f0c5
Add 'Maybe' PIL / image tensor conversions in case image alread in tensor format
2024-07-08 13:43:51 -07:00
Steffen Schneider
c01a47c9e7
Fix typo in type annotations in timm.models.hrnet
2024-07-08 00:53:16 +02:00
Daniel Suess
197c10463b
Fix jit.script breaking with features_fx
2024-06-28 03:58:51 +00:00
Ross Wightman
b751da692d
Add latest ix (xavier init for mqa) hybrid medium & large weights for MobileNetV4
2024-06-24 13:49:55 -07:00
Ross Wightman
d4d4d84fda
Dev version 1.0.8.dev0
2024-06-24 11:34:13 -07:00
Ross Wightman
f8342a045a
Merge pull request #2213 from huggingface/florence2
...
Fix #2212 map florence2 image tower to davit with a few changes
2024-06-24 11:01:08 -07:00
Sejik
c33a001397
Fix typo
2024-06-24 21:54:38 +09:00
Ross Wightman
02d0f27721
cleanup davit padding
2024-06-22 12:06:46 -07:00
Ross Wightman
c715c724e7
Fix tracing by removing float cast, should end up float anyways
2024-06-22 08:35:30 -07:00
Ross Wightman
fb58a73033
Fix #2212 map florence2 image tower to davit with a few changes
2024-06-21 15:31:29 -07:00
Ross Wightman
b28945ff05
Version 1.0.7, prep for release
2024-06-18 16:19:43 -07:00
Ross Wightman
fb13e6385e
Merge pull request #2203 from huggingface/more_mobile
...
Add mobilenet edgetpu defs for exp, add ol mobilenet v1 back for comp…
2024-06-18 15:20:01 -07:00
Ross Wightman
16e082e1c2
Add mobilenetv4 hybrid-large weights
2024-06-17 11:08:31 -07:00
Ross Wightman
e41125cc83
Merge pull request #2209 from huggingface/fcossio-vit-maxpool
...
ViT pooling refactor
2024-06-17 07:51:12 -07:00
Ross Wightman
a22466852d
Add 2400 epoch mobilenetv4 small weights, almost at paper, rounds to 73.8
2024-06-16 10:51:00 -07:00
Ross Wightman
b1a6f4a946
Some missed reset_classifier() type annotations
2024-06-16 10:39:27 -07:00
Ross Wightman
71101ebba0
Refactor vit pooling to add more reduction options, separately callable
2024-06-14 23:16:58 -07:00
Ross Wightman
a0bb5b4a44
Missing stem_kernel_size argument in EfficientNetFeatures
2024-06-14 13:39:31 -07:00
Fernando Cossio
9567cf6d84
Feature: add option global_pool='max' to VisionTransformer
...
Most of the CNNs have a max global pooling option. I would like to extend ViT to have this option.
2024-06-14 15:24:54 +02:00
Ross Wightman
9613c76844
Add mobilenet edgetpu defs for exp, add ol mobilenet v1 back for completeness / comparison
2024-06-13 17:33:04 -07:00
Ross Wightman
22de845add
Prepping for final MobileCLIP weight locations ( #2199 )
...
* Prepping for final MobileCLIP weight locations
* Update weight locations to coreml-projects
* Update mobileclip weight locations with final apple org location
2024-06-13 16:55:49 -07:00
Ross Wightman
575978ba55
Add mnv4_conv_large 384x384 weight location
2024-06-13 12:58:04 -07:00
Ross Wightman
7b5f17d1bd
Update README.md, bump dev version 1.0.6
2024-06-12 12:35:44 -07:00
Ross Wightman
e42e453128
Fix mmnv4 conv_large weight link, reorder mnv4 pretrained cfg for proper precedence
2024-06-12 11:16:49 -07:00
Ross Wightman
7b0a5321cb
Merge pull request #2198 from huggingface/openai_clip_resnet
...
Mapping OpenAI CLIP Modified ResNet weights -> ByobNet.
2024-06-12 09:33:30 -07:00
Ross Wightman
57adc1acc8
Fix rotary embed version of attn pool. Bit of cleanup/naming
2024-06-11 23:49:17 -07:00
Ross Wightman
cdc7bcea69
Make 2d attention pool modules compatible with head interface. Use attention pool in CLIP ResNets as head. Make separate set of GAP models w/ avg pool instead of attn pool.
2024-06-11 21:32:07 -07:00
Ross Wightman
c63da1405c
Pretrained cfg name mismatch
2024-06-11 21:16:54 -07:00
Ross Wightman
88efca1be2
First set of MobileNetV4 weights trained in timm
2024-06-11 18:53:01 -07:00
Ross Wightman
30ffa152de
Fix load of larger ResNet CLIP models, experimenting with making AttentionPool *the* head, seems to fine-tune better, one less layer.
2024-06-10 12:07:14 -07:00
Ross Wightman
5e9ff5798f
Adding pos embed resize fns to FX autowrap exceptions
2024-06-10 12:06:47 -07:00
Ross Wightman
f0fb471b26
Remove separate ConvNormActAa class, merge with ConvNormAct
2024-06-10 12:05:35 -07:00
Ross Wightman
5efa15b2a2
Mapping OpenAI CLIP Modified ResNet weights -> ByobNet. Improve AttentionPool2d layers. Fix #1731
2024-06-09 16:54:48 -07:00
Ross Wightman
7702d9afa1
ViTamin in_chans !=3 weight load fix
2024-06-07 20:39:23 -07:00
Ross Wightman
66a0eb4673
Experimenting with tiny test models, how small can they go and be useful for regression tests?
2024-06-07 16:09:25 -07:00
Ross Wightman
5ee06760dc
Fix classifier input dim for mnv3 after last changes
2024-06-07 13:53:13 -07:00
Ross Wightman
a5a2ad2e48
Fix consistency, testing for forward_head w/ pre_logits, reset_classifier, models with pre_logits size != unpooled feature size
...
* add test that model supports forward_head(x, pre_logits=True)
* add head_hidden_size attr to all models and set differently from num_features attr when head has hidden layers
* test forward_features() feat dim == model.num_features and pre_logits feat dim == self.head_hidden_size
* more consistency in reset_classifier signature, add typing
* asserts in some heads where pooling cannot be disabled
Fix #2194
2024-06-07 13:53:00 -07:00
Ross Wightman
4535a5412a
Change default serialization for push_to_hf_hub to 'both'
2024-06-07 13:40:31 -07:00
Ross Wightman
5cce2185e1
Update version.py
2024-06-07 13:13:23 -07:00
Ross Wightman
7ccb10ebff
Disable efficient_builder debug flag
2024-06-06 21:50:27 -07:00
Ross Wightman
ad026e6e33
Fix in_chans switching on create
2024-06-06 17:56:14 -07:00
Ross Wightman
fc1b66a51d
Fix first conv name for mci vit-b
2024-06-06 13:42:26 -07:00
Ross Wightman
88a1006e02
checkpoint filter fns with consistent name, add mobileclip-b pretrained cfgs
2024-06-06 12:38:52 -07:00
Ross Wightman
7d4ada6d16
Update ViTamin model defs
2024-06-06 09:16:43 -07:00
Ross Wightman
cc8a03daac
Add ConvStem and MobileCLIP hybrid model for B variant. Add full norm disable support to ConvNormAct layers
2024-06-06 09:15:27 -07:00
Ross Wightman
3c9d8e5b33
Merge remote-tracking branch 'origin/efficientnet_x' into fastvit_mobileclip
2024-06-05 17:35:15 -07:00
Ross Wightman
5756a81c55
Merge remote-tracking branch 'origin/Beckschen-vitamin' into fastvit_mobileclip
2024-06-05 15:20:54 -07:00
Ross Wightman
58591a97f7
Enable features_only properly
2024-06-04 16:57:16 -07:00
Ross Wightman
1b66ec7cf3
Fixup ViTamin, add hub weight reference
2024-06-03 17:14:03 -07:00
Ross Wightman
b2c0aeb0ec
Merge branch 'main' of https://github.com/Beckschen/pytorch-image-models into Beckschen-vitamin
2024-06-02 14:16:30 -07:00
Ross Wightman
7f96538052
Add missing lkc act for mobileclip fastvits
2024-05-31 11:59:51 -07:00
Ross Wightman
a503639bcc
Add mobileclip fastvit model defs, support extra SE. Add forward_intermediates API to fastvit
2024-05-30 10:17:38 -07:00
Ross Wightman
5fa6efa158
Add anti-aliasing support to mobilenetv3 and efficientnet family models. Update MobileNetV4 model defs, resolutions. Fix #599
...
* create_aa helper function centralized for all timm uses (resnet, convbnact helper)
* allow BlurPool w/ pre-defined channels (expand)
* mobilenetv4 UIB block using ConvNormAct layers for improved clarity, esp with AA added
* improve more mobilenetv3 and efficientnet related type annotations
2024-05-27 22:06:22 -07:00
Ross Wightman
5dce710101
Add vit_little in12k + in12k-ft-in1k weights
2024-05-27 14:56:03 -07:00
Ross Wightman
3c0283f9ef
Fix reparameterize for NextViT. Fix #2187
2024-05-27 14:48:58 -07:00
Ross Wightman
4ff7c25766
Pass layer_scale_init_value to Mnv3Features module
2024-05-24 16:44:50 -07:00
Ross Wightman
a12b72b5c4
Fix missing head_norm arg pop for feature model
2024-05-24 15:50:34 -07:00
Ross Wightman
7fe96e7a92
More MobileNet-v4 fixes
...
* missed final norm after post pooling 1x1 PW head conv
* improve repr of model by flipping a few modules to None when not used, nn.Sequential for MultiQueryAttention query/key/value/output
* allow layer scaling to be enabled/disabled at model variant level, conv variants don't use it
2024-05-24 15:09:29 -07:00
Ross Wightman
28d76a97db
Mixed up kernel size for last blocks in mnv4-conv-small
2024-05-24 11:50:42 -07:00
Ross Wightman
0c6a69e7ef
Add comments to MNV4 model defs with block variants
2024-05-23 15:54:05 -07:00
Ross Wightman
cb33956b20
Fix some mistakes in mnv4 model defs
2024-05-23 14:24:32 -07:00
Ross Wightman
70176a2dae
torchscript typing fixes
2024-05-23 11:43:05 -07:00
Ross Wightman
2a1a6b1236
Adding missing attention2d.py
2024-05-23 11:06:32 -07:00
Ross Wightman
cee79dada0
Merge remote-tracking branch 'origin/main' into efficientnet_x
2024-05-23 11:01:39 -07:00
Ross Wightman
6a8bb03330
Initial MobileNetV4 pass
2024-05-23 10:49:18 -07:00
Ross Wightman
e748805be3
Add regex matching support to AttentionExtract. Add return_dict support to graph extractors and use returned output in AttentionExtractor
2024-05-22 14:33:39 -07:00
Ross Wightman
44f72c04b3
Change node/module name matching for AttentionExtract so it keeps outputs in order. #1232
2024-05-22 13:45:25 -07:00
Ross Wightman
84cb225ecb
Add in12k + 12k_ft_in1k vit_medium weights
2024-05-20 15:52:46 -07:00
Ross Wightman
4634c3e134
Version 1.0.4.dev0
2024-05-20 15:52:27 -07:00
Beckschen
7a2ad6bce1
Add link to model weights on Hugging Face
2024-05-17 06:51:35 -04:00
Beckschen
530fb49e7e
Add link to model weights on Hugging Face
2024-05-17 06:48:59 -04:00
Fernando Cossio
9b11801cb4
Credit earlier work with the same idea.
...
Hi, this earlier work has the same name and idea behind this layer. It could be useful for readers to keep both links here if they want to see the effects of introducing this layer on a very different domain. 😄
2024-05-16 22:50:34 +02:00
Ross Wightman
cb0e4391be
Release 1.0.3
2024-05-15 11:06:22 -07:00
Ross Wightman
27fd2f35d3
Merge pull request #2181 from huggingface/Delaunay-dist-backend
...
Delaunay dist backend flag
2024-05-15 10:00:59 -07:00
Ross Wightman
e57625e814
Tweak dist_backend to use device_type (before possible :)
2024-05-15 08:49:25 -07:00
Ross Wightman
6ca92570f7
Merge branch 'patch-1' of https://github.com/Delaunay/pytorch-image-models into Delaunay-dist-backend
2024-05-15 08:40:58 -07:00
Ross Wightman
cd0e7b11ff
Merge pull request #2180 from yvonwin/main
...
Remove a duplicate function in mobilenetv3.py
2024-05-15 07:54:17 -07:00
Ross Wightman
83aee5c28c
Add explicit GAP (avg pool) variants of other SigLIP models.
2024-05-15 07:53:19 -07:00
yvonwin
58f2f79b04
Remove a duplicate function in mobilenetv3.py: `_gen_lcnet` is repeated in mobilenetv3.py.Remove the duplicate code.
2024-05-15 17:59:34 +08:00
Ross Wightman
7b3b11b63f
Support loading of paligemma weights into GAP variants of SigLIP ViT. Minor tweak to npz loading for packed transformer weights.
2024-05-14 15:44:37 -07:00
Beckschen
df304ffbf2
the dataclass init needs to use the default factory pattern, according to Ross
2024-05-14 15:10:05 -04:00
Ross Wightman
cc5f2f6f70
version 1.0.2dev0
2024-05-13 15:25:15 -07:00
Ross Wightman
3bfd036b58
Add normalize flag to transforms factory, allow return of non-normalized native dtype torch.Tensors
2024-05-13 15:23:25 -07:00
Ross Wightman
a69863ad61
Merge pull request #2156 from huggingface/hiera
...
WIP Hiera implementation.
2024-05-13 14:58:12 -07:00
Setepenre
8848dad362
Update distributed.py
2024-05-13 16:55:42 -04:00
Ross Wightman
f7aa0a1a71
Add missing vit_wee weight
2024-05-13 12:05:47 -07:00
Ross Wightman
7a4e987b9f
Hiera weights on hub
2024-05-13 11:43:22 -07:00
Ross Wightman
23f09af08e
Merge branch 'main' into efficientnet_x
2024-05-12 21:31:08 -07:00
Ross Wightman
c838c4233f
Add typing to reset_classifier() on other models
2024-05-12 11:12:00 -07:00
Ross Wightman
3e03b2bf3f
Fix a few more hiera API issues
2024-05-12 11:11:45 -07:00
Ross Wightman
211d18d8ac
Move norm & pool into Hiera ClassifierHead. Misc fixes, update features_intermediate() naming
2024-05-11 23:37:35 -07:00
Ross Wightman
2ca45a4ff5
Merge remote-tracking branch 'upstream/main' into hiera
2024-05-11 15:43:05 -07:00
Ross Wightman
1d3ab176bc
Remove debug / staging code
2024-05-10 22:16:34 -07:00
Ross Wightman
aa4d06a11c
sbb vit weights on hub, testing
2024-05-10 17:15:01 -07:00
Ross Wightman
3582ca499e
Prepping weight push, benchmarking.
2024-05-10 14:14:06 -07:00
Ross Wightman
2bfa5e5d74
Remove JIT activations, take jit out of ME activations. Remove other instances of torch.jit.script. Breaks torch.compile and is much less performant. Remove SpaceToDepthModule
2024-05-06 16:32:49 -07:00
Beckschen
99d4c7d202
add ViTamin models
2024-05-05 02:50:14 -04:00
Ross Wightman
07535f408a
Add AttentionExtract helper module
2024-05-04 14:10:00 -07:00
Ross Wightman
45b7ae8029
forward_intermediates() support for byob/byoanet models
2024-05-04 14:06:52 -07:00
Ross Wightman
c4b8897e9e
attention -> attn in davit for model consistency
2024-05-04 14:06:11 -07:00
Ross Wightman
cb57a96862
Fix early stop for efficientnet/mobilenetv3 fwd inter. Fix indices typing for all fwd inter.
2024-05-04 10:21:58 -07:00
Ross Wightman
01dd01b70e
forward_intermediates() for MlpMixer models and RegNet.
2024-05-04 10:21:03 -07:00
Ross Wightman
f8979d4f50
Comment out time local files while testing new vit weights
2024-05-03 20:26:56 -07:00
Ross Wightman
c719f7eb86
More forward_intermediates() updates
...
* add convnext, resnet, efficientformer, levit support
* remove kwargs only for fn so that torchscript isn't broken for all :(
* use reset_classifier() consistently in prune
2024-05-03 16:22:32 -07:00
Ross Wightman
301d0bb21f
Stricter check on pool_type for adaptive pooling module. Fix #2159
2024-05-03 16:16:51 -07:00
Ross Wightman
d6da4fb01e
Add forward_intermediates() to efficientnet / mobilenetv3 based models as an exercise.
2024-05-02 14:19:16 -07:00
Ross Wightman
c22efb9765
Add wee & little vits for some experiments
2024-05-02 10:51:35 -07:00
Ross Wightman
67332fce24
Add features_intermediate() support to coatnet, maxvit, swin* models. Refine feature interface. Start prep of new vit weights.
2024-04-30 16:56:33 -07:00