Commit Graph

1975 Commits (dc18cda2e74962d28df4c16c667745df03ffa0b5)
 

Author SHA1 Message Date
Ross Wightman dc18cda2e7 efficientvit (mit) msa attention q/k/v ops need to be in float32 to train w/o NaN 2023-08-20 11:49:36 -07:00
Ross Wightman e6aeb91ac1
Merge pull request #1923 from huggingface/yehuitang-Add-GhostNetV2
ghostnetv2 cleanup
2023-08-20 02:27:58 -07:00
Ross Wightman be4e0d8f76 Update attrib comment to include v2 2023-08-19 23:39:09 -07:00
Ross Wightman 126a58e563 Combine ghostnetv2 with ghostnet, reduec redundancy, add weights to hf hub. 2023-08-19 23:33:43 -07:00
Ross Wightman 3f320a9e57 Merge branch 'Add-GhostNetV2' of github.com:yehuitang/pytorch-image-models into yehuitang-Add-GhostNetV2 2023-08-19 22:07:54 -07:00
Ross Wightman 7c2728c6fe
Merge pull request #1919 from ChengpengChen/main
Add RepGhost models and weights
2023-08-19 16:26:45 -07:00
Ross Wightman 69e0ca2e36 Weights on hf hub, bicubic yields slightly better eval 2023-08-19 16:25:45 -07:00
Ross Wightman b8011565bd
Merge pull request #1894 from seefun/master
add two different EfficientViT models
2023-08-19 09:24:14 -07:00
Ross Wightman 7d7589e8da Fixing efficient_vit torchscript, fx, default_cfg issues 2023-08-18 23:23:11 -07:00
Ross Wightman 58ea1c02c4 Add fixed_input_size flag to msra efficient_vit 2023-08-18 16:48:17 -07:00
Ross Wightman c28324a150 Update efficient_vit (msra), hf hub weights 2023-08-18 16:45:37 -07:00
Ross Wightman e700a32626 Cleanup of efficient_vit (mit), tweak eps for better AMP behaviour, formatting/cleanup, weights on hf hub 2023-08-18 16:06:07 -07:00
方曦 00f670fa69 fix bug in ci for efficientvits 2023-08-17 14:40:17 +08:00
Chengpeng Chen e7f97cb5ce Fix typos RepGhost models 2023-08-16 14:27:45 +08:00
Chengpeng Chen d1d0193615 Add RepGhost models and weights 2023-08-16 11:54:53 +08:00
Ross Wightman 24fc720d34
Merge pull request #1918 from Mingosnake/fix-efficientformerv2-typo
Fix typo in efficientformer_v2
2023-08-15 20:04:35 -07:00
Minseo Kang 7938f28542 Fix typo in efficientformer_v2 2023-08-16 03:29:01 +09:00
yehuitang e4babe7372
Add GhostNetV2 2023-08-13 18:23:52 +08:00
yehuitang b407794e3a
Add GhostNetV2 2023-08-13 18:20:27 +08:00
yehuitang fc865282e5
Add ghostnetv2.py 2023-08-13 18:16:26 +08:00
Ross Wightman da75cdd212
Merge pull request #1900 from huggingface/swin_maxvit_resize
Add support for resizing swin transformer, maxvit, coatnet at creation time
2023-08-11 15:05:28 -07:00
Ross Wightman 78a04a0e7d
Merge pull request #1911 from dsuess/1910-fixes-batchnormact-fx
Register norm_act layers as leaf modules
2023-08-11 14:34:16 -07:00
Ross Wightman 379998ed5d
Merge pull request #1914 from developer0hye/developer0hye-patch-1
Set self.num_features to neck_chans if neck_chans > 0 for vision_transformer_sam
2023-08-11 11:23:57 -07:00
Yonghye Kwon 2048f6f20f
set self.num_features to neck_chans if neck_chans > 0 2023-08-11 13:45:06 +09:00
Ross Wightman 3a44e6c602 Fix #1912 CoaT model not loading w/ return_interm_layers 2023-08-10 11:15:58 -07:00
Daniel Suess 986de90360
Register orm_act layers as leaf modules 2023-08-10 15:37:26 +10:00
Ross Wightman c692715388 Some RepVit tweaks
* add head dropout to RepVit as all models have that arg
* default train to non-distilled head output via distilled_training flag (set_distilled_training) so fine-tune works by default w/o distillation script
* camel case naming tweaks to match other models
2023-08-09 12:41:12 -07:00
Ross Wightman c153cd4a3e Add more advanced interpolation method from BEiT and support non-square window & image size adaptation for
* beit/beit-v2
* maxxvit/coatnet
* swin transformer
And non-square windows for swin-v2
2023-08-08 16:41:16 -07:00
Ross Wightman f6771909ff
Merge pull request #1903 from twmht/fix_num_classes
fix num_classes not found in repvit
2023-08-07 16:35:36 -07:00
alec.tu bb2b6b5f09 fix num_classes not found 2023-08-07 15:16:03 +08:00
Ross Wightman 1dab536cb1 Fix torch.fx for swin padding change 2023-08-05 13:09:55 -07:00
Ross Wightman 7c0f492dbb Fix type annotation for torchscript 2023-08-04 23:03:52 -07:00
Ross Wightman 7790ea709b Add support for resizing swin transformer img_size and window_size on init and load from pretrained weights. Add support for non-square window_size to both swin v1/v2 2023-08-04 22:10:46 -07:00
Ross Wightman 81089b10a2 Remove unecessary LongTensor in EfficientFormer. Possibly maybe fix #1878 2023-08-03 16:38:53 -07:00
Ross Wightman 4224529ebe Version 0.9.5 prep for release. README update 2023-08-03 15:16:46 -07:00
Ross Wightman d138a9bf88 Add gluon hrnet small weights, fix #1895 2023-08-03 12:15:04 -07:00
Ross Wightman 76d166981d Fix missing norm call in Mlp forward (not used by default, but can be enabled for normformer MLP scale). Fix #1851 fix #1852 2023-08-03 11:36:30 -07:00
Ross Wightman 8e4480e4b6 Patch and pos embed resample done in float32 always (cast to float and back). Fix #1811 2023-08-03 11:32:17 -07:00
Ross Wightman 150356c493 Fix unfortunate selecsls case bug caused by aggressive IDE rename 2023-08-03 10:37:06 -07:00
Ross Wightman 6e8c53d0d3 Comment out beit url, no longer valid as now require long query string, leave for reference, must use HF hub now. 2023-08-03 10:00:46 -07:00
方曦 a56e2bbf19 fix efficientvit_msra pretrained load 2023-08-03 18:44:38 +08:00
方曦 e94c60b546 efficientvit_msra refactor 2023-08-03 17:45:50 +08:00
方曦 047bab6ab2 efficientvit_mit stage refactor 2023-08-03 14:59:35 +08:00
方曦 e8fb866ccf fix efficientvit_msra pool 2023-08-02 14:40:01 +08:00
方曦 43443f64eb fix efficientvits 2023-08-02 14:12:37 +08:00
方曦 82d1e99e1a add efficientvit(msra) 2023-08-01 18:51:08 +08:00
方曦 b91a77fab7 add EfficientVit (MIT) 2023-08-01 12:42:21 +08:00
Ross Wightman 3b8ef3f32f
Merge pull request #1890 from Separius/patch-1
use float in resample_abs_pos_embed_nhwc
2023-07-28 21:47:26 -07:00
Sepehr Sameni 40a518c194
use float in resample_abs_pos_embed_nhwc
since F.interpolate doesn't always support BFloat16
2023-07-28 16:01:42 -07:00
Ross Wightman 8cb0ddac45 Update README, version 0.9.4dev0 2023-07-27 17:07:31 -07:00