Ross Wightman
dc18cda2e7
efficientvit (mit) msa attention q/k/v ops need to be in float32 to train w/o NaN
2023-08-20 11:49:36 -07:00
Ross Wightman
e6aeb91ac1
Merge pull request #1923 from huggingface/yehuitang-Add-GhostNetV2
...
ghostnetv2 cleanup
2023-08-20 02:27:58 -07:00
Ross Wightman
be4e0d8f76
Update attrib comment to include v2
2023-08-19 23:39:09 -07:00
Ross Wightman
126a58e563
Combine ghostnetv2 with ghostnet, reduec redundancy, add weights to hf hub.
2023-08-19 23:33:43 -07:00
Ross Wightman
3f320a9e57
Merge branch 'Add-GhostNetV2' of github.com:yehuitang/pytorch-image-models into yehuitang-Add-GhostNetV2
2023-08-19 22:07:54 -07:00
Ross Wightman
7c2728c6fe
Merge pull request #1919 from ChengpengChen/main
...
Add RepGhost models and weights
2023-08-19 16:26:45 -07:00
Ross Wightman
69e0ca2e36
Weights on hf hub, bicubic yields slightly better eval
2023-08-19 16:25:45 -07:00
Ross Wightman
b8011565bd
Merge pull request #1894 from seefun/master
...
add two different EfficientViT models
2023-08-19 09:24:14 -07:00
Ross Wightman
7d7589e8da
Fixing efficient_vit torchscript, fx, default_cfg issues
2023-08-18 23:23:11 -07:00
Ross Wightman
58ea1c02c4
Add fixed_input_size flag to msra efficient_vit
2023-08-18 16:48:17 -07:00
Ross Wightman
c28324a150
Update efficient_vit (msra), hf hub weights
2023-08-18 16:45:37 -07:00
Ross Wightman
e700a32626
Cleanup of efficient_vit (mit), tweak eps for better AMP behaviour, formatting/cleanup, weights on hf hub
2023-08-18 16:06:07 -07:00
方曦
00f670fa69
fix bug in ci for efficientvits
2023-08-17 14:40:17 +08:00
Chengpeng Chen
e7f97cb5ce
Fix typos RepGhost models
2023-08-16 14:27:45 +08:00
Chengpeng Chen
d1d0193615
Add RepGhost models and weights
2023-08-16 11:54:53 +08:00
Ross Wightman
24fc720d34
Merge pull request #1918 from Mingosnake/fix-efficientformerv2-typo
...
Fix typo in efficientformer_v2
2023-08-15 20:04:35 -07:00
Minseo Kang
7938f28542
Fix typo in efficientformer_v2
2023-08-16 03:29:01 +09:00
yehuitang
e4babe7372
Add GhostNetV2
2023-08-13 18:23:52 +08:00
yehuitang
b407794e3a
Add GhostNetV2
2023-08-13 18:20:27 +08:00
yehuitang
fc865282e5
Add ghostnetv2.py
2023-08-13 18:16:26 +08:00
Ross Wightman
da75cdd212
Merge pull request #1900 from huggingface/swin_maxvit_resize
...
Add support for resizing swin transformer, maxvit, coatnet at creation time
2023-08-11 15:05:28 -07:00
Ross Wightman
78a04a0e7d
Merge pull request #1911 from dsuess/1910-fixes-batchnormact-fx
...
Register norm_act layers as leaf modules
2023-08-11 14:34:16 -07:00
Ross Wightman
379998ed5d
Merge pull request #1914 from developer0hye/developer0hye-patch-1
...
Set self.num_features to neck_chans if neck_chans > 0 for vision_transformer_sam
2023-08-11 11:23:57 -07:00
Yonghye Kwon
2048f6f20f
set self.num_features to neck_chans if neck_chans > 0
2023-08-11 13:45:06 +09:00
Ross Wightman
3a44e6c602
Fix #1912 CoaT model not loading w/ return_interm_layers
2023-08-10 11:15:58 -07:00
Daniel Suess
986de90360
Register orm_act layers as leaf modules
2023-08-10 15:37:26 +10:00
Ross Wightman
c692715388
Some RepVit tweaks
...
* add head dropout to RepVit as all models have that arg
* default train to non-distilled head output via distilled_training flag (set_distilled_training) so fine-tune works by default w/o distillation script
* camel case naming tweaks to match other models
2023-08-09 12:41:12 -07:00
Ross Wightman
c153cd4a3e
Add more advanced interpolation method from BEiT and support non-square window & image size adaptation for
...
* beit/beit-v2
* maxxvit/coatnet
* swin transformer
And non-square windows for swin-v2
2023-08-08 16:41:16 -07:00
Ross Wightman
f6771909ff
Merge pull request #1903 from twmht/fix_num_classes
...
fix num_classes not found in repvit
2023-08-07 16:35:36 -07:00
alec.tu
bb2b6b5f09
fix num_classes not found
2023-08-07 15:16:03 +08:00
Ross Wightman
1dab536cb1
Fix torch.fx for swin padding change
2023-08-05 13:09:55 -07:00
Ross Wightman
7c0f492dbb
Fix type annotation for torchscript
2023-08-04 23:03:52 -07:00
Ross Wightman
7790ea709b
Add support for resizing swin transformer img_size and window_size on init and load from pretrained weights. Add support for non-square window_size to both swin v1/v2
2023-08-04 22:10:46 -07:00
Ross Wightman
81089b10a2
Remove unecessary LongTensor in EfficientFormer. Possibly maybe fix #1878
2023-08-03 16:38:53 -07:00
Ross Wightman
4224529ebe
Version 0.9.5 prep for release. README update
2023-08-03 15:16:46 -07:00
Ross Wightman
d138a9bf88
Add gluon hrnet small weights, fix #1895
2023-08-03 12:15:04 -07:00
Ross Wightman
76d166981d
Fix missing norm call in Mlp forward (not used by default, but can be enabled for normformer MLP scale). Fix #1851 fix #1852
2023-08-03 11:36:30 -07:00
Ross Wightman
8e4480e4b6
Patch and pos embed resample done in float32 always (cast to float and back). Fix #1811
2023-08-03 11:32:17 -07:00
Ross Wightman
150356c493
Fix unfortunate selecsls case bug caused by aggressive IDE rename
2023-08-03 10:37:06 -07:00
Ross Wightman
6e8c53d0d3
Comment out beit url, no longer valid as now require long query string, leave for reference, must use HF hub now.
2023-08-03 10:00:46 -07:00
方曦
a56e2bbf19
fix efficientvit_msra pretrained load
2023-08-03 18:44:38 +08:00
方曦
e94c60b546
efficientvit_msra refactor
2023-08-03 17:45:50 +08:00
方曦
047bab6ab2
efficientvit_mit stage refactor
2023-08-03 14:59:35 +08:00
方曦
e8fb866ccf
fix efficientvit_msra pool
2023-08-02 14:40:01 +08:00
方曦
43443f64eb
fix efficientvits
2023-08-02 14:12:37 +08:00
方曦
82d1e99e1a
add efficientvit(msra)
2023-08-01 18:51:08 +08:00
方曦
b91a77fab7
add EfficientVit (MIT)
2023-08-01 12:42:21 +08:00
Ross Wightman
3b8ef3f32f
Merge pull request #1890 from Separius/patch-1
...
use float in resample_abs_pos_embed_nhwc
2023-07-28 21:47:26 -07:00
Sepehr Sameni
40a518c194
use float in resample_abs_pos_embed_nhwc
...
since F.interpolate doesn't always support BFloat16
2023-07-28 16:01:42 -07:00
Ross Wightman
8cb0ddac45
Update README, version 0.9.4dev0
2023-07-27 17:07:31 -07:00