mirror of
https://github.com/huggingface/pytorch-image-models.git
synced 2025-06-03 15:01:08 +08:00
Add in12k fine-tuned convnext_xxlarge
This commit is contained in:
parent
855719fca6
commit
d3e83a190f
@ -31,6 +31,7 @@
|
||||
* DINOv2 'register' ViT model weights added
|
||||
* Add `quickgelu` ViT variants for OpenAI, DFN, MetaCLIP weights that use it (less efficient)
|
||||
* Improved typing added to ResNet, MobileNet-v3 thanks to [Aryan](https://github.com/a-r-r-o-w)
|
||||
* ImageNet-12k fine-tuned (from LAION-2B CLIP) `convnext_xxlarge`
|
||||
* 0.9.9 release
|
||||
|
||||
### Oct 20, 2023
|
||||
|
@ -828,6 +828,10 @@ default_cfgs = generate_default_cfgs({
|
||||
hf_hub_id='timm/',
|
||||
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, num_classes=11821,
|
||||
input_size=(3, 384, 384), pool_size=(12, 12), crop_pct=1.0, crop_mode='squash'),
|
||||
'convnext_xxlarge.clip_laion2b_soup_ft_in12k': _cfg(
|
||||
hf_hub_id='timm/',
|
||||
mean=OPENAI_CLIP_MEAN, std=OPENAI_CLIP_STD, num_classes=11821,
|
||||
input_size=(3, 256, 256), pool_size=(8, 8), crop_pct=1.0),
|
||||
|
||||
# CLIP original image tower weights
|
||||
'convnext_base.clip_laion2b': _cfg(
|
||||
|
Loading…
x
Reference in New Issue
Block a user