mirror of
https://github.com/open-mmlab/mmpretrain.git
synced 2025-06-03 14:59:18 +08:00
* Add cosine cool down lr updater * Use ema hook * Update decay mult * Update configs. * Update T2T-ViT readme and format all readme * Update swin readme * Update tnt readme * Add docstring for `CosineAnnealingCooldownLrUpdaterHook`. * Update t2t readme and metafile
Shufflenet v2: Practical guidelines for efficient cnn architecture design
Abstract
Currently, the neural network architecture design is mostly guided by the indirect metric of computation complexity, i.e., FLOPs. However, the direct metric, e.g., speed, also depends on the other factors such as memory access cost and platform characterics. Thus, this work proposes to evaluate the direct metric on the target platform, beyond only considering FLOPs. Based on a series of controlled experiments, this work derives several practical guidelines for efficient network design. Accordingly, a new architecture is presented, called ShuffleNet V2. Comprehensive ablation experiments verify that our model is the state-of-the-art in terms of speed and accuracy tradeoff.

Citation
@inproceedings{ma2018shufflenet,
title={Shufflenet v2: Practical guidelines for efficient cnn architecture design},
author={Ma, Ningning and Zhang, Xiangyu and Zheng, Hai-Tao and Sun, Jian},
booktitle={Proceedings of the European conference on computer vision (ECCV)},
pages={116--131},
year={2018}
}
Results and models
ImageNet-1k
Model | Params(M) | Flops(G) | Top-1 (%) | Top-5 (%) | Config | Download |
---|---|---|---|---|---|---|
ShuffleNetV2 1.0x | 2.28 | 0.149 | 69.55 | 88.92 | config | model | log |