Abstract

We introduce an extremely computation-efficient CNN architecture named ShuffleNet, which is designed specially for mobile devices with very limited computing power (e.g., 10-150 MFLOPs). The new architecture utilizes two new operations, pointwise group convolution and channel shuffle, to greatly reduce computation cost while maintaining accuracy. Experiments on ImageNet classification and MS COCO object detection demonstrate the superior performance of ShuffleNet over other structures, e.g. lower top-1 error (absolute 7.8%) than recent MobileNet [12] on ImageNet classification task, under the computation budget of 40 MFLOPs. On an ARM-based mobile device, ShuffleNet achieves ~13× actual speedup over AlexNet while maintaining comparable accuracy.

Keywords

FLOPSComputer scienceComputationConvolutional neural networkMobile deviceSpeedupConvolution (computer science)Artificial neural networkArtificial intelligenceParallel computingAlgorithm

Affiliated Institutions

Related Publications

Publication Info

Year
2018
Type
preprint
Pages
6848-6856
Citations
8394
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

8394
OpenAlex
677
Influential
6264
CrossRef

Cite This

Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin et al. (2018). ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition , 6848-6856. https://doi.org/10.1109/cvpr.2018.00716

Identifiers

DOI
10.1109/cvpr.2018.00716
arXiv
1707.01083

Data Quality

Data completeness: 84%