Abstract

Currently, the neural network architecture design is mostly guided by the indirect metric of computation complexity, i.e., FLOPs. However, the direct metric, e.g., speed, also depends on the other factors such as memory access cost and platform characterics. Thus, this work proposes to evaluate the direct metric on the target platform, beyond only considering FLOPs. Based on a series of controlled experiments, this work derives several practical guidelines for efficient network design. Accordingly, a new architecture is presented, called ShuffleNet V2. Comprehensive ablation experiments verify that our model is the state-of-the-art in terms of speed and accuracy tradeoff.

Keywords

Computer scienceFLOPSMetric (unit)ArchitectureComputationPerformance metricComputer engineeringDistributed computingComputer architectureParallel computingAlgorithm

Affiliated Institutions

Related Publications

Publication Info

Year
2018
Type
book-chapter
Pages
122-138
Citations
6005
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

6005
OpenAlex
611
Influential

Cite This

Ningning Ma, Xiangyu Zhang, Hai-Tao Zheng et al. (2018). ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. Lecture notes in computer science , 122-138. https://doi.org/10.1007/978-3-030-01264-9_8

Identifiers

DOI
10.1007/978-3-030-01264-9_8
PMID
41328267
PMCID
PMC12664865
arXiv
1807.11164

Data Quality

Data completeness: 79%