Abstract

This paper aims to accelerate the test-time computation of convolutional neural networks (CNNs), especially very deep CNNs [1] that have substantially impacted the computer vision community. Unlike previous methods that are designed for approximating linear filters or linear responses, our method takes the nonlinear units into account. We develop an effective solution to the resulting nonlinear optimization problem without the need of stochastic gradient descent (SGD). More importantly, while previous methods mainly focus on optimizing one or two layers, our nonlinear method enables an asymmetric reconstruction that reduces the rapidly accumulated error when multiple (e.g., ≥ 10) layers are approximated. For the widely used very deep VGG-16 model [1] , our method achieves a whole-model speedup of 4 × with merely a 0.3 percent increase of top-5 error in ImageNet classification. Our 4 × accelerated VGG-16 model also shows a graceful accuracy degradation for object detection when plugged into the Fast R-CNN detector [2] .

Keywords

Convolutional neural networkSpeedupComputer scienceStochastic gradient descentArtificial intelligenceComputationDeep learningNonlinear systemObject detectionDetectorFocus (optics)Pattern recognition (psychology)AlgorithmGradient descentArtificial neural networkParallel computing

Affiliated Institutions

Related Publications

Publication Info

Year
2015
Type
article
Volume
38
Issue
10
Pages
1943-1955
Citations
830
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

830
OpenAlex

Cite This

Xiangyu Zhang, Jianhua Zou, Kaiming He et al. (2015). Accelerating Very Deep Convolutional Networks for Classification and Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence , 38 (10) , 1943-1955. https://doi.org/10.1109/tpami.2015.2502579

Identifiers

DOI
10.1109/tpami.2015.2502579