Abstract

This paper aims to accelerate the test-time computation of deep convolutional neural networks (CNNs). Unlike existing methods that are designed for approximating linear filters or linear responses, our method takes the nonlinear units into account. We minimize the reconstruction error of the nonlinear responses, subject to a low-rank constraint which helps to reduce the complexity of filters. We develop an effective solution to this constrained nonlinear optimization problem. An algorithm is also presented for reducing the accumulated error when multiple layers are approximated. A whole-model speedup ratio of 4× is demonstrated on a large network trained for ImageNet, while the top-5 error rate is only increased by 0.9%. Our accelerated model has a comparably fast speed as the "AlexNet" [11], but is 4.7% more accurate.

Keywords

SpeedupComputer scienceConvolutional neural networkNonlinear systemComputationAlgorithmConstraint (computer-aided design)Rank (graph theory)Mathematical optimizationArtificial intelligenceMathematicsParallel computing

Affiliated Institutions

Related Publications

Network In Network

Abstract: We propose a novel deep network structure called In Network (NIN) to enhance model discriminability for local patches within the receptive field. The conventional con...

2014 arXiv (Cornell University) 1037 citations

Publication Info

Year
2015
Type
preprint
Citations
289
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

289
OpenAlex

Cite This

Xiangyu Zhang, Jianhua Zou, Ming Xiang et al. (2015). Efficient and accurate approximations of nonlinear convolutional networks. . https://doi.org/10.1109/cvpr.2015.7298809

Identifiers

DOI
10.1109/cvpr.2015.7298809