Abstract

In this paper, we introduce a new channel pruning method to accelerate very deep convolutional neural networks. Given a trained CNN model, we propose an iterative two-step algorithm to effectively prune each layer, by a LASSO regression based channel selection and least square reconstruction. We further generalize this algorithm to multi-layer and multi-branch cases. Our method reduces the accumulated error and enhance the compatibility with various architectures. Our pruned VGG-16 achieves the state-of-the-art results by 5× speed-up along with only 0.3% increase of error. More importantly, our method is able to accelerate modern networks like ResNet, Xception and suffers only 1.4%, 1.0% accuracy loss under 2× speedup respectively, which is significant.

Keywords

SpeedupComputer sciencePruningConvolutional neural networkAlgorithmResidual neural networkMean squared errorChannel (broadcasting)Artificial intelligenceArtificial neural networkDeep neural networksDeep learningPattern recognition (psychology)MathematicsParallel computingTelecommunications

Affiliated Institutions

Related Publications

Publication Info

Year
2017
Type
preprint
Pages
1398-1406
Citations
2474
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2474
OpenAlex

Cite This

Yihui He, Xiangyu Zhang, Jian Sun (2017). Channel Pruning for Accelerating Very Deep Neural Networks. , 1398-1406. https://doi.org/10.1109/iccv.2017.155

Identifiers

DOI
10.1109/iccv.2017.155