Abstract

We present a simple but powerful architecture of convolutional neural network, which has a VGG-like inference-time body composed of nothing but a stack of 3 × 3 convolution and ReLU, while the training-time model has a multi-branch topology. Such decoupling of the training-time and inference-time architecture is realized by a structural re-parameterization technique so that the model is named RepVGG. On ImageNet, RepVGG reaches over 80% top-1 accuracy, which is the first time for a plain model, to the best of our knowledge. On NVIDIA 1080Ti GPU, RepVGG models run 83% faster than ResNet-50 or 101% faster than ResNet-101 with higher accuracy and show favorable accuracy-speed trade-off compared to the state-of-the-art models like EfficientNet and RegNet. The code and trained models are available at https://github.com/megvii-model/RepVGG.

Keywords

Computer scienceInferenceResidual neural networkConvolution (computer science)Convolutional neural networkCode (set theory)Decoupling (probability)Artificial intelligenceArchitectureFLOPSSimple (philosophy)Parallel computingPattern recognition (psychology)AlgorithmComputer engineeringArtificial neural networkProgramming language

Affiliated Institutions

Related Publications

Publication Info

Year
2021
Type
article
Pages
13728-13737
Citations
2124
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2124
OpenAlex
195
Influential
1929
CrossRef

Cite This

Xiaohan Ding, Xiangyu Zhang, Ningning Ma et al. (2021). RepVGG: Making VGG-style ConvNets Great Again. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 13728-13737. https://doi.org/10.1109/cvpr46437.2021.01352

Identifiers

DOI
10.1109/cvpr46437.2021.01352
arXiv
2101.03697

Data Quality

Data completeness: 88%