Abstract

The ability to learn richer network representations generally boosts the performance of deep learning models. To improve representation-learning in convolutional neural networks, we present a multi-branch architecture, which applies channel-wise attention across different network branches to leverage the complementary strengths of both feature-map attention and multi-path representation. Our proposed Split-Attention module provides a simple and modular computation block that can serve as a drop-in replacement for the popular residual block, while producing more diverse representations via cross-feature interactions. Adding a Split-Attention module into the architecture design space of RegNet-Y and FBNetV2 directly improves the performance of the resulting network. Replacing residual blocks with our Split-Attention module, we further design a new variant of the ResNet model, named ResNeSt, which outperforms EfficientNet in terms of the accuracy/latency trade-off.

Keywords

Computer scienceLeverage (statistics)ResidualModular designFeature learningArtificial intelligenceConvolutional neural networkBlock (permutation group theory)Representation (politics)ArchitectureComputationLatency (audio)Deep learningResidual neural networkTheoretical computer scienceComputer engineeringAlgorithm

Affiliated Institutions

Related Publications

Publication Info

Year
2022
Type
article
Pages
2735-2745
Citations
1177
Access
Closed

External Links

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

1177
OpenAlex

Cite This

Hang Zhang, Chongruo Wu, Zhongyue Zhang et al. (2022). ResNeSt: Split-Attention Networks. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) , 2735-2745. https://doi.org/10.1109/cvprw56347.2022.00309

Identifiers

DOI
10.1109/cvprw56347.2022.00309