Abstract

Semantic image segmentation is a basic street scene understanding task in autonomous driving, where each pixel in a high resolution image is categorized into a set of semantic labels. Unlike other scenarios, objects in autonomous driving scene exhibit very large scale changes, which poses great challenges for high-level feature representation in a sense that multi-scale information must be correctly encoded. To remedy this problem, atrous convolution[14]was introduced to generate features with larger receptive fields without sacrificing spatial resolution. Built upon atrous convolution, Atrous Spatial Pyramid Pooling (ASPP)[2] was proposed to concatenate multiple atrous-convolved features using different dilation rates into a final feature representation. Although ASPP is able to generate multi-scale features, we argue the feature resolution in the scale-axis is not dense enough for the autonomous driving scenario. To this end, we propose Densely connected Atrous Spatial Pyramid Pooling (DenseASPP), which connects a set of atrous convolutional layers in a dense way, such that it generates multi-scale features that not only cover a larger scale range, but also cover that scale range densely, without significantly increasing the model size. We evaluate DenseASPP on the street scene benchmark Cityscapes[4] and achieve state-of-the-art performance.

Keywords

Artificial intelligenceComputer sciencePyramid (geometry)Computer visionSegmentationPoolingPattern recognition (psychology)Feature (linguistics)Scale (ratio)Convolution (computer science)Image resolutionFeature extractionMathematicsGeographyCartographyArtificial neural networkGeometry

Related Publications

Publication Info

Year
2018
Type
article
Citations
1566
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1566
OpenAlex
130
Influential
1207
CrossRef

Cite This

Maoke Yang, Kun Yu, Chi Zhang et al. (2018). DenseASPP for Semantic Segmentation in Street Scenes. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition . https://doi.org/10.1109/cvpr.2018.00388

Identifiers

DOI
10.1109/cvpr.2018.00388

Data Quality

Data completeness: 77%