Affiliations 

  • 1 Department of Electrical Engineering, Yuan Ze University, Taoyuan 32003, Taiwan
  • 2 Department of Computer Science & Information Engineering, National Central University, Taoyuan City 32001, Taiwan
  • 3 Department of Computer Science, University Tunku Abdul Rahman, Kampar 31900, Malaysia
Sensors (Basel), 2020 May 21;20(10).
PMID: 32455537 DOI: 10.3390/s20102907

Abstract

Semantic segmentation of street view images is an important step in scene understanding for autonomous vehicle systems. Recent works have made significant progress in pixel-level labeling using Fully Convolutional Network (FCN) framework and local multi-scale context information. Rich global context information is also essential in the segmentation process. However, a systematic way to utilize both global and local contextual information in a single network has not been fully investigated. In this paper, we propose a global-and-local network architecture (GLNet) which incorporates global spatial information and dense local multi-scale context information to model the relationship between objects in a scene, thus reducing segmentation errors. A channel attention module is designed to further refine the segmentation results using low-level features from the feature map. Experimental results demonstrate that our proposed GLNet achieves 80.8% test accuracy on the Cityscapes test dataset, comparing favorably with existing state-of-the-art methods.

* Title and MeSH Headings from MEDLINE®/PubMed®, a database of the U.S. National Library of Medicine.