Title
Semantic Image Segmentation with Improved Position Attention and Feature Fusion
Abstract
Encoder-decoder structure is an universal method for semantic image segmentation. However, some important information of images will lost with the increasing depth of convolutional neural network (CNN), and the correlation between arbitrary pixels will get worse. This paper designs a novel image segmentation model to obtain dense feature maps and promote segmentation effects. In encoder stage, we employ ResNet-50 to extract features, and then add a spatial pooling pyramid (SPP) to achieve multi-scale feature fusion. In decoder stage, we provide an improved position attention module to integrate contextual information effectively and remove the trivial information through changing the construction way of attention matrix. Furthermore, we also propose the feature fusion structure to generate dense feature maps by preforming element-wise sum operation on the upsampling features and corresponding encoder features. The simulation results illustrate that the average accuracy and mIOU on CamVid dataset can reach 90.7% and 63.1% respectively. It verifies the effectiveness and reliability of the proposed method.
Year
DOI
Venue
2020
10.1007/s11063-020-10240-9
NEURAL PROCESSING LETTERS
Keywords
DocType
Volume
Semantic image segmentation,Spatial pooling pyramid,Improved position attention,Feature fusion,Dense feature map
Journal
52.0
Issue
ISSN
Citations 
SP1.0
1370-4621
0
PageRank 
References 
Authors
0.34
0
3
Name
Order
Citations
PageRank
Hegui Zhu1465.73
Yan Miao200.34
Xiangde Zhang39115.32