Title
Context Prior for Scene Segmentation
Abstract
Recent works have widely explored the contextual dependencies to achieve more accurate segmentation results. However, most approaches rarely distinguish different types of contextual dependencies, which may pollute the scene understanding. In this work, we directly supervise the feature aggregation to distinguish the intra-class and inter-class context clearly. Specifically, we develop a Context Prior with the supervision of the Affinity Loss. Given an input image and corresponding ground truth, Affinity Loss constructs an ideal affinity map to supervise the learning of Context Prior. The learned Context Prior extracts the pixels belonging to the same category, while the reversed prior focuses on the pixels of different classes. Embedded into a conventional deep CNN, the proposed Context Prior Layer can selectively capture the intra-class and inter-class contextual dependencies, leading to robust feature representation. To validate the effectiveness, we design an effective Context Prior Network (CPNet). Extensive quantitative and qualitative evaluations demonstrate that the proposed model performs favorably against state-of-the-art semantic segmentation approaches. More specifically, our algorithm achieves 46.3% mIoU on ADE20K, 53.9% mIoU on PASCAL-Context, and 81.3% mIoU on Cityscapes. Code is available at https://git.io/ContextPrior.
Year
DOI
Venue
2020
10.1109/CVPR42600.2020.01243
CVPR
DocType
ISSN
Citations 
Conference
IEEE Conference on Computer Vision and Pattern Recognition 2020
4
PageRank 
References 
Authors
0.41
24
6
Name
Order
Citations
PageRank
Changqian Yu1224.46
Jingbo Wang2213.43
Changxin Gao318838.01
Gang Yu438219.85
Chunhua Shen54817234.19
Nong Sang647572.22