Title
Learning contextual superpixel similarity for consistent image segmentation
Abstract
This paper addresses the problem of image segmentation by iterative region aggregations starting from an initial superpixel decomposition. Classical approaches for this task compute superpixel similarity using distance measures between superpixel descriptor vectors. This usually poses the well-known problem of the semantic gap and fails to properly aggregate visually non-homogeneous superpixels that belong to the same high-level object. This work proposes to use random forests to learn the merging probability between adjacent superpixels in order to overcome the aforementioned issues. Compared to existing works, this approach learns the fusion rules without explicit similarity measure computation. We also introduce a new superpixel context descriptor to strengthen the learned characteristics towards better similarity prediction. Image segmentation is then achieved by iteratively merging the most similar superpixel pairs selected using a similarity weighting objective function. Experimental results of our approach on four datasets including DAVIS 2017 and ISIC 2018 show its potential compared to state-of-the-art approaches.
Year
DOI
Venue
2020
10.1007/s11042-019-08391-6
Multimedia Tools and Applications
Keywords
Field
DocType
Context description, Superpixels similarity, Machine learning, Random forests, Image segmentation, Region-growing
Computer vision,Pattern recognition,Computer science,Image segmentation,Artificial intelligence
Journal
Volume
Issue
ISSN
79
3
1380-7501
Citations 
PageRank 
References 
0
0.34
0
Authors
5
Name
Order
Citations
PageRank
Mahaman Sani Chaibou100.34
Pierre-Henri Conze2337.73
Karim Kalti3208.58
Mohamed Ali Mahjoub48332.74
B. Solaiman5768.79