Title
Boundary Detection Benchmarking: Beyond F-Measures
Abstract
For an ill-posed problem like boundary detection, human labeled datasets play a critical role. Compared with the active research on finding a better boundary detector to refresh the performance record, there is surprisingly little discussion on the boundary detection benchmark itself. The goal of this paper is to identify the potential pitfalls of today's most popular boundary benchmark, BSDS 300. In the paper, we first introduce a psychophysical experiment to show that many of the "weak" boundary labels are unreliable and may contaminate the benchmark. Then we analyze the computation of f-measure and point out that the current benchmarking protocol encourages an algorithm to bias towards those problematic "weak" boundary labels. With this evidence, we focus on a new problem of detecting strong boundaries as one alternative. Finally, we assess the performances of 9 major algorithms on different ways of utilizing the dataset, suggesting new directions for improvements.
Year
DOI
Venue
2013
10.1109/CVPR.2013.276
CVPR
Keywords
Field
DocType
boundary detection benchmarking,active research,strong boundary,ill-posed problem,new problem,beyond f-measures,better boundary detector,popular boundary benchmark,boundary detection benchmark,boundary label,new direction,boundary detection,computer vision,benchmarking,image segmentation,algorithm design and analysis,classification algorithms,reliability,benchmark testing,edge detection,f measure,detectors
Data mining,Computer science,Edge detection,Boundary detection,Artificial intelligence,Detector,Machine learning,Benchmarking,Computation
Conference
Volume
Issue
ISSN
2013
1
1063-6919
Citations 
PageRank 
References 
17
0.77
21
Authors
3
Name
Order
Citations
PageRank
Xiaodi Hou1206972.53
Alan L. Yuille2103391902.01
Christof Koch37248973.47