Title
A unifying representation for pixel-precise distance estimation
Abstract
We propose a new representation of distance information that is independent from any specific acquisition device, based on the size of portrayed subjects. In this alternative description, each pixel of an image is associated with the size, in real life, of what it represents. Using our proposed representation, datasets acquired with different devices can be effortlessly combined to build more powerful models, and monocular distance estimation can be performed on images acquired from devices that were never used during training. To assess the advantages of the proposed representation, we used it to train a fully convolutional neural network that predicts with pixel-precision the size of different subjects depicted in the image, as a proxy for their distance. Experimental results show that our representation, allowing the combination of heterogeneous training datasets, makes it possible for the trained network to gain better results at test time.
Year
DOI
Venue
2019
10.1007/s11042-018-6568-2
Multimedia Tools and Applications
Keywords
Field
DocType
Distance estimation, Depth estimation, Perspective geometry, Convolutional neural network
Computer vision,Pattern recognition,Convolutional neural network,Computer science,Artificial intelligence,Pixel,Monocular,Perspective (geometry)
Journal
Volume
Issue
ISSN
78.0
10
1573-7721
Citations 
PageRank 
References 
1
0.38
22
Authors
3
Name
Order
Citations
PageRank
Simone Bianco122624.48
Marco Buzzelli2294.91
Raimondo Schettini31476154.06