Title
Deep Multispectral Semantic Scene Understanding of Forested Environments Using Multimodal Fusion.
Abstract
Semantic scene understanding of unstructured environments is a highly challenging task for robots operating in the real world. Deep Convolutional Neural Network architectures define the state of the art in various segmentation tasks. So far, researchers have focused on segmentation with RGB data. In this paper, we study the use of multispectral and multimodal images for semantic segmentation and develop fusion architectures that learn from RGB, Near-InfraRed channels, and depth data. We introduce a first-of-its-kind multispectral segmentation benchmark that contains 15, 000 images and 366 pixel-wise ground truth annotations of unstructured forest environments. We identify new data augmentation strategies that enable training of very deep models using relatively small datasets. We show that our UpNet architecture exceeds the state of the art both qualitatively and quantitatively on our benchmark. In addition, we present experimental results for segmentation under challenging real-world conditions. Benchmark and demo are publicly available at http://deepscene.cs.uni-freiburg.de.
Year
DOI
Venue
2016
10.1007/978-3-319-50115-4_41
Springer Proceedings in Advanced Robotics
Keywords
Field
DocType
Semantic segmentation,Convolutional neural networks,Scene understanding,Multimodal perception
Computer vision,Segmentation,Convolutional neural network,Multispectral image,Communication channel,Control engineering,Ground truth,Multispectral segmentation,Artificial intelligence,RGB color model,Engineering,Robot
Conference
Volume
ISSN
Citations 
1
2511-1256
3
PageRank 
References 
Authors
0.43
0
4
Name
Order
Citations
PageRank
Abhinav Valada15910.54
Gabriel Leivas Oliveira22259.70
Thomas Brox37866327.52
W Burgard4144381393.44