Title
Progressively diffused networks for semantic visual parsing.
Abstract
Recent deep models advance the task of semantic visual parsing by increasing the depth of networks and the resolution (size) of the predicted labelmaps. However, the contextual information within each layer and between layers is not fully explored. Long Short Term Memory Networks(LSTM) that learn to propagate information is well-suited to model pixels dependencies with respect to spacial locations within layers and depths across layers. Unlike previous LSTM-based methods that tend to enhance representation of each pixel only by involving the information from adjacent area. This work proposes Progressively Diffused Networks (PDNs) to deal with complex semantic parsing tasks. It can explore spatial dependencies in a larger field that represents the rich contextual information among pixels. The proposed model has three appealing properties. First, it enables information to be progressively broadcast across feature maps by stacking multiple diffusion layers. Second, in each layer, multiple convolutional LSTMs are adopted to generate a series of feature maps with different ranges of contexts. Third, in each LSTM unit, a special type of atrous filters are designed to capture the short range and long range dependencies from various neighbors. Extensive experiments demonstrate the effectiveness of PDNs to substantially improve the performances of existing LSTM-based models.
Year
DOI
Venue
2019
10.1016/j.patcog.2019.01.011
Pattern Recognition
Keywords
Field
DocType
Visual understanding,Image segmentation,Recurrent neural networks,Representation learning
Broadcasting,Contextual information,Pattern recognition,Long short term memory,Artificial intelligence,Pixel,Parsing,Mathematics,Stacking
Journal
Volume
Issue
ISSN
90
1
0031-3203
Citations 
PageRank 
References 
2
0.36
37
Authors
6
Name
Order
Citations
PageRank
Ruimao Zhang132518.86
Wei Yang21897.87
Zhanglin Peng3264.43
Pengxu Wei492.87
Xiaogang Wang59647386.70
Liang Lin63007151.07