Title
Improved Multimodal Representation Learning with Skip Connections.
Abstract
Multimodal Deep Boltzmann Machines (DBMs) have demonstrated huge successes in multimodal representation learning tasks. During inference, DBMs function as Recurrent Neural Nets (RNNs) because of the intractable distributions. To learn the parameters, optimizations can alternatively be operated on these surrogate RNNs with "truncated message passing". As a consequence, the gradient will propagate through a long chain without any local guidance which can potentially affects the optimization procedure. In this paper, we address this problem by adding skip connections during back-propagation while keeping the forward propagation (inference) untouched. With skip connections, we implicitly assign local "targets" for the states of intermediate inference loops to approach. Applied to different training criteria on different data sets, we demonstrate the proposed algorithms can consistently help to train better models while at a lower cost of training time. Experimental results show that our algorithms can achieve state-of-the-art performance on the Multimedia Information Retrieval (MIR) Flickr data set.
Year
DOI
Venue
2017
10.1145/3123266.3123332
MM '17: ACM Multimedia Conference Mountain View California USA October, 2017
Keywords
Field
DocType
Multimodal, Generative Model, Skip Connection
Computer vision,Data set,Recurrent neural nets,Computer science,Inference,Multimedia information retrieval,Artificial intelligence,Message passing,Machine learning,Feature learning,Generative model
Conference
ISBN
Citations 
PageRank 
978-1-4503-4906-2
1
0.39
References 
Authors
24
4
Name
Order
Citations
PageRank
Ning Zhang1152.42
Yu Cao210014.01
Benyuan Liu31534101.09
Yan Luo414819.24