Abstract | ||
---|---|---|
Multimodal Deep Boltzmann Machines (DBMs) have demonstrated huge successes in multimodal representation learning tasks. During inference, DBMs function as Recurrent Neural Nets (RNNs) because of the intractable distributions. To learn the parameters, optimizations can alternatively be operated on these surrogate RNNs with "truncated message passing". As a consequence, the gradient will propagate through a long chain without any local guidance which can potentially affects the optimization procedure. In this paper, we address this problem by adding skip connections during back-propagation while keeping the forward propagation (inference) untouched. With skip connections, we implicitly assign local "targets" for the states of intermediate inference loops to approach. Applied to different training criteria on different data sets, we demonstrate the proposed algorithms can consistently help to train better models while at a lower cost of training time. Experimental results show that our algorithms can achieve state-of-the-art performance on the Multimedia Information Retrieval (MIR) Flickr data set.
|
Year | DOI | Venue |
---|---|---|
2017 | 10.1145/3123266.3123332 | MM '17: ACM Multimedia Conference
Mountain View
California
USA
October, 2017 |
Keywords | Field | DocType |
Multimodal, Generative Model, Skip Connection | Computer vision,Data set,Recurrent neural nets,Computer science,Inference,Multimedia information retrieval,Artificial intelligence,Message passing,Machine learning,Feature learning,Generative model | Conference |
ISBN | Citations | PageRank |
978-1-4503-4906-2 | 1 | 0.39 |
References | Authors | |
24 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ning Zhang | 1 | 15 | 2.42 |
Yu Cao | 2 | 100 | 14.01 |
Benyuan Liu | 3 | 1534 | 101.09 |
Yan Luo | 4 | 148 | 19.24 |