Title
Label Distribution For Multimodal Machine Learning
Abstract
Multimodal machine learning (MML) aims to understand the world from multiple related modalities. It has attracted much attention as multimodal data has become increasingly available in real-world application. It is shown that MML can perform better than single-modal machine learning, since multi-modalities containing more information which could complement each other. However, it is a key challenge to fuse the multi-modalities in MML. Different from previous work, we further consider the side-information, which reflects the situation and influences the fusion of multi-modalities. We recover multimodal label distribution (MLD) by leveraging the side-information, representing the degree to which each modality contributes to describing the instance. Accordingly, a novel framework named multimodal label distribution learning (MLDL) is proposed to recover the MLD, and fuse the multimodalities with its guidance to learn an in-depth understanding of the jointly feature representation. Moreover, two versions of MLDL are proposed to deal with the sequential data. Experiments on multimodal sentiment analysis and disease prediction show that the proposed approaches perform favorably against state-of-the-art methods.
Year
DOI
Venue
2022
10.1007/s11704-021-0611-6
FRONTIERS OF COMPUTER SCIENCE
Keywords
DocType
Volume
multimodal machine learning, label distribution learning, sentiment analysis, disease prediction
Journal
16
Issue
ISSN
Citations 
1
2095-2228
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Yi Ren100.34
xu ning22515.72
Miaogen Ling300.34
Xin Geng4155783.54