A Survey of Multimodal Learning: Methods, Applications, and Future

Yuan Yuan, Zhaojian Li, Bin Zhao

科研成果: 期刊稿件文章同行评审

2 引用 (Scopus)

摘要

The multimodal interplay of the five fundamental senses-Sight, Hearing, Smell, Taste, and Touch-provides humans with superior environmental perception and learning skills. Adapted from the human perceptual system, multimodal machine learning tries to incorporate different forms of input, such as image, audio, and text, and determine their fundamental connections through joint modeling. As one of the future development forms of artificial intelligence, it is necessary to summarize the progress of multimodal machine learning. In this article, we start with the form of a multimodal combination and provide a comprehensive survey of the emerging subject of multimodal machine learning, covering representative research approaches, the most recent advancements, and their applications. Specifically, this article analyzes the relationship between different modalities in detail and sorts out the key issues in multimodal research from the application scenarios. Besides, we thoroughly reviewed state-of-The-Art methods and datasets covered in multimodal learning research. We then identify the substantial challenges and potential developing directions in this field. Finally, given the comprehensive nature of this survey, both modality-specific and task-specific researchers can benefit from this survey and advance the field.

源语言英语
文章编号167
期刊ACM Computing Surveys
57
7
DOI
出版状态已出版 - 20 2月 2025

指纹

探究 'A Survey of Multimodal Learning: Methods, Applications, and Future' 的科研主题。它们共同构成独一无二的指纹。

引用此