Back to Search
Start Over
A survey of transformer-based multimodal pre-trained modals.
- Source :
-
Neurocomputing . Jan2023, Vol. 515, p89-106. 18p. - Publication Year :
- 2023
-
Abstract
- • Multimodal Pre-trained models with document layout, vision-text and audio-text domains as input. • Collection of common multimodal downstream applications with related datasets. • Modality feature embedding strategies. • Cross-modality alignment pre-training tasks for different multimodal domains. • Variations of the audio-text cross-modal learning architecture. With the broad industrialization of Artificial Intelligence(AI), we observe a large fraction of real-world AI applications are multimodal in nature in terms of relevant data and ways of interaction. Pre-trained big models have been proven as the most effective framework for joint modeling of multi-modality data. This paper provides a thorough account of the opportunities and challenges of Transformer-based multimodal pre-trained model (PTM) in various domains. We begin by reviewing the representative tasks of multimodal AI applications, ranging from vision-text and audio-text fusion to more complex tasks such as document layout understanding. We particularly address the new multi-modal research domain of document layout understanding. We further analyze and compare the state-of-the-art Transformer-based multimodal PTMs from multiple aspects, including downstream applications, datasets, input feature embedding, and model architectures. In conclusion, we summarize the key challenges of this field and suggest several future research directions. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 09252312
- Volume :
- 515
- Database :
- Academic Search Index
- Journal :
- Neurocomputing
- Publication Type :
- Academic Journal
- Accession number :
- 160031079
- Full Text :
- https://doi.org/10.1016/j.neucom.2022.09.136