Back to Search Start Over

Cyclic Autoencoder for Multimodal Data Alignment Using Custom Datasets.

Authors :
Zhenyu Tang
Jin Liu
Chao Yu
Wang, Y. Ken
Source :
Computer Systems Science & Engineering; 2021, Vol. 39 Issue 1, p37-54, 18p
Publication Year :
2021

Abstract

The subtitle recognition under multimodal data fusion in this paper aims to recognize text lines from image and audio data. Most existing multimodal fusion methods tend to be associated with pre-fusion as well as post-fusion, which is not reasonable and difficult to interpret. We believe that fusing images and audio before the decision layer, i.e., intermediate fusion, to take advantage of the complementary multimodal data, will benefit text line recognition. To this end, we propose: (i) a novel cyclic autoencoder based on convolutional neural network. The feature dimensions of the two modal data are aligned under the premise of stabilizing the compressed image features, thus the high-dimensional features of different modal data are fused at the shallow level of the model. (ii) A residual attention mechanism that helps us improve the performance of the recognition. Regions of interest in the image are enhanced and regions of disinterest are weakened, thus we can extract the features of the text regions without further increasing the depth of the model (iii) a fully convolutional network for video subtitle recognition. We choose DenseNet-121 as the backbone network for feature extraction, which effectively enabling the recognition of video subtitles in complex backgrounds. The experiments are performed on our custom datasets, and the automatic and manual evaluation results show that our method reaches the state-of-the-art. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
02676192
Volume :
39
Issue :
1
Database :
Supplemental Index
Journal :
Computer Systems Science & Engineering
Publication Type :
Academic Journal
Accession number :
161570260
Full Text :
https://doi.org/10.32604/csse.2021.017230