Back to Search
Start Over
Prediction of evoked expression from videos with temporal position fusion.
- Source :
-
Pattern Recognition Letters . Aug2023, Vol. 172, p245-251. 7p. - Publication Year :
- 2023
-
Abstract
- • Temporal convolutional network with temporal position fusion for estimating evoked expressions in video. • Address the inconsistency of time steps during training and testing with temporal position fusion. • State-of-the-art result on Evoked Expressions from Videos (EEV) dataset. • Made huge improvements with temporal position fusion on MediaEval 2018 dataset. This paper introduces an approach for estimating evoked categories expression from videos with the temporal position fusion. Pre-trained models on large-scale datasets in computer vision and audio signals were used to extract the deep representation for timestamps in the video. A temporal convolution network, rather than an RNN-like architecture, was applied to explore temporal relationships due to its advantage in memory consumption and parallelism. Furthermore, to address the noise labels, the temporal position was fused with the deep learned feature to ensure the network differentiates the time steps when noise labels were removed from the training set. This technique helps the system gain a considerable improvement compared to other methods. We conducted experiments on EEV, a large-scale dataset for evoked expression from videos, and achieved a score of 0.054 in terms of Pearson correlation coefficient as a state-of-the-art result. Further experiments on a sub set of LIRIS-ACCEDE dataset - MediaEval 2018 benchmark, also demonstrated the effectiveness of our approach. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 01678655
- Volume :
- 172
- Database :
- Academic Search Index
- Journal :
- Pattern Recognition Letters
- Publication Type :
- Academic Journal
- Accession number :
- 169814888
- Full Text :
- https://doi.org/10.1016/j.patrec.2023.07.002