Back to Search Start Over

Knowledge-driven Egocentric Multimodal Activity Recognition.

Authors :
YI HUANG
XIAOSHAN YANG
JUNYU GAO
JITAO SANG
CHANGSHENG XU
Source :
ACM Transactions on Multimedia Computing, Communications & Applications; Dec2020, Vol. 16 Issue 4, p1-21, 21p
Publication Year :
2020

Abstract

Recognizing activities from egocentric multimodal data collected by wearable cameras and sensors, is gaining interest, as multimodal methods always benefit from the complementarity of different modalities. However, since high-dimensional videos contain rich high-level semantic information while low-dimensional sensor signals describe simple motion patterns of the wearer, the large modality gap between the videos and the sensor signals raises a challenge for fusing the raw data. Moreover, the lack of large-scale egocentric multimodal datasets due to the cost of data collection and annotation processes makes another challenge for employing complex deep learning models. To jointly deal with the above two challenges, we propose a knowledge-driven multimodal activity recognition framework that exploits external knowledge to fuse multimodal data and reduce the dependence on large-scale training samples. Specifically, we design a dual-GCLSTM (Graph Convolutional LSTM) and a multi-layer GCN (Graph Convolutional Network) to collectively model the relations among activities and intermediate objects. The dual-GCLSTM is designed to fuse temporal multimodal features with top-down relation-aware guidance. In addition, we apply a co-attention mechanism to adaptively attend to the features of different modalities at different timesteps. The multi-layer GCN aims to learn relation-aware classifiers of activity categories. Experimental results on three publicly available egocentric multimodal datasets show the effectiveness of the proposed model. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
15516857
Volume :
16
Issue :
4
Database :
Complementary Index
Journal :
ACM Transactions on Multimedia Computing, Communications & Applications
Publication Type :
Academic Journal
Accession number :
148931799
Full Text :
https://doi.org/10.1145/3409332