Back to Search Start Over

STTG-net: a Spatio-temporal network for human motion prediction based on transformer and graph convolution network

Authors :
Lujing Chen
Rui Liu
Xin Yang
Dongsheng Zhou
Qiang Zhang
Xiaopeng Wei
Source :
Visual Computing for Industry, Biomedicine, and Art, Vol 5, Iss 1, Pp 1-15 (2022)
Publication Year :
2022
Publisher :
SpringerOpen, 2022.

Abstract

Abstract In recent years, human motion prediction has become an active research topic in computer vision. However, owing to the complexity and stochastic nature of human motion, it remains a challenging problem. In previous works, human motion prediction has always been treated as a typical inter-sequence problem, and most works have aimed to capture the temporal dependence between successive frames. However, although these approaches focused on the effects of the temporal dimension, they rarely considered the correlation between different joints in space. Thus, the spatio-temporal coupling of human joints is considered, to propose a novel spatio-temporal network based on a transformer and a gragh convolutional network (GCN) (STTG-Net). The temporal transformer is used to capture the global temporal dependencies, and the spatial GCN module is used to establish local spatial correlations between the joints for each frame. To overcome the problems of error accumulation and discontinuity in the motion prediction, a revision method based on fusion strategy is also proposed, in which the current prediction frame is fused with the previous frame. The experimental results show that the proposed prediction method has less prediction error and the prediction motion is smoother than previous prediction methods. The effectiveness of the proposed method is also demonstrated comparing it with the state-of-the-art method on the Human3.6 M dataset.

Details

Language :
English
ISSN :
25244442
Volume :
5
Issue :
1
Database :
Directory of Open Access Journals
Journal :
Visual Computing for Industry, Biomedicine, and Art
Publication Type :
Academic Journal
Accession number :
edsdoj.43b6399ef1ae4f30bef8ca92bfd0b800
Document Type :
article
Full Text :
https://doi.org/10.1186/s42492-022-00112-5