1. Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space
- Author
-
Kosmas Dimitropoulos, Dimitrios Konstantinidis, Ilias Papastratis, and Petros Daras
- Subjects
General Computer Science ,Computer science ,business.industry ,cross-modal learning ,Feature extraction ,General Engineering ,Space (commercial competition) ,Sign language ,computer.software_genre ,continuous sign language recognition ,deep-learning ,joint latent space ,Modal ,Encoding (memory) ,Computer vision ,General Materials Science ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Artificial intelligence ,Focus (optics) ,Joint (audio engineering) ,business ,Representation (mathematics) ,lcsh:TK1-9971 ,computer ,Natural language processing - Abstract
Continuous Sign Language Recognition (CSLR) refers to the challenging problem of recognizing sign language glosses and their temporal boundaries from weakly annotated video sequences. Previous methods focus mostly on visual feature extraction neglecting text information and failing to effectively model the intra-gloss dependencies. In this work, a cross-modal learning approach that leverages text information to improve vision-based CSLR is proposed. To this end, two powerful encoding networks are initially used to produce video and text embeddings prior to their mapping and alignment into a joint latent representation. The purpose of the proposed cross-modal alignment is the modelling of intra-gloss dependencies and the creation of more descriptive video-based latent representations for CSLR. The proposed method is trained jointly with video and text latent representations. Finally, the aligned video latent representations are classified using a jointly trained decoder. Extensive experiments on three well-known sign language recognition datasets and comparison with state-of-the-art approaches demonstrate the great potential of the proposed approach.
- Published
- 2020
- Full Text
- View/download PDF