Back to Search Start Over

Introduction and Analysis of an Event-Based Sign Language Dataset

Authors :
Pablo Negri
Bernabe Linares-Barranco
Teresa Serrano-Gotarredona
Ajay Vasudevan
Source :
FG
Publication Year :
2020
Publisher :
IEEE, 2020.

Abstract

Human gestures recognition is a complex visual recognition task where motion across time distinguishes the type of action. Automatic systems tackle this problem using complex machine learning architectures and training datasets. In recent years, the use and success of robust deep learning techniques was compatible with the availability of a great number of these sets. This paper presents SL-Animals-DVS, an event-based action dataset captured by a Dynamic Vision Sensor (DVS). The DVS records humans performing sign language gestures of various animals as a continuous spike flow at very low latency. This is especially suited for sign language gestures which are usually made at very high speeds. We also benchmark the recognition performance on this data using two state-of-the-art Spiking Neural Networks (SNN) recognition systems. SNNs are naturally compatible to make use of the temporal information that is provided by the DVS where the information is encoded in the spike times. The dataset has about 1100 samples of 58 subjects performing 19 sign language gestures in isolation at different scenarios, providing a challenging evaluation platform for this emerging technology.

Details

Database :
OpenAIRE
Journal :
2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020)
Accession number :
edsair.doi...........40b912b3ad96fada59056077648ead60
Full Text :
https://doi.org/10.1109/fg47880.2020.00069