Sign Language Fingerspelling Recognition using Synthetic Data

Autor/a: FOWLEY, Frank; VENTRESQUE, Anthony
Año: 2021
Editorial: 29th Irish Conference on Artificial Intelligence and Cognitive Science (AICS)
Tipo de código: Copyright
Soporte: Digital


Medios de comunicación y acceso a la información » Nuevas Tecnologías


Sign Language Recognition (SLR)is a Computer Vision (CV) and Machine Learning (ML) task, with potential applications that would be beneficial to the Deaf community, which includes not only deaf per- sons but also hearing people who use Sign Languages. SLR is particularly challenging due to the lack of training datasets for CV and ML models, which impacts their overall accuracy and robustness. In this paper, we explore the use of synthetic images to augment a dataset of fingerspelling signs and we evaluate whether this could be used to reliably increase the performance of an SLR system. Our model is based on a pretrained con- volutional network, fine-tuned using synthetic images, and tested using a corpus dataset of real recordings of native signers. An accuracy of 71% recognition was achieved using skeletal wireframe image training datasets and using the MediaPipe pose estimation model in the test pipeline. This compares favourably with state-of-the-art CV models which achieve up to 62% accuracy with “in-the-wild” fingerspelling test datasets.