Sign Language Production With Avatar Layering: A Critical Use Case over Rare Words

Autor/a: KIM, Jung-Ho; HWANG, Eui Jun; CHO, Sukmin; LEE, Du Hui; PARK, Jong C.
Año: 2022
Editorial: European Language Resources Association
Tipo de código: Copyright
Soporte: Digital


Medios de comunicación y acceso a la información » Nuevas Tecnologías


Sign language production (SLP) is the process of generating sign language videos from spoken language expressions. Since sign languages are highly under-resourced, existing vision-based SLP approaches suffer from out-of-vocabulary (OOV) and test-time generalization problems and thus generate low-quality translations. To address these problems, we introduce an avatar-based SLP system composed of a sign language translation (SLT) model and an avatar animation generation module. Our Transformer-based SLT model utilizes two additional strategies to resolve these problems: named entity transformation to reduce OOV tokens and context vector generation using a pretrained language model (e.g., BERT) to reliably train the decoder. Our system is validated on a new Korean-Korean Sign Language (KSL) dataset of weather forecasts and emergency announcements. Our SLT model achieves an 8.77 higher BLEU-4 score and a 4.57 higher ROUGE-L score over those of our baseline model. In a user evaluation, 93.48% of named entities were successfully identified by participants, demonstrating marked improvement on OOV issues.

En Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022), pp. 1519–1528.