@inproceedings{f014b7978c88418c83e44455c7b28916,
title = "ArSL21L: Arabic Sign Language Letter Dataset Benchmarking and an Educational Avatar for Metaverse Applications",
abstract = "It is complicated for the PwHL (people with hearing loss) to make a relationship with social majority, which naturally demands an interactive auto computer systems that have ability to understand sign language. With a trending Metaverse applications using augmented reality (AR) and virtual reality (VR), it is easier and interesting to teach sign language remotely using an avatar that mimics the gesture of a person using AI (Artificial Intelligence)-based system. There are various proposed methods and datasets for English SL (sign language); however, it is limited for Arabic sign language. Therefore, we present our collected and annotated Arabic Sign Language Letters Dataset (ArSL21L) consisting of 14202 images of 32 letter signs with various backgrounds collected from 50 people. We benchmarked our ArSL21L dataset on state-of-the-art object detection models, i.e., 4 versions of YOLOv5. Among the models, YOLOv5l achieved the best result with COCOmAP of 0.83. Moreover, we provide comparison results of classification task between ArSL2018 dataset, the only Arabic sign language letter dataset for classification task, and our dataset by running classification task on in-house short video. The results revealed that the model trained on our dataset has a superior performance over the model trained on ArSL2018. Moreover, we have created our prototype avatar which can mimic the ArSL (Arabic Sign Language) gestures for Metaverse applications. Finally, we believe, ArSL21L and the ArSL avatar will offer an opportunity to enhance the research and educational applications for not only the PwHL, but also in general real and virtual world applications. Our ArSL21L benchmark dataset is publicly available for research use on the Mendeley.",
keywords = "Arabic sign language, YOLOv5, benchmarking, deep learning, image classification, metaverse, object detection",
author = "Ganzorig Batnasan and Munkhjargal Gochoo and Otgonbold, {Munkh Erdene} and Fady Alnajjar and Shih, {Timothy K.}",
note = "Publisher Copyright: {\textcopyright} 2022 IEEE.; 13th IEEE Global Engineering Education Conference, EDUCON 2022 ; Conference date: 28-03-2022 Through 31-03-2022",
year = "2022",
doi = "10.1109/EDUCON52537.2022.9766497",
language = "???core.languages.en_GB???",
series = "IEEE Global Engineering Education Conference, EDUCON",
publisher = "IEEE Computer Society",
pages = "1814--1821",
editor = "Mohammed Jemni and Ilhem Kallel and Abdeljalil Akkari",
booktitle = "Proceedings of the 2022 IEEE Global Engineering Education Conference, EDUCON 2022",
}