Submitted:
01 August 2024
Posted:
02 August 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
- Translation: The system efficiently translates videos of Israeli Sign Language signs and gestures into written Hebrew.
- Learning: The system allows users to practice ISL and receive immediate feedback on the accuracy of their signed words. Utilizing a web camera, the application captures the signs, processes them through a trained classifier, and provides users with information on the correctness of their signing.
2. Related Work
3. Signsability Dataset
,
, and
, with 200 samples each. The signs for these letters are static, thus images are sufficient to capture and recognize them. The videos, on the other hand, are for words that involve motion and dynamic gestures in their sign language pronunciation. The 400 collected videos capture five distinct words, with 80 samples each. Each video was recorded using a simple web camera. The average duration of the sample videos is 3 seconds, which are then split into 30 frames.4. Methodology
4.1. Landmarks Extraction
4.2. Preprocessing
4.3. Training
4.3.1. Static Signs Recognition - MediaPipe with a Simple NN
, (b)
, and (c)
, with 200 images per class. By leveraging the robust capabilities of MediaPipe, we were able to utilize the numerical representation of the extracted landmarks to construct a relatively simple yet effective neural network. We opted to develop the model from scratch, obviating the need for reliance on pretrained models. The network architecture comprises three dense layers with ReLU activation functions, having two dropout layers between them. The model contains a total of 1,114 trainable parameters. For the loss function, we selected sparse categorical cross-entropy, enabling the model to classify a single word for each sample.4.3.2. Dynamic Signs Recognition - MediaPipe with RNN
5. Results and Discussions
5.1. Static Signs Recognition Results
5.2. Dynamic Signs Recognition Results
5.3. Website Application
- User-Friendly Navigation: The site features a clear and intuitive layout, allowing users to easily access different sections and functionalities. Menu options and interactive elements are strategically placed to facilitate smooth exploration and learning.
- Responsive Design: The website is optimized for various devices, including desktops, tablets, and smartphones. This ensures that users can engage with the learning materials from any device, providing flexibility and convenience.
- Multimedia Integration: To enhance the learning experience, the site integrates multimedia resources such as instructional videos, animated demonstrations, and example gestures. These resources help users understand and practice sign language more effectively.
- Personalized Learning Experience: The website offers personalized features such as progress tracking dashboards. These features are designed to meet individual learning needs and preferences, making the experience more engaging and effective.
6. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
| 1 | The study was conducted as a part of the undergraduate project for a B.Sc. degree in Software Engineering. |
| 2 | The application will be provided at no cost to any individual who requests it via email. We will gladly fulfill all such requests. |
| 3 | The ’Signsability’ dataset is available for download on Kaggle: https://kaggle.com/datasets/fe1ca51ef2c13347756617aedfc611d5699035cdc74ca72e89a357e466f8ebb5
|
| 4 | |
| 5 |
URL: http://129.159.153.217.
Development of the site is ongoing; additional features and improvements are constantly added.
|
References
- Chengk, K. Chengk, K. American Sign Language vs Israeli Sign Language. StartASL 2022, pp. 2–5.
- Murali, L.; Ramayya, L.; Santosh, V.A. Sign Language Recognition System Using Convolutional Neural Network and ComputerVision. International Journal of Engineering Innovations in Advanced Technology 2022, 4, 138–141. [Google Scholar]
- Obi, Y.; Claudio, K.S.; Budiman, V.M.; Achmad, S.; Kurniawan, A. Sign language recognition system for communicating to people with disabilities. Procedia Computer Science 2023, 216, 13–20. [Google Scholar] [CrossRef]
- Balaha, M.M.; El-Kady, S.; Balaha, H.M.; Salama, M.; Emad, E.; Hassan, M.; Saafan, M.M. A vision-based deep learning approach for independent-users Arabic sign language interpretation. Multimedia Tools and Applications 2023, 82, 6807–6826. [Google Scholar] [CrossRef]
- Gogoi, J.B.S.D.A.B.A.C.D. Real-time Assamese Sign Language Recognition using MediaPipe and Deep Learning. Elsevier 2023, pp. 1386–1393.
- Shamitha, S.H.; Badarinath, K. Sign Language Recognition Utilizing LSTM And Mediapipe For Dynamic Gestures Of Indian Sign Language. International Journal for Multidisciplinary Research 2023, 5, 138–152. [Google Scholar]





,
,
.| Class | P | R | F1-score |
|---|---|---|---|
![]() |
1 | 0.99 | 099 |
![]() |
0.98 | 0.92 | 0.95 |
![]() |
0.91 | 0.99 | 0.95 |
| Average | 0.96 | 0.96 | 0.96 |
| Class | English equivalent | P | R | F1-score |
|---|---|---|---|---|
![]() |
hello | 0.98 | 0.97 | 0.97 |
![]() |
thanks | 0.99 | 1 | 0.99 |
![]() |
bye | 0.89 | 0.96 | 0.92 |
![]() |
yes | 0.92 | 0.94 | 0.93 |
![]() |
no | 0.87 | 0.91 | 0.89 |
| Average | 0.95 | 0.96 | 0.94 | |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).




