Javascript must be enabled to continue!
Sign Language Recognition with Multimodal Sensors and Deep Learning Methods
View through CrossRef
Sign language recognition is essential in hearing-impaired people’s communication. Sign language recognition is an important concern in computer vision and has been developed with rapid progress in image recognition technology. However, sign language recognition using a general monocular camera has problems with occlusion and recognition accuracy in sign language recognition. In this research, we aim to improve accuracy by using a 2-axis bending sensor as an aid in addition to image recognition. We aim to achieve higher recognition accuracy by acquiring hand keypoint information of sign language actions captured by a monocular RGB camera and adding sensor assist. To improve sign language recognition, we need to propose new AI models. In addition, the amount of dataset is small because it uses the original data set of our laboratory. To learn using sensor data and image data, we used MediaPipe, CNN, and BiLSTM to perform sign language recognition. MediaPipe is a method for estimating the skeleton of the hand and face provided by Google. In addition, CNN is a method that can learn spatial information, and BiLSTM can learn time series data. Combining the CNN and BiLSTM methods yields higher recognition accuracy. We will use these techniques to learn hand skeletal information and sensor data. Additionally, the 2-axis Bending sensor glove data support training AI model. Using these methods, we aim to improve the recognition accuracy of sign language recognition by combining sensor data and hand skeleton data. Our method performed better than using skeletal information, achieving 96.5% accuracy in Top-1.
Title: Sign Language Recognition with Multimodal Sensors and Deep Learning Methods
Description:
Sign language recognition is essential in hearing-impaired people’s communication.
Sign language recognition is an important concern in computer vision and has been developed with rapid progress in image recognition technology.
However, sign language recognition using a general monocular camera has problems with occlusion and recognition accuracy in sign language recognition.
In this research, we aim to improve accuracy by using a 2-axis bending sensor as an aid in addition to image recognition.
We aim to achieve higher recognition accuracy by acquiring hand keypoint information of sign language actions captured by a monocular RGB camera and adding sensor assist.
To improve sign language recognition, we need to propose new AI models.
In addition, the amount of dataset is small because it uses the original data set of our laboratory.
To learn using sensor data and image data, we used MediaPipe, CNN, and BiLSTM to perform sign language recognition.
MediaPipe is a method for estimating the skeleton of the hand and face provided by Google.
In addition, CNN is a method that can learn spatial information, and BiLSTM can learn time series data.
Combining the CNN and BiLSTM methods yields higher recognition accuracy.
We will use these techniques to learn hand skeletal information and sensor data.
Additionally, the 2-axis Bending sensor glove data support training AI model.
Using these methods, we aim to improve the recognition accuracy of sign language recognition by combining sensor data and hand skeleton data.
Our method performed better than using skeletal information, achieving 96.
5% accuracy in Top-1.
Related Results
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
<p><em><span style="font-size: 11.0pt; font-family: 'Times New Roman',serif; mso-fareast-font-family: 'Times New Roman'; mso-ansi-language: EN-US; mso-fareast-langua...
Increased life expectancy of heart failure patients in a rural center by a multidisciplinary program
Increased life expectancy of heart failure patients in a rural center by a multidisciplinary program
Abstract
Funding Acknowledgements
Type of funding sources: None.
INTRODUCTION Patients with heart failure (HF)...
Development of a multimodal imaging system based on LIDAR
Development of a multimodal imaging system based on LIDAR
(English) Perception of the environment is an essential requirement for the fields of autonomous vehicles and robotics, that claim for high amounts of data to make reliable decisio...
Sign Language Linguistics
Sign Language Linguistics
Sign language linguistics is one of the younger areas of linguistic research, having been a field in its own right only since the 1960s, when the first research investigating sign ...
Where Can I Sign My Language?: A Systematic Literature Review (2000–2019) of Young People's Access to Sign Language Socialisation Spaces in the Nordic Countries
Where Can I Sign My Language?: A Systematic Literature Review (2000–2019) of Young People's Access to Sign Language Socialisation Spaces in the Nordic Countries
This article presents the results from a systematic literature review of Nordic research studies from 2000 to 2019, addressing the question of children and adolescents' access to s...
Imagined worldviews in John Lennon’s “Imagine”: a multimodal re-performance / Visões de mundo imaginadas no “Imagine” de John Lennon: uma re-performance multimodal
Imagined worldviews in John Lennon’s “Imagine”: a multimodal re-performance / Visões de mundo imaginadas no “Imagine” de John Lennon: uma re-performance multimodal
Abstract: This paper addresses the issue of multimodal re-performance, a concept developed by us, in view of the fact that the famous song “Imagine”, by John Lennon, was published ...
Developing Amharic Sign Language Recognition Model for Amharic Characters Using Deep Learning Approach
Developing Amharic Sign Language Recognition Model for Amharic Characters Using Deep Learning Approach
Abstract
Hearing-impaired people use Sign Language to communicate with each other as well as with other communities. Usually, they are unable to communicate with normal peo...
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
Multimodal sentiment analysis is an essential task in natural language processing which refers to the fact that machines can analyze and recognize emotions through logical reasonin...

