Javascript must be enabled to continue!
Sign Language Recognition with Multimodal Sensors and Deep Learning Methods
View through CrossRef
Sign language recognition is essential in hearing-impaired people’s communication. Sign language recognition is an important concern in computer vision and has been developed with rapid progress in image recognition technology. However, sign language recognition using a general monocular camera has problems with occlusion and recognition accuracy in sign language recognition. In this research, we aim to improve accuracy by using a 2-axis bending sensor as an aid in addition to image recognition. We aim to achieve higher recognition accuracy by acquiring hand keypoint information of sign language actions captured by a monocular RGB camera and adding sensor assist. To improve sign language recognition, we need to propose new AI models. In addition, the amount of dataset is small because it uses the original data set of our laboratory. To learn using sensor data and image data, we used MediaPipe, CNN, and BiLSTM to perform sign language recognition. MediaPipe is a method for estimating the skeleton of the hand and face provided by Google. In addition, CNN is a method that can learn spatial information, and BiLSTM can learn time series data. Combining the CNN and BiLSTM methods yields higher recognition accuracy. We will use these techniques to learn hand skeletal information and sensor data. Additionally, the 2-axis Bending sensor glove data support training AI model. Using these methods, we aim to improve the recognition accuracy of sign language recognition by combining sensor data and hand skeleton data. Our method performed better than using skeletal information, achieving 96.5% accuracy in Top-1.
Title: Sign Language Recognition with Multimodal Sensors and Deep Learning Methods
Description:
Sign language recognition is essential in hearing-impaired people’s communication.
Sign language recognition is an important concern in computer vision and has been developed with rapid progress in image recognition technology.
However, sign language recognition using a general monocular camera has problems with occlusion and recognition accuracy in sign language recognition.
In this research, we aim to improve accuracy by using a 2-axis bending sensor as an aid in addition to image recognition.
We aim to achieve higher recognition accuracy by acquiring hand keypoint information of sign language actions captured by a monocular RGB camera and adding sensor assist.
To improve sign language recognition, we need to propose new AI models.
In addition, the amount of dataset is small because it uses the original data set of our laboratory.
To learn using sensor data and image data, we used MediaPipe, CNN, and BiLSTM to perform sign language recognition.
MediaPipe is a method for estimating the skeleton of the hand and face provided by Google.
In addition, CNN is a method that can learn spatial information, and BiLSTM can learn time series data.
Combining the CNN and BiLSTM methods yields higher recognition accuracy.
We will use these techniques to learn hand skeletal information and sensor data.
Additionally, the 2-axis Bending sensor glove data support training AI model.
Using these methods, we aim to improve the recognition accuracy of sign language recognition by combining sensor data and hand skeleton data.
Our method performed better than using skeletal information, achieving 96.
5% accuracy in Top-1.
Related Results
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
<p><em><span style="font-size: 11.0pt; font-family: 'Times New Roman',serif; mso-fareast-font-family: 'Times New Roman'; mso-ansi-language: EN-US; mso-fareast-langua...
Development of a multimodal imaging system based on LIDAR
Development of a multimodal imaging system based on LIDAR
(English) Perception of the environment is an essential requirement for the fields of autonomous vehicles and robotics, that claim for high amounts of data to make reliable decisio...
Sign Language Linguistics
Sign Language Linguistics
Sign language linguistics is one of the younger areas of linguistic research, having been a field in its own right only since the 1960s, when the first research investigating sign ...
Imagined worldviews in John Lennon’s “Imagine”: a multimodal re-performance / Visões de mundo imaginadas no “Imagine” de John Lennon: uma re-performance multimodal
Imagined worldviews in John Lennon’s “Imagine”: a multimodal re-performance / Visões de mundo imaginadas no “Imagine” de John Lennon: uma re-performance multimodal
Abstract: This paper addresses the issue of multimodal re-performance, a concept developed by us, in view of the fact that the famous song “Imagine”, by John Lennon, was published ...
Where Can I Sign My Language?: A Systematic Literature Review (2000–2019) of Young People's Access to Sign Language Socialisation Spaces in the Nordic Countries
Where Can I Sign My Language?: A Systematic Literature Review (2000–2019) of Young People's Access to Sign Language Socialisation Spaces in the Nordic Countries
This article presents the results from a systematic literature review of Nordic research studies from 2000 to 2019, addressing the question of children and adolescents' access to s...
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
Multimodal sentiment analysis is an essential task in natural language processing which refers to the fact that machines can analyze and recognize emotions through logical reasonin...
Sign language dactyl recognition based on machine learning algorithms
Sign language dactyl recognition based on machine learning algorithms
In the course of our research work, the American, Russian and Turkish sign languages were analyzed. The program of recognition of the Kazakh dactylic sign language with the use of ...
Sign Language to Text Conversion
Sign Language to Text Conversion
Abstract: Sign language, being one of the oldest and most natural forms of communication, serves as a crucial means of expression for individuals with hearing and speech impairment...

