Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Deep Learning-Based Feature Extraction for Speech Emotion Recognition

View through CrossRef
Emotion recognition from speech signals is an important and challenging component of Human-Computer Interaction. In the field of speech emotion recognition (SER), many techniques have been utilized to extract emotions from speech signals, including many well-established speech analysis and classification techniques. This model can be built by using various methods such as RNN, SVM, deep learning, cepstral coefficients, and various other methods, out of which SVM normally gives us the highest accuracy. We propose a model that can identify emotions present in the speech, which can be identified by various parameters such as pitch, speaking rate, speech time, and frequency patterns. Emotion detection in digitized speech contains 3 components: Signal processing, Feature extraction, and Classification. The model first tries to remove the background noises then extract the features present in the speech and classify it into a single emotion. This model is capable of identifying seven different emotions that can be found in human speech. We can use different classifiers like GMM and HMM to classify features such as Spectral Subtraction, Wiener Filtering, Adaptive Filtering, and Deep Learning Techniques. This model can be used in various fields such as healthcare, security, psychology, medicine, education, and entertainment.
Title: Deep Learning-Based Feature Extraction for Speech Emotion Recognition
Description:
Emotion recognition from speech signals is an important and challenging component of Human-Computer Interaction.
In the field of speech emotion recognition (SER), many techniques have been utilized to extract emotions from speech signals, including many well-established speech analysis and classification techniques.
This model can be built by using various methods such as RNN, SVM, deep learning, cepstral coefficients, and various other methods, out of which SVM normally gives us the highest accuracy.
We propose a model that can identify emotions present in the speech, which can be identified by various parameters such as pitch, speaking rate, speech time, and frequency patterns.
Emotion detection in digitized speech contains 3 components: Signal processing, Feature extraction, and Classification.
The model first tries to remove the background noises then extract the features present in the speech and classify it into a single emotion.
This model is capable of identifying seven different emotions that can be found in human speech.
We can use different classifiers like GMM and HMM to classify features such as Spectral Subtraction, Wiener Filtering, Adaptive Filtering, and Deep Learning Techniques.
This model can be used in various fields such as healthcare, security, psychology, medicine, education, and entertainment.

Related Results

Robust speech recognition based on deep learning for sports game review
Robust speech recognition based on deep learning for sports game review
Abstract To verify the feasibility of robust speech recognition based on deep learning in sports game review. In this paper, a robust speech recognition model is bui...
Exploring Speech Emotion Recognition in Tribal Language with Deep Learning Techniques
Exploring Speech Emotion Recognition in Tribal Language with Deep Learning Techniques
Emotion is fundamental to interpersonal interactions since it assists mutual understanding. Developing human-computer interactions and a related digital product depends heavily on ...
What about males? Exploring sex differences in the relationship between emotion difficulties and eating disorders
What about males? Exploring sex differences in the relationship between emotion difficulties and eating disorders
Abstract Objective: While eating disorders (ED) are more commonly diagnosed in females, there is growing awareness that men also experience ED and may do so in a different ...
Introduction: Autonomic Psychophysiology
Introduction: Autonomic Psychophysiology
Abstract The autonomic psychophysiology of emotion has a long thought tradition in philosophy but a short empirical tradition in psychological research. Yet the past...
Analyzing Noise Robustness of Cochleogram and Mel Spectrogram Features in Deep Learning Based Speaker Recogntion
Analyzing Noise Robustness of Cochleogram and Mel Spectrogram Features in Deep Learning Based Speaker Recogntion
Abstract The performance of speaker recognition is very well in a clean dataset or without mismatch between training and test set. However, the performance is degraded with...
Depth-aware salient object segmentation
Depth-aware salient object segmentation
Object segmentation is an important task which is widely employed in many computer vision applications such as object detection, tracking, recognition, and ret...
Speech, communication, and neuroimaging in Parkinson's disease : characterisation and intervention outcomes
Speech, communication, and neuroimaging in Parkinson's disease : characterisation and intervention outcomes
<p dir="ltr">Most individuals with Parkinson's disease (PD) experience changes in speech, voice or communication. Speech changes often manifest as hypokinetic dysarthria, a m...

Back to Top