Research Methods Speech Recognition
Research Methods Speech Recognition
Speech
Recognition in ML
An Application of Machine Learning
1
[2]
Speech
Recognition in ML
Converting spoken language to text using
computational models.
Key Applications:
• Virtual Assistants (e.g., Siri, Alexa)
• Healthcare Diagnostics
• Emotion Analysis for Mental Health
2
[2]
3
Core Speech Recognition
Aspects
Ensure high precision in Enable rapid analysis and Go beyond basic transcription by
transcribing spoken language, response for applications that identifying emotional tones and
addressing challenges like rely on immediate feedback, such sentiments in speech, useful for
accents, background noise, and as virtual assistants and applications in mental health and
diverse dialects. interactive customer support. user experience enhancement.
4
[1]
EDNUS
Emotion Detection of Neurological disorder
Using Speech
EDNUS demonstrates the application of speech
recognition in detecting emotions, which is
crucial for supporting individuals with
neurological disorders.
Provides real-time emotional insights for better
support in patient care.
Focuses on speech-based emotion recognition,
addressing a key need in mental health.
Offers a non-invasive, accessible approach for
monitoring emotional states.
Its use of deep autoencoders and ensemble
models addresses challenges in capturing
subtle emotional cues, providing an innovative
approach within healthcare, unlike other
proposed models.
5
[1]
6
Classification of
Emotions
[3]
7
Workflow Algorithm/ SLM / DAE
8
Super Learner Model
The Logistic Regression meta-learner
combines predictions from base models
to produce a final, more accurate output,
effectively optimizing each model’s
contribution to improve overall
performance.
9
Feature Extraction using DAE
[5]
10
Results
With and Without Auto Encoder
Accuracy Score 84 76
F1 Score 84 76
Jaccard Score 72 61
11
Our Traction
Benefits of using Deep Auto Encoder
INFERENCE
IMPACT
[4]
12
Summary of Key
Takeaways
• Accurate Emotion Detection: EDNUS effectively captures emotional
nuances in speech using deep autoencoders for feature extraction.
13
Back to Agenda
References
[1] https://www.shutterstock.com
[2] https://chatgpt.com
[3] https://www.canva.com
[4] https://colab.research.google.com
[5] https://doi.org/10.1007/978-981-19-2130-8_42
14
Thank you!
15