A Survey of Information Technology Applications To Treat Fear of
A Survey of Information Technology Applications To Treat Fear of
DOI: https://doi.org/10.31185/wjcm.Vol1.Iss1.8
Received: January 2021; Accepted: February 2021; Available online: March 2021
Keywords: fear, phobias, treatment, computerbased toolsPublic speaking started to gain much attention when it comes
to phobias, which is anxiety for new presenters. In some cases, specialists consider that avoiding the phenomenon which
causes the phobia is sufficient treatment; in others, the exact opposite, being gradually exposed to the object of fear may
lead to a cure. We have to start looking for other psychotherapeutic methods, innovative ones, to help people surpass their
immense fears and improve their ability to give presentations. The current article presents a survey on discovering fear
and anxiety when preventing and treating it and analyses their utility as tools for learning how to overcome this type of
phobias, thus improving presentation ability. Using IT-based solutions for treating presented this fear, especially anxiety
for new presenters. The current methods of dealing with the fear of public speaking will be reviewed, as well as Clarify the
technology (tools, systems, and applications) based used for detecting and treatment. We will analyze research that studies
how to detect fear and the ways to treat it, the concept behind their mechanism and the possibility of exploiting them in
presentations. therefore, the paper debates these IT instruments and applications in this field. Based on the results of the
survey, we will propose an appropriate mechanism for detecting degrees and types of fear when presenting presentations
and their treatment.
1. INTRODUCTION
Modern life often involves situations where we are required to speak in public, both in our professional lives, for
instance when presenting results of our work in front of colleagues, and in our working life, such as when teaching and
giving presentations.
Given the prevalence of public speaking situations in modern professional and personal life, it is natural that some
individuals want to improve their ability to speak in public. Additionally, anxiety about public speaking is very common,
and some people experience an uncontrollable amount of stress when preparing or public speaking. These two cases
require the development of some methods and tools to support assessing people’s ability to speak to the public, training
them in public speaking skills and reducing anxiety and stress during public speaking [1, 2].
Emotion is defined as a conscious mental reaction subjectively experienced and directed towards a specific object,
accompanied by physiological and behavioral changes in the body. The field of affective computing aims to enhance the
interaction between the human and the machines by identifying emotions and designing applications that automatically
adapt to these changes. [2]
Effective computing is a study of systems or devices that can identify and simulate emotions and their treatment meth-
ods. This field is applicable to education, medicine, social sciences, entertainment and so on. The purpose of emotional
computing is to improve user experience and quality of life, which is why various emotional models have been proposed
over the years and effective mathematical models have been applied to extract, categorize and analyze emotions. [3]
It has drawn the attention of researchers from interdisciplinary domains, being at the confluence of psychology,
medicine, and computer science. With applications in education, cognitive-behavioral sciences, healthcare and entertain-
ment, affective computing deals with recognizing and modeling human emotions in a way that would improve overall user
experience. To classify emotions, some separate models and dimensions have been proposed and applied over the years.
Discrete patterns of influence depend on the existence of a set of fundamental feelings from which the most complex
feelings are derived. Dimensional models rely on a multidimensional space where each axis represents the values of the
emotional component [4, 5].
Public speaking started to gain much attention when it comes to phobias, which is anxiety for new presenters. De-
pending on the survey results we will use new integrated artificial intelligence techniques, we will propose computational
models that detect models and levels of emotion and voice of those suffering from a phobia. As a goal, develop a phobias
treatment system that will automatically determine fear levels and adjust exposure according to the current effective user
situation.
Examination and designing of human conduct are basic for human-driven systems to anticipate the result of commu-
nication between social, and to improve the connection between people or among people and PCs. Human conduct is
communicated and saw regarding verbal and viewable prompts (for example hand and body signals, facial expression).
These conduct signs can be caught and handled to anticipate the result of social connections. Open talking is a significant
part of the human correspondence. A decent speaker is lucid, has persuading non-verbal communication, and regularly,
can significantly influence individuals. While the accomplishment of open talking generally relies upon the substance
of the discussion, and the speaker’s verbal conduct, non-verbal (viewable) signals, for example, motions and physical
appearance play asignifi control in public speaking.
Our paper consists of two sections: the first section is the introduction, in the second part we Show and analyzes a group
of research that discusses the IT applications to detect and treat fear of public speaking, and the last section summarizes
conclusion.
41
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
Another epitome of the method is regulation for characterizing speech. The method involves a PC system having a
(CPU), an inputting device, one memory for saving information indicative of a talking signal, and an outputting device.
The PC method likewise contains a rationale for getting and examining the discourse signal, the rationale for partitioning
the discourse sign, and rationale for taking one characteristic from the discourse signal. The system likewise includes a
database of discourse signals and statistics attainable of a computer for comparing the Voice signal and an outputting tool
coupled to the PC for advising a user of the passionate situation that disclosed in the Voice signal.
42
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
for a lift pitch. This examination followed a semi exploratory set-up where they investigated the learning impacts of the
feedback gave by the.PT.
The Presentation Trainer.
is a multimodal instrument intended to help the exercise of public talking abilities, on granting the learner the notes
about various parts of nonverbal correspondence? It tracks the client’s voice and body to explain his present showing. In
light of this showing the Presentation Trainer chooses the kind of interference that be displayed as notes to the client.
feedback was developed system thinking about the outcomes from past examinations that show how troublesome it is for
users to see and effectively decipher continuous notes while rehearsing their talking. to introducing the learner experience
assessment of users who utilized the Presentation Trainer to rehearse for a lift pitch, demonstrating that the input gave by
the Presentation Trainer important impacts on education. studies have confirmed that notes given from a mentor impacts
the improvement of public talking aptitudes [9] and that the extent of this impact relies upon how these notes were given to
the user. A significant factor that influences the improvement of these aptitudes is the time where notes are given. like the
nonverbal connection of the user, quick notes have demonstrated to be effective and productive [10]. Thusly the adaptation
of the PT depicted here can examine the user’s presentation, and manner select nonverbal connection to be displayed as
notes.
1. Hand/Body Segmentation: The essential strategy of hand division is to identify and understand the hand area in the
picture which is got hand motions, and subtract them from the surroundings.
2. Gesture Modeling: In the period of hand movement assessment, sundry hand moves and motions are assembled and
registered to send them as traineeship and testing information to make a model that will be utilized during order.
3. Gesture Classification: hand motion estimate, which will be utilized for preprocessing and advantages conclusion.
1. Identifying Filler words: In request to recognize the filler words, the application should transform talking to write,
and the procedure is done through the sound copy [16].
43
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
2. Sound segment of the discourse in ’.wav’ form is embed into the app. Volume variety is specified. [19], [20], [21], [22], [23], [24].
4. Distinguishing the vocal variety and make a chart of volume variation. [25].
6. The system converts audio to text format to show differences in sound and size [26].
• The effect of speaker normalization (SN) is also studied, which removes the mean of features and normalizes them
to unit variance. Experiments are under a speaker-independent condition.
• Additionally, a feature selection technique is assessed to obtain good features from the set of features extracted in.
The rest of the chapter is organized as follows. In the next section, they researchers start by introducing the nature of
speech emotions. Section 3 describes features they extracted from a speech signal. A feature selection method and machine
learning algorithms used for SER are presented. Section 4 reports on the databases they used and presents the simulation
results obtained using different features and different machine learning (ML) paradigms. Section 5 closes this chapter by
analyses and conclusion.
44
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
45
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
also been identified as one of the most influential interventions in learning .Having a human tutor providing us with high
quality feedback whenever they have time to practice our skills is neither an affordable nor a feasible solution. In our effort
to study an affordable solution for this feedback availability challenge, they explored the topic ‘public speaking skills’.
Where they followed a design-based research methodology developing different prototypes of the Presentation Trainer
(PT). The PT is an example of an automated feedback tool that tracks the learners’ voice and body. It provides them with
feedback about their nonverbal communication, with the purpose to support them with the development of their public
speaking skills.
In this article researchers describe the current version of the PT, and present the user experience evaluation of a study,
where participants had to prepare themselves for an elevator pitch. This study followed a quasi-experimental set-up where
they explored the learning effects of the feedback provided by the PT.
3. CONCLUSION
In this paper, we are presenting a survey of information technology applications to treat the fear of public speaking.
There are many types of applications were used different methods to treat fear of public speaking that are presented in
this survey. which were many types of methods used that have yielded satisfactory results. Through our in-depth study,
46
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
we note that some of these methods need to be further developed in order to obtain accurate results. We suggest that data
mining techniques that are more appropriate be used in order to avoid defects, one research [31] used the KNN algorithm
and that the main disadvantage of this algorithm is that it is a lazy learner, that is, it does not learn anything from the
training data and simply uses the same training data for classification. This can be addressed using the deep learning
algorithm for better and accurate results.
FUNDING
None
ACKNOWLEDGEMENT
None
CONFLICTS OF INTEREST
The author declares no conflict of interest.
REFERENCES
[1] F. Dermody, A. Sutherland, and M. Farren A Multi-modal System for Public Speaking Pilot Study on Evaluation of Real-Time Feedback, vol. 1,
pp. 499–501, 2015.
[2] M. Chollet, W. Torsten, L. Morency, and S. Scherer A Multimodal Corpus for the Assessment of Public Speaking Ability and Anxiety, pp. 488–495.
[3] O. Bălan Emotions classification based on biophysical signals and machine learning techniques, 2019.
[4] W. Torsten and S. Scherer Automatic Assessment and Analysis of Public Speaking Anxiety : A Virtual Audience Case Study.
[5] F. Moldoveanu Dimensions and Machine Learning Techniques, pp. 1–18, 2019.
[6] Greatest.coaches.in.NFL.History, 2015.
[7] J. Hattie and & H Timperley, “The power of feedback”,” Review of Educational Research Journal, pp. 81–112, 2007.
[8] T. Anderson and J. Shattuck, “Design-Based Research A Decade of Progress in Education Research?,” Educational Researcher Journl, pp. 16–25.
[9] D. Kerby and & J Romine, “Develop Oral Presentation Skills Through Accounting Curriculum Design and Course-Embedded Assessment,”
Business, vol. 85, pp. 172–179, 2009.
[10] P. King, J. Young, and & R Behnke, “Public speaking performance improvement as a function of information processing in immediate and
delayed feedback interventions,” Communication Education Journal, vol. 49, pp. 365–374, 2000.
[11] “Every Toastmasters Journey Starts with A Single Speech,” Available: [Accessed, 2017.
[12] K. Lee, H. Hon, and R. Reddy, “An overview of the SPHINX speech recognition system,” IEEE Transactions on Acoustics, Speech, and Signal
Processing, vol. 38, pp. 35–45, 1990.
[13] J. Goodman, A bit of progress in language modeling. 2001.
[14] W. Walker, P. Lamere, P. Kwok, B. Raj, R. Singh, E. Gouvea, P. Wolf, and J. Woelfel, “Sphinx-4: A flexible open source framework for speech
recognition,” Sun Microsystems Inc, 2004.
[15] . . Microsoft and B. Speech 2017.
[16] G. Bohouta, Comparision Speech Recognition Systems. 2017.
[17] 2017.
[18] . . Spellchecker 2017.
[19] " Mathworks, Matlab, and Mathworks 2017.
[20] D. Wood, “Sound: Definition, Influences, Pitch & Volume,” Study.com, 2017.
[21] O. C. M. L. Six TarsosDSP, a Real-Time Audio Processing, 2014.
[22] A. D. Cheveigné and H. Kawahara, “YIN, a fundamental frequency estimator for speech and music,” The Journal of the Acoustical Society of
America, vol. 111, 2002.
[23] P. Mcleod and G. Wyvill, “A Smarter Way to Find Pitch,” Proceedings of the International Computer Music Conference (ICMC 2005), 2005.
[24] M. J. Ross, H. L. Shaffer, A. Cohen, R. Freudberg, and H. J. Manley, “Average Magnitude Difference Function Pitch Extractor,” IEEE Trans. on
Acoustics, Speech, and Signal Processing, vol. 22, pp. 353–362, 1974.
[25] P. Jahoda MPAndroidChart, 2017.
[26] “microsoft Cognitive Services 2017.
[27] L. Kerkeni, Y. Serrestou, M. Mbarki, K. Raoof, M. A. Mahjoub, and C. Cleder Automatic Speech Emotion Recognition Using Machine Learning,
Social Media and Machine Learning, Alberto Cano, IntechOpen, 2019.
[28] A. Davletcharova, S. Sugathan, B. Abraham, and A. P. James, “Detection and Analysis of Emotion from Speech Signals,” the Second International
Symposium on Computer Vision and the Internet, vol. 58, 2015.
[29] T. Giraud, “Multimodal Expressions of Stress during a Public Speaking Task: Collection, Annotation and Global Analyses,” in Humaine
Association Conference on Affective Computing and Intelligent Interaction, pp. 417–422, 2013.
[30] J. Schneider, D. Börner, P. V. Rosmalen, and M. Specht, “Presentation Trainer, your Public Speaking Multimodal Coach,” in Proceedings of the
2015 ACM on International Conference on Multimodal Interaction (ICMI ’15), pp. 539–546, Association for Computing Machinery.
[31] D. K. nska and A. Pelikant, “Recognition of Human Emotion from a Speech Signal Based on Plutchik’s Model”,” International Journal of
Electronics and Telecommunications, vol. 58, no. 2, pp. 2300–1933, 2012.
[32] S. Shangavi, S. Jeyamaalmarukan, A. Jathevan, M. Umatharsini, and P. Samarasinghe, “Self-Speech Evaluation with Speech Recognition and
Gesture Analysis,” in 2018 National Information Technology Conference (NITC), pp. 1–7, 2018.
47
Qays Algahreeb et al. , Wasit Journal of Computer and Mathematics Science, Vol. 0 No. 0 (2021) p. 40-48
[33] N. Damodar, H. Y. Vani, and A. M. A, “Voice Emotion Recognition using CNN and Decision Tree,” International Journal of Innovative Technology
and Exploring Engineering (IJITEE), pp. 2278–3075, 2019.
[34] V. A. Petrushin and B. Grove Detecting emotions using voice signal analysis, 2007.
[35] T. Giraud, J. Hua, and A. Delaborde, “Multimodal Expressions of Stress during a Public Speaking Task: Collection, Annotation and Global
Analyses,” in Humaine Association Conference on Affective Computing and Intelligent Interaction, pp. 417–422, 2013.
[36] J. Schneider, D. Boerner, P, and M. Specht Presentation Trainer, your Public Speaking Multimodal Coach.
[37] S. Shangavi, S. Jeyamaalmarukan, A. Jathevan, and M Umatharsini Pradeepa Samarasinghe, Self-Speech Evaluation with Speech Recognition
and Gesture Analysis, 2018.
48