Paper 5
Paper 5
net/publication/361108119
CITATIONS READS
15 12,214
3 authors, including:
Radhey Shyam
Babu Banarasi Das University
40 PUBLICATIONS 231 CITATIONS
SEE PROFILE
All content following this page was uploaded by Radhey Shyam on 06 June 2022.
Abstract– Facial emotional expression is a part of face various facial extraction features, emotional databases,
recognition, it has always been an easy task for humans, classifier algorithms, and so on.[4] The classical FER
but achieving the same with a computer algorithm is consists of two main steps: feature extraction and emotion
challenging. With the recent and continuous recognition. In addition, image pre-processing, including
advancements in computer vision and machine learning, face detection, cropping, and resizing. Face detection crops
it is possible to detect emotions in images, videos, etc. A the facial region after removing the backdrop and non-face
face expression recognition method based on the Deep areas. Finally, the retrieved characteristics are used to
Neural Networks especially the convolutional neural classify emotions, which is commonly done with the help of
network (CNN) and an image edge detection is proposed. neural networks (NN) and other machine learning
The edge of each layer of the image is retrieved in the approaches. The challenge of facial emotion recognition is
convolution process after the facial expression image is to automatically recognize facial emotion states with high
normalized. To maintain the texture picture's edge accuracy. Therefore, it is challenging to find the similarity
structure information, the retrieved edge information is of the same emotional state between different people since
placed on each feature image. In this research, several they may express the same emotional state in various ways.
datasets are investigated and explored for training As an example, the expression may vary in different
expression recognition models. The purpose of this situations such as the individual’s mood, skin colour, age,
paper is to make a study on face emotion detection and and the environment surrounding. Generally, FER is
recognition via Machine learning algorithms and deep separated into three major stages as shown in Figure 1: (i)
learning. This research work will present deeper insights Face Detection, (ii)Feature Extraction, and (iii) Emotion
into Face emotion detection and Recognition. It will also Classification.
highlight the variables that have an impact on its
efficacy.
176
International Journal of Engineering Applied Sciences and Technology, 2022
Vol. 7, Issue 1, ISSN No. 2455-2143, Pages 176-179
Published Online May 2022 in IJEAST (http://www.ijeast.com)
177
International Journal of Engineering Applied Sciences and Technology, 2022
Vol. 7, Issue 1, ISSN No. 2455-2143, Pages 176-179
Published Online May 2022 in IJEAST (http://www.ijeast.com)
inputs to the classification module which finally categorizes A. Support Vector Machine (SVM): SVM is one of the
different emotions. famous statistical techniques employed in machine learning
Feature extraction will be divided into two categories which to analyse data used for classification and multivariate
are; (i) feature base and (ii) appearance base. analysis. SVM used different kernel functions to map data
in input space into high-dimensional feature spaces.
A. Convolutional Neural Network (CNN) Currently, CNN is
one of the foremost mainstream approaches to deep learning B. Neural Network (NN): NN executes a nonlinear
techniques. It uses a variation of multilayer perceptron reduction of the input dimensionality. It makes a statistical
designed to want minimal pre-processing. It gets its name determination regarding the category of the observed
from the type of hidden layers it has. Convolutional layers, expression. Every output unit will estimate the probability
pooling layers, fully connected layers, and normalising of the examined expression belonging to the associated
layers are common components of a CNN's hidden layers. category [5].
[2]
1. Inception-V1 toV3
The Inception network was a significant step forward in the
evolution of CNN classifiers. It is a 22-layer design with a
total of 5M parameters. It employed numerous techniques to
improve performance, both in terms of speed and precision.
This is frequently utilised in machine learning applications
[20, 21]. Inception V2, It is the 24M parameter successor to
Inception-V1. Inception-v3 is a popular image recognition
model that has been shown to achieve more than 78.1
Fig 2: Emotion Detection Process percent accuracy on the Image Net dataset. However, it is
not widely utilised. [1]
V. DATASET
To perform an experiment on FER, a regular database is
required. The information will be perceived as primary or
secondary. A primary dataset consumes an extended period
to be completed with dataset collection. For study in FER, a
Fig 3: Image Classification spread of datasets available currently There are few datasets
available for the emotion recognition problem; among those,
Karolinska Directed Emotional Faces (KDEF)and Japanese
Female facial features (JAFFE) datasets are well-known and
regarded during this study. The dataset’s images are divided
into seven main emotion categories [3]. The KDEF dataset
(also refer as KDEF for simplicity, henceforth) was
developed by Karolinska Institute, Sweden. Specifically, the
aim of the dataset was to use for perception memory
emotional attention, and backward masking experiment. The
Fig 4: Convolution Filter Operation dataset contains 4900 photos of 70 people, each of whom is
depicted in seven different emotional states.
3. Expression Classification
This stage is performed by a classifier. There are various VI. RESULT AND DISCUSSION
classifications methods accustomed extract expressions. To analyse the performance of the algorithm, the FER-2013
expression dataset was used initially. Dataset had only 7178
Supervised Learning- with 412 posers, causing accuracy to reach up to 55%
Supervised learning is a way of training a system using maximum. To overcome the problem of low efficiency,
labelled data. The tagged data serves as a manager. The multiple datasets were downloaded from the Internet, and
model is given both inputs and outputs to learn from. also author’s own pictures of different expressions were
Following that, the model would forecast for a new data included. As the number of images in the dataset increases,
point. Classification and regression are the two types of the accuracy also increased. We kept 70% of 11K dataset
supervised learning. [3] images as training and 30% of dataset images as testing
images. The number of layers and the number of filters, for
178
International Journal of Engineering Applied Sciences and Technology, 2022
Vol. 7, Issue 1, ISSN No. 2455-2143, Pages 176-179
Published Online May 2022 in IJEAST (http://www.ijeast.com)
179