Feature Extraction and Classification Methods of Facial Expression: A Surey
Feature Extraction and Classification Methods of Facial Expression: A Surey
Corresponding Author:
Moe Moe Htay,
Faculty of Computer Science,
University of Computer Studies, Mandalay (UCSM),
Patheingyi, Myanmar.
Email: [email protected]
1. INTRODUCTION
In Artificial Intelligent era, facial expression recognition (FER) is interesting and challenging task
with the problems of limited dataset, different environments, pose, occlusion, person variation etc. FER
systems have been applied many systems such as human-computer-interaction (HCI), games, animation of
data-driven, surveillance, clinical monitoring etc., [1]. Ekman and Friesen, psychologists from America defined
six universal facial expressions: fear, happiness, anger, disgust, surprise, and sadness and also explored Action
Units based facial action coding system (FACS) to describe facial features of expressions [2]. Facial
expressions convey nonverbal communication cues that play a significant role in interpersonal relations. Some
literatures work adding on other emotions neutral, contempt, and many compound facial emotions. Some
researchers employed on handcrafted features extracted using algorithms and others employed on complicated
features extracted using deep learning methods. In this paper, we explored the feature extraction methods,
feature descriptors, classification methods, methods of feature dimension reduction, frameworks of the facial
expression recognition system and the comparison of the results. The remainder of the paper is organized as
follows. In section 2, Literature of current FER system. Typical FER system is shown in Section 3. After that,
two types feature of facial images is discussed in section 4, and section 5 described facial databases for FER
system. Section 6 describes the problem statement of FER system. In the last section, conclusion and future
work is presented.
Feature extraction and classification methods of facial expression: a surey… (Moe Moe Htay)
28 ISSN: 2722-3221
Used DCNN adding data augmentation, cross entropy and L2 multi-class SVM [13]. In [14], weighted
center regression adaptive feature mapping (W-CR-AFM) for feature distribution and CNN for feature training
on CK+, Radbound Faces database (RaFD), Amsterdam dynamic facial expression set (ADFES) and
proprietary database. Different of other papers, spatial normalization and feature enhancement preprocessing
methods are used. The recognition obtained 89.84%, 96.27%, 92.70% for CK+, RaFD and ADFES
respectively. Address illumination problem of real-world facial images using fast fourier transform and contrast
limited adaptive histogram equalization (FFT+CLAHE) for poor illumination and then applied merged binary
pattern code (MBPC). PCA is used as a method of feature dimension reduction and k-NN as a classifier on
SFEW dataset [15]. Released a new database iCV-MEFED at FG work-shop. Multi-modality CNN is compared
with CNN for micro emotion recognition in the paper. The proposed network extracted firstly visual and
geometrical information of features then concatenated these into a long vector. The feature vector is fed to the
hinge loss layer. The framework is better performance than CNN with the misclassification of 80.212137 using
caffe [16]. Also proposed another three works of the work-shop. The first winner method using CNN with
geometric representation of landmark displacement leading better results compared with texture-only
information. The recognition accuracy achieves 51.84% for seven expressions and 13.7% for compound
emotion with the performance of average time 1.57ms using GPU or 30ms using CPU [17].
Employed deep emotional attention model using cross channel CNN by adding attention modulator
on the bimodal face and body (FABO) benchmark database. The system applied CNN to learn the location of
face expressions in a cluttered scene. The study has shown that the experimentation of one expression attention
mechanism and two expression attention mechanism. The accuracy of the framework with attention is better
than that of without attention [18]. Proposed a robust facial landmark extraction method by combining data-
driven of fully convolution network (FCN) and model-driven of pre-trained point distribution model (PDM)
with three steps estimation-correction-tuning (ECT). The computation of response maps of global landmark
estimation is trained by FCN and then the maximum points of the maps are fitted with PDM to generate initial
facial shape. In the final, a weighted version of regularized landmark mean-shift (RLMS) is applied to fine-
tune the facial shape iteratively [19].
Designed to learn NN architecture with three loss functions fully supervised, weekly supervised and
hybrid regularization. The experimentation of the proposed model has achieved promising results on CK+,
JAFFE under lab-environment and SFEW in the wild [20]. Proposed transductive deep transfer learning
(TDTL) architecture to address the problem of cross-database non-frontal facial expression recognition
applying VGGface 16-Net on BU-3DEF and Multi-PIE datasets. The study found that feature representation
with VGG network is better than traditional handcrafted features such like SIFT and LBP to represent
complicated features [21]. [22] Also used the two datasets for the experimentation to address the problem of
cross-domain and cross-view of facial expressions using transductive transfer regularized least-square
regression (TTRLSR) model, color SIFT (CSIFT) features with 49 landmarks and SVM classifiers. The two
databases have only four identical categories neutral, surprise, happy and disgust. The experimentation of the
study conducted two kinds cross-domain and same view and cross-view and same domain. PCA algorithm also
applied to reduce the features dimension.
The studies in references [3, 5-7] classified six universal emotions as happiness, angry, sadness,
surprise, fear, and disgust. In [9, 13, 15, 23-24] have classified one more class as neutral and [8, 17, 23] have
done contempt class. All of eight classes have been classified by the studies in [11, 10, 16]. However, [21] and
[22] have worked on neutral, happiness, surprise and disgust expressions. Chen et al. [4] employed with 5
classes of GEMEP-FERA 2011 database and 7 classes of CK+ and AFEW. Li et al. [25] explained seven basic
emotions and 11 compound emotions sadly angry, sadly surprised, sadly fearful, happily surprised, happily
disgusted, sadly disgusted, fearfully surprised, fearfully angry, angrily surprised, angrily disgusted and
disgustedly surprised. Ferreira et al. [20] has worked classification 6 universal classes of JAFFE, SFEW with
classes of 6 basic and neutral, and CK+ with 8 classes including contempt.
5. FACIAL DATASETS
Facial expression datasets have two types of creation of images: posed expressions images and
spontaneous expressions images datasets. Researchers acquired facial images in three ways such as peak
expression images only, image sequences portraying an emotion from neutral to its peak, and video clips with
Feature extraction and classification methods of facial expression: a surey… (Moe Moe Htay)
30 ISSN: 2722-3221
emotional annotations. The two widely used datasets are CK+ and JAFFE [26-29]. The real-world facial
databases are FER-2013, FERG-DB, SFEW2.0 (static facial expression in the wild), RAF-DB (real world
affective face database) and AffectNet database. Sample images of basic facial expression are described in
Table. 1 for each dataset.
CK+
JAFFE
FER-2013
FERG-
DB
SFEW
RAF-DB
AffectNet
6. PROBLEM STATEMENT
FER system is need to develop under the problem of illumination, lighting, pose, aging, occlusion for
the real-world expression classification system. The major challenges of the study include:
Most of researches classify basic emotions but fine-grain emotion is relatively small.
The reaearch works on mocro-expression and compound emotion recognition system are limited.
Mathematical model is needed to be developed for extraction more discriminant features facial images in
the wild.
Real time facial expression recognition systems should be developed to meet practical application.
Deep learning model also need to create for improving facial feature extraction and classification.
REFERENCES
[1] Kalsum, Tehmina, Anwar, Syed, Majid, Muhammad, Ali, Sahibzada. “Emotion Recognition from Facial Expressions
using Hybrid Feature Descriptors.” IET Image Processing. vol. 12, no. 6, January 2018.
[2] P. Ekman, W. V. Friesen. “Facial action coding system a technique for the measurement of facial movement.” Palo
Alto: Consulting Psychologists Press, pp. 271-302, 1978.
[3] A. Majumder, L. Behera and V. K. Subramanian. “Automatic Facial Expression Recognition System Using Deep
Network-Based Data Fusion.” in IEEE Transactions on Cybernetics, vol. 48, no. 1, pp. 103-114, Jan. 2018.
[4] J. Chen, Z. Chen, Z. Chi and H. Fu. “Facial Expression Recognition in Video with Multiple Feature Fusion.” in IEEE
Transactions on Affective Computing, vol. 9, no. 1, pp. 38-50, 1 Jan.-March 2018.
[5] Yang, Dongri, Abeer Alsadoon, P. W. Chandana Prasad, Ashutosh Kumar Singh and Amr Elchouemi. “An Emotion
Recognition Model Based on Facial Recognition in Virtual Learning Environment.” Procedia Computer Science.
vol. 125, pp. 2-10, 2018.
[6] T. Kalsum, S. M. Anwar, M. Majid, B. Khan and S. M. Ali. “Emotion recognition from facial expressions using
hybrid feature descriptors.” in IET Image Processing, vol. 12, no. 6, pp. 1004-1012, 2018.
[7] C. Qi et al. “Facial Expressions Recognition Based on Cognition and Mapped Binary Patterns.” in IEEE Access, vol.
6, pp. 18795-18803, 2018.
[8] A. M. M. Shabat and J. Tapamo. “Angled local directional pattern for texture analysis with an application to facial
expression recognition.” in IET Computer Vision, vol. 12, no. 5, pp. 603-608, 8 2018
Feature extraction and classification methods of facial expression: a surey… (Moe Moe Htay)
32 ISSN: 2722-3221
[9] N. P. Nirmala Sreedharan, B. Ganesan, R. Raveendran, P. Sarala, B. Dennis and R. Boothalingam R. “Grey Wolf
optimisation-based feature selection and classification for facial emotion recognition.” in IET Biometrics, vol. 7, no.
5, pp. 490-499, 2018.
[10] Zeng, N., Zhang, H., Song, B., Liu, W., Li, Y., Dobaie, A. M. “Facial expression recognition via learning deep sparse
autoencoders.” Neurocomputing, vol. 273, pp. 643-649, 2018.
[11] Y. Tang, X. M. Zhang and H. Wang. “Geometric-Convolutional Feature Fusion Based on Learning Propagation for
Facial Expression Recognition.” in IEEE Access, vol. 6, pp. 42532-42540, 2018.
[12] Mayya, V., Pai, R. M., & Pai, M. M., “Automatic facial expression recognition using DCNN.” Procedia Computer
Science, vol. 93, pp. 453-461, 2016.
[13] D. V. Sang, N. Van Dat and D. P. Thuan. “Facial expression recognition using deep convolutional neural networks.”
2017 9th International Conference on Knowledge and Systems Engineering (KSE), pp. 130-135, 2017.
[14] B. Wu and C. Lin. “Adaptive Feature Mapping for Customizing Deep Learning Based Facial Expression Recognition
Model.” in IEEE Access, vol. 6, pp. 12451-12461, 2018.
[15] Munir, A., Hussain, A., Khan, S. A., Nadeem, M., Arshid, S. “Illumination invariant facial expression recognition
using selected merged binary patterns for real world images.” Optic; vol. 158, pp. 1016-1025, 2018.
[16] Guo, J., Zhou, S., Wu, J., Wan, J., Zhu, X., Lei, Z., & Li, S. Z. “Multi-modality network with visual and geometrical
information for micro emotion recognition.” In Automatic face and Gesture Recognition (FG 2017), 12th IEEE
International Conference, pp.814-819, 2017.
[17] J. Guo et al. “Dominant and Complementary Emotion Recognition From Still Images of Faces.” in IEEE Access, vol.
6, pp. 26391-26403, 2018.
[18] Barros, P., Parisi, G.I., Weber, C., Wermter S. “Emotion-modulated attention improves expression recognition: A
deep learning model.” Neurocomputing, vol. 253, pp. 104-114, 2017.
[19] H. Zhang, Q. Li, Z. Sun and Y. Liu, "Combining Data-Driven and Model-Driven Methods for Robust Facial
Landmark Detection," in IEEE Transactions on Information Forensics and Security, vol. 13, no. 10, pp. 2409-2422,
Oct. 2018.
[20] P. M. Ferreira, F. Marques, J. S. Cardoso and A. Rebelo. “Physiological Inspired Deep Neural Networks for Emotion
Recognition.” in IEEE Access, vol. 6, pp. 53930-53943, 2018.
[21] Yan, K., Zheng, W., Zhang, T., Zong, Y., Cui, Z. “Cross-database non-frontal facial expression recognition based on
transductive deep transfer learning.” arXiv preprint arXiv: 1811.12774, 2018.
[22] W. Zheng, Y. Zong, X. Zhou and M. Xin. “Cross-Domain Color Facial Expression Recognition Using Transductive
Transfer Subspace Learning.” in IEEE Transactions on Affective Computing, vol. 9, no. 1, pp. 21-37, 2018.
[23] Tautkute, I., Trzcinski, T., and Bielski, A. “I Know How You Feel: Emotion with Facial Landmarks.” arXiv: preprint
arXiv: 1805.00326, 2018.
[24] B. Wu and C. Lin. “Adaptive Feature Mapping for Customizing Deep Learning Based Facial Expression Recognition
Model.” in IEEE Access, vol. 6, pp. 12451-12461, 2018.
[25] S. Li and W. Deng. “Reliable Crowdsourcing and Deep Locality-Preserving Learning for Unconstrained Facial
Expression Recognition.” in IEEE Transactions on Image Processing, vol. 28, no. 1, pp. 356-370, Jan. 2019.
[26] C. Loob et al. “Dominant and Complementary Multi-Emotional Facial Expression Recognition Using C-Support
Vector Classification.” 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG
2017), Washington, DC, pp. 833-838, 2017.
[27] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar and I. Matthews. “The Extended Cohn-Kanade Dataset
(CK+): A complete dataset for action unit and emotion-specified expression.” 2010 IEEE Computer Society
Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, pp. 94-101, 2010.
[28] Dhall, A., Goecke, R., Lucey, S., & Gedeon, T. Static facial expressions in the wild: data and experiment protocol.
CVHCI Google Scholar. [Online] https://fipa.cs.kit.edu/download/SFEW.pdf.
[29] Lyons, M. J., Akamatsu, S., Kamachi, M., Gyoba, J., & Budynek, J. “The Japanese female facial expression (JAFFE)
database.” In Proceedings of third international conference on automatic face and gesture recognition, pp. 14-16,
1998.
[30] Goodfellow I., Erhan D., Carrier PL., Courville A., Mirza M., Hamner B., Cukierski W., Tang Y., Lee DG., Zhou
Y., Ramaiah C., Feng F., Li R., Wang X., Athanasakis D., Shawe-Taylor J., Milakov M., Park J., Ionescu R., Popescu
M., Grozea C., Bergstra J., Xie J., Romaszko L., Xu B., Chaung Z., ans Bengio Y. “Challenges in Representation
Learning: A report on three machine learning contests.” International Conference on Neural Information Procession
Springer Berlin Heidelberg, 2013.
[31] Aneja, D., Colburn, A., Faigin, G., Shapiro, L., Mones, B. “Modeling stylized character expressions via deep
learning.” In Asian Conference on Computer Vision, Springer, Cham, pp. 136-135, 2016.