0% found this document useful (0 votes)
11 views

Paper 11-A Survey On Deep Learning Face Age Estimation Model

Uploaded by

yusrafaisalcs
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
11 views

Paper 11-A Survey On Deep Learning Face Age Estimation Model

Uploaded by

yusrafaisalcs
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

(IJACSA) International Journal of Advanced Computer Science and Applications,

Vol. 12, No. 11, 2021

A Survey on Deep Learning Face Age Estimation


Model: Method and Ethnicity
Hadi A. Dahlan
Faculty of Information Science and Technology
National University of Malaysia
43600 UKM Bangi, Selangor, Malaysia

Abstract—Face age estimation is a type of study in computer inaccurate age classification. Meanwhile, deep learning is
vision and pattern recognition. Designing an age estimation or another approach that could help algorithms improve the
classification model requires data as training samples for the computer's ability to discover common facial aging traits (e.g.
machine to learn. Deep learning method has improved estimation aging wrinkles) within vast amounts of data and classify the
accuracy and the number of deep learning age estimation models facial image into its correct age. However, face age databases
developed. Furthermore, numerous datasets availability is mostly have limited ethnic subjects, only one or two ethnicities
making the method an increasingly attractive approach. and may result in ethnic bias during age estimation, thus
However, face age databases mostly have limited ethnic subjects, impeding progress in understanding face age estimation.
only one or two ethnicities and may result in ethnic bias during
age estimation, thus impeding progress in understanding face age In this study, the review on face age estimation/
estimation. This paper reviewed available face age databases, classification/distribution examined problems regarding:
deep learning age estimation models, and discussed issues related
to ethnicity when estimating age. The review revealed changes in 1) What face databases are frequently used in the age
deep learning architectural designs from 2015 to 2020, frequently estimation study, and how many different ethnics are in those
used face databases, and the number of different ethnicities databases?
considered. Although model performance has improved, the 2) What deep learning technique is used in facial aging
widespread use of specific few multi-races databases, such as the
research? How did the technique change through time? And
MORPH and FG-NET databases, suggests that most age
estimation studies are biased against non-Caucasians/non-white do they account for different ethnicities in their studies?
subjects. Two primary reasons for face age research’s failure to 3) What are the most used deep learning network
further discover and understand ethnic traits effects on a architecture and what are their strengths and weaknesses?
person’s facial aging process: lack of multi-race databases and 4) How to obtain more face images of people of different
ethnic traits exclusion. Additionally, this study presented a ethnicities in the time of restrictions (e.g. due to quarantine)?
framework for accounting ethnic in face age estimation research
and several suggestions on collecting and expanding multi-race Accordingly, this study surveyed the available face age
databases. The given framework and suggestions are also databases, the most used database in this type of research, and
applicable for other secondary factors (e.g. gender) that affect the deep learning techniques used for the face age estimation (or
face age progression and may help further improve future face distribution; or classification) model design. More than 50
age estimation research. papers (2015-2020) that used the deep learning method for face
age studies were reviewed in this study. The aim of this paper is
Keywords—Deep learning; face age estimation; face database; to survey the different deep learning face age estimation
ethnicity bias methods and how they account for different ethnicities. By
I. INTRODUCTION understanding the different deep learning face age estimation
methods and the problem related to ethnic bias in their face age
Facial aging is a complex biological process. Most estimation, we can discover significant racial traits that could
researchers in the computer vision and the pattern recognition help distinguish unique aging patterns used to solve racial face
fields have already found multiple ways to extract information age estimation problems in real-life applications. Moreover, a
from the face for age estimation/classification. However, not all framework for studying CNN face age estimation while
information extracted can help the system learn. When the considering the ethnicities of the subjects is included in this
system learned from only a specific ethnic sample, it may not paper to help guide future face age estimation studies that use
estimate/classify the age of other ethnic subjects correctly, even either the deep learning approach or the standard machine
after the face age estimation system improved. learning approach.
Earlier face aging models combined extractors and The remainder of this paper is structured as follows:
classifiers to extract specific aging features and accurately Section 2 mention several related works regarding deep learning
classify the facial image into its correct age. The downside of and early face age estimation; Section 3 explains the human
this approach is that the data needed for learning are usually facial aging and differences in process between several races;
structured and quantitatively limited; too little or too much data Section 4 surveys the face age image databases that can be used
could lead to models learning incorrect patterns, resulting in for facial age estimation studies and shows the quantities of

86 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

each race in each database (if any); Section 5 explains the face age-invariant features, difficulties in capturing a person's face
age estimation model and reviews the different deep learning aging progression in a controlled/uncontrolled environment, and
techniques proposed between 2015-2020 as well as the capturing >100 face images of different ethnic people in equal
databases used. The importance of ethnic traits in age quantities can be time-consuming and costly. Nonetheless, it is
estimation is also highlighted; Section 6 discusses the relevant undeniable that the facial aging process differs between races;
open issues regarding ethnic characteristics; Section 7 discusses therefore, ethnicities should be considered in future research
several possible solutions to solve the problems, Section 8 when experimenting with the next CNN age estimation model.
presents the conclusions and Section 9 mentions the future Moreover, analyses on the ethnic age difference can contribute
directions. to a better understanding of human facial aging.
II. RELATED WORK III. HUMAN FACIAL AGING – ETHNICITIES
The deep learning model has two primary processes: 1) Face features and expression are fundamental ways of
training and 2) inferring. The training phase is the process of human communication. Many studies have observed the facial
labelling large quantities of data (i.e. identifying and appearance and examined ways to apply the knowledge to real-
memorising the data matching characteristics). Meanwhile, in world applications. One of these studies is face age estimation,
the inferring phase, the deep learning model decides on the label which is research on estimating a person's age based on facial
for the new data using the knowledge gained from the earlier appearance observations. Over the years, multiple facial traits
training phase. Manual feature extraction on the data is help determine a person's age, including the shape of the face,
unnecessary because the model’s neural network architecture skin texture, skin features, and skin colour contrast [5, 6]. The
can learn the feature directly from the data, eliminating the need two predetermined features are as follows: 1) face shape
for data labelling. This learning feature is advantageous when change, particularly the cranium bones that grow with time.
working on large quantities of unstructured data (multiple This process predominately occurs during childhood to
formats like text and pictures). Recently, deep learning, such as adulthood transition; 2) development of wrinkles or face texture
convolutional neural network (CNN), has become well-known as facial muscle weakens due to decreased elasticity. This
in the image processing and pattern recognition fields for its process occurs during the transition from adulthood to the
capability to 'learn' from a large number of images and perform senior stage [7, 8].
specific tasks accurately. The deep learning method can fit the
parameters of multi-layered networks of nodes to the vast
amounts of data before extrapolating outputs from new inputs.
Knowing the commonly used network designs in face age
estimation studies and their strengths and weaknesses would be
interesting enough.
Recently, face age estimation studies using the deep
learning approach to estimate a person’s age based on aging
features, such as the facial skull shape and aging wrinkle, have
increased. These aging features are a person's regular facial Fig. 1. Different Ethnic Facial Aging Features for Four Women Aged Over
60 Years Old. from Left to Right: Caucasian, East Asian, Latino/Hispanic,
aging changes that occur through the years. Nevertheless, and African (All Images were Taken from [13]).
considering ethnicity in age estimation can pose a different
problem since each ethnicity/race has been confirmed to have a
different rate of facial aging [1, 2, 3, 4]. For example, a 20-year-
old White subject would look older than a 20-year-old Asian
because of their facial bones and skin structures differences [2].
For the CNN model to learn correctly, many datasets containing
multiple races with equal ratios are needed.
Although many face databases are available for age
estimation, most are racially biased and have just only one or 1
two significant ethnicities. Unbalanced ethnic samples can
create problems as age estimation models depend solely on
these databases. A bias might occur, for example, when
estimating the age of an Asian subject if the majority of
ethnicities available in a database are Caucasians/White due to
the differences in facial structure and rate of skin aging [1, 2]. In
most previous face age estimation/classification/ distribution
studies, all sample databases were used for training and testing
while utilising different deep learning methods that match their Fig. 2. Facial Feature and Aging Difference for Adult Caucasian (Top Left)
research aim(s) and main objective(s). However, ethnic traits and Asian (Bottom left), while the Baby’s Face for the Caucasian is on the
are usually ignored, resulting in very few analyses of racial Top Right and Bottom right for the Asian (Images were taken from [2],
traits' effects on the face age estimation process. A few reasons Except for the Caucasian Baby, from [14]).
for this exclusion: researchers mainly consider racial traits as

87 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

Internal and external forces act upon the outer and inner skin IV. FACE AGE DATABASE
as a person age, causing some level of damage and changing the Designing face age estimation models require many samples
skin’s appearance. As demonstrated in [9, 10], the older skin for training and testing. Several studies collected face samples
was perceived to have a different colour contrast and luminosity and then made them available to the public so that others might
than the younger skin. Healthy young skin, which is plumper use them in their research. Furthermore, the shared database
and emits radiant colour, has a smooth and uniformly fine may serve as a benchmark against which other models can be
texture that reflects light evenly. Meanwhile, aged skin tends to compared and improved. Table I shows the face databases with
be rough and dry with more wrinkles, freckles, and age spots age information or labels (henceforth, called Face Age
and emits dull colour [11, 12]. However, ethnicities can affect Database). Only two databases captured face images in a
these aging rates because of differences in skull structure and controlled environment (MORPH and FACES). In contrast, the
skin type [1] (see Fig. 1). For instance, the skin of a Caucasian rest captured the face image in either a partially controlled or
subject will gradually have more aging wrinkles when uncontrolled environment. Meanwhile, the FG-NET database
compared to an Asian subject as the age increases from 20 to 39 has the most undersized samples and subjects, while the
years old. This phenomenon is due to the different skull and IMDB+WIKI database offers the most samples and subjects.
skin structures of each ethnic. Caucasians have a significant
angular face, while Asians tend to be broader and less angular, Table I reveals that most of the subjects in the databases are
similar to a baby's broad face [2] (see Fig. 2). Due to this Caucasian/White, whereas Table II provides the ethnic count.
broader facial structure, soft-tissue loss in Asians is seen and Correspondingly, the ethnic percentage is shown in Fig. 3,
felt to a lesser extent. Another example is between the which reveals very few databases with non-Caucasians/non-
Caucasians and the African-Americans' skin. Black skin’s White ethnic (White = 80%; Black = 3%; Asian = 8%; and
epidermis contains a thicker stratum corneum with more active Others = 9%). This gap creates an imbalance in the databases
fibroblasts than the Caucasians, making them less affected by when ethnicity is considered to estimate the age of non-
photo aging [3, 4]. Although black skin does not tend to get fine Caucasian/non-White races. Moreover, not all the databases
lines like white skin, it does get folded when getting older. Such have ethnic information (e.g. IMDB+WIKi, FERET, and
information should be considered to design a more accurate age Webface). The lack of ethnic labels can make it difficult for
estimation model which can specify proper age face age model researchers to divide samples into their
estimation/classification knowledge when dealing with specific appropriate ethnicity, eventually treated as one of their research
ethnic subjects. limitations.

TABLE I. SUMMARY OF FACE DATABASES WITH AVAILABLE INFORMATION


Environment
Year Database Samples Age Range Ethnic
C UC
1998 FERET [21] 14,126 samples; 1,199 subjects √ √ Not mentioned (real age) Not mentioned
2002 FG-NET [22] 1,002 samples; 82 subjects √ 0-69 (real age) All White/Caucasian
African-American:89; Caucasian:435;
2004 LIFESPAN [23] 1, 142 samples; 575 subjects √ √ 18-93 (age group)
Others:52
2005 FRGC [24] 44,278 samples; 568 subjects √ √ 16-77 (real age) White:386; Asian:125; Others:57
White-Black ratios 4:1; Others-very
2006 MORPH [25] 55,134 samples; 13,618 subjects √ 16-77 (real age)
small
2008 YGA [26] 8,000 samples; 1,600 subjects √ 0-93 (real age) Not mentioned
2009 GROUPS [27] 28,231 samples; 28,231 subjects √ 0-66+ (age group) Not mentioned
2010 FACES [28] 2, 052 samples; 171 subjects √ 19-80 (age group) All White/Caucasian
2012 Webface [29] 59, 930 samples √ 1-80 (real age) Not mentioned
2014 Adience [30] 26,580 images; 2,284 subjects √ 0-60 (age group) Not mentioned
2014 CACD [31] 160,000 samples; 2,000 subjects √ 16-62 (real age) Not mentioned
2015 Chalearn 2015 [32] 4, 699 samples √ Not mentioned (real age) Not mentioned
2016 Chalearn 2016 [33] 7, 591 samples √ Not mentioned (real age) Not mentioned
2017 AgeDB [34] 16,516 samples; 570 subjects √ 1-101 (real age) Not mentioned
2018 IMDB+WIKI [35] 523,051 samples; 20,284+ √ 0-100 (real age) Not mentioned
2007 Iranian face [15] 3,600 samples; 616 subjects √ 2-85 (real age) All Iranian
2013 IMFDB [16] 34,512 samples; 100 subjects √ Not mentioned (age group) All Indian
2016 AFAD [17] 164,432 samples √ 15-40 (real age) All Asian
Caucasian: 6,686; Asian: 674; Afro-
2017 APPA-REAL [36] 7,591 samples; 7,000+ subjects √ 0-95 (real age)
American: 231
(C – captured or collected in a controlled environment; UC – captured or collected in an uncontrolled environment)

88 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

to analyse unstructured data; they require additional tasks to


TABLE II. ETHNIC COUNT (BASED ON TABLE I) divide the problem and later recombine the results to form a
conclusion, which takes time and resources. Nevertheless, the
Ethnic Count Subjects (Approx. ≈)
deep learning method’s known capability and strength have
White 7760 shifted the face aging system approach.
Black 320 A. Deep Learning Approach
Asian 799 The rise of deep learning in image processing and machine
Others 825 learning has also impacted face age estimation. Better age
estimation performance is strictly associated with the depth of
the used network in the deep learning method, and it has
become the generalist network adopted for feature extraction,
including deep architectures that require a considerable amount
of image samples, such as AlexNet, VGG-Net, VGG-Face,
GoogLeNet, and Residual Networks (ResNet) [37]. VGG-Net
has been reported to be one of the most effective deep learning
architectures for age estimation. Notwithstanding, new studies
continue to propose deep architecture designs for improving
model accuracy when processing a sample of subjects' faces
captured in an uncontrolled environment.
Deep learning face age research can be classified into three
types: 1) classification age (CA) - classify the face age with
several classes equal to the number of the considered age
groups; 2) estimation age (EA) - estimate age using a regressor;
and 3) distribution age (DA) - a modified CA strategy obtained
by substituting the one-hot encoding vector with a statistical
Fig. 3. Ethnic Percentage from All Databases (Based on known Data from distribution centred on the estimated age [37]. Furthermore, the
Table I). deep learning approach is much more accurate than other older
machine learning methods at estimating age from sample
Some studies, however, have collected particular ethnic images captured in the wild (uncontrolled environment).
subjects with age information, such as Bastanfard A. et al. Nevertheless, if the subject's ethnicity in the dataset is not
(Iranian face database) [15]; Setty S. et al. (Indian Movie Face considered, the ethnic bias will persist.
database - IMFDB) [16]; and Niu Z. et al. (Asian Face Age
Dataset - AFAD) [17]. Furthermore, there are ethnic-specific B. Deep Learning Model Method and Ethnicity Bias in
databases that can be used for face age research (see Table I Database
coloured in grey). However, no studies have used these When searching for papers on face age research, this study
databases for deep learning age estimation research in the past focused on research that used the deep learning method from
six years; these databases are either not considered benchmark 2015-2020. Deep learning has the potential to revolutionise
databases or less known by the face age estimation community. computer science and machine learning. Furthermore, data
biases are becoming more important with the rise of more
V. FACE AGE DATABASE ESTIMATION MODEL powerful machine learning, which deep learning takes
In one of the earliest face age model studies, Kwon and advantage of when dealing with large amounts of data. The
Lobo [18] classified age into three categories: infant, adult, and search was conducted using a variety of web search engines,
senior, and used simple feature extraction and machine learning including Google Scholar and Web of Science.
for face age classification. Subsequently, computer science and Table III displays the search results, which include the
pattern recognition researchers introduced various age following information: publisher, year of publication, network
classification/estimation methods [19, 20]. Earlier machine architecture, domain area, selected databases, and ethnicity
learning methods typically included one (or more) feature consideration. From 2015-2016, the most commonly used
extractor and one (or more) age classifier (or estimator). The network architectures were well-known general architectures
feature extractors can be holistic (e.g. whole facial shape), local such as GoogleNet, VGG-Net, and DCNN (or Deep-CNN). As
(e.g. aging wrinkle), or both. The selection of feature extractors the year progressed, an increasing number of studies began to
is influenced by the database used, with most of the sample design the architectural network or modify the general CNN
quantity used by these methods being less than that of the deep network architecture to fit their research objectives. As a result,
learning approach. the network design became more complex to produce a more
Previous machine learning approaches can produce precise accurate novel model (e.g. by combining multiple CNN
estimation (or classification) using just one or two databases, networks to create a hybrid network). In the research domain
but are confined to those databases and could give an erroneous area, there have been 33 EA studies, 20 CA studies, and only 8
estimation if a wild sample is used for testing instead. DA studies. However, there is no significant preference
Moreover, it is difficult for most machine learning approaches between the research domain and databases used in the studies.

89 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

Therefore, it can be inferred that most of these databases can be used in all deep learning areas: EA, CA, and DA.
TABLE III. DEEP LEARNING FACE AGE RESEARCH AREA AND AGE DATABASES USED FOR TRAINING AND TESTING (FROM 2015-2020)

Domain Area Database Model Account


Ref. Publisher Year Network Architecture
CA DA EA M C15 I A F C16 W G P O Ethnic
[38] IEEE 2015 GoogLeNet √ √ √ √ √ √ No
[39] IEEE 2015 VGG-Net √ √ √ No
[40] IEEE 2015 GoogLeNet √ √ √ √ √ √ √ No
[41] IEEE 2015 VGG-Net & GoogleNet √ √ No
[42] IEEE 2015 VGG-Net & Novel arch. √ √ √ √ √ √ No
[43] IEEE 2015 DLA √ √ √ No
[44] IEEE 2015 Tree kernel adaptive CNN √ √ √ No
[45] Elsevier 2015 LeNet √ √ √ No
[46] IEEE 2015 Novel arch. √ √ No
[47] IEEE 2015 DCNN-H-3NNR √ √ √ √ No
[48] IEEE 2016 DCNN √ √ √ √ √ No
[49] IEEE 2016 VGG-Net √ √ No
[50] IEEE 2016 Novel arch. √ √ √ No
[51] IEEE 2016 VGG-Net √ √ No
[52] IEEE 2016 Compact-CNN √ √ √ √ Yes
[53] Elsevier 2016 DCNN √ √ No
GilNet; AlexNet; VGG-
[54] IEEE 2016 √ √ No
Net
[55] IEEE 2016 VGG-Net √ √ √ √ √ No
[56] IEEE 2016 DADL √ √* √ √** No
[57] IEEE 2016 VGG-Net √ √ √ √ No
[58] IEEE 2016 VGG-Net √ √ √ No
[59] IEEE 2016 DCNN √ √ √ No
[60] IEEE 2016 Novel arch. √ √ √ √ No
[61] IEEE 2017 AGEn & MO-CNN √ √ √ √ √ √ √ No
[62] IEEE 2017 Multitask CNN √ √ √ √ √ √ No
[63] IEEE 2017 ODFL & ODL √ √ √ √ No
[64] Elsevier 2017 GA-DFL √ √ √ √ No
[65] Elsevier 2017 VGG-Net CNN+LDAE √ √ √ √ √ √ No
[66] Elsevier 2017 Novel arch. √ √ √ √ √ Yes
[67] Elsevier 2017 D2C √ √ √ No
[68] PMLR 2017 R-SAAFc2 √ √* √ √ √ √ No
[69] IEEE 2017 Deep-ROR √ √ √ No
[70] IEEE 2017 DMTL √ √ √ Yes
[71] IEEE 2017 M-LSDML √ √ √ √ √ √ √ √ No
[72] IEEE 2017 DMTL √ √ √ √ √ Yes
[35] Springer 2018 VGG-Net √ √ √ √ √ √ √ No
[73] IEEE 2018 VGG-Net-GPR √ √ √ √ √ Yes
[74] IEEE 2018 ELM √ √ √ √ √ Yes
[75] ALM-DL 2018 ScatNet √ √ No
[76] IEEE 2018 CMT-deep network √ √ √ √ No
[77] Elsevier 2018 DAG-CNNs √ √ √ √ No
[78] Springer 2019 CNN+triplet ranking √ √ √ √ √ No
[79] Elsevier 2019 DeepAge √ √ √ No
[80] IEEE 2019 SADAL √ √ √ No
[81] IEEE 2019 Novel Arch. √ √ No
[82] IEEE 2019 Multitasks-AlexNet √ √ √ No
[83] IEEE 2019 ODFL & ODL √ √ √ √ √ No
[84] IEEE 2020 SADAL & VDAL √ √ √ √ No
[85] Elsevier 2020 LRN √ √ √ √ √ Yes
[86] IEEE 2020 CR-MT √ √ √ √ √ No
[87] SYMMETRY 2020 MA-SFV2 √ √ √ √ No
[88] IEEE 2020 DOEL-groups √ √ √ √ √ √ No
[89] IEEE 2020 MSFCL √ √ √ √ √ No
(CA – Classification Age; DA – Distribution Age; EA – Estimation Age; M – MORPH; C15 - ChaLearn2015; I – IMDB+WKI; A – Adience; F – FG-NET; C16 - ChaLearn2016; W – Webface; G – GROUP; P –
Private DB; O – Others; ‘√*’ - variation of ChaLearn2015; ‘√**’- variation of chalearn2016).

90 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

samples (1,002 samples) compared to other databases. FG-NET


has been used in face age research since around 2005 [5],
making it one of the most well-known databases used primarily
for comparing model performance in the face age research
community. Despite this, the majority of its subjects are
Caucasians/Whites. When used in the CNN model, a small
dataset should be fine-tuned or pre-trained with another
database with large sample size, such as IMDB+WIKI. The
IMDB+WIKI [35] dataset contains images with one or more
people in them, as well as annotations for researchers’ reference
when there are multiple people in one image. However, there is
no proper explanation for which annotation refers to which
person in the image of multiple people. Therefore, studies
primarily use this database for pre-training deep networks due
to its large sample size. Because of the lack of annotation, no
model performance results for IMDB+WIKI are shown in Table
IV, which reveals the model performance on studies based on
their selected databases. Although the IMDB+WIKI database
Fig. 4. Breakdown of Databases used in Deep Learning Face Age samples contain multiple ethnicities, no annotation for a
Estimation Research (based on Table III). ‘ * ’ Means that there are Multiple
Versions of the Databases Included in the Count (M – MORPH; C15 -
subject's ethnicity is available.
ChaLearn2015; I – IMDB+WKI; A – Adience; F – FG-NET; C16 - Adience [30], a database for gender and age group
ChaLearn2016; W – Webface; G – GROUP; P – Private DB)
classification, comes in fourth place, with subjects drawn from
Meanwhile, Fig. 4 depicts the most commonly used real-world conditions. Its sources are mostly Flickr albums
databases for face age research (derived from Table III), uploaded from smartphone devices. This database was made
indicating that MORPH [25] is the most commonly used available to the general public under the Creative Commons
database because it has the highest sample count (55,134 (CC) licence. Meanwhile, in fifth place is Webface [29], a
samples) captured in a controlled environment. Because of this database collected for the experiments of a PhD thesis, and in
advantage, the MORPH database is the best benchmark sixth place is GROUP [27], a collection of images of people
database for comparing the CNN model performance with other captured in a group (hence the name) that includes age group
models since it can lessen the influence of unwanted factors that and gender information. Nevertheless, none of these three
may affect the overall estimation results. The MORPH databases has a record of the subject's ethnicity.
database, on the other hand, has an unbalanced ratio of races in Although some studies used/included their own database,
its dataset (refer to Table I), which can lead to ethnic bias when these private databases [42, 52, 55, 65] contain no information
estimating age. about the subject’s ethnicity. Furthermore, some of them were
The second most used face age database is ChaLearn2015 only used to fine-tune network models [55], [65]. CACD [31],
[32], which was explicitly developed for the ICCV 2015 LIFESPAN [23], LFW [90], FACES [28], FRGC [24], AFAD
ChaLearn Looking at People Apparent Age Estimation [17], and FERET [21] are the remaining databases used in the
Challenge [32]. This challenge event was a competition to build age model. Only a few of these were used in the face age deep
the best appearance age estimation model, and most of the learning research (categorised as ‘Others’ in Table III).
authors of the research surveyed in this study competed in it. Meanwhile, some of the studies used a different database for
ChaLearn2016 [33], the fifth most used database, is the pre-training their models (e.g. face detection in images) than the
second/expanded version of the ChaLearn2015 database. The one used for age estimation, such as the CelebFaces Attributes
lack of ethnicity records for subjects in both ChaLearn (CelebA) [91] and ImageNet databases [92]. The CelebA
databases makes analysing the effect of ethnicity on a model's database was built using the CelebFaces [91] face verification
overall performance difficult, even though both databases have database with face attribute annotations. ImageNet, on the other
a diverse set of races. hand, is a database for object classification and detection. Both
The FG-NET [22] database comes in third place, with databases lack age information and were primarily used for pre-
images captured in uncontrolled real-life conditions that are not training/fine-tuning the network model in these studies [39, 52,
equally distributed across age groups and has the lowest 57, 69, 70, 74].

TABLE IV. SUMMARY OF BEST CNN MODEL PERFORMANCE ON SELECTED DATABASES (FROM 2015-2020)

Best CNN Model Performance on Selected Databases Account


Ref. Year
Database_used(performance_measurement) Ethnic
[38] 2015 C15(MAE = 3.33); C15-testset(e-error = 0.27) No
[39] 2015 C15-validset(MAE = 3.22); C15-testset(e-error = 0.260) No
[40] 2015 C15-validset(e-error = 0.309); C15-testset(e-error = 0.290) No
[41] 2015 C15-validset(MAE = 3.29; e-error = 0.285); C15-testset(e-error = 0.287) No
[42] 2015 C15-validset(e-error = 0.338); C15-testset(e-error = 0.306) No

91 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

[43] 2015 M(MAE = 4.77); F(MAE = 4.26) No


[44] 2015 M(MAE = 3.61); W(MAE = 7.27) No
[45] 2015 M(MAE = 3.88); FRGC(MAE = 3.31) No
[46] 2015 A(AEM = 50.7±5.1, AEO = 84.7±2.2) No
[47] 2015 C15-validset(e-error = 0.359); C15-testset(e-error = 0.373) No
[48] 2016 A(AEM = 52.88±6%, AEO = 88.45±2.2); C15-validset(e-error = 0.297) No
[49] 2016 C16-validset(MAE = 3.85, e-error = 0.330); C16-testset(e-error = 0.370) No
[50] 2016 M(MAE = 3.27); AFAD(MAE = 3.34) No
[51] 2016 M(MAE = 3.45) No
[52] 2016 M(MAE = 3.23); P(acc. = 88%) Yes
[53] 2016 G(AEM = 56%, AEO = 92%) No
[54] 2016 A(mean AEM = 57.9%) No
[55] 2016 C15-validset(e-error = 0.261); C15-testset(e-error = 0.241) No
[56] 2016 C15-validset(MAE = 1.76, e-error = 0.134); C15-testset(e-error = 0.321) No
[57] 2016 C16-validset(e-error = 0.240); C16-testset(e-error = 0.336) No
[58] 2016 C16-testset(e-error = 0.367) No
[59] 2016 A(acc. = 42%); FERET(acc. = 86.4%) No
[60] 2016 M(MAE = 2.78); F(MAE = 2.80) No
M(MAE = 2.52); F(MAE = 2.96); CACD (ave. MAE = 4.68); C15-validset(MAE = 3.21, e-error = 0.28); C15-testset(MAE =
[61] 2017 No
2.94, e-error = 0.264); C16-testset(MAE = 3.82, e-error = 0.310)
[62] 2017 F(MAE = 2.00); C15-validset(e-error = 0.293) No
[63] 2017 M(MAE = 2.92); F(MAE = 3.71); C15-validset(MAE = 3.95, e-error = 0.312) No
[64] 2017 M(MAE = 3.25); F(MAE = 3.93); C15-validset(MAE = 4.21, e-error = 0.369) No
[65] 2017 F(MAE = 2.84); M(MAE = 2.35); P(MAE = 4.33); C16 (e-error = 0.241) No
[66] 2017 M(ave. MAE =2.96); W(ave. MAE = 5.75) Yes
[67] 2017 M(ave. MAE =3.06); W(ave. MAE = 6.104) No
[68] 2017 F(MAE = 3.01 MAE); A(AEM = 67.3, AEO = 97.4) No
[69] 2017 A(AEM = 67.34 ± 3.56%, AOE = 97.51 ± 0.67%) No
[70] 2017 M(acc. = 85.30 ± 0.6%) Yes
[71] 2017 M(MAE = 2.89); F(MAE = 3.31); A(AEM = 60.20±5.3%, AEO = 93.70± 2.3%); C15-validset(e-error = 0.315) No
[72] 2017 M(MAE = 3.00); LFW(MAE = 4.50) Yes
[35] 2018 M(MAE = 2.68); F(MAE = 3.09); CACD(MAE = 4.79); A(AEM = 64.00±4.2%, AEO = 96.60±0.9%) No
[73] 2018 M(MAE = 2.93); CACD (MAE = 5.22); C15-validset(MAE = 3.30, e-error = 0.290) Yes
[74] 2018 M(MAE = 2.61); A(AEM = 66.49 ± 5.08%); C16-validset(MAE = 3.67, e-error = 0.325); C16-testset(e-error = 0.368) Yes
[75] 2018 LIFESPAN(MAE = 4.01); FACES (MAE = 5.95) No
[76] 2018 M(MAE = 2.89); F(MAE = 3.43) No
[77] 2018 M(MAE = 2.81); F(MAE = 3.05) No
[78] 2019 M(MAE = 2.87); A(AEM = 63.10 ± 1%, AEO = 96.7 ± 0.4%) No
[79] 2019 M(MAE = 2.87); F(MAE = 7.08) No
[80] 2019 M(MAE = 2.75); F(MAE = 3.67) No
[81] 2019 C15-testset(MAE = 6.031, e-error = 0.441) No
[82] 2019 Wiki(MAE = 5.47); UTKFace(MAE = 9.54); AgeDB(MAE = 10.01) No
[83] 2019 M(MAE = 2.92); F(MAE = 3.71); APPARENT-AGE(MAE = 3.95) No
[84] 2020 M(MAE = 2.57); F(MAE = 2.98); C15(MAE = 3.58, e-error = 0.285) No
[85] 2020 M(MAE = 1.90); C15-validset(MAE = 3.05, e-error = 0.274); MegaAge-Asian(CA(7) = 91.64) Yes
[86] 2020 M(ave. MAE = 2.36); CACD(MAE = 4.48); Webface(MAE = 5.67) No
[87] 2020 M(MAE = 2.68); F(MAE = 3.81) No
M(MAE = 2.75), F(MAE = 3.44); AgeDB(MAE = 5.69); C15-validset(MAE = 2.93, e-error = 0.258); C15-testset(MAE = 2.71, e-
[88] 2020 No
error = 0.247)
M(MAE = 2.73); F(MAE = 2.71); A(AEM = 65.3%, AEO = 96.3%); MEGAAGE-ASIAN(MAE = 2.81, CA(3)(62.89%),
[89] 2020 No
CA(5)(82.46%))
M – MORPH; C15 - ChaLearn2015; A – Adience; F – FG-NET; C16 - ChaLearn2016; W – Webface; G – GROUPS. Note that other databases that are not shown in this table but shown in Table III (e.g. I –
IMDB+WKI) were used for pre-training by the studies.

92 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

C. Deep Learning Technique Strengths and Weaknesses research aim and objectives, such as the problem(s) to solve that
A review of the different deep learning architectural can help improve face age classification/estimation/distribution.
networks used in previous studies revealed several techniques The problems include face detection, landmark localisation,
that are frequently used for face age estimation. Table V optimisation, regression, classification, feature extraction,
summarises the network architectures frequently used in the age residual learning part, sampling technique, layer size (depth and
estimation studies reviewed, as well as their strengths and width), discriminative distance, learning speed, training and/or
weaknesses. As previously stated, the main goal of deep testing process and others. This study identified several known
learning face age estimation is to find the best method for network architectures that were frequently used in comparison
learning the face aging features from a large sample of data and to the others [93]. Among these network architectures are the
then use the information to distinguish the different ages of test following:
subjects. Each study’s architecture was chosen based on its
TABLE V. SUMMARY OF NETWORK ARCHITECTURES MOSTLY USED BY AGE ESTIMATION STUDIES IN THIS SURVEY

Author(s) that
Background Learning
Architecture Strength Weakness Used the
(referred from [93]) Methodology
Architecture
- Problem to detect all aging
- Small and simple design.
features. Require extensive training.
- Invented in 1998 by Yann Lecun. Spatial - A good introduction to
LeNet - Speed and accuracy are [45]
- First popular CNN architecture. exploitation neural networks for
outperformed by newer network
beginners.
architecture.
- Introduced in 2012 at the - Using GPUs for training
ImageNet Large Scale Visual leads to faster training of
- Authors require to find design
Recognition Challenge. models.
Spatial solutions on how to compete with
AlexNet - Uses ReLu, dropout and overlap - ReLu helps lessen the [54, 82]
exploitation other newer network architectures
pooling. loss of features and
that are more accurate and faster.
- First major CNN model that used improve model training
GPUs for training. speed.
- Homogenous topology.
- Smaller kernels. - Computationally expensive as
- Visual geometric group (VGG) - Good architecture for more layer increases.
[35, 39, 42, 49, 51,
was introduced in 2014. Spatial benchmarking face age - Face age estimation studies need
VGG-Net 54, 55, 57, 58, 65,
- It groups multiple convolution exploitation estimation to consider the vanishing gradient
73]
layers with smaller kernel sizes. - Pre-trained networks for problem to improve the estimation
VGG-Net are freely performance.
available.
- Researchers at Google - Trains faster than VGG-
introduced GoogleNet in 2014. Net.
- Heterogeneous topology design
- Introduced block concept. - Smaller pre-trained size
Spatial require face age estimation studies
GoogleNet - Split transform and merge idea. than VGG-Net. [38, 41, 43]
exploitation to make thorough customisation -
- In a single layer, multiple types - Training network has
from module to module.
of ‘feature extractors’ are present many options to solve
to help the network perform better. tasks.
- Capable of skipping
learned feature(s),
reducing training time and
- Introduced in 2015.
improve accuracy.
- Residual learning.
Depth + - Solve the vanishing - Computationally expensive as
ResNet - Identity mapping-based skip [69, 71]
multi-path gradient problem faced by more layer increases.
connections.
VGG-Net.
- Possible to train very
deep networks and
generalise well.
[42, 43, 44, 46, 47,
- Specialise in learning 48, 50, 52, 53, 54,
Most designs were face representation for 56, 59, 60, 61, 62,
- Cater to a very specific
expanded/modified/or built from different ages. 63, 64, 65, 66, 67,
problem(s).
Novel Arch. scratch based on the previously - - Improving several parts 68, 69, 70, 71, 72,
- Time-consuming when building
available architectures (e.g. of the network based on 73, 74, 75, 76, 77,
from scratch.
AlexNet, VGG-Net, etc.) the study’s aim and 78, 79, 80, 81, 82,
objectives. 83, 84, 85, 86, 87,
88, 89]

93 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

1) LeNet: Yann Lecun invented the LeNet architecture in module. This study discovered that only [38, 40, 41] used this
1998 to perform optical character recognition (OCR), and its network architecture.
design is smaller and simpler than the rest of the network 5) ResNet; ResNet was introduced in 2015 and provides
architectures. For beginners, this network is a good way to residual learning to help solve the vanishing gradient problem
learn neural networks and can be used for face age estimation (from the VGG-Net architecture). Furthermore, ResNet can
studies, such as in [45]. However, due to its simple design, the have a deeper network (more layers) than VGG-Net while
network requires additional improvement that the designer avoiding performance degradation. ResNet is a concept in
must build from scratch if used for face age estimation. It is which if a feature has already been learned, it can be skipped
also outclassed by newer models in terms of speed and and focus can be given to newer features, thereby improving
accuracy when used as is, with no modifications. training time and accuracy. On the other hand, the ResNet
2) AlexNet: Alex Krizhevsky introduced the AlexNet structure design is primarily concerned with how deep the
architecture in 2012, and it was the first major CNN model to structure should be. If ResNet is chosen for face age
use graphics processing units (GPUs) for training, which aided estimation, the designer must consider how the network
in training speed. Meanwhile, ReLu, dropout and overlap should be structured to learn multiple aging features. Adding
pooling were used to reduce feature loss and improve training more layers is one of the common ideas. However, this could
speed. This architecture design was used in [54], [82] for face result in a longer learning time for the model (it can take
age classification and estimation, respectively. Their accuracy several weeks); therefore, the designer must also account for
performance, however, was inferior to that of the model that this. This study discovered that only a few face age estimation
used the LeNet network design [45] (see Table IV). This studies used Resnet architecture/concept in their design [69,
implies that, even though AlexNet is a newer network than 71].
LeNet, proper modification, structuring, and organisation of 6) New Arch: Is a network architecture created by
the AlexNet network are still required to achieve the best face expanding previous architectures, modifying them, or building
age estimation (or classification) performance. the network from scratch. These architectures were created
3) VGG-Net: Introduced in 2014, the VGG model specifically to find the best network approach for learning
improves training accuracy by improving its depth structure. how to best estimate age. For example, a facial image with a
The addition of more layers with smaller kernels increases specific age can be affected by facial variations caused by
nonlinearity, which is good for deep learning. This study external factors, such as lighting, which can lead to a
discovered that VGG-Net is the most commonly used network neighbouring age category being predicted as the final bias.
model among the many available (11 papers). One of the The study in [80] attempted to address this problem by
possible explanations is that the VGG pre-trained networks are proposing a network composed of a generator that could
freely available online. Although it is the best architecture for generate discriminative hard-examples (taken from extracted
benchmarking on the face age estimation task, the features done by a deep CNN) to complement the training
performance obtained by studies that used this model is not space for robust feature learning and a discriminator that could
the best, but it is also not the worst. This could be due to the determine the authenticity of the generated sample using a
vanishing gradient problem, one of the main challenges faced pre-trained age ranker [80]. This approach offers designers the
when using VGG-Net, which occurs when the number of ‘freedom’ to create the best solution to a given problem. The
layers exceeds 20, causing the model to fail to converge to the designs can be based on available networks and further
minimum error percentage. When this happens, the learning modified to their preferences, rather than being limited to the
rate slows to the point where no changes are made to the original design architecture. This study found that most of the
model’s weights. Furthermore, using VGG-Net can be time- previous studies, particularly those conducted in 2020, tend to
consuming because the training process can exceed a week, propose their own architectural network design. However, one
especially if it was built from scratch. As a result, when using major drawback of this design approach is that the designer
the VGG-Net network for face age estimation, users must may take a long time to modify/create networks when
address the vanishing gradient problem as well as the training compared to using available networks.
time.
D. Model Performance Evaluation
4) GoogleNet: A class of architecture designed by Google
researchers that won ImageNet 2014. Instead of a sequential Multiple protocols and performance calculations were used
in the studies to evaluate model performance. Table IV shows
architecture design, GoogleNet opted for a split transform and
the performance of the CNN models used in the studies on the
merge design, in which a single layer can have multiple types databases that they were tested on. The evaluation protocol is a
of “feature extractors”. In addition, GoogleNet has a smaller method for studies to determine the optimal number of training
pre-trained size and trains faster than VGG-Net [93]. One and testing datasets for their chosen databases. Meanwhile, the
drawback of GoogleNet is that almost every module must be performance calculation allows studies to compare the
customised. As a result, when designing a face age estimation estimation/classification/ distribution accuracy of their own
using GoogleNet, users must customise from module to model to that of others. Because of the numerous ways for
designing protocols and performance calculations, problems

94 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

arise when performance is compared on the same database but year; the improvement might be valid for MORPH since most
using different evaluation methods, resulting in a unanimous of the standard deviation obtained is low (< 0.3). However, the
'agreement' from most of the studies that specific performance implication for FG-NET may be invalid because only a few
calculation(s) should be used for comparison's sake for a studies used this database in 2015-2016. Most of the standard
specific database. Among the performance calculations used to deviation for 2017-2020 is high (> 0.4), meaning that the MAE
evaluate the accuracy of the face age deep learning model are: results obtained by the different studies are too wide apart.
Among other databases, the performance of the MOPRH
1) Mean absolute error (MAE): a widely used database appears to be the best. The samples captured in a
performance evaluation for age estimation studies that controlled environment help the models to better identify aging
measures the error between the predicted and actual ages. features because unwanted factors are absent (e.g occlusion).
MORPH, FG-NET, ChaLearn2015, and ChaLearn2016 are Meanwhile, the low quantity (1,002 images) and low quality
examples of databases that used this evaluation method. The (old images captured in an uncontrolled environment) samples
model performance improves as the MAE value decreases. of FG-NET might hinder the CNN model learning process in
2) E-Error: is the performance calculation used in the studies. Nonetheless, some studies were able to obtain low
apparent age estimation. This evaluation metric was used to MAE values using the FG-NET database: [62] MAE = 2.00 and
compare the performance of studies that used ChaLearn2015 [89] MAE = 2.71.
[32] and ChaLearn2016 [33] datasets. The lower the e-Error, Regarding publishers, from 2015-2020 (see Fig. 6), IEEE is
the better the performance. the publisher with the highest reviewed papers in this study.
3) Accuracy of an exact match (AEM): a method of Elsevier is in second place, and Springer is in third. The bar
calculating accuracy that calculates the percentage of correctly chart in Fig. 6 shows that the number of published papers
estimated/classified age per the total number of test images increased in 2016, but then declined until 2018, and then
remained relatively low until 2020. The figure seems to imply
used. This type of evaluation metric was used by the Adience
that the deep learning approach is becoming less attractive to
database. The higher the AEM value, the better the the face age research community, but this is most likely not the
performance. Some studies went so far as to include the case. When a more robust, advanced, and practical deep
standard deviation value in their evaluation. learning technique becomes available, a resurgence may occur.
4) Accuracy error of one age category (AEO): Is another
type of evaluation metric used on the Adience database, in
which errors of one age group are also included as correct age
classifications. The higher the AEO value obtained, the better
the overall model performance.
5) Cumulative score (CS): is defined as the percentage of
images with an error of no more than a certain number of
years. The evaluation is usually shown as a curve on a graph
(which is not depicted in this paper), with the x-axis
representing the error level in years and the y-axis
representing the cumulative score (in percentage value). This
type of evaluation performance was sometimes combined with
the MAE evaluation method in studies that used MORPH, FG-
NET, and other earlier year databases. Meanwhile, studies that
used the MegaAge-Asian database present some of their
results in terms of CA(θ), where θ is the allowable age error
corresponding to the cumulative accuracy, which several of
them are shown in Table IV.
Because the studies reviewed from 2015-2020 (see
Table IV) used different databases, analysing and comparing
their performance progress was difficult. Therefore, only the
most frequently used databases were chosen and averaged to
create a line chart depicting the performance progress of face
age research from 2015-2020. Fig. 5 illustrates the average
yearly performance for two different databases: MORPH and
FG-NET. As shown in Fig. 5, the MAE values for the MORPH
database decreased from 2015-2020, but not for FG-NET. The
chart may imply that models applied to the MORPH database
improved over six years, whereas FG-NET did not. Table VI Fig. 5. Line Chart Shows Face Age Research's Performance Progress from
shows the average MAE and its standard deviation for each 2015-2020 for MORPH and FG-NET (based on Table VI).

95 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

estimation, stating that having more samples for a specific


ethnic can increase age estimation accuracy. These studies
suggest that the CNN model can be further improved if the
overall framework takes ethnicity into account first. When a
large number of samples are available, CNN models perform
better at discovering significant aging features, as this also
improves the learning of racial aging traits.
Meanwhile, other studies on face age estimation that used
machine learning rather than CNN have demonstrated the
importance of ethnic aging traits. Ricanek et al. [94], for
example, investigated the ethnicity of the subject and introduced
the least angle regression (LAR) method, which was conducted
on three databases: MORPH, FG-NET, and PAL, with five
races included (African-American, Asian, Caucasian, Hispanic,
and Indian). In another study, Akinyemi and Onifade [95]
improved the performance of their model by incorporating
Fig. 6. Publishers and Frequency of Publication of Deep Learning Face Age ethnic parameters for African and Caucasian people into the
Estimation Papers, Including Publication Year (based on Table III). GroupWise age ranking model. The FG-NET and FAGE
databases were used in their experiment. FAGE is a locally
TABLE VI. PERFORMANCE PROGRESS OF FACE AGE RESEARCH FROM collected dataset of 238 images of 209 black (African)
2015-2020 FOR MORPH AND FG-NET DATABASES
individuals aged 0 to 41 years.
MORPH FG-NET
Year Shin et al. [96] presented an age estimation system that
MAE (std. deviation) MAE (std. deviation)
considered ethnic differences for Asians and Non-Asians using
2015 4.09 (0.61) 4.26 (-)
CNN and support vector machine (SVM). The proposed age
2016 3.18 (0.28) 2.8 (-) estimation system outperformed the standard system when
2017 2.87 (0.29) 3.11 (0.63) trained on an ethnicity-biased database. The study relied on
2018 2.78 (0.14) 3.19 (0.21) LFW [90] and its samples, with Asians in the datasets
2019 2.85 (0.07) 4.82 (1.96) consisting of Korean, Japanese, and Chinese web celebrities
2020 2.50 (0.33) 3.24 (0.49) [96]. Several other studies, however, were unable to investigate
their approach further due to a lack of multiple races in their
VI. OPEN ISSUES datasets [94, 95, 96]. Hence, the importance of having more
As mentioned in the facial aging section, different ethnic race variety in databases is demonstrated.
subjects age differently, which means that a 20-year-old White Table III shows that the databases in the deep learning age
subject would look older than a 20-year-old Asian. More estimation model mostly favour Caucasian subjects. The race
research into the effects of ethnicity on face age estimation is variety in the databases is imbalanced; in most databases, the
needed. However, most studies focus only on primary aging Caucasian/White subjects are always the majority, while other
features, such as face shape and aging wrinkles, and ignore races are either underrepresented or missing. Moreover, some of
secondary ones, such as racial facing aging traits. There are two the databases with large samples and multiple races have no
possible reasons why studies did not take ethnic traits into information on the subjects’ ethnicity. There are only a few
account. The first is the perception that secondary aging features ethnic-specific databases, such as AFAD, IMFDB, and Iranian
are non-essential for better model performance. A few CNN Face available. It would be beneficial to have more multiple
face age estimation studies have disproved the perception that ethnic databases with large samples and races that are evenly
race is unimportant. The second possible reason is that the lack distributed.
of race variety in most databases causes researchers to overlook
racial traits as one of the aging estimation problems in the first VII. DISCUSSION AND SUGGESTION
place. The first problem to address is the negative perception that
Among papers reviewed, only seven considered ethnicity ethnicity is not a critical aging factor. Researchers should be
traits in the face age estimation experiment [52, 66, 70, 72, 73, informed more about the importance of ethnic traits in the aging
74, 85]. Studies in [52, 66, 70] considered ethnicity in the model face; thus, this paper aimed to raise awareness on this to others.
learning performance and discovered that it does improve age Moreover, face age estimation research should be expanded;
estimation. However, these studies did not investigate the more researchers should consider the secondary aging traits
influence of racial traits on effectiveness in facial age when building CNN face age estimation models. The research
estimation. Meanwhile, the study in [73] inferred that scope should not be limited to primary aging features (face
performance would improve if both gender and race shape and aging wrinkles) but also expanded to secondary
information were included. Another study [74] discovered that features that can help distinguish unique aging traits that occur
gender and race could easily affect its age estimation model. only in specific races. One suggestion is to create a framework
Combining all of the age, gender, and race features further for organising different racial samples in a database before
improved the age estimation performance of the model. Lastly, being used for a CNN model. The steps of the framework are as
[85] explored the impact of ethnicity and gender on age follow:

96 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

1) First, decide on the number of races to be included in


the study and then collect as many samples as possible for
each race while ensuring the samples are similar in quantity.
This may require the creation of multiple databases with
various ethnicities (e.g. using MORPH [25], IMFDB [16],
Iranian face [15], and MEGAAGE-ASIAN [89] databases
together). Because CNN would be the model approach, having
a large sample size would not be an issue for the CNN
learning process - it is required. The study must also decide
whether to use all samples or specific ones based on the
research aim and objective(s).
2) Next, apply the necessary image processing to the
sample images, such as face detection, face landmark, and
face alignment.
3) Each database’s estimation performance is evaluated
using an evaluation protocol. Multiple ethnic subjects from
chosen databases are mixed and segregated into specific
training and test sets when accounting for ethnicity in age
estimation. The ethnic effect analysis requires two protocols:
one that considers ethnicity and one that does not. The first
protocol requires different ethnic subjects within these sets to
be divided equally in quantity. Training and testing, for
example, take up 80% and 20% of the total samples,
respectively. When the samples are made up of two races (e.g.
Caucasian and Asian), half of the training samples should be Fig. 7. Framework for Organising Different Racial Samples for CNN Face
Age Estimation Study - Subject's Ethnicity Accounted for.
Caucasian and the other half should be Asian. Similarly with
the test samples – half is Caucasian, and another half is Asian. Although many public databases are available for face age
The second protocol is similar to the first, except that the estimation studies, very few are non-Caucasian/non-White
different ethnic subjects are split randomly rather than equally. databases. Accordingly, two suggestions could enable the
4) Afterwards, run the samples into the CNN model and collection of more ethnic-specific samples; either for private or
analyse the result in terms of the ethnicity's effect on the public use:
overall age estimation. Search for any significant finding 1) Organise an ethnic-biased age estimation contest and
regarding the ethnicities traits that can be exploited in future develop an ethnic-specific dataset like how it was done for the
age estimation studies. ICCV 2015 ChaLearn [32] challenge dataset. This approach
Fig. 7 shows the proposed framework for studying CNN can help increase ethnic-biased age estimation studies from
face age estimation while considering the ethnicities of the contestants and ethnic-specific database usage (e.g. AFAD,
subjects. This framework can guide future face age estimation IMFDB, and Iranian database). These databases may become
studies that use either the deep learning approach or the benchmark databases themselves later on.
standard machine learning approach. 2) In dangerous times, such as the current COVID-19
The review of the papers revealed that most studies did not pandemic, most work and communication are now done
consider using other ethnic-specific databases (e.g. Iranian [15], online. Governments, businesses, educational institutions,
Indian [16]), even though these databases are available for use medical institutions, and others are now using communication
(see Table I). Benchmark databases like MORPH and FG-NET platforms for videoconferencing, online meetings, workspace
are more preferred because it is safer since these databases are chat, online classes, and even file sharing. One of the
frequently used and have long been used for comparison; thus, communication platform's primary functions is video
making it easy to perform comparative analysis. Nevertheless, streaming, which can accommodate up to nine people (or
using only the same benchmark databases and ignoring other
more) concurrently. This video streaming function allows
available ethnic-specific databases can pose a risk, which will
hinder the face age estimation research's progress in researchers to organise a video conference for a group of
understanding the overall ethnic factor in facial aging process. volunteers to collect ethnic-specific samples for face age
Suppose various databases are continuously and increasingly studies by capturing volunteers' face images during video
used throughout the years. There will be enough results to allow streaming. Researchers must first decide whether to collect
meaningful comparison between studies, resulting in new samples in a controlled/uncontrolled environment, for
benchmark databases that can be used and compared in the example, by requesting volunteers to standardise their
future. background colours (use one colour) and stand still while
researchers prepare to capture their faces (controlled

97 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

environment). Researchers must also decide whether to It would be interesting to develop the suggested model
capture a single face image or multiple faces at once. framework with different ethnic races for face age recognition.
However, the size and quality of the faces in the video may Significant racial traits might be discovered, which can further
differ between users. Therefore, this should be taken into distinguish the aging processes between different ethnic people.
This discovery could further improve the understanding of
account when trying to use this approach to collect samples
racial aging traits, particularly concerning the face and the
from volunteers, which can be co-workers or students (if the development of a model that can learn and identify those traits.
researcher is also an educator). Moreover, additional Additionally, using the suggested sample collection method to
information about volunteers, such as their age and ethnicity, collect and capture own samples may help ease the collection
can be directly requested and recorded for research purposes. process. Aside from face age studies, the collected face
Microsoft Teams, Zoom, and Google Meet are some of the images/samples can also be used for other facial image studies,
communication platforms that are available for use. Fig. 8 such as emotion recognition and ethnic recognition. These
shows an example of captured face images using Microsoft suggestions, however, are beyond the scope of this study and
Teams (single face or multiple faces). will be considered in future research.

When collecting samples, likely, some people would not be VIII. CONCLUSION
willing to help or give any personal information. Therefore, The analysis in this paper focused on ethnic consideration in
proper planning on target subjects selection before collecting the dataset used for the last six years for accurate age estimation
their face images is required. using the deep learning approach. This paper specifically
analysed 53 papers on deep learning face age estimation, model
performance, selected databases, and whether or not any face
ethnicity traits analysis was performed when estimating age.
This paper also highlighted 19 database papers that promote the
use of publicly available databases for age estimation research,
as well as information on multiple database ethnicities.
Although the deep learning approach improves face age
estimation over time, it can be further enhanced by
understanding how ethnicity affects face age estimation and
designing an evaluation protocol that takes the subjects’ ethnic
traits into account. Moreover, a sizeable multi-racial database is
needed for the investigation of aging in different ethnic groups.
Therefore, it is crucial to collect the necessary information to
create an extensive database with well-distributed age and
ethnic labels. Suggestions for capturing samples were also
provided to help researchers in increasing their ethnic-specific
samples for private or public use.
IX. FUTURE DIRECTION
Making the collected ethnic-specific samples public and
sharing them via web image collection sites can increase
interest in conducting more ethnicity-based face age estimation
research. More robust deep learning face age estimation models
can be developed by performing more such studies, sample
collection, and analyses in the future. Future research could also
discover significant racial traits that could help distinguish
unique aging patterns used to solve racial face age estimation
problems in real-life applications. Proper planning and key
considerations must be made when collecting samples, such as
ensuring personal data privacy or a subject’s consent.
Additionally, it would be good to reiterate the benefit of having
more samples for studies beyond facial age recognition.
ACKNOWLEDGMENT
The authors are grateful to the Faculty of Information
Fig. 8. Samples of Face Images - Captured using Microsoft Teams (Images Science and Technology, The National University of Malaysia,
taken from [97]).
for supporting and contributing to this study; under the grant
code GGPM-2019-038.

98 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

REFERENCES [23] M. Minear, & D. C. Park, “A lifespan database of adult facial stimuli,”
[1] N. A. Vashi, M. B. D. C. Maymone, & R. V. Kundu, “Aging differences Behavior Research Methods, Instruments, & Computers, vol. 36, no. 4,
in ethnic skin,” The Journal of clinical and aesthetic dermatology, vol. 9, pp. 630-633, 2004.
no. 1, pp. 31, 2016. [24] P. J. Phillips, et al., “Overview of the face recognition grand challenge,”
[2] Y. Shirakabe, Y. Suzuki, & S. M. Lam, “A new paradigm for the aging in 2005 IEEE computer society conference on computer vision and
Asian face,” Aesthetic plastic surgery, vol. 27, no. 5, pp. 397-402, 2003. pattern recognition (CVPR'05), Vol. 1, pp. 947-954, June, 2005.
[3] A. E. Brissett, & M. C. Naylor, “The aging african-american face,” [25] K. Ricanek, & T. Tesafaye, “Morph: A longitudinal image database of
Facial Plastic Surgery, vol. 26, no. 2, pp. 154-163, 2010. normal adult age-progression,” in IEEE 7th International Conference on
Automatic Face and Gesture Recognition (FGR06), pp. 341-345, April,
[4] M. O. Harris, “The aging face in patients of color: Minimally invasive 2006.
surgical facial rejuvenation—A targeted approach,” Dermatologic
therapy, vol. 17, no. 2, pp. 206-211, 2004. [26] Y. Fu & T. S. Huang, “Human age estimation with regression on
discriminative aging manifold,” in IEEE Transactions on Multimedia,
[5] M. G. Rhodes, “Age estimation of faces: A review,” Applied Cognitive vol. 10 no. 4, pp. 578-584, 2008.
Psychology: The Official Journal of the Society for Applied Research in
Memory and Cognition, vol. 23, no. 1, pp. 1-12, 2009. [27] A. C. Gallagher, & T. Chen, “Understanding images of groups of
people,” in 2009 IEEE Conference on Computer Vision and Pattern
[6] M. S. Zimbler, M. S. Kokoska, & J. R. Thomas, “Anatomy and Recognition, pp. 256-263, June, 2009.
pathophysiology of facial aging,” Facial plastic surgery clinics of North
America, vol. 9, no. 2, pp. 179-87, 2001. [28] N. C. Ebner, M. Riediger, & U. Lindenberger, “FACES—A database of
facial expressions in young, middle-aged, and older women and men:
[7] N. Ramanathan, R. Chellappa, & S. Biswas, “Age progression in human Development and validation,” Behavior research methods, vol. 42, no. 1,
faces: A survey,” Journal of Visual Languages and Computing, vol. 15, pp. 351-362, 2010.
3349-3361, 2009.
[29] S. Zheng, “Visual image recognition system with object-level image
[8] G. Y., Guo, & T. S. Huang, “Age synthesis and estimation via faces: A representation,” Doctoral dissertation, 2012.
survey,” IEEE transactions on pattern analysis and machine intelligence,
vol. 32, no. 11, pp. 1955-1976, 2010. [30] E. Eidinger, R. Enbar, & T. Hassner, “Age and gender estimation of
unfiltered faces,” in IEEE Transactions on Information Forensics and
[9] R. Russell, et al. "Facial contrast is a cue for perceiving health from the Security, vol. 9, no. 12, pp. 2170-2179, 2014.
face," in Journal of Experimental Psychology: Human Perception and
Performance, vol 42, no. 9, pp. 1354, 2016. [31] B. C. Chen, C. S. Chen, & W. H. Hsu, “Cross-age reference coding for
age-invariant face recognition and retrieval,” in European conference on
[10] C. Trojahn, G. Dobos, A. Lichterfeld, U. Blume-Peytavi, & J. Kottner, computer vision, Springer, Cham, pp. 768-783, September, 2014.
“Characterizing facial skin aging in humans: disentangling extrinsic
from intrinsic biological phenomena,” in BioMed research international, [32] S. Escalera, et al., “Chalearn looking at people 2015: Apparent age and
2015. cultural event recognition datasets and results,” In Proceedings of the
IEEE International Conference on Computer Vision Workshops, pp. 1-9,
[11] M. S. Zimbler, M. S. Kokoska, & J. R. Thomas, “Anatomy and 2015.
pathophysiology of facial aging,” Facial plastic surgery clinics of North
America, vol. 9, no. 2, pp. 179-87, 2001. [33] S. Escalera, et al., “Chalearn looking at people and faces of the world:
Face analysis workshop and challenge 2016,” In Proceedings of the
[12] T. Igarashi, K. Nishino, & S. K. Nayar, “The appearance of human skin: IEEE Conference on Computer Vision and Pattern Recognition
A survey,” Foundations and Trends® in Computer Graphics and Vision, Workshops, pp. 1-8, 2016.
vol. 3, no. 1, pp. 1-95, 2007.
[34] S. Moschoglou, A. Papaioannou, C. Sagonas, J. Deng, I. Kotsia, & S.
[13] N. A. Vashi, M. B. D. C. Maymone, & R. V. Kundu, “Aging differences Zafeiriou. “Agedb: the first manually collected, in-the-wild age
in ethnic skin,” The Journal of clinical and aesthetic dermatology, vol. 9, database,” In Proceedings of the IEEE Conference on Computer Vision
no. 1, pp. 31, 2016 and Pattern Recognition Workshops, pp. 51-59, 2017.
[14] Stock photo caucasian baby. 2021. Crello. [online] Available at:
[35] R. Rothe, R. Timofte, & L. Van Gool, “Deep expectation of real and
<https://crello.com/unlimited/stock-photos/166742418/stock-photo-
apparent age from a single image without facial landmarks,” in
caucasian-baby/> [Accessed 22 July 2021].
International Journal of Computer Vision, vol. 126, no. 2-4, pp. 144-
[15] A. Bastanfard, M. A. Nik, & M. M. Dehshibi, “Iranian face database 157, 2018.
with age, pose and expression,” in Machine Vision, pp. 50-55, 2007.
[36] E. Agustsson, et al., “Apparent and real age estimation in still images
[16] S. Setty, et al., “Indian movie face database: a benchmark for face with deep residual regressors on appareal database,” in 2017 12th IEEE
recognition under wide variations,” in 2013 Fourth National Conference International Conference on Automatic Face & Gesture Recognition (FG
on Computer Vision, Pattern Recognition, Image Processing and 2017), pp. 87-94, May, 2017.
Graphics (NCVPRIPG), pp. 1-5, December, 2013.
[37] V. Carletti, A. Greco, G. Percannella, & M. Vento, “Age from faces in
[17] Z. Niu, M. Zhou, L. Wang, X. Gao, & G. Hua, “Ordinal regression with the deep learning revolution,” in IEEE transactions on pattern analysis
multiple output cnn for age estimation,” in Proceedings of the IEEE and machine intelligence, 2019.
conference on computer vision and pattern recognition, pp. 4920-4928,
[38] X. Liu, et al. “Agenet: Deeply learned regressor and classifier for robust
2016.
apparent age estimation,” in Proceedings of the IEEE International
[18] Y. H. Kwon & N. da Vitoria Lobo. “Age classification from facial Conference on Computer Vision Workshops, pp. 16-24, 2015.
images,” Computer vision and image understanding, vol. 74, no. 1, pp.
[39] R. Rothe, R. Timofte, & L.Van Gool, “Dex: Deep expectation of
1-21, 1999.
apparent age from a single image,” in Proceedings of the IEEE
[19] R. Angulu, J. R. Tapamo, & A. O. Adewumi. “Age estimation via face International Conference on Computer Vision Workshops, pp. 10-15,
images: a survey,” EURASIP Journal on Image and Video Processing, 2015.
2018, no. 1, pp. 42, 2018.
[40] Y. Zhu, Y. Li, G. Mu, & G. Guo, “A study on apparent age estimation,”
[20] O. F. Osman, & M. H. Yap. “Computational intelligence in automatic in Proceedings of the IEEE International Conference on Computer
face age estimation: A survey,” IEEE Transactions on Emerging Topics Vision Workshops, pp. 25-31, 2015.
in Computational Intelligence, vol. 3, no. 3, pp. 271-285, 2018.
[41] Z. Kuang, C. Huang, & W. Zhang, “Deeply learned rich coding for
[21] P. J. Phillips, H. Wechsler, J. Huang, & P. J. Rauss, “The FERET cross-dataset facial age estimation,” in Proceedings of the IEEE
database and evaluation procedure for face-recognition algorithms,” International Conference on Computer Vision Workshops, pp. 96-101,
Image and vision computing, vol. 16, no. 5, pp. 295-306, 1998. 2015.
[22] A. Lanitis, C. J. Taylor, & T. F. Cootes, “Toward automatic simulation [42] X. Yang, et al. “Deep label distribution learning for apparent age
of aging effects on face images,” in IEEE Transactions on pattern estimation,” in Proceedings of the IEEE international conference on
Analysis and machine Intelligence, vol. 24, no. 4, pp. 442-455, 2002. computer vision workshops, pp. 102-108, 2015.

99 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

[43] X. Wang, R. Guo, & C. Kambhamettu, “Deeply-learned feature for age on pattern analysis and machine intelligence, vol. 40, no. 11, pp. 2610-
estimation,” in 2015 IEEE Winter Conference on Applications of 2623, 2017.
Computer Vision, pp. 534-541, January, 2015. [62] R. Ranjan, S. Sankaranarayanan, C. D. Castillo, & R. Chellappa, “An
[44] S. Li, J. Xing, Z. Niu, S. Shan, & S. Yan, “Shape driven kernel all-in-one convolutional neural network for face analysis,” in 2017 12th
adaptation in convolutional neural network for robust facial traits IEEE International Conference on Automatic Face & Gesture
recognition,” in Proceedings of the IEEE Conference on Computer Recognition (FG 2017), pp. 17-24, May, 2017.
Vision and Pattern Recognition, pp. 222-230, 2015. [63] H. Liu, J. Lu, J. Feng, & J. Zhou, “Ordinal deep learning for facial age
[45] I. Huerta, C. Fernández, C. Segura, J. Hernando, & A. Prati, “A deep estimation,” in IEEE Transactions on Circuits and Systems for Video
analysis on age estimation,” Pattern Recognition Letters, vol. 68, pp. Technology, vol. 29, no. 2, pp. 486-501, 2017.
239-249, 2015. [64] H. Liu, J. Lu, J. Feng, & J. Zhou, “Group-aware deep feature learning
[46] G. Levi, & T. Hassner, “Age and gender classification using for facial age estimation,” in Pattern Recognition, vol. 66, pp. 82-94,
convolutional neural networks,” in Proceedings of the IEEE conference 2017.
on computer vision and pattern recognition workshops, pp. 34-42, 2015. [65] G. Antipov, M. Baccouche, S. A. Berrani, & J. L. Dugelay, “Effective
[47] R. Ranjan, S. Zhou, J. Cheng Chen, A. Kumar, A. Alavi, V. M. Patel, & training of convolutional neural networks for face-based gender and age
R. Chellappa, “Unconstrained age estimation with deep convolutional prediction,” in Pattern Recognition, vol. 72, pp. 15-26, 2017.
neural networks,” in Proceedings of the IEEE International Conference [66] J. Xing, K. Li, W. Hu, C. Yuan, & H. Ling, “Diagnosing deep learning
on Computer Vision Workshops, pp. 109-117, 2015. models for high accuracy age estimation from a single image,” in Pattern
[48] J. C. Chen, A. Kumar, R. Ranjan, V. M. Patel, A. Alavi, & R. Recognition, vol. 66, pp. 106-116, 2017.
Chellappa, “A cascaded convolutional neural network for age estimation [67] K. Li, J. Xing, W. Hu, & S. J. Maybank, “D2C: Deep cumulatively and
of unconstrained faces,” in 2016 IEEE 8th International Conference on comparatively learning for human age estimation,” in Pattern
Biometrics Theory, Applications and Systems (BTAS), pp. 1-8, Recognition, vol. 66, pp. 95-105, 2017.
September, 2016.
[68] L. Hou, D. Samaras, T. M. Kurc, Y. Gao, & J. H. Saltz, “Convnets with
[49] F. Gurpinar, H. Kaya, H. Dibeklioglu, & A. Salah, “Kernel ELM and smooth adaptive activation functions for regression,” in Proceedings of
CNN based facial age estimation,” in Proceedings of the IEEE machine learning research, vol. 54, pp. 430, 2017.
conference on computer vision and pattern recognition workshops, pp.
[69] K. Zhang, et al., “Age group and gender estimation in the wild with deep
80-86, 2016.
ror architecture,” IEEE Access, vol. 5, pp. 22492-22503, 2017.
[50] Z. Niu, M. Zhou, L. Wang, X. Gao, & G. Hua, “Ordinal regression with
[70] F. Wang, H. Han, S. Shan, & X. Chen, “Deep multitask learning for
multiple output cnn for age estimation,” in Proceedings of the IEEE
joint prediction of heterogeneous face attributes,” in 2017 12th IEEE
conference on computer vision and pattern recognition, pp. 4920-4928,
International Conference on Automatic Face & Gesture Recognition (FG
2016.
2017), pp. 173-179, May, 2017.
[51] R. Rothe, R. Timofte, & L. Van Gool, “Some like it hot-visual guidance
[71] H. Liu, J. Lu, J. Feng, & J. Zhou, “Label-sensitive deep metric learning
for preference prediction,” in Proceedings of the IEEE conference on
computer vision and pattern recognition, pp. 5553-5561, 2016. for facial age estimation,” IEEE Transactions on Information Forensics
and Security, vol. 13, no. 2, pp. 292-305, 2017.
[52] Y. Yang, F. Chen, X. Chen, Y. Dai, Z. Chen, J. Ji, & T. Zhao, “Video
[72] H. Han, A. K. Jain, F. Wang, S. Shan, & X. Chen, “Heterogeneous face
system for human attribute analysis using compact convolutional neural
attribute estimation: A deep multitask learning approach,” IEEE
network,” in 2016 IEEE International Conference on Image Processing
(ICIP), pp. 584-588, September, 2016. transactions on pattern analysis and machine intelligence, vol. 40, no.
11, pp. 2597-2609, 2017.
[53] Y. Dong, Y. Liu, & S. Lian, “Automatic age estimation based on deep
learning algorithm,” Neurocomputing, vol. 187, pp. 4-10, 2016. [73] J. Wan, Z. Tan, Z. Lei, G. Guo, & S. Z. Li, “Auxiliary demographic
information assisted age estimation with cascaded structure,” IEEE
[54] G. Ozbulak, Y. Aytar, & H. K. Ekenel, “How transferable are CNN- transactions on cybernetics, vol. 48, no. 9, pp. 2531-2541, 2018.
based features for age and gender classification?” in 2016 International
[74] M. Duan, K. Li, & K. Li, “An Ensemble CNN2ELM for Age
Conference of the Biometrics Special Interest Group (BIOSIG), pp. 1-6,
September, 2016. Estimation,” IEEE Transactions On Information Forensics And Security,
vol. 13, no. 3, 2018.
[55] G. Antipov, M. Baccouche, S. A. Berrani, & J. L. Dugelay, “Apparent
[75] H. F. Yang, B. Y. Lin, K. Y. Chang, & C. S. Chen, “Joint estimation of
age estimation from face images combining general and children-
age and expression by combining scattering and convolutional
specialized deep learning models,” in Proceedings of the IEEE
networks,” ACM Transactions on Multimedia Computing,
conference on computer vision and pattern recognition workshops, pp.
Communications, and Applications (TOMM), vol. 14, no. 1, pp. 9. 2018.
96-104, 2016.
[56] Z. Huo, X. Yang, C. Xing, Y. Zhou, P. Hou, J. Lv, & X. Geng, “Deep [76] B. Yoo, Y. Kwak, Y. Kim, C. Choi, & J. Kim, “Deep facial age
age distribution learning for apparent age estimation,” in Proceedings of estimation using conditional multitask learning with weak label
the IEEE conference on computer vision and pattern recognition expansion,” IEEE Signal Processing Letters, vol. 25, no. 6, pp. 808-812,
workshops, pp. 17-24, 2016. 2018.
[57] M. Uricár, R. Timofte, R. Rothe, J. Matas, & L. Van Gool, “Structured [77] S. Taheri, & Ö. Toygar, “On the use of DAG-CNN architecture for age
output svm prediction of apparent age, gender and smile from deep estimation with multi-stage features fusion,” Neurocomputing, vol. 329,
features,” in Proceedings of the IEEE Conference on Computer Vision pp. 300-310, 2019.
and Pattern Recognition Workshops, pp. 25-33, 2016. [78] W. Im, S. Hong, S. E. Yoon, & H. S. Yang, “Scale-Varying Triplet
[58] R. Can Malli, M. Aygun, & H. Kemal Ekenel, “Apparent age estimation Ranking with Classification Loss for Facial Age Estimation,” in
using ensemble of deep learning models,” in Proceedings of the IEEE Computer Vision – ACCV 2018. Lecture Notes in Computer Science,
Conference on Computer Vision and Pattern Recognition Workshops, vol. 11365. Springer, Cham, 2018.
pp. 9-16, 2016. [79] O. Sendik, and Y. Keller. “DeepAge: Deep Learning of face-based age
[59] B. Hebda, & T. Kryjak, “A compact deep convolutional neural network estimation,” in Signal Processing: Image Communication, vol. 78, pp.
architecture for video based age and gender estimation,” in 2016 368-375, 2019.
Federated Conference on Computer Science and Information Systems [80] S. Penghui, L. Hao, W. Xin, Y. Zhenhua, & S. Wu. “Similarity-aware
(FedCSIS), pp. 787-790, September, 2016. deep adversarial learning for facial age estimation,” in 2019 IEEE
[60] Z. Hu, Y. Wen, J. Wang, M. Wang, R. Hong, & S. Yan, “Facial age International Conference on Multimedia and Expo (ICME), pp. 260-265,
estimation with age difference,” in IEEE Transactions on Image July, 2019.
Processing, vol. 26, no. 7, pp. 3087-3097, 2016. [81] C. Miron, V. Manta, R. Timofte, A. Pasarica, & R. I. Ciucu, “Efficient
[61] Z. Tan, J. Wan, Z. Lei, R. Zhi, G. Guo, & S. Z. Li, “Efficient group-n convolutional neural network for apparent age prediction,” in 2019 IEEE
encoding and decoding for facial age estimation,” in IEEE transactions

100 | P a g e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 12, No. 11, 2021

15th International Conference on Intelligent Computer Communication environments,” Workshop on Faces in 'Real-Life' Images: Detection,
and Processing (ICCP), pp. 259-262, September, 2019. Alignment, and Recognition, October, 2008.
[82] N. Savov, M. Ngo, S. Karaoglu, H. Dibeklioglu, & T. Gevers, “Pose and [91] Z. Liu, P. Luo, X. Wang, & X. Tang, “Deep learning face attributes in
Expression Robust Age Estimation via 3D Face Reconstruction from a the wild,” in Proceedings of the IEEE international conference on
Single Image,” in Proceedings of the IEEE International Conference on computer vision, pp. 3730-3738, 2015.
Computer Vision Workshops, pp. 0-0, 2019. [92] O. Russakovsky, et al., “Imagenet large scale visual recognition
[83] H. Liu, J. Lu, J. Feng, & J. Zhou. “Ordinal Deep Learning for Facial challenge,” International journal of computer vision, vol. 115, no. 3, pp.
Age Estimation,” IEEE Transactions on Circuits and Systems for Video 211-252, 2015.
Technology, vol. 29, no. 2, pp. 486-501, 2019. [93] A. Khan, A. Sohail, U. Zahoora, & A. S. Qureshi. “A survey of the
[84] H. Liu, P. Sun, J. Zhang, S. Wu, Z. Yu, & X. Sun. “Similarity-Aware recent architectures of deep convolutional neural networks,” Artificial
and Variational Deep Adversarial Learning for Robust Facial Age Intelligence Review, vol. 53, no. 8, 5455-5516, 2020.
Estimation,” IEEE Transactions on Multimedia, 2020. [94] K. Ricanek, Y. Wang, C. Chen, & S. J. Simmons, “Generalized multi-
[85] P. Li, Y. Hu, X. Wu, R. He, & Z. Sun. “Deep label refinement for age ethnic face age-estimation,” in 2009 IEEE 3rd International Conference
estimation,” Pattern Recognition, vol. 100, p. 107178, 2020. on Biometrics: Theory, Applications, and Systems, pp. 1-6, September,
[86] N. Liu, F. Zhang, & F. Duan. “Facial Age Estimation Using a Multitask 2009.
Network Combining Classification and Regression,” IEEE Access, vol. [95] J. D. Akinyemi, & O. F. Onifade, “An ethnic-specific age group ranking
8, pp. 92441-92451, 2020. approach to facial age estimation using raw pixel features,” in 2016
[87] X. Liu, Y. Zou, H. Kuang, & X. Ma. “Face Image Age Estimation Based IEEE Symposium on Technologies for Homeland Security (HST), pp. 1-
on Data Augmentation and Lightweight Convolutional Neural 6, May, 2016.
Network,” Symmetry, vol. 12, no. 1, pp. 146, 2020. [96] M. Shin, J. H. Seo, & D. S. Kwon, “Face image-based age and gender
[88] J. C. Xie, & C. M. Pun. “Deep and Ordinal Ensemble Learning for estimation with consideration of ethnic difference,” in 2017 26th IEEE
Human Age Estimation From Facial Images,” IEEE Transactions on International Symposium on Robot and Human Interactive
Information Forensics and Security, vol. 15, pp. 2361-2374, 2020. Communication (RO-MAN), pp. 567-572. 2017.
[89] M. Xia, X. Zhang, L. Weng, & Y. Xu. “Multi-Stage Feature Constraints [97] Microsoft.com. 2021. Online Meeting Software, Video Conferencing |
Learning for Age Estimation,” IEEE Transactions on Information Microsoft Teams. [online] Available at:
Forensics and Security, vol. 15, pp. 2417-2428, 2020. <https://www.microsoft.com/en-my/microsoft-teams/online-meetings>
[Accessed 3 May 2021].
[90] G. B. Huang, M. Mattar, T. Berg, & E. Learned-Miller, “Labeled faces
in the wild: A database for studying face recognition in unconstrained

101 | P a g e
www.ijacsa.thesai.org

You might also like