Speeding Up Composite Differential Evolution for Structural Optimization Using Neural Networks
Speeding Up Composite Differential Evolution for Structural Optimization Using Neural Networks
To cite this article: Tran-Hieu Nguyen & Anh-Tuan Vu (2022) Speeding up Composite
Differential Evolution for structural optimization using neural networks, Journal of Information
and Telecommunication, 6:2, 101-120, DOI: 10.1080/24751839.2021.1946740
1. Introduction
Structural optimization (SO) plays an important role in the field of architecture, engineer-
ing, and construction. Applying SO in the design phase not only reduces the construction
cost but also saves the consumption of natural resources, thereby minimizing the impact
on the environment. Therefore, SO has gained great attention in recent years. SO is a kind
of constrained optimization problem, in which the final solution must satisfy a set of
design constraints. Checking design constraints is often computationally expensive
because it requires conducting finite element analysis. In the past three decades, evol-
utionary algorithms have been preferred for this task due to their capability to cope
with the multimodal objective function as well as nonlinear constraints. Many evolution-
ary algorithms have been used to optimize structures, for example, GA (Rajeev &
CONTACT Tran-Hieu Nguyen [email protected] National University of Civil Engineering, 55 Giai Phong Road,
Hanoi, Vietnam
© 2021 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/
licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly
cited.
102 T.-H. NGUYEN AND A.-T. VU
Krishnamoorthy, 1992), ES (Cai & Thierauf, 1996), DE (Wang et al., 2009). Particularly, many
researchers have selected the DE algorithm for solving structural optimization because it
is simple and easy to implement.
Basically, DE has four operators: initialization, mutation, crossover, and selection, in
which the mutation operator strongly affects the performance of the DE algorithm.
There exist many mutation strategies, for example, ‘rand/1’, ‘rand/2’, ‘best/1’, ‘best/2’,
‘current-to-rand/1’, and ‘current-to-best/1’, and each of them has its own advantage.
The ‘rand/1’ strategy is strong in exploring new space but it is weak in the local exploi-
tation. Consequently, the convergence speed of the ‘rand/1’ strategy is very slow. On the
contrary, the ‘current-to-best/1’ has a strong exploitation capacity but it is easy to fall
local optimum. It is noted that the balance of the global exploration and the local
exploitation is the key factor to ensure the performance of a metaheuristic algorithm.
Some techniques have been introduced for this purpose. For instance, the ‘current-to-
pbest’ mutation strategy was presented by Zhang and Sanderson (2009), in which the
best vector xbest at the current generation in the ‘current-to-best’ strategy is replaced
by the vector xpbest randomly selected from the top p × NP individuals. This mutation
strategy is able to avoid trapping to a local optimum. This idea was then applied in
the Adaptive DE algorithm when optimizing truss structures (Bureerat & Pholdee,
2016). Another technique, called ‘opposition-based mutation’ was introduced by Rahna-
mayan et al. (2008) to ensure that the difference is always oriented towards the better
vector. The opposition-based mutation was then employed for solving truss structure
optimization problems (Pham, 2016). Furthermore, it can be observed that in the
sooner generations, the diversity of the population plays an important role while the
ability of the local exploitation is needed in the later generations. Thus, an adaptive
mutation scheme was proposed (Ho-Huu et al., 2016) where one of two mutation strat-
egies ‘rand/1’ and ‘current-to-best/1’ is adaptively selected during the optimization
process.
Another approach is to use multiple mutation strategies simultaneously with the aim
of balancing global exploration and local exploitation. Mallipeddi et al. (2011) developed
the EPSDE algorithm, in which three strategies ‘best/2’, ‘rand/1’, ‘current-to-rand/1’ are
selected to form a pool. The scaling factor F varies from 0.4–0.9 with the step of 0.1
and the crossover rate Cr is taken a value from the list of (0.1–0.9) with the step of 0.1.
Initially, each target vector in the initial population is associated with a mutation strategy
and a control parameter setting from the pools. If the trial vector is better than the target
vector, the trial vector is kept to the next generation and the combination of the current
mutation strategy and parameters are saved into the memory. Otherwise, a new trial
vector is produced using a new combination of mutation strategy and control parameters
from pools or the memory. In this way, the probability of producing offspring which is
better than its parent gradually increases during the evolution.
Wang and Cai (2011) proposed the (µ + λ)-Constrained Differential Evolution ((µ + λ)-
CDE) in which three mutation strategies (i.e. ‘rand/1’, ‘rand/2’, and ‘current-to-best/1’)
are used simultaneously to solve constrained optimization problems. The experimental
results show that the (µ + λ)-CDE algorithm is very competitive when it outperformed
other algorithms in benchmark functions. This algorithm was later improved into the
(µ + λ)-ICDE (Jia et al., 2013). In this study, a new proposed mutation strategy called
‘current-to-rand/best/1’ was used instead of the ‘current-to-best/1’ strategy.
JOURNAL OF INFORMATION AND TELECOMMUNICATION 103
In both studies above, three offspring are generated by using three different mutation
strategies but the scaling factor and the crossover rate are fixed during the optimization
process. Obviously, these parameters also strongly influence the performance of the
algorithm. A large value of the scaling factor F increases the population diversity over
the design space while a small value of F speeds up the convergence. The influence of
the crossover rate Cr is similar. Based on this observation, Wang et al. (2011) proposed
a new algorithm called Differential Evolution with composite trials and control parameters
(CoDE) for solving unconstrained optimization problems. In details, the mutation strat-
egies including ‘rand/1/bin’, ‘rand/2/bin’, and ‘current-to-rand/1’ and three pairs of
control parameters [F = 1.0, Cr = 0.1], [F = 1.0, Cr = 0.9], [F = 0.8, Cr = 0.2] are randomly com-
bined to generate three trial vectors. The tests on 25 benchmark functions proposed in
the CEC 2005 show that CoDE is better than seven algorithms including JADE, jDE,
SaDE, EPSDE, CLPSO, CMA-ES, and GL-25. Following this trend, the C2oDE is developed
for constrained optimization problems (Wang et al., 2019). Three strategies used in the
C2oDE include ‘current-to-rand/1’, modified ‘rand-to-best/1/bin’, and ‘current-to-best/1/
bin’. Additionally, at each generation, the scaling factor and the crossover rate are ran-
domly selected a value from the pool Fpool = [0.6, 0.8, 1.0] and Crpool = [0.1, 0.2, 1.0],
respectively. The C2oDE is proved to be as good as or better than other state-of-the-art
algorithms.
Although these above modifications greatly improve the efficiency of the DE, the com-
putation time is also increased by approximately three times due to the expansion of the
population. For problems where the fitness evaluation is very costly, for example, struc-
tural optimization, the use of the C2oDE is not beneficial. In recent years, machine learning
(ML) has been increasingly applied in the field of structural engineering. The results of
some previous studies have shown that ML can accurately predict the behaviour of struc-
tures (Kim et al., 2020; Nguyen & Vu, 2021; Truong et al., 2020). This leads to a potential
solution to reduce the computation cost of the composite DE. In fact, despite generating
three trial vectors, only the best of them is used to compare with their parent vector. By
building ML-based surrogate models to identify the best trial vector, the number of exact
fitness evaluations can be greatly reduced and the overall computation time is conse-
quently shortened.
In the literature, there exist a few studies relating to this topic. Mallipeddi and Lee
(2012) employed a Kriging model to improve the performance of EPSDE. The Kriging sur-
rogate model was utilized with the aim of generating a competitive trial vector with less
computational cost. Krempser et al. (2012) generated three trial vectors using ‘best/1/bin’,
‘target-to-best/1/bin’ and ‘target-to-rand/1/bin’, then a Nearest Neighbors-based surro-
gate model was used to detect the best one. The advantage of the Nearest Neighbors
algorithm is that it does not require the training process as other machine learning algor-
ithms, thus, data is always updated. Although there exist many ML algorithms, the com-
parative study indicates that neural networks are a good choice for regression tasks due to
a trade-off between accuracy and computation cost (Nguyen & Vu, 2021). In the study pre-
sented at the 12th international conference on computational collective intelligence
(ICCCI 2020) (Nguyen & Vu, 2020), surrogate models developed based on neural networks
are integrated into the DE algorithm to reduce the optimization time. However, due to the
error of the surrogate model, the optimal results found by the surrogate assisted DE algor-
ithm and the conventional DE algorithm always have a little difference.
104 T.-H. NGUYEN AND A.-T. VU
This paper is an extended version of the ICCCI 2020 paper. In this work, neural networks
do not directly replace the exact fitness evaluation. Alternately, neural networks are
trained to predict the constraint violation degree of trial vectors. The trial vector having
the smallest predicted value is then selected for the knockout selection with the target
vector. The final selection is still based on the exact fitness evaluation, thus ensuring to
find the true optimum. The proposed approach takes full advantage of the CoDE algor-
ithm but it still keeps the optimization time as fast as the conventional DE algorithm.
The rest of the paper is organized as follows. Section 2 states the optimization problem
for truss structures and the basics of the CoDE algorithm. Neural networks are briefly pre-
sented in Section 3. The proposed approach is introduced in Section 4. Three test pro-
blems are carried out in Section 5 to demonstrate the efficiency of the proposed
approach. Section 6 concludes this paper.
where: Ai is the cross-sectional area of the ith member; ALi and AUi are the lower bound and
the upper bound of the cross-sectional area of the ith member; ρi and li are the density
and the length of the ith member, respectively; σi and sallow j are the calculated stress
and the allowable stress of ith member, respectively; δj and dj allow
are the calculated dis-
placement and the allowable displacement of jth node, respectively; n is the number of
truss members; and m is the number of free nodes.
erated over the design space. Each target vector contains D components xi (0) = {xi,1 (0) (0)
, xi,2 ,
… , x(0)
i,D}. NP is the population size, D is the number of design variables, and xi,j
(0)
is the jth
JOURNAL OF INFORMATION AND TELECOMMUNICATION 105
component of the ith vector. At generation t, each target vector produces a mutant vector
using the mutation operator. Some widely used mutation operators are presented as
follows.
where: xi (t) is the target vector; vi (t) is the mutant vector; xr1(t), xr2(t), xr3(t), xr4(t), xr5(t) are vectors
which are randomly selected from the population; F is the scaling factor; xbest (t) is the best
target vector of the population at generation t.
Then, a binomial crossover operator is employed to produce a trial vector as follows:
(t)
vi,j if j = K or rand[0, 1] ≤ Cr
ui,j =
(t)
(4)
xi,j(t) otherwise
where: ui,j (t), vi,j (t), xi,j (t) are the jth component of the trial vector, the mutant vector, and
the target vector, respectively; K is an integer randomly selected from 1 to D; rand[0,1]
is a uniformly distributed random number on the interval [0,1]; Cr is the crossover rate.
Next, the trial vector ui (t) and the target vector xi (t) are compared, and the better one is
kept for the (t + 1)th generation. The cycle of three steps (i.e. mutation, crossover, and
selection) is performed repeatedly until the stopping condition is reached.
2.2.2. Composite DE
CoDE was proposed by Wang et al. (2011) for unconstrained optimization problems. The
basic idea behind the CoDE is to randomly combine three mutation strategies with three
pairs of F and Cr for producing three trial vectors. Three mutation strategies used in the
CoDE are ‘rand/1/bin’, ‘rand/2/bin’, and ‘current-to-rand/1’. Three fixed control parameter
settings are [F = 1.0, Cr = 0.1], [F = 1.0, Cr = 0.9], [F = 0.8, Cr = 0.2]. Through experimental
results, the CoDE achieves outstanding performance.
A variation of Composite DE for constrained optimization problems called C2oDE was
introduced in 2018 (Wang et al., 2019). Some modifications in the C2oDE can be summar-
ized as follows. First of all, three mutation strategies used in C2oDE are: ‘current-to-rand/1’,
‘current-to-best/1/bin’, and modified ‘rand-to-best/1/bin’. This modification improves the
trade-off between diversity and convergence when searching over design space.
The second difference is that in the C2oDE, the scaling factor F and the crossover rate Cr
are taken from two pools Fpool and Crpool, respectively. According to the authors’ rec-
ommendation, the scaling factor pool consists of three values Fpool = [0.6,0.8,1.0] while
the crossover rate pool also contains three values Crpool = [0.1,0.2,1.0].
The final modification is carried out in the selection step. Obviously, the balance
between minimizing the degree of constraint violation and minimizing the objective
function is very important in constrained optimization problems. Therefore, the authors
106 T.-H. NGUYEN AND A.-T. VU
employ two constraint handling techniques simultaneously. The feasible rule proposed
by Deb (2000) is utilized to select the best trial vector among three ones while the selec-
tion between the best trial vector and the target vector is based on the ϵ-constrained
method (Takahama, 2006). The details of the feasible rule and the ϵ-constrained
method are presented in Section 2.3. Additionally, a brief illustration of the CoDE and
the C2oDE is displayed in Figure 1.
. Among two feasible solutions, the one having lower objective function value is
preferred.
. Among two infeasible solutions, the one having smaller constraint violation is
preferred.
It can be seen that in Deb’s rule, the satisfaction of constraints is the first priority, fol-
lowed by the minimization of the objective function. Takahama et al. (2006) proposed the
ϵ-constrained method as follows.
In the ϵ-constrained method, the constraint violation ϕ(x) is defined as the maximum of
all constraints or the sum of all constraints.
f(x) = max {maxj {0, gj (x)}} (6a)
f(x) = max {0, gj (x)}p (6b)
j
In FFNN, neurons are organized into many layers including the input layer, the hidden
layers, and the output layers. The information moves through the FFNN in a one-way
direction from the input layer through hidden layers to the output layers. The simplest
kind of FFNN is a ‘single-layer perceptron’ which contains only one hidden layer. Neural
networks having more than one hidden layer are called ‘multilayer perceptron’ or
‘deep neural networks’ with the aim of improving the performance of the models. A
typical architecture of a deep neural network is shown in Figure 2.
It can be expressed as follows. The neuron of jth layer uj (j = 1,2, … ,J ) receives a sum of
input xi (i = 1,2, … ,I) which is multiplied by the weight wji and will be transformed into the
input of the neuron in the next layer:
I
xj = f (uj ) = f w ji xi (8)
i=1
where f(.) is the activation function. The most commonly used activation functions are:
tanh, sigmoid, softplus, and rectifier linear unit (ReLU).
The loss function is used to measure the error between predicted values and true
values. The loss function is chosen depending on the type of task. For regression tasks,
the loss functions can be Mean Squared Error (MSE), Mean Absolute Error (MAE), or
Root Mean Squared Error (RMSE). To adapt NN for better predictions, the network must
be trained. The training process is essentially finding a set of weights in order to minimize
the loss function. In other words, it is an optimization process. In the field of machine
learning, the commonly used optimization algorithms are stochastic gradient descent
(SGD) or Adam optimizer. An effective technique, namely ‘backpropagation’, developed
by Rumelhart et al. (1986) is normally used for training, in which weights can be
updated with the gradient descent method as follows:
∂E
wij(t+1) = wij(t) − h (9)
∂wij(t)
where: E is the loss function between the predicted and the true values; η is the learning
rate.
4. Proposed approach
The proposed approach, called surrogate assisted-CoDE (SA-CoDE), is presented in Figure
3, in which the optimization process is separated into two phases.
In the first phase, the CoDE algorithm is implemented as usual. All trial vectors pro-
duced in this phase are checked design constraints. All information including details of
trial vectors as well as their constraint violation degrees is saved into memory. Data col-
lected during this phase are then used to train a neural network-based surrogate model.
The surrogate model aims to predict the degree of constraint violation for new trial
vectors produced in the next phase.
In the second phase, three trial vectors are still produced as in the CoDE algorithm.
However, these trial vectors are then passed through the surrogate model which has
just been trained in the previous phase. As described above, there always exists an
error between the predicted value and the true value. Therefore, three trial vectors are
compared based on the ϵ-level comparison to cover the inaccuracy of the surrogate
model. The best one continues to go to the pairwise comparison with its target vector.
At this step, the exact fitness evaluation is used and the better one is retained for the next
generation.
The CoDE algorithm employs [3 × (iter1 + iter2-1) + 1] × NP times of exact fitness evalu-
ations while the proposed approach needs only [3 × (iter1-1) + iter2 + 1] × NP fitness
evaluations where iter1 and iter2 are the numbers of generations of the first and the
second phase, respectively. Obviously, the number of exact fitness evaluations is signifi-
cantly reduced.
5. Experiments
5.1. Benchmark problems
Three famous benchmark problems in the field of SO are used to demonstrate the effec-
tiveness of the proposed approach. The first problem is a planar 10-bar truss structure, the
second problem is a spatial 25-bar truss structure, while the remaining one is a spatial 72-
bar truss structure.
truss structure has been studied by many researchers (Camp & Farshchin, 2014; Pham,
2016).
5.2. Setup
Three above problems are optimized using both CoDE and SA-CoDE with the same par-
ameters as follows: NP = 20, Fpool = [0.6, 0.8, 1.0], and Crpool = [0.1, 0.2, 1.0]. Three mutation
strategies used in this study are ‘rand/1/bin’, ‘rand/2/bin’, and ‘current-to-rand/1’. A large
penalty term is added to the objective function when any constraint is violated ensuring
all infeasible vectors are omitted during the optimization process. The ϵ value for the ϵ-
level comparison is fixed to 0.2. The optimization is finished after maxiter = 500
generations.
Particularly for the SA-CoDE, the number of generations of the first phase is iter1 = 50
while the number of generations of the second phase is iter2 = maxiter-iter1 = 450. It
means the dataset collected after the first phase contains [3 × (iter1−1)+1] × NP = 2960
samples. However, for some samples, the truss structures are close to being unstable,
leading to the large values of the constraint violation. These samples are considered out-
liers and should be cleaned up from the training data. Therefore, all samples having a con-
straint violation greater than 0.5 are deleted from the dataset.
Neural networks developed in all problems have the same architecture of (D-10D-10D-
10D-1). The input layer consists of D neurons where D is the number of design variables
(D = 10 for the 10-bar truss problem, D = 8 for the 25-bar truss problem, and D = 16 for the
72-bar truss problem). Three networks have also three hidden layers with 10D neurons per
layer. There exists only one neuron in the output layer. Inputs of the surrogate models are
the cross-sectional areas of truss members while the output is the constraint violation of
such structure. The activation function used in this study is ReLU and the loss function is
MSE. Neural networks are trained for 1000 epochs with a batch size of 10. The Adam algor-
ithm is utilized to minimizing the loss function of neural networks. Before training, the
input data should be normalized by dividing to the upper bounds of members’ cross-sec-
tional areas.
In addition, the standard DE algorithm is also implemented with the mutation strategy
‘current-to-best/2/bin’. This variant is known as one of the powerful mutation strategies of
the DE algorithm. The standard DE algorithm is implemented two times with the popu-
lation size NP is set to be 20 (called the DE(20)) and 30 (called the DE(30)). Other par-
ameters are as follows: the scaling factor F = 0.8, the crossover rate Cr = 0.9, maxiter = 500.
The experiments are carried out using the programming language Python. The finite
element code for structural analysis is also written in Python based on the direct
stiffness method. Neural network-surrogate models are developed using the ‘MLPRegres-
sor’ feature from the open-source machine learning library Scikit-learn (Pedregosa et al.,
2011). For all cases, 25 independents runs are performed. The computation is performed
on a personal computer with the following configuration: CPU Intel Core i5-5257 2.7 GHz,
RAM 8.00 Gb.
5.3. Results
Figures 7–9 plot the convergence histories for the 10-bar truss, the 25-bar truss, and the
72-bar truss, respectively.
The results obtained from the CoDE, the SA-CoDE, and the DE for 25 runs are reported
in terms of best, mean, worst, and standard deviation (SD) values in Tables 5–7. The
optimal designs found by the Differential Evolution with opposition-based mutation
114 T.-H. NGUYEN AND A.-T. VU
Figure 7. Convergence histories of the CoDE and the SA-CoDE for the 10-bar truss.
and nearest neighbor comparison (ODE-NNC) (Pham, 2016), and the modified Teaching-
Learning Based Optimization (TLBO) (Camp & Farshchin, 2014) are also included for the
comparison. Moreover, the required numbers of exact fitness evaluations are given in
Table 8.
Based on the obtained results, it can be pointed out some observations as follows. In
general, the CoDE attains the best results among these algorithms. In detail, the best
weights of the 10-bar truss, the 25-bar truss, and the 72-bar truss found by the CoDE
Figure 8. Convergence histories of the CoDE and the SA-CoDE for the 25-bar truss.
JOURNAL OF INFORMATION AND TELECOMMUNICATION 115
Figure 9. Convergence historyies of the CoDE and the SA-CoDE for the 72-bar truss.
are 5060.854, 545.163, and 379.617 lb, respectively. These values are slightly lower than
the best weights found by the ODE-NNC and the TLBO. In addition, the best results
found by the DE(30) are the same as the CoDE but the standard deviations of the DE
(30) are larger than the corresponding values of the CoDE. It indicates that the CoDE is
more stable than the DE(30). For the same population size of 20, the results obtained
by the DE(20) are worse than those of the CoDE. The reason is that the CoDE employs
three mutation strategies at the same time, leading to the strong ability to search over
the design space.
Nonetheless, the convergence speed of the CoDE is very slow. With the same control
parameters, the CoDE performs 29,960 exact fitness evaluations while the numbers for DE
(20) and DE(30) are 10,000 times and 15,000 times, respectively. In other words, the CoDE
is approximately 3 times slower than the DE(20) and 2 times slower than the DE(30).
Moreover, the SA-CoDE has also shown good results. The optimal weights of the 10-bar
truss and the 25-bar truss found by the SA-CoDE are 5060.854 and 545.163 lb, respectively.
These values are the same as the results of the CoDE and the DE(30). For the 72-bar truss,
the best weight of the SA-CoDE equals 379.623 lb, which is slightly larger than the CoDE
and the DE(30) (379.617 lb). But the advantage of SA-CoDE is its convergence speed. As
shown in Table 8, the SA-CoDE performs only 11,960 evaluations while the required
numbers of exact fitness evaluations of the DE(30) and the CoDE are 15,000 times and 29,960
times, respectively. It means the SA-CoDE is more than 2.5 times faster than the CoDE.
Comparing with previous studies, it is clearly seen that the SA-CoDE is better than the
TLBO in terms of the best weight as well as the computation cost. The optimal weight of
the SA-CoDE is better than the ODE-NNC for the 10-bar truss but worse than ODE-NNC for
the 72-bar truss. However, the numbers of fitness evaluations carried out by the ODE-NNC
are much smaller than those of the SA-CoDE for all three structures.
In summary, the experiment results show that the proposed approach SA-CoDE is a
promising solution. Compared to the CoDE, the SA-CoDE accomplishes the same
results but it saves the computation cost by reducing the number of exact fitness
evaluations.
Other parameters are the same for all cases. The statistical results of four cases are given in
Table 9. It can be found that the computation cost of the first case (iter1 = 10) is the smal-
lest but the optimal weight is larger than those of the remaining cases. When the number
of generations of the first stage reaches 25, the best weight achieves 5060.854 lb. The
result does not improve despite increasing the number of training samples. Overall, the
recommended value of the parameter iter1 is in the range [25,50].
6. Conclusions
The study proposed an approach, called SA-CoDE, to enhance the CoDE algorithm in struc-
tural optimization. By using neural network-based surrogate models, the convergence
speed is increased while still maintaining the powerful search capacity. This is a meaningful
result, especially for computationally expensive optimization problems like structural
optimization. The performance of the SA-CoDE is investigated through three well-known
truss optimization problems. In all problems, the SA-CoDE provides similar results but it
uses fewer exact fitness evaluations than the CoDE. Quantitatively, the SA-CoDE reduces
the number of exact fitness evaluations by about 60% compared to the CoDE.
118 T.-H. NGUYEN AND A.-T. VU
This work carries out surrogate models built using neural networks. In the future,
methods to improve the accuracy of the surrogate model will be studied, and other struc-
tural optimization problems will be investigated.
Acknowledgments
The first author was funded by Vingroup Joint Stock Company and supported by the Domestic Ph.D.
Scholarship Programme of Vingroup Innovation Foundation (VINIF), Vingroup Big Data Institute
(VINBIGDATA) code VINIF.2020.TS.134.
Disclosure statement
No potential conflict of interest was reported by the author(s).
Funding
This research is funded by the Domestic Ph.D. Scholarship Programme of Vingroup Innovation
Foundation (VINIF), code VINIF.2020.TS.134.
Notes on contributors
Tran-Hieu Nguyen is currently a lecturer at National University of Civil Engineering (NUCE), Hanoi,
Vietnam and PhD student in NUCE. He received the MSc degree in 2012 at NUCE. His main research
interests are structural optimization, artificial intelligence and steel structures. Email: hieunt2@nuce.
edu.vn.
Anh-Tuan Vu received his PhD in 2009 at Bauhaus University Weimar, Germany. He is currently the
head of the Department of Steel and Timber Constructions, NUCE. His main research interests are
structural optimization, steel structures, steel and concrete composite structures. Email: tuanva@
nuce.edu.vn.
ORCID
Tran-Hieu Nguyen http://orcid.org/0000-0002-1446-5859
References
Bureerat, S., & Pholdee, N. (2016). Optimal truss sizing using an adaptive differential evolution algor-
ithm. Journal of Computing in Civil Engineering, 30(2), 04015019. https://doi.org/10.1061/
(ASCE)CP.1943-5487.0000487
Cai, J., & Thierauf, G. (1996). Evolution strategies for solving discrete optimization problems.
Advances in Engineering Software, 25(2-3), 177–183. https://doi.org/10.1016/0965-9978
(95)00104-2
Camp, C. V., & Farshchin, M. (2014). Design of space trusses using modified teaching–learning based
optimization. Engineering Structures, 62-63, 87–97. https://doi.org/10.1016/j.engstruct.2014.01.
020
Deb, K. (2000). An efficient constraint handling method for genetic algorithms. Computer Methods in
Applied Mechanics and Engineering, 186(2-4), 311–338. https://doi.org/10.1016/S0045-7825
(99)00389-8
JOURNAL OF INFORMATION AND TELECOMMUNICATION 119
Ho-Huu, V., Nguyen-Thoi, T., Vo-Duy, T., & Nguyen-Trang, T. (2016). An adaptive elitist differential
evolution for optimization of truss structures with discrete design variables. Computers &
Structures, 165, 59–75. https://doi.org/10.1016/j.compstruc.2015.11.014
Jia, G., Wang, Y., Cai, Z., & Jin, Y. (2013). An improved (μ+ λ)-constrained differential evolution for
constrained optimization. Information Sciences, 222, 302–322. https://doi.org/10.1016/j.ins.2012.
01.017
Kim, S. E., Vu, Q. V., Papazafeiropoulos, G., Kong, Z., & Truong, V. H. (2020). Comparison of machine
learning algorithms for regression and classification of ultimate load-carrying capacity of steel
frames. Steel and Composite Structures, 37(2), 193–209. https://doi.org/10.12989/scs.2020.37.2.193
Krempser, E., Bernardino, H. S., Barbosa, H. J., & Lemonge, A. C. (2012). Differential evolution assisted
by surrogate models for structural optimization problems. In Proceedings of the international con-
ference on computational structures technology (CST). Civil-Comp Press (Vol. 49).
Mallipeddi, R., & Lee, M. (2012, June). Surrogate model assisted ensemble differential evolution
algorithm. In 2012 IEEE congress on evolutionary computation (pp. 1–8). IEEE.
Mallipeddi, R., Suganthan, P. N., Pan, Q. K., & Tasgetiren, M. F. (2011). Differential evolution algorithm
with ensemble of parameters and mutation strategies. Applied Soft Computing, 11(2), 1679–1696.
https://doi.org/10.1016/j.asoc.2010.04.024
Nguyen, T. H., & Vu, A. T. (2020). Using neural networks as surrogate models in differential evolution
optimization of truss structures. In International conference on computational collective intelli-
gence (pp. 152–163). Springer, Cham.
Nguyen, T. H., & Vu, A. T. (2021). A comparative study of machine learning algorithms in predicting
the behavior of truss structures. In Kumar R, Quang N.H, Solanki V.J, Cardona M, & Pattnaik P.K
(Eds.), Research in intelligent and computing in engineering (pp. 279–289). Springer, Singapore.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer,
P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M., &
Duchesnay, E. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning
Research, 12(Oct), 2825–2830.
Pham, H. A. (2016). Truss sizing optimization using enhanced differential evolution with opposition-
based mutation and nearest neighbor comparison. Journal of Science and Technology in Civil
Engineering (STCE)-NUCE, 10(5), 3–10.
Rahnamayan, S., Tizhoosh, H. R., & Salama, M. M. (2008). Opposition-based differential evolution. IEEE
Transactions on Evolutionary Computation, 12(1), 64–79. https://doi.org/10.1109/TEVC.2007.894200
Rajeev, S., & Krishnamoorthy, C. S. (1992). Discrete optimization of structures using genetic algor-
ithms. Journal of Structural Engineering, 118(5), 1233–1250. https://doi.org/10.1061/(ASCE)0733-
9445(1992)118:5(1233)
Rosenblatt, F. (1958). The perceptron: A probabilistic model for information storage and organiz-
ation in the brain. Psychological Review, 65(6), 386–408. https://doi.org/10.1037/h0042519
Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986). Learning representations by back-propagating
errors. Nature, 323(6088), 533–536. https://doi.org/10.1038/323533a0
Storn, R., & Price, K. (1997). Differential evolution–a simple and efficient heuristic for global optim-
ization over continuous spaces. Journal of Global Optimization, 11(4), 341–359. https://doi.org/10.
1023/A:1008202821328
Takahama, T., Sakai, S., & Iwane, N. (2006, October). Solving nonlinear constrained optimization pro-
blems by the ϵ constrained differential evolution. In 2006 IEEE International conference on systems,
man and cybernetics (Vol. 3, pp. 2322–2327). IEEE.
Truong, V. H., Vu, Q. V., Thai, H. T., & Ha, M. H. (2020). A robust method for safety evaluation of steel
trusses using Gradient Tree Boosting algorithm. Advances in Engineering Software, 147, 102825.
https://doi.org/10.1016/j.advengsoft.2020.102825
Wang, Y., & Cai, Z. (2011). Constrained evolutionary optimization by means of (μ+ λ)-differential
evolution and improved adaptive trade-off model. Evolutionary Computation, 19(2), 249–285.
https://doi.org/10.1162/EVCO_a_00024
Wang, Y., Cai, Z., & Zhang, Q. (2011). Differential evolution with composite trial vector generation
strategies and control parameters. IEEE Transactions on Evolutionary Computation, 15(1), 55–66.
https://doi.org/10.1109/TEVC.2010.2087271
120 T.-H. NGUYEN AND A.-T. VU
Wang, B. C., Li, H. X., Li, J. P., & Wang, Y. (2019). Composite differential evolution for constrained evol-
utionary optimization. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 49(7), 1482–
1495. https://doi.org/10.1109/TSMC.2018.2807785
Wang, Z., Tang, H., & Li, P. (2009, December). Optimum design of truss structures based on differ-
ential evolution strategy. In 2009 international conference on information engineering and compu-
ter science (pp. 1–5). IEEE.
Zhang, J., & Sanderson, A. C. (2009). JADE: Adaptive differential evolution with optional external
archive. IEEE Transactions on Evolutionary Computation, 13(5), 945–958. https://doi.org/10.1109/
TEVC.2009.2014613