Algorithms 12 00080 PDF
Algorithms 12 00080 PDF
Article
An Improved Squirrel Search Algorithm for Global
Function Optimization
Yanjiao Wang and Tianlin Du *
School of Electrical Engineering, Northeast Electric Power University, Jilin 132012, China;
[email protected]
* Correspondence: [email protected]
Received: 10 March 2019; Accepted: 12 April 2019; Published: 17 April 2019
Abstract: An improved squirrel search algorithm (ISSA) is proposed in this paper. The proposed
algorithm contains two searching methods, one is the jumping search method, and the other is
the progressive search method. The practical method used in the evolutionary process is selected
automatically through the linear regression selection strategy, which enhances the robustness of
squirrel search algorithm (SSA). For the jumping search method, the ‘escape’ operation develops
the search space sufficiently and the ‘death’ operation further explores the developed space, which
balances the development and exploration ability of SSA. Concerning the progressive search method,
the mutation operation fully preserves the current evolutionary information and pays more attention
to maintain the population diversity. Twenty-one benchmark functions are selected to test the
performance of ISSA. The experimental results show that the proposed algorithm can improve the
convergence accuracy, accelerate the convergence speed as well as maintain the population diversity.
The statistical test proves that ISSA has significant advantages compared with SSA. Furthermore,
compared with five other intelligence evolutionary algorithms, the experimental results and statistical
tests also show that ISSA has obvious advantages on convergence accuracy, convergence speed
and robustness.
Keywords: the squirrel search algorithm; the jumping search method; the progressive search method;
linear regression selection strategy
1. Introduction
Optimization is one of the most common problems in the engineering field, and with the
development of new technology, the problems that need to be optimized have gradually turn to large
scale, multi peak and nonlinear approaches. The intelligence evolutionary algorithm is a mature global
optimization method with high robustness and wide applicability. The fact that the evolutionary
process is not constrained by search space and does not require other auxiliary information means
that the intelligence evolutionary algorithm can deal with complex problems effectively, which are
too difficult to be solved by the traditional optimization algorithms [1,2]. The applications of the
intelligence evolutionary algorithms have covered system control, machine design and engineering
planning, for example [3–7].
The intelligence evolutionary algorithms can be divided into the evolutionary heuristic algorithms,
the physical heuristic algorithms and the group heuristic algorithms, according to their inspiration. The
evolutionary heuristic algorithms originate from the genetic evolution process, with the representative
algorithms described as follows: The genetic algorithm imitates Darwin’s theory of natural selection
and finds the optimal solution by selection, crossover and mutation [8]. Similarly, the essence of the
differential evolutionary algorithm is the genetic algorithm based on real coding; the mutation operation
modifies each individual according to the difference vectors of population [9]. In the covariance-matrix
adaptation evolution strategy, the direction of mutation steps of a population is directly described
by the covariance matrix, where the search range of the next generation is increased or decreased
adaptively. The individuals produced by sampling are optimized through the iterative loop [10]. The
clonal selection algorithm is based on the clonal selection theory; the fitness value corresponds to the
cell affinity, and the optimization process imitates the affinity maturation process of cells with low
antigen affinity [11]. Given that human beings have higher survivability because they are good at
observing and drawing experience from others’ habits, the social cognitive optimization algorithm
was proposed, with better solutions being selected by the imitating process and new solutions being
produced by the observing process [12]. Imitating the toning process of musicians, the melody search
is aimed at finding the best melody of continuity, the harmony memory considering rate controls the
search range of each solution (harmony) and the pitch adjusting rate produces a local perturbation
of the new solution [13]. The teaching-learning-based optimization algorithm is proposed through
imitating the teachers’ teaching and the students’ learning process. The teacher is the individual with
the best grade (fitness value) and the other individuals in the population are students. In order to
improve the grades of the whole class, each student studies from the teacher in the teaching stage
and the students learn from each other in learning stage [14]. There are three relationships among
living things: mutualism, commensalism and parasitism; creatures can benefit themselves by any of
these relationships. The symbiotic organisms search was proposed according to this phenomenon.
Each individual interacts with other individuals during the optimization and the better individuals
are retained after each interaction [15]. The mouth brooding fish algorithm simulates the symbiotic
interaction strategies adopted by organisms to survive and propagate in the ecosystem. The proposed
algorithm uses the movement, dispersion and protection behavior of a mouth brooding fish as an
update mode, and the individuals in the algorithm are updated after these three stages to find the best
possible answer [16].
The physical heuristic algorithms are inspired by physical phenomena, with the representative
algorithms as follows: The freedom of molecules increases after a solid melts, and the temperature
needs to drop slowly to return to stable solids with minimum energy. Simulated annealing takes the
fitness value as the energy of the solid, with the energy decreasing gradually with the optimization
proceeding and the optimal solution being found [17]. The gravitational search algorithm is based on
the law of universal gravitation—for each individual the fitness value represents its resultant force
produced by all the individuals in the population [18]. The magnetic optimization algorithm is inspired
by the theory of magnetic field, where the resultant forces of individuals are changed by the field
strength and the distance among individuals. The acceleration, the velocity and positions of individuals
are also updated, with the individuals reaching the optimum values gradually [19]. Considering the
refraction that occurs when light travels from a light scattering medium to a denser medium, the ray
optimization algorithm was proposed. For each individual, the normal vector is determined by its
optimal solution and the global optimal solution; the optimal solution can be found with the exit rays
close to normal [20]. The kinetic energy of gas molecules takes the energy of gas as the fitness value.
When the pressure remains unchangeable and the temperature decreases, the molecules gradually
accumulate to the position where the temperature is the lowest and the kinetic energy is the smallest in
the container [21]. Inspired by the physical phenomenon of water evaporation, the water evaporation
optimization algorithm was proposed. The factors that affect the water evaporation rate are taken
as the fitness values. According to the water evaporation rate model, the evaporation probability
matrix was considered as the individual renewal probability. Considering that the aggregated forms
of water molecules are different, the algorithm is divided into a monolayer evaporation phase in the
early evolutionary stage and a droplet evaporation phase in the later evolutionary stage [22]. The
lightning attachment procedure optimization algorithm simulates the lightning formation process,
which takes the test points between cloud and ground as individuals and the corresponding electrical
fields represent fitness values. The three evolutionary operations—downward pilot, upward pilot
Algorithms 2019, 12, 80 3 of 29
and branch fading—imitate the downward leader movement, the upward leader propagation and the
discharge of lightning, respectively [23].
The group heuristic algorithms mainly simulate biological habits in nature. The representative
algorithms are as follows: The ant colony optimization algorithm was proposed according to the
way that ants leave pheromones on their path during movement, with better paths having more
pheromones, and thus better paths have greater possibilities to be chosen by ants. As a result, more
and more pheromones will be left on those paths, and the optimal solution will be found with the
increasing concentration of pheromones [24]. The particle swarm optimization algorithm is inspired by
the behavior of birds seeking food. For each individual, the position is updated by its current speed, its
optimal position and the global best position [25]. The artificial bee colony algorithm was proposed by
imitating honeybee foraging behavior. The whole population is divided into three groups: the leading
bees, the following bees and the detecting bees. The leading bees are responsible for producing a new
honey source, while the following bees search greedily near better honey sources. If the quality of the
honey source remains unchanged after many iterations, the leading bees will change to detecting bees
and continue to search for a high quality honey source [26]. The social spider optimization regards the
whole search space as the spiders’ attached web, with the spiders’ positions as the possible solutions of
the optimization problem, and the corresponding weights representing the fitness values of individuals.
Female and male subpopulations produce offspring through their respective cooperation and mating
behavior [27]. The selfish herd theory proves that when animals encounter predators, each individual
increases its survival possibilities by aggregating with other individuals in the herd, whether this
approach affects the survival probability of other individuals or not. According to this theory, the
selfish herd optimizer was proposed, wherein each individual updates the location in this way to
obtain a greater probability of survival [28]. Inspired by the foraging process of hummingbirds, the
hummingbirds optimization algorithm was proposed. The hummingbird can search according to its
cognitive behavior without interacting with other individuals in a self-searching phase. In addition to
searching through experience, hummingbirds can also search by using various dominant individuals
as guidance information in a guided-search phase, with the two phases cooperating to promote the
population evolution [29].
A large number of experimental results show that the intelligence optimization algorithms can
obtain, exact or approximate an optimal solution to large-scale optimization problems in a limited time
frame. However, there are also disadvantages, such as the convergence speed being not fast enough
and easily falling into the local optimal. Therefore, scholars have put forward various new intelligence
evolutionary algorithms.
In 2018, the squirrel search algorithm (SSA) [30] was proposed by Jain M. The algorithm imitates
the dynamic jumping strategies and the gliding characters of flying squirrels. The mathematical
model mainly consists of the location of a food source and the appearance of predators. The whole
optimization process includes the summer phase and the winter phase. However, similar to other
intelligent evolutionary algorithms, SSA also has some shortcomings, such as low convergence accuracy
and slow convergence speed [31,32]. According to SSA, the single winter search method of the global
search ability is not enough, which makes the algorithm easily fall into local optimal. Furthermore, the
random summer search method decreases the convergence speed, and the convergence precision is
also reduced. In order to improve the convergence precision and the convergence speed, this paper
proposed an improved squirrel search algorithm (ISSA). The proposed algorithm includes the jumping
search method and the progressive search method. When the squirrels meet with predators, the
‘escape’ and ‘death’ operations are introduced into the jumping search method and the ‘mutation’
operation is introduced into the progressive search method. ISSA also chooses the suitable search
method through the linear regression selection strategy during the optimization process. Twenty-one
benchmark functions are used to evaluate the performance of the proposed algorithm. The experiments
contain three parts: the influence of the parameter on ISSA, the comparison of the proposed methods
and SSA and the comparison of ISSA and five other improved evolutionary algorithms.
Algorithms 2019, 12, 80 4 of 29
The remaining sections are arranged as follows: Section 2 reviews the basic SSA. Section 3 presents
the proposed ISSA. The experiments and results analysis are reported in Section 4. Section 5 concludes
this paper.
FSi represents the i-th individual, (i = 1 . . . N); rand is a random number between 0 and 1; D is the
dimension of the problem.
hg is the constant valued 8; sf is the constant valued 18; tan(ϕ) represents the gliding angle which can
be calculated by Formula (5):
D
tan(ϕ) = (5)
L
D is the drag force and L is the lift force which can be calculated by calculated by Formulas (6) and (7),
respectively:
1
D= (6)
2ρV 2 SCD
1
L= (7)
2ρV 2 SCL
ρ, V, S and CD are all the constants which are equal to 1.204 kg m−3 , 5.25 ms−1 , 154 cm2 and 0.6,
respectively; CL is a random number between 0.675 and 1.5.
10e−6
Smin = (9)
(365)t/(T/2.5)
T is the maximum number of iterations, if Stc < Smin , winter is over and the season turns to summer,
otherwise, the season is unchanged. When the season turns to summer, all the individuals who glide to
Fh stay at the updated location, and all the individuals who glide to Fa and do not meet with predators
relocate their positions by Formula (10):
+1
FStinew = FSL + Le0 vy(n) × (FSU − FSL ) (10)
Le0 vy is the random walk model whose step obey the Le0 vy distribution and can be calculated by
Formula (11):
ra × σ
Le0 vy(x) = 0.01 × 1
(11)
|rb | β
β is the constant valued 1.5; σ can calculated by Formula (12):
1
πβ β
Γ(1+β)×sin( 2 )
σ=( β−1 )
1+β
Γ( 2 )×β×2
( 2 ) (12)
whereΓ(x) = (x − 1)!
Start
Population
Initialization
N
W inter is over?
Y
The season turns to summer and
random ly update the
corres ponding individuals
N
t<T?
Y
End
Figure
Figure 1. The
1. The procedureof
procedure of the
the standard
standard squirrel
squirrelsearch algorithm
search (SSA).
algorithm (SSA).
3. The
3. The improved
improved Squirrel
Squirrel SearchAlgorithm
Search Algorithm
3.1. Motivation
3.1. Motivation
A large
A large number
number of experiments
of experiments have
have proventhat
proven thatdifferent
differentevolutionary
evolutionary strategies
strategiesare
aresuitable
suitable for
for different problems, and also that the requirements are also different with the development of
different problems, and also that the requirements are also different with the development of evolution.
evolution. In the early stage of optimization, individuals are distributed dispersedly in the search
In the early stage of optimization, individuals are distributed dispersedly in the search space, and
space, and there are still large distances among the individuals with better fitness values. Thus it is
thereimportant
are still large distances
to maintain the among theofindividuals
diversity withtobetter
the population developfitness values.space
the search Thussufficiently.
it is important
to maintain
Meanwhile, the convergence speed should be improved as well. In the later stage of optimization, the
the diversity of the population to develop the search space sufficiently. Meanwhile,
convergence speed
the difference should
among the be improved
individuals areas well. In the
increasingly laterand
shorter, stage
thusofthe
optimization,
main work is the difference
to search
among the individuals
around are increasingly
the elite individuals to improveshorter, and thus
the convergence the main
speed. worktoisthis,
In addition to search
in orderaround the elite
to prevent
the algorithm
individuals from falling
to improve into a localspeed.
the convergence optimal,Inthe diversity
addition of theinpopulation
to this, also needs
order to prevent the to be
algorithm
supplemented.
from falling into a local optimal, the diversity of the population also needs to be supplemented.
Considering the analysis above, an improved squirrel search algorithm (ISSA) is proposed in
Considering the analysis above, an improved squirrel search algorithm (ISSA) is proposed in this
this paper to improve the performance and the robustness of SSA. The proposed algorithm includes
paper to improve the performance and the robustness of SSA. The proposed algorithm includes the
the jumping search method and the progressive search method, both of having an independent
jumping search method and the progressive search method,St both Smin
of having an independent winter
winter search strategy for the early evolutionary stage when t ≥ Sc and summer search strategy for
search strategy for the early evolutionary stage when S c min and summer search strategy for the
S ct S min
later the later evolutionary
evolutionary stage when
stage when Stc < Smin . Algorithm
. Algorithm 1 shows
1 shows thedetailed
the detailedsteps
steps of
of ISSA:
ISSA:
Algorithm
Algorithm 1. Pseudo
1. Pseudo CodeCode of ISSA
of ISSA
Input: pop
Input: pop
Output: fbest (fbest is the best fitness value optimized by the algorithm)
Output: fbest (fbest is the best fitness value optimized by the algorithm)
for t = 1 to T (T is the total generation of the algorithm to be executed)
for t = 1 to T (T is the total generation of the algorithm to be executed)
evaluate the fitness values of the population
evaluate the fitness values of the population
update the population through the jumping search method introduced in 3.2
update theifpopulation
t == T/n (nthrough the substages
is the total jumping search
of the method introduced indetails
whole optimization, Sectionin3.2
3.4)
if t == calculate
T/n (n is the the corresponding linear regression equations introduced in 3.43.4)
total substages of the whole optimization, details in Section
calculate
if twothe or corresponding
more calculatedlinear regression
slopes equations introduced in Section 3.4
are positive
if two or more calculated slopes are positive
continue optimizing through the progressive search method introduced in 3.3
continueelseoptimizing through the progressive search method introduced in Section 3.3
else continue optimizing through the jumping search method mentioned in 3.2
continue
endoptimizing through the jumping search method mentioned in Section 3.2
end
end
end
Algorithms 2019, 12, 80 7 of 29
If r < Pdp , FSi is endangered, FSi is considered to be dead and generates a new one by Formula (14):
FStj+1 = FSnew
j × (0.5 + rand) (17)
In order to satisfy the corresponding requirement, a new summer search strategy was proposed.
The proposed strategy searches around the elite individual carefully and supplements the population
diversity to make up the disadvantages introduced above. The details are as follows:
If r ≥ Pdp , FSi is safe, the position is updated by Formula (18):
If r < Pdp , FSi is endangered, FSi is considered to be dead and a new one will be generated by
Formula (19). Furthermore, all the threatened individuals are considered to be dead and the new ones
will be generated by Formula (20):
In the formula above, N (0,1) is a random number which obeys the standard normal distribution.
Fh is the best individual found so far. Formula (18) takes Fh as the base vector and takes the
differential vector between Fh and FSi as the disturbance. Due to the fact that differences among the
individuals in summer (Stc < Smin ) are smaller than those in winter (Stc ≥ Smin ), the essence of Formula
(18) is to search finely around Fh and retain the current evolutionary information. In Formula (19), the
random number is generated by 0.5(N(0,1)) scattered in [0.5−2 ,0.52 ] but close to 1 in greater possibilities,
thus the individuals generated by Formula (19) are more similar to FSi , while the individuals generated
by Formula (20) distribute in [0.5Fh ,1.5Fh ] uniformly. Therefore, the search space of Formulas (19) and
(20) are smaller than the random relocation shown in Formula (10). In addition, Formula (19) pays
more attention to retain the current evolutionary information and Formula (20) pays more attention to
developing the search space.
Le0 vy(x) is calculated by Formulas (11) and (12), which makes the individuals search in a short distance
with greater possibilities and search in a long distance occasionally.
If r < Pdp , FSi is endangered, select a dimension randomly and mutate it in the range of L and U.
Algorithms 2019, 12, 80 10 of 29
substage is finished. There are three linear regression equations needed to be calculated: the best
fitness value of the first half substage, the second half substage and the whole substage. If two or
more equations’ slopes are positive, the best fitness value may fluctuate or become larger, that is,
the population diversity
Algorithms2019,
2019,12,
12,xxFOR
needs
FORPEER
PEERREVIEW
to be supplemented. The jumping search method is not suitable
REVIEW 11 of
of 31
31
for
Algorithms 11
the current problem. Therefore, the progressive search method is selected to finish the evolution.
currentthe
Otherwise,
current problem.
population
problem. Therefore, the progressive
diversity
Therefore, the progressive
is abundant search
enough,
search method
method andisiswill
selected
continue
selected to finish
to finish the evolution.
evolving
the evolution.
through the
jumpingOtherwise,
search the
Otherwise, the population
method population diversity
to converge
diversity isis abundant
faster. abundant enough,
Figure 2enough, and will
is the illustration
and will continue
continue evolving
of the linear
evolving throughselection
regression
through the
the
jumping search method to converge faster. Figure 2 is the illustration
strategy introduced above. Taking n = 10 and t = 0.2 T as an example, calculate the linear regression
jumping search method to converge faster. Figure 2 is the illustration of
of the
the linear
linear regression
regression selection
selection
strategyintroduced
introducedabove. above.Taking
Takingnn==10 10and
andtt==0.20.2TTas
strategy
equations of the three regions: [0.1 T, 0.15 T], [0.15 T, as
0.2anexample,
an T]example,
and [0.1 calculatethe
calculate
T, 0.2 T].thelinear
linearregression
It can regression
be seen from
equations of
equations of the
the three
three regions:
regions: [0.1
[0.1 T,
T, 0.15
0.15 T],
T], [0.15
[0.15 T,T, 0.2
0.2 T]
T] and
and [0.1
[0.1 T,T, 0.2
0.2 T].
T]. ItItcan
can bebe seen
seen from
from
Figure 2 that the slope of ab is negative, and the slopes of bc and ac are positive, which means that the
Figure22that
Figure thatthe
theslope
slopeofofababisisnegative,
negative,and andthe
theslopes
slopesof ofbcbcand
andacacare
arepositive,
positive,which
whichmeans
meansthat that
best fitness
thebest
value
bestfitness
will
fitnessvalue
fluctuate
valuewill
or become
willfluctuate
fluctuateor orbecome
larger,
becomelarger,
therefore,
larger,therefore,
select
therefore,select
the
selectthe
progressive
theprogressive
progressivesearch
searchmethod
method to
searchmethod
the
finishtothe evolution. Figure 3 is the procedure
finish the evolution. Figure 3 is the procedure of ISSA. of ISSA.
to finish the evolution. Figure 3 is the procedure of ISSA.
The image
The image of
of liner
liner regression
regression between
between bb and
and cc
cc
aa
bb
Figure
Figure 2.The
The
2. 2.
Figure The illustrationof
illustration
illustration ofthe
of thelinear
linearregression
linear regression selection
regressionselection
selectionstrategy.
strategy.
strategy.
Start
Start
Populationinitialization
Population initialization
Updatethe
Update thepopulation
populationthrough
through
thejumping
the jumpingsearch
searchmethod
method
NN t/(0.1T)==0?
t/(0.1T)==0?
YY
Calculatethe
Calculate theequation
equationof
oflinear
linearregression
regressionof
ofthe
the
correspondingphase
corresponding phase
NN Twoor
Two ormore
moreslopes
slopes
arepositive?
are positive?
YY
Updatethe
Update thepopulation
populationthrough
through Updatethe
Update thepopulation
populationthrough
through
thejumping
the jumpingsearch
searchmethod
method theprogressive
the progressivesearch
searchmethod
method
NN
NN Meetthe
Meet theterminating
terminating Meetthe
Meet theterminating
terminating
condition?
condition? condition?
condition?
YY YY
End
End
n
F
0 5 10 15 20
F2 0±0 0±0 0±0 0±0 0±0
6.2502e−04 ± 1.7129e−10 ± 3.5323e−08 ± 7.7188e−07 ± 7.5656e−07 ±
F6
1.9021e−04 9.2634e−11 1.2626e−09 2.8698e−08 1.9151e−08
F10 −7.6632 ± 0.7985 −9.4261 ± 0.0680 −9.2260 ± 0.7780 −8.1919 ± 1.3054 −8.0299 ± 1.1781
F16 6.2909e−163 ± 0 7.4432e−162 ± 0 1.0302e−164 ± 0 2.0455e−165 ± 0 7.1391e−164 ± 0
F17 0±0 0±0 0±0 0±0 0±0
1.8504e−15 ± 4.4095e−08 ± 2.1945e−15 ± 2.9976e−15 ± 3.1826e−15 ±
F18
1.0135e−15 1.9301e−07 8.2481e−16 6.5204e−15 7.1639e−15
4.3.2. Comparison
4.3.2. Comparisonof
ofPopulation
PopulationDiversity
Diversityon
onFour
Four Methods
Methods
In order
In order to to compare
compare the the population
population diversity
diversityofof the
the four
fourmethods
methodsintuitively,
intuitively, all
all the
the population
population
sizes are 30. Table 3 is the comparison of the individuals’ distribution on the 2-dimensional
sizes are 30. Table 3 is the comparison of the individuals’ distribution on the 2-dimensional unimodal unimodal
function F2
function F2 and
and the
the 2-dimensional
2-dimensional multimodal
multimodal function
functionF5 F5 when
whenthethe convergence
convergence accuracy
accuracyisis up
up to to
10 −6 , 10 −8 and 10 −10 . The points in the figures refer to the individuals’ positions, and the blue, green,
10 , 10 and 10 . The points in the figures refer to the individuals’ positions, and the blue, green,
−6 −8 −10
black and
black and red
red points
points refer
refer toto SSA,
SSA, the
the jumping
jumping search
search method,
method, thethe progressive
progressive search
search method
method and and
ISSA, respectively.
ISSA, respectively. The The abscissa
abscissaand andthe
theordinate
ordinateof ofeach
eachfigure
figureare
arethe
thesearch
searchrange
rangeforforthe
thefunction;
function;
for F2 the search range is from −10
for F2 the search range is from −10 to 10 and for F5 the search range is from −100 to 100. TheThe
to 10 and for F5 the search range is from −100 to 100. titletitle
of
of each
each figurefigure is the
is the name nameof F2oforF2F5.
orTable
F5. Table 4 shows
4 shows the variance
the variance of theof the population’s
population’s fitnessfitness
valuesvalues
when
when the algorithms converge to about −4 of benchmark functions. The functions contain the high
the algorithms converge to about 10−4 10of benchmark functions. The functions contain the high
dimensional unimodal functions F11 and
dimensional unimodal functions F11 and F12 and the high F12 and the high dimensional
dimensional multimodal
multimodal functions F20 F20
functions and
F21. F21.
and The number
The number beforebefore
‘±’ is the variance’s
‘±’ is mean and
the variance’s mean theand
number after ‘±’after
the number is the‘±’
variance’s standard
is the variance’s
deviation. ‘/’ represents the algorithm failure to converge to 10−4 after evaluating 30,000 times.
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms 2019,
2019,
2019,
2019,
2019,
Algorithms 12,
12,
2019,
2019, 12,
12,
12,xx
12,
x
12, xxxFOR
FOR
FOR FOR
FOR PEER
PEER
PEER
PEER
PEER REVIEW
REVIEW
REVIEW
REVIEW
REVIEW 15 15
1515
15
of of
ofof
of
31 31
313131
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms 2019,
2019,
2019,
2019,
Algorithms
Algorithms 12,
2019,
2019,
2019, 12,
12,
12,
12,
xx
12,
12,xxxxFOR
x
x FOR
FOR
FOR
xFOR
FOR
FOR
FORPEER
PEER
PEER
PEER
PEER
PEER
PEER
PEERREVIEW
REVIEW
REVIEW
REVIEW
REVIEW
REVIEW
REVIEW
REVIEW 1515
15
15
15
15
15of
15of
of
of
of
of
of31
of31
31
31
31
31
3131
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms
Algorithms2019,
2019,
2019,
2019,
2019,
2019,12,
12,
12,
12,
2019, x FOR
12,
12,
12, xxx
xFOR xFOR
FOR
FORPEER
FOR
FOR PEER
PEER
PEER
FOR REVIEW
PEER
PEER
PEERREVIEW
REVIEW
REVIEW
REVIEW
REVIEW
PEER REVIEW
REVIEW 15
1515
15of
15
ofof
15
15
15of31
of
of31
of
of
31 31
31
31
31
31
standard
standard
standard
standard
standard
standard
standard deviation.
deviation.
deviation.
deviation.
deviation.
deviation.
deviation.‘/’ ‘/’
‘/’‘/’
‘/’‘/’ represents
represents
represents
represents
represents
‘/’ representsthe
the
represents the
the
the
the
the algorithm
algorithm
algorithm
algorithm
algorithm
algorithm
algorithm failure
failure
failure
failure
failure
failure
failureto to
toto
toto
to converge
converge
converge
converge
converge
converge
converge to to
toto
toto
to1010
1010
10
−4−4
10
10
−4−4
−4
−4 after
after
after
after
after
−4
−4 after
after evaluating
evaluating
evaluating
evaluating
evaluating
evaluating
evaluating 30,000
30,000
30,000
30,000
30,000
30,000
30,000
standard
standard
standard
standard
standard
standard deviation.
deviation.
deviation. ‘/’
deviation.
deviation.
standard deviation. ‘/’
‘/’
deviation. ‘/’
‘/’represents
represents
represents
represents
represents the
the
thethe
thealgorithm
algorithm
algorithm
algorithm
algorithm failure
failure
failure to
failure
failure to to
to
toconverge
converge
converge to
converge
converge to to
to
10 10
10
to 10
−4−4
10 −4 −4after
after
after
−4
after evaluating
evaluating
evaluating
after
evaluating 30,000
30,000
30,000
evaluating
30,000 30,000
standard ‘/’‘/’
deviation. ‘/’
‘/’ represents
represents
represents the
the
the algorithm
algorithm
algorithm failure
failure
failuretototo
to converge
converge
convergetototo
to1010
10
10 −4after
after
after evaluating
evaluating
evaluating 30,000
30,000
30,000
−4 −4−4
−4
standard
standard
standard
times.
standard
times.
times.
times.
Algorithms
times.
times.
times. deviation.
deviation.
deviation.
2019,deviation.
12, 80 ‘/’
‘/’
‘/’represents
represents
represents
representsthe
thethe
thealgorithm
algorithm
algorithm
algorithm failure
failure
failure
failure to
to
toconverge
converge
converge
converge to
to
to 10
10
10
−4−4
−4 after
after
after evaluating
evaluating
after evaluating 30,000
30,000
evaluating
15 30,000
30,000
of 29
times.
times.
times.
times.
times.
times.
times.
times.
times.
times.
times.
times.
Table
Table
Table
Table
Table
Table
Table3.
Table 3.
3.3.
3.
The
The
3.
3.
3. The
The
The
The
The
The comparison
comparison
comparison
comparison
comparison
comparison
comparison
comparisonof of
ofof
ofof
the
the
of
of the
the
the
the
the
the individuals’
individuals’
individuals’
individuals’
individuals’
individuals’ distribution.
distribution.
distribution.
distribution.
distribution.
individuals’
individuals’ distribution.
distribution.
distribution.
Table
Table
Table
Table
Table 3.
3.
Table
3.
Table
Table 3.3.The
3.
The
The
TheThe comparison
comparison
comparison
comparison
The
comparison of
of
comparison
of of
ofofthe
the
the
the the individuals’
individuals’
individuals’
individuals’
the distribution.
distribution.
distribution.
distribution.
individuals’
individuals’ distribution.
distribution.
Table
Table
Table 3.3.
Table
Table 3.
3.The
3.
3.
3.
The The
The
The
The
Thecomparison
comparison
comparison
comparison
comparison
comparisonofof
comparison of
ofof
ofofthe
thethethe
the
the individuals’
individuals’
individuals’
individuals’
individuals’
the individuals’ distribution.
distribution.
distribution.
distribution.
distribution.
individuals’ distribution.
distribution.
Fun
Fun
Fun
Fun
Fun
Fun
Fun Method
Method
Method
Method
Method
Method
Method
Fun
Fun
Fun
FunFun
Fun
Fun
Fun Method
Method
Method
Method
Method
Method
Method
Method
Method
Fun Fun
Fun
Fun
Accuracy
Fun
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy SSA
SSA
SSA
SSA
SSA
SSA
SSA Jumping
Jumping
Jumping
Jumping
Jumping
Jumping
Jumping Search
Search
Search
Search
Search
SearchMethod
Method
SearchMethod
Method
Progressive
Progressive
Progressive
Progressive
Progressive
Progressive
Progressive Search
Search
Search
Search
Search
Search
Search ISSA
ISSA
ISSA
ISSAISSA
ISSA
ISSA
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy
Accuracy SSA
SSA
SSA
SSA
SSA
SSA
SSA
SSA
SSA Jumping
Jumping
Jumping
Jumping
Jumping
Jumping Search
Search
Search
Jumping Search
Search
Jumping
JumpingSearch
Search
Search
Search Progressive
Progressive
Progressive
Progressive
Progressive Search
Search
Search
Progressive Search
Search
Progressive
Progressive
Progressive Search
Search
Search
Search ISSA
ISSA
ISSA
ISSAISSA
ISSA
ISSA
ISSA
ISSA
Accuracy
Accuracy
Accuracy
Accuracy SSA
SSA
SSA
SSA Jumping
Jumping
Jumping Search
Search
Jumping Search Progressive
Search Progressive
Progressive Search
Search
Progressive Search
Search ISSA
ISSA
ISSA
ISSA
10 10
1010
10
10
10
−6−6
−6
−6−6−6
−6
−6
10
10
1010
−6−6
1010
10
10
−6
−6
−6
−6
−6
−6
−6
−6
1010
10
10
10−6
−6 −6
−6
F2 10
1010
10 −8
−8−8−8
F2F2
F2
F2F2
F2
F2 10
1010
1010
−8−8
10
−8−8
10
−8
−8
−8
−8
−8 −8
F2
F2
F2F2
F2F2
F2
F2 10
1010
10
10
10
−8
10
10
10
−8
−8
−8
−8
−8
−8−8
F2
F2F2
F2
10 10
1010
10
10
−10
10
−10
−10
−10
−10
−10
−10
−10
−10
10
10
1010
10
−10
1010
10
10
−10
−10
−10
−10
−10−10
−10
−10
1010
10
10
−10−10
−10
−10
10 10
1010
10
10
10
−6−6
−6
−6−6−6
−6
−6
−6
10
10
1010
−6−6
10
1010
10
10
−6
−6
−6
−6 −6
−6
−6
1010
10
10−6
−6 −6
−6
F5 10
1010
10 −8
−8−8
−8−8
F5
F5F5
F5
F5
F5
F5 1010
10
10
10 10
−8−8
1010
−8
−8−8
−8
−8
−8 −8
F5
F5
F5F5
F5F5
F5
F5 1010
10
10
−8
1010
10
−8−8
−8
10−8−8
−8 −8
F5
F5F5
F5
10 −10
10 10
1010
10
10
−10
10
−10
−10
−10
−10
−10
−10
−10
10
10
1010
−10
10−10
−10
−10−10
1010
10
10
−10−10
−10
−10
10
10
10
−10−10
−10
−10
Algorithms 2019, 12, 80 16 of 29
Method
Name
Jumping Search
SSA Progressive Search Method ISSA
Method
1.2837e+05 ± 2.5981e−04 ±
F11 1.1115e+04 ± 3.3201e+03 0.0017 ± 0.0015
7.3449e+04 2.4867e−04
5.7955e+04 ± 7.1891e−05 ±
F12 1.3107e+03 ± 543.9195 7.2366e−04 ± 6.0848e−04
2.1202e+04 4.6142e−05
6.0584e−05 ±
F20 / 1.2767 ± 0.2946 2.2921e−04 ± 1.7366e−04
2.5350e−05
8.5775e−05 ±
F21 / 7.2621e+04 ± 6.0527e+04 6.2753e−04 ± 4.7156e−04
8.4886e−05
It can be seen from Table 3 that SSA has the most dispersed individual distribution because of the
individuals’ re-initialization. For the three proposed strategies mentioned in this paper, individuals of
the progressive search method distribute most dispersedly. ISSA takes second place, and the jumping
search method has the most concentrated distribution. The progressive search method mutates
individuals on a certain dimension, which will make more individuals distribute on lines x1 = 0 and
x2 = 0. As a result, the benchmark functions have a great chance to converge to the optimal value.
SSA re-initializes the whole individual, which makes the individuals distribute in the search space
irregularly. From the analysis above, the progressive search method has the best population diversity
while the jumping search method performs worst when the population evaluates to the fixed accuracy.
Data in Table 4 show that the variances of SSA and the progressive search method are much larger
than the variances of the jumping method and ISSA. Excessive population diversity slows down the
convergence speed and some functions cannot converge to the fixed accuracy. Meanwhile, for the
population diversity of the three proposed strategies in this paper, the progressive search method
performs best, ISSA takes second place and the jumping search method performs worst.
Method
Name
Jumping Search Progressive
SSA ISSA
Method Search Method
−1 ± 1.0909e−16
F1 −1 ± 0 (=) −1 ± 0 (=) −1 ± 3.9171e−16
(=)
4.4212e−21 ± 9.7986e−149 ±
F2 0 ± 0 (=) 0±0
1.5032e−20 (−) 5.3669e−148 (−)
7.4015e−18 ±
F3 0 ± 0 (=) 0 ± 0 (=) 0±0
4.0540e−17 (−)
1.0191e−20 ±
F4 0 ± 0 (=) 0 ± 0 (=) 0±0
2.9625e−20 (−)
1.2019e−20 ±
F5 0 ± 0 (=) 0 ± 0 (=) 0±0
4.4285e−20 (−)
1.4402e−18 ± 2.6295e−33 ± 4.9331e−12 ±
F6 0.0202 ± 0.0238 (−)
3.0624e−18 (+) 1.4403e−32 (+) 1.6068e−11
−1.8468 ± 0.0838 −1.7911 ± 0.0116 −1.8013 ± −1.8013 ±
F7
(−) (−) 6.8344e−16 (=) 8.5739e−16
F8 0 ± 0 (=) 0 ± 0 (=) 0 ± 0 (=) 0±0
−4.6459 ± −4.4543 ± 0.1818 −4.6808 ± 0.1336
F9 −4.6617 ± 0.0304
2.7201e−15 (−) (−) (+)
−9.5333 ± 0.2598 −8.2037 ± 0.3744
F10 −9.6602 ± 0 (+) −9.5882 ± 0.0891
(−) (−)
2.7472e−13 ± 2.1318e−16 ±
F11 0 ± 0 (=) 0±0
1.1640e−12 (−) 1.1676e−15 (−)
8.0478e−13 ± 1.9301e−15 ±
F12 0 ± 0 (=) 0±0
4.3612e−12 (−) 1.0572e−14 (−)
2.0435e−10 ± 3.5232e−11 ±
F13 0 ± 0 (=) 0±0
1.0129e−09 (−) 1.9297e−10 (−)
F14 0.5472 ± 2.7161 (−) 0 ± 0 (=) 0.0865 ± 0.4149 (−) 0±0
7.2194e−08 ±
F15 0 ± 0 (=) 0.0241 ± 0.1321 (−) 0±0
6.4247e−08 (−)
1.4322e−11 ±
F16 0.0978 ± 0.4017 (−) 0 ± 0 (=) 0±0
7.8443e−11 (−)
9.7812e+03 ± 2.3545e−15 ±
F17 0 ± 0 (=) 0±0
3.5168e+04 (−) 1.2896e−14 (−)
1.9614e−16 ±
F18 0.6323 ± 0.3594 (−) 0 ± 0 (+) 0.0129 ± 0.0650 (−)
3.6843e−16
F19 0.1099 ± 0.0548 (−) 0 ± 0 (=) 0.0233 ± 0.0773 (−) 0±0
9.1222e−13 ±
F20 0.0025 ± 0.0109 (−) 0 ± 0 (=) 0±0
4.9964e−12 (−)
7.3093e−05 ±
F21 8.6605 ± 38.6332 (−) 0 ± 0 (=) 0±0
2.2255e−05 (−)
Table 5 shows that the convergence precision of the jumping search method, the progressive
method and the ISSA are all obviously better than SSA. Meanwhile, compared with ISSA, the numbers
of benchmark functions with the better mean, the worse mean and the equivalent mean of SSA are
1, 18, and 2, respectively; the corresponding numbers of the jumping search method are 1, 4 and 16,
respectively; the corresponding numbers of the progressive search method are 3, 12 and 6, respectively.
In order to compare the differences of each method, a Friedman test was taken to check the data in
Table 5 [36]. The specific process is shown below:
The Friedman test ranks the algorithms for each data set separately; k refers to the number of
algorithms and n refers to the number of data sets of each algorithm. The results are shown in Table 6,
k
and is calculated as follows: χ3r = nk(12 R2j − 3n(k + 1) = 21×4×12(4+1) (752 + 442 + 512 + 402 ) − 3 ×
P
k +1)
j=1
21 × (4 + 1) = 21.0571; α = 0.05, df = 4 − 1 = 3 at the 5% significant level and χ30.05 = 7.81 < 21.0571
Algorithms 2019, 12, 80 18 of 29
according to the Chi-square distribution table. Therefore, the four methods are considered to have
significant differences at the 5% significance level.
Rank
Function
SSA Jumping Search Progressive Search ISSA
F1 1.5 3 1.5 4
F2 4 1.5 3 1.5
F3 4 2 2 2
F4 4 2 2 2
F5 4 2 2 2
F6 2 4 1 3
F7 4 3 1 2
F8 2.5 2.5 2.5 2.5
F9 3 4 1 2
F10 3 4 1 2
F11 4 1.5 3 1.5
F12 4 1.5 3 1.5
F13 4 1.5 3 1.5
F14 4 1.5 3 1.5
F15 3 1.5 4 1.5
F16 4 1.5 3 1.5
F17 4 1.5 3 1.5
F18 4 1 3 2
F19 4 1.5 3 1.5
F20 4 1.5 3 1.5
F21 4 1.5 3 1.5
Total rank 75 44 51 40
Average rank 3.5714 2.0952 2.4286 1.9048
Sort 1 3 2 4
To further compare the performance of the four methods, assuming that the convergence
performance of ISSA is better than the other three methods, a Holm test was carried out and the results
are shown in Table 7:
It can be seen from Table 7 that P1 < α(k − 1), P2 > α(k − 2), P3 > α(k − 3). The original hypothesis
is rejected at the 5% significance level. Therefore, compared with SSA, ISSA has significantly better
performance. ISSA has a smaller average rank, though it does not outperform the progressive search
method and the jumping search method.
In conclusion, the jumping search method has the best convergence speed and the progressive
search method performs best on maintaining the population diversity. Compared with SSA, both have
obvious advantages in convergence accuracy. ISSA combines the two methods together, improves the
convergence speed and the convergence accuracy and maintains the population diversity as well. ISSA
can find the global optimal of more benchmark functions and has the best comprehensive performance.
Algorithms 2019, 12, 80 19 of 29
Table 8 shows the optimization results of low dimensional functions (F1–F10) and Table 11 shows
the ones of high dimensional functions (F11–F21). Best, Worst, Mean and SD respectively represent the
best fitness value, the worst fitness value, the mean fitness value and the standard deviation obtained
by 30 independent executions. R represents the times of the algorithm converges to the appointed
precision. The appointed precision is 10−8 for the benchmark functions, whose optimal is 0. For the
benchmark functions F1, F7, F9 and F10 whose optimal is not equal to 0, the appointed precision is
−0.6, −1.6, −3.6 and −8.6 respectively.
Table 8. Cont.
It can be seen from Table 8 that MDE can converge to the optimal on F1, F3, F4, F5, F6, F7 and
F8, but the performances of F7 and F8 are not stable enough; distABC only converges to the optimal
stably on F4 and has a certain probability to converge to the optimal on F3 and F5; IGSA/PSO has no
stable convergence performance on all the functions but still has a certain probability to converge to
the optimal on F1, F3, F4 and F5; ADN-RSN-PSO has the worst performance without any function
convergences to the optimal; PSO-GWO can converge stably to the optimal on F3, F4, F5 and F8, and
there is also a certain probability for F7 to converge to the optimal. For ISSA, all the low dimensional
functions can converge to the optimal and have the stable convergence performances on F2, F3, F4,
F5 and F8. Aside from this, ISSA obtains the minimum mean on all functions except F6. In order to
compare the differences of each method, we used the Friedman test to check the data in Table 8. The
specific process is shown below, and the results can be obtained by Table 9:
Algorithms 2019, 12, 80 21 of 29
Rank
Function
MDE distABC IGSA/PSO AND-RSN-PSOPSO-GWO ISSA
F1 1 5 4 6 3 2
F2 3 6 4 5 2 1
F3 2 4 5 6 2 2
F4 2.5 2.5 5 6 2.5 2.5
F5 2 5 4 6 2 2
F6 1 6 2 4 5 3
F7 2 6 5 4 3 1
F8 3 5 4 6 1.5 1.5
F9 2 6 3 5 4 1
F10 2 6 3 5 4 1
Total rank 20.5 51.5 39 53 29 17
Average rank 2.05 5.15 3.9 5.3 2.9 1.7
Sort 5 2 3 1 4 6
k
χ5r = 12
R2j − 3n(k + 1) = 12
(20.52 + 51.52 + 392 + 532 + 292 + 172 ) − 3 × 10 × (6 +
P
nk(k+1) 10×6×(6+1)
j=1
1) = 33.7857; α = 0.05, df = 6 − 1 = 5 at the 5% significant level and χ50.05 = 11.07 < 33.7857 according
to the Chi-square distribution table. Therefore, the six algorithms are considered to have significant
differences at the 5% significance level.
To further compare the performance of the six algorithms, assume that the convergence
performance of ISSA is better than the other five algorithms. A Holm test was carried out and
the results are shown in Table 10:
It can be seen from Table 10 that P1 < α(k − 1), P2 < α(k − 2), P3 < α(k − 3), P4 > α(k − 4),
P5 > α(k − 5); the original hypothesis is rejected at the 5% significance level. Therefore, compared
with AND-RSN-PSO, distABC and IGSA/PSO, ISSA has significantly better performance. ISSA
has a smaller average rank, though it does not outperform PSO-GWO and MDE. In summary,
compared with five other algorithms, the proposed algorithm ISSA has better performance on the low
dimensional functions.
Table 11 shows the convergence results on the high dimensional functions. It can be seen that
MDE can converge to a certain precision on F11, F12, F13, F16, F18 and F20, but the obtained precision
has obvious gaps compared with ISSA; distABC almost has no efficient convergence performance on
the high dimensional functions; IGSA/PSO can converge to a certain precision on F11, F12 and F13,
but the results are not good enough; ADN-RSN-PSO can converge to a certain precision on all the
high dimensional functions except on F18, but the obtained precision is far from ISSA; PSO-GWO
can converge to a better precision except on F18, whose convergence result is not good enough and
on F21 which cannot converge efficiently, with the obtained convergence precision still being worse
than that of ISSA. For ISSA, all the functions can converge to the optimal except F15, F16 and F20.
Algorithms 2019, 12, 80 22 of 29
Aside from this, the best fitness value, the worst fitness value, the mean fitness value and the standard
deviation of F11, F12, F13, F14, F17, F19 and F21 are all equal to zero and keep unchanged even if the
dimensions become higher. The best fitness value of F18 is also equal to zero no matter whether the
dimension is 30, 50 or 100, and the deviation is much smaller than the other five algorithms for F15,
F16 and F20, which cannot converge to the optimal. The best fitness value, the worst fitness value, the
mean fitness value and the standard deviation are still much better than the other five algorithms. In
order to compare the differences of each method, a Friedman test is taken to check the data in Table 11
when the dimension is 30. The specific process is shown below:
k
It can be obtained by: χ5r = 12
R2j − 3n(k + 1) = 12
(372 + 552 + 462 + 552 + 272 +
P
nk(k+1) 11×6×(6+1)
j=1
112 ) − 3 × 11 × (6 + 1) = 38.7403 with the results shown in Table 12; α = 0.05, df = 6 − 1 = 5 at the 5%
significant level and χ50.05 = 11.07 < 38.7403 according to the Chi-square distribution table. Therefore,
the six algorithms are considered to have significant differences at the 5% significance level.
Rank
Function
MDE distABC IGSA/PSO AND-RSN-PSOPSO-GWO ISSA
F11 3 5 4 6 2 1
F12 3 5 4 6 2 1
F13 3 5 4 6 2 1
F14 5 6 3 4 2 1
F15 5 6 4 3 2 1
F16 3 4 5 6 2 1
F17 3 4 5 6 2 1
F18 2 5 4 6 3 1
F19 3 5 6 4 2 1
F20 3 5 4 6 2 1
F21 4 5 3 2 6 1
Total rank 37 55 46 55 27 11
Average rank 3.3636 5 4.1818 5 2.4545 1
Sort 4 1 3 1 5 6
Algorithms 2019, 12, 80 26 of 29
To further compare the performance of the six algorithms, assume that the convergence
performance of ISSA is better than the other five methods. A Holm test is carried out and the
results Algorithms
are shown 2019,in
12, Table
x FOR PEER
13: REVIEW 27 of 31
To further compare the performance of the six algorithms, assume that the convergence
Table 13. The results of the Holm test.
performance of ISSA is better than the other five methods. A Holm test is carried out and the results
are shown in Table 13: q
k(k+1)
q
6×(6+1)
i Algorithm z = (Ri − R6 )/ 6n = (Ri − R6 )/ 6×11 = Pi α/(k−i)
(RThe
Table 13. R6 )/0.7977
i − results of the Holm test.
1 AND-RSN-PSO (5k− 1)/0.7977 = 5.0144
(k 1) 6 (6 1) 0 0.01
i Algorithm z = (Ri − R6)/(5 −6n 1)/0.7977
= (Ri − R=6)/5.0144
6 11 = (Ri − R6)/0.7977 Pi / (k i)
2 distABC 0 0.01
3 1 IGSA/PSO
AND-RSN-PSO 1)/0.7977== 3.9887
(4.1818(5−−1)/0.7977 5.0144 0
6e−05 0.01 0.0167
4 2 MDE distABC 1)/0.7977== 2.9630
(3.3636(5−−1)/0.7977 5.0144 0
0.0030 0.01 0.025
5 3 PSO-GWO
IGSA/PSO (4.1818
(2.4545 − 1)/0.7977 = 1.8234
− 1)/0.7977 = 3.9887 6e−05
0.068 0.0167 0.05
4 MDE (3.3636 − 1)/0.7977 = 2.9630 0.0030 0.025
5 PSO-GWO (2.4545 − 1)/0.7977 = 1.8234 0.068 0.05
It can be seen from Table 13 that P1 < α(k − 1), P2 < α(k − 2), P3 < α(k − 3), P4 < α(k − 4),
P5 > α(k − 5It),canandbetheseenoriginal
from hypothesis
Table 13 thatis rejected at, the
P1 (k 1) P2 5% P (level.
( ksignificance
2)
, 3
P4 ( k 4)compared
k 3) , Therefore, ,
with AND-RSN-PSO, distABC, IGSA/PSO and MDE, ISSA has a significantly
P5 ( k 5) , and the original hypothesis is rejected at the 5% significance level. Therefore, better performance. ISSA
compared
has a smaller average rank,distABC,
with AND-RSN-PSO, though it does notand
IGSA/PSO outperform
MDE, ISSAPSO-GWO. In summary,
has a significantly the ISSA proposed
better performance.
in this ISSA
paper hashas obvious
a smaller advantages
average on convergence
rank, though precision
it does not outperform and stability
PSO-GWO. in terms
In summary, of the high
the ISSA
proposed in this paper has obvious advantages on convergence precision
dimensional functions. Figure 5a–u shows the convergence curves of the algorithms mentioned and stability in terms of the above
high dimensional functions. Figure 5a–u shows the convergence curves
on F1–F21 in order to compare the convergence performances intuitively. The yellow, cyan, purple, of the algorithms mentioned
above on F1–F21 in order to compare the convergence performances intuitively. The yellow, cyan,
green, black, blue and red curves refer to MDE, distABC, IGSA/PSO, AND-RSN-PSO, PSO-GWO, SSA
purple, green, black, blue and red curves refer to MDE, distABC, IGSA/PSO, AND-RSN-PSO, PSO-
and ISSA,
GWO, respectively.
SSA and ISSA, Therespectively.
abscissa ofTheeach figure
abscissa of is thefigure
each iteration
is theand the ordinate
iteration of eachoffigure
and the ordinate each and is
the bestfigure
valueand found so far. The title of each figure is given the name of F1–F21.
is the best value found so far. The title of each figure is given the name of F1–F21.
Figure 5. Cont.
Algorithms 2019, 12, 80 (g) (h) (i) 28 of 31 27 of 29
functions and the statistical tests show that the proposed algorithm can promote the convergence
speed, improve the convergence accuracy and maintain the population diversity at the same time.
Furthermore, ISSA has obvious advantages in convergence performances compared with other five
intelligent evolutionary algorithms.
Author Contributions: Y.J.W. designed the experiments and revised the paper; T.L.D. shared the ideas and mainly
wrote the paper.
Funding: This research was funded by the National Natural Science Foundation of China under grants NO.61501107
and NO.61603073, and the Project of Scientific and Technological Innovation Development of Jilin NO.201750227
and NO.201750219.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. He, J.; Lin, G. Average Convergence Rate of Evolutionary Algorithms. IEEE Trans. Evol. Comput. 2016, 20,
316–321. [CrossRef]
2. Chugh, T.; Sindhya, K.; Hakanen, J. A survey on handling computationally expensive multiobjective
optimization problems with evolutionary algorithms. Soft Comput. 2017, 23, 3137–3166. [CrossRef]
3. Bhattacharyya, B.; Raj, S. Swarm intelligence based algorithms for reactive power planning with Flexible AC
transmission system devices. Int. J. Electr. Syst. 2016, 78, 158–164. [CrossRef]
4. Laina, R.; Lamzouri, F.E.-Z.; Boufounas, E.-M.; El Amrani, A.; Boumhidi, I. Intelligent control of a DFIG
wind turbine using a PSO evolutionary algorithm. Procedia Comput. Sci. 2018, 127, 471–480. [CrossRef]
5. Manjunath, P.G.C.; Krishna, P.; Parappagoudar, M.B.; Vundavilli, P.R. Multi-Objective Optimization of
Squeeze Casting Process using Evolutionary Algorithms. IJSIR 2016, 7, 55–74.
6. Ntouni, G.D.; Paschos, A.E.; Kapinas, V.M.; Karagiannidis, G.K.; Hadjileontiadis, L.J. Optimal detector
design for molecular communication systems using an improved swarm intelligence algorithm. Micro Nano
Lett. 2018, 13, 383–388. [CrossRef]
7. Liu, M.; Zhang, F.; Ma, Y.; Pota, H.R.; Shen, W. Evacuation path optimization based on quantum ant colony
algorithm. Adv. Eng. Inform. 2016, 30, 259–267. [CrossRef]
8. Yang, J.H.; Honavar, V. Feature Subset Selection Using a Genetic Algorithm. IEEE Intell. Syst. 1998, 13, 44–49.
[CrossRef]
9. Storn, R.; Price, K. Differential Evolution—A Simple and Efficient Heuristic for global Optimization over
Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [CrossRef]
10. Hansen, N.; Ostermeier, A. Completely Derandomized Self-Adaptation in Evolution Strategies. Evol. Comput.
2001, 9, 159–195. [CrossRef]
11. De Castro, L.N.; Von Zuben, F.J. Learning and Optimization Using the Clonal Selection Principle. IEEE Trans.
Evol. Comput. 2002, 6, 239–251. [CrossRef]
12. Xie, X.F.; Zhang, W.J.; Yang, Z.L. Social cognitive optimization for nonlinear programming problems. In
Proceedings of the 2002 International Conference on Machine Learning and Cybernetics, Beijing, China, 4–5
November 2002; pp. 779–783.
13. Ashrafi, S.M.; Dariane, A.B. A novel and effective algorithm for numerical optimization: Melody Search
(MS). In Proceedings of the 2011 11th International Conference on Hybrid Intelligent Systems, Malacca,
Malaysia, 5–8 December 2011; pp. 109–114.
14. Rao, R.V.; Patel, V. An elitist teaching-learning-based optimization algorithm for solving complex constrained
optimization problems. Int. J. Ind. Eng. Comput. 2012, 3, 535–560. [CrossRef]
15. Cheng, M.Y.; Prayogo, D. Symbiotic Organisms Search: A new metaheuristic optimization algorithm. Comput.
Struct. 2014, 139, 98–112. [CrossRef]
16. Jahani, E.; Chizari, M. Tackling global optimization problems with a novel algorithm—Mouth Brooding Fish
algorithm. Appl. Soft Comput. 2018, 62, 987–1002. [CrossRef]
17. Kirkpatrick, S.; Gelatt, C.D., Jr.; Vecchi, M.P. Optimization by Simulated Annealing. Science 1983, 220, 671–680.
[CrossRef]
18. Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. GSA: A Gravitational Search Algorithm. Inf. Sci. 2009, 179,
2232–2248. [CrossRef]
Algorithms 2019, 12, 80 29 of 29
19. Mirjalili, S.; Sadiq, A.S. Magnetic Optimization Algorithm for training Multi Layer Perceptron. In Proceedings
of the IEEE 3rd International Conference on Communication Software and Networks, Xi’an, China, 27–29
May 2011; pp. 42–46.
20. Kaveh, A.; Khayatazad, M. A new meta-heuristic method: Ray Optimization. Comput. Struct. 2012, 112–113,
283–294. [CrossRef]
21. Moein, S.; Logeswaran, R. KGMO: A swarm optimization algorithm based on the kinetic energy of gas
molecule. Inf. Sci. 2014, 275, 127–144. [CrossRef]
22. Kaveh, A.; Bakhshpoori, T. Water Evaporation Optimization: A novel physically inspired optimization
algorithm. Comput. Struct. 2016, 167, 69–85. [CrossRef]
23. Nematollahi, A.F.; Rahiminejad, A.; Vahidi, B. A novel physical based meta-heuristic optimization method
known as Lightning Attachment Procedure Optimization. Appl. Soft Comput. 2017, 59, 596–621. [CrossRef]
24. Colorni, A. Distributed Optimization by Ant Colonies. In Proceedings of the 1st European Conference on
Artificial Life, Paris, France, 11–13 December 1991; pp. 134–142.
25. Kennedy, J.; Eberhart, R.C. Particle swarm optimization. In Proceedings of the IEEE International Conference
on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948.
26. Basturk, B.; Karaboga, D. An artificial bee colony (ABC) algorithm for numeric function optimization. In
Proceedings of the Swarm Intelligence Symposium, Indianapolis, IN, USA, 12–14 May 2006; pp. 687–697.
27. Cuevas, E.; Cienfuegos, M. A swarm optimization algorithm inspired in the behavior of the social-spider.
Expert. Syst. Appl. 2013, 40, 6374–6384. [CrossRef]
28. Fausto, F.; Cuevas, E.; Valdivia, A. A global optimization algorithm inspired in the behavior of selfish herds.
Biosystems 2017, 160, 39–55. [CrossRef] [PubMed]
29. Zhang, Z.; Huang, C.; Huang, H.; Tang, S.; Dong, K. An optimization method: Hummingbirds optimization
algorithm. J. Syst. Eng. Electron. 2018, 29, 168–186. [CrossRef]
30. Jain, M.; Singh, V.; Rani, A. A novel nature-inspired algorithm for optimization: Squirrel search algorithm.
Swarm Evol. Comput. 2018, 44, 148–175. [CrossRef]
31. Zhang, Y.; Liu, M. An improved genetic algorithm encoded by adaptive degressive ary number. Soft Comput.
2018, 22, 6861–6875. [CrossRef]
32. Gomes, W.C.; dos Santos Filho, R.C.; de Sales Junior, C.D.S. An Improved Artificial Bee Colony Algorithm
with Diversity Control. In Proceedings of the 2018 Brazilian Conference on Intelligent Systems, Sao Paulo,
Brazil, 22–25 October 2018; pp. 19–24.
33. Wang, S.; Li, Y.; Yang, H. Self-adaptive differential evolution algorithm with improved mutation strategy.
Appl. Intell. 2017, 47, 644–658. [CrossRef]
34. Kiran, M.S.; Hakli, H.; Gunduz, M. Artificial bee colony algorithm with variable search strategy for continuous
optimization. Inf. Sci. 2015, 300, 140–157. [CrossRef]
35. Feng, X.; Xu, H.; Wang, Y. The social team building optimization algorithm. Appl. Soft Comput. 2018, 2, 1–22.
[CrossRef]
36. Demišar, J.; Schuurmans, D. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn.
Res. 2006, 7, 1–30.
37. Li, X.; Yin, M. Modified differential evolution with self-adaptive parameters method. J. Comb. Optim. 2016,
31, 546–576. [CrossRef]
38. Xiao, J.; Niu, Y.; Chen, P. An improved gravitational search algorithm for green partner selection in virtual
enterprises. Neurocomputing 2016, 217, 103–109. [CrossRef]
39. Babaoglu, I. Artificial bee colony algorithm with distribution-based update rule. Appl. Soft Comput. 2015, 34,
851–861. [CrossRef]
40. Sun, W.; Lin, A.; Yu, H. All-dimension neighborhood based particle swarm optimization with randomly
selected neighbors. Inf. Sci. 2017, 405, 141–156. [CrossRef]
41. Teng, Z.J.; Lv, J.L.; Guo, L.W. An improved hybrid grey wolf optimization algorithm. Soft Comput. 2018, 22,
1–15. [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).