Intelligent Optimization Principles Algorithms And Applications Changhe Li instant download
Intelligent Optimization Principles Algorithms And Applications Changhe Li instant download
https://ebookbell.com/product/intelligent-optimization-
principles-algorithms-and-applications-changhe-li-58439466
https://ebookbell.com/product/network-optimization-in-intelligent-
internet-of-things-applications-principles-and-challenges-payal-
khurana-batra-58745762
https://ebookbell.com/product/intelligent-optimization-modelling-in-
energy-forecasting-weichiang-hong-55251802
https://ebookbell.com/product/intelligent-optimization-of-mold-design-
and-process-parameters-in-injection-molding-1st-ed-mehdi-
moayyedian-7321250
https://ebookbell.com/product/intelligent-optimization-and-control-of-
complex-metallurgical-processes-1st-ed-2020-min-wu-10806138
Configurable Intelligent Optimization Algorithm Design And Practice In
Manufacturing 1st Edition Fei Tao
https://ebookbell.com/product/configurable-intelligent-optimization-
algorithm-design-and-practice-in-manufacturing-1st-edition-fei-
tao-4931874
https://ebookbell.com/product/learning-and-intelligent-
optimization-16th-international-conference-lion-16-milos-island-
greece-june-510-2022-revised-selected-papers-dimitris-e-simos-49114524
https://ebookbell.com/product/learning-and-intelligent-optimization-
second-international-conference-lion-2007-ii-trento-italy-
december-812-2007-selected-papers-1st-edition-sameh-alshihabi-2039740
https://ebookbell.com/product/learning-and-intelligent-
optimization-5th-international-conference-lion-5-rome-italy-
january-1721-2011-selected-papers-1st-edition-olcay-taner-yldz-2456114
https://ebookbell.com/product/learning-and-intelligent-
optimization-15th-international-conference-lion-15-athens-greece-
june-2025-2021-revised-selected-papers-lecture-notes-in-computer-
science-1st-ed-2021-dimitris-e-simos-editor-37583830
Changhe Li · Shoufei Han
Sanyou Zeng · Shengxiang Yang
Intelligent
Optimization
Principles, Algorithms and
Applications
Intelligent Optimization
Changhe Li • Shoufei Han • Sanyou Zeng •
Shengxiang Yang
Intelligent Optimization
Principles, Algorithms and Applications
Changhe Li Shoufei Han
School of Artificial Intelligence School of Artificial Intelligence
Anhui University of Science Anhui University of Sciences
and Technology & Technology
Hefei, Anhui, China Hefei, Anhui, China
Jointly published with China University of Geosciences Press, Hubei Province, P.R. China
The print edition is not for sale in Mainland China. Customers from Mainland China please order the
print book from: China University of Geosciences Press.
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
vii
viii Preface
(3) Researchers in the domain of intelligent optimization. For each topic of this
book, we present the recent development and propose several challenging issues
that should be addressed in this domain. Potential solutions to these challenging
issues are also discussed.
The overall structure of this book contains four parts. Part I presents the
fundamentals of optimization from Chaps. 1 to 3. Part II presents evolutionary
computation algorithms from Chaps. 4 to 5. Part III presents optimization techniques
from Chaps. 6 to 8. Part IV presents advanced topics and applications. The four parts
are followed by the references with nearly 400 entries. We may have inevitably
missed some and please feel free to send us an email in case a really important one
is forgotten.
This book is distributed as part of the OFEC project, which is an open-source
project that aims to develop a common platform for students, users, and researchers
to quickly design and visualize their algorithms (see Appendix B). OFEC has
collected a plenty of algorithms and problems and is constantly maintained and
updated. The source code of the project and other materials of the book can be
obtained by following the link https://gitlab.com/ofec-release.
This book could not have been possible without the support of many people. The
authors thank the Intelligent Optimization and Learning team for their hard work on
the research. Furthermore, the authors thank Yong He, Min Wu, Chuanke Zhang,
Michalis Mavrovouniotis, Trung Thanh Nguyen, Carlos A. Coello Coello, Ming
Yang, Xin Yao, Kay Chen Tan, Boyang Qu, Jing Liang, Juan Zou, Qingfu Zhang,
Yew-Soon Ong, Jinliang Ding, Ponnuthurai Nagaratnam Suganthan, Miqing Li,
Juergen Branke, Xiaodong Li, Hui Wang, Aimin Zhou, Xinye Cai, Qunfeng Liu,
Ke Li, Yong Liu, Bob McKay, and Yuping Chen for their sincere support.
I would like to thank my wife, Fengxia Wang, for her endless love and care for
the family and children. I also would like to thank my parents and my parents-in-law
for their support and encouragement.
The genesis of this book is deeply rooted in the desire to honor the memory of
Lishan Kang. Lishan Kang made significant contributions to the fields of parallel
computing and evolutionary computation. His enduring scientific spirit serves as a
perpetual source of inspiration, motivating us to continually explore the frontiers of
knowledge.
The publication of this book was supported by the National Natural Science
Foundation of China under Grant 62076226, in part by the Hubei Provincial Natural
Science Foundation of China under Grant 2023AFA049.
ix
Contents
xi
xii Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
Acronyms
xxi
xxii Acronyms
This part introduces the basics of optimization, including the relationship between
optimization and machine learning, categories of optimization problems and opti-
mization algorithms, the fitness landscape, and computational complexity. Several
typical canonical optimization algorithms, i.e., numerical optimization algorithms,
state space search algorithms, and single-solution-based random search algorithms,
are introduced.
Chapter 1
Introduction
Temperature in Celsius
25
20
15
10
0
0 25 50 75 100 125 150 175
Cricket chirps per minute
. y ' = wx + b (1.1)
where .y ' means temperature in Celsius, the label we predict; x means the number
of chirps of crickets per minute, the input feature;w is the slope of the line or the
weight of the feature; and b is y -intercept or the bias.
This equation is the model we need to build. w and b are its model parameters,
and now, our job is to find the right values for them. In supervised learning, we build
our model by examining a set of examples and trying to find a model that minimizes
loss. So, how to evaluate the loss? A simple and popular way is to use mean square
error (MSE). MSE is the average squared loss per example over the entire dataset.
1 https://developers.google.com/machine-learning/crash-course/descending-into-ml/linear-
regression
2 https://en.wikipedia.org/wiki/Dolbear%27s_law
1.2 Optimization Problems 5
To calculate MSE, we sum up all squared losses for each example and then divide
by the number of examples.
1 ⎲
MSE =
. (y − prediction(x))2 (1.2)
|D|
(x,y)∈D
1 ⎲
MSE =
. (y − wx − b)2 (1.3)
|D|
(x,y)∈D
Now we can examine this equation, all the x and y are known for us, and now,
the problem is very explicit to find the right values of w and b, which make MSE
minimal. The final problem here is a pure optimization problem. Once we solve this
problem, we can get the parameters needed in our model in Eq. (1.1), i.e., we have
built the model.
Generally speaking, as we have seen in the previous example, optimization lies
at the heart of most machine learning algorithms [1]. No matter how complex the
problems look, most machine learning problems can be reduced to optimization
problems. The modeler formulates the problem by selecting an appropriate family
of models. Then the model is typically trained by solving an optimization problem
that optimizes the parameters of the model given a selected loss function.
subject to
gj (x) ≤ 0,
. j = 1, · · · , N
hk (x) = 0, k = 1, · · · , P
where .fi (x) are objective functions to be minimized or maximized; they are
measures to quantify the performance of a solution.
In manufacturing, we want to maximize profit and minimize production cost,
while in data fitting, we want to minimize the difference error between observed
data and predicted results like Eq. (1.3). .x are variables represented for a solution.
The aim is to find solutions that have the best performance in terms of the objective
functions. In manufacturing, they may be the amount of each resource consumed
in production process, whereas in data fitting, they may be the parameters of a
model like the weights and bias defined in Eq. (1.3). .gj (x) ≤ 0 and .hk (x) = 0 are
called inequality constraints and equality constraints, respectively. They describe the
relationships among variables and define the possible range of values for variables.
A solution is feasible only if it satisfies all constraints; otherwise, the solution is
infeasible.
The term minimization problem refers to minimizing the objective function,
whereas the maximization problem refers to maximizing the objective function,
which can be defined as negating the minimized objective function. If both .N and
P are zero, the problem is an unconstrained optimization problem.
If M is larger than one, which means that there are more than one objectives
that need to be optimized synchronously, then the problem is called multi-objective
optimization problem. A simple example of a multi-objective optimization problem
is building a computer: we need to use limited money to balance performance,
appearance, comfort, and other conflicting indicators; the variables are configura-
tions of hardware components; the constraints may be the limited budget, or we
only want a 32-inch display.
The optimization problem can be divided into two types according to the type of
input: continuous optimization and discrete optimization. Here is an example to
help us understand continuous optimization and discrete optimization.
A chemical company has two factories .F1 and .F2 and three retailers .R1 , R2 , R3 .
Each factory .Fi produces .ai tons of a certain chemical product each week; each
retailer .Rj demands .bj tons of the product weekly. .cij is the cost of shipping one
ton of the product from factory .Fi to retailer .Rj . Figure 1.2 shows the problem.
The problem is to determine how much of the product to ship from each factory to
all retailers to satisfy all the requirements and minimize the total cost. The variables
1.2 Optimization Problems 7
of the problem are .xij , .i = 1, 2, j = 1, 2, 3, where .xij is the amount of the product
shipped from .Fi to .Rj . The problem can be formulated as
⎲
. min cij xij (1.5)
ij
subject to
⎲
3
. xij ≤ ai , i = 1, 2
j =1
⎲
2
xij ≥ bj , j = 1, 2, 3
i=1
xij ≥ 0, i = 1, 2, j = 1, 2, 3
A deterministic algorithm refers to an algorithm that will always produce the same
output given a particular input. It is often used in a problem where the objective
function is clear and the solution can be efficiently searched with a direction-guided
strategy. For example, the gradient information is normally used as the search
direction for continuous optimization problems that are smooth and particularly
differentiable. However, when the relationship between a solution and its objective
value is not obvious or calculating its objective value is very time-consuming or
the problem has multiple objectives or all possible solutions cannot be examined
by each due to the large number, particularly for problems where the number of
possible solutions increases exponentially as the number of variables increases, then
probabilistic algorithms come into play.
A probabilistic optimization algorithm refers to an algorithm that takes random
inputs during the search process, which are typically uniformly and independently
distributed. A typical characteristic of the probabilistic optimization algorithm is
that the algorithm may produce different outputs even when given the same input
due to randomness. However, we can get a stable output as long as the randomization
process is pseudo-random, which is usually driven by a random seed generator. The
randomness is typically used to help the algorithm discover new solutions and makes
the algorithm less sensitive to modeling errors. The algorithm normally produces a
satisfactory solution within an affordable amount of time. Furthermore, the injected
randomness can help the algorithm avoid trapping into local optima and eventually
find a global optimum.
Indeed, this randomization principle is known to be a simple yet effective way
to achieve good performance uniformly across many sorts of problems, particularly
for complex problems mentioned above. The probabilistic algorithm belongs to the
family of Monte Carlo method devised in the 1930s. Now, it has become one of
the most important research branches of optimization since it was proposed. Due
to the successful application of these intelligent optimization algorithms for solving
complicated optimization problems where deterministic algorithms fail, a variety of
probabilistic algorithms have been implemented in some commercial software, such
1.3 Optimization Algorithms 9
3 https://tomopt.com/tomlab/products/midaco/
4 http://www.midaco-solver.com/
5 http://iosotech.com/index.htm
10 1 Introduction
References
1. Bennett, K.P., Parrado-Hernández, E.: The interplay of optimization and machine learning
research. J. Mach. Learn. Res. 7(46), 1265–1281 (2006)
2. Cowling, P., Kendall, G., Soubeiga, E.: A hyperheuristic approach to scheduling a sales summit.
In: International Conference on the Practice and Theory of Automated Timetabling, pp. 176–
190. Springer, Berlin (2000)
3. Glover, F.W., Kochenberger, G.A.: Handbook of Metaheuristics. Kluwer, Dordrecht (2003)
4. Michalewicz, Z., Fogel, D.B.: How to Solve It: Modern Heuristics. Springer Science & Business
Media, Berlin (2000)
5. Pearl, J.: Heuristics: Intelligent Search Strategies for Computer Problem Solving. Addison
Wesley, Reading (1984)
Chapter 2
Fundamentals
The concept of fitness landscapes, originating from theoretical biology more than 90
years ago [15], has played an important role in optimization. It is a valuable tool for
analyzing the optimization process, especially for studying the relationship between
the performance of optimization algorithms and the characteristics of optimization
problems [12]. In practice, it attempts to draw an analogy with real landscapes to
gain a better understanding of how and where algorithms operate [7].
A fitness landscape is a mapping from every possible solution, which represents
a candidate solution of a problem to be solved, into a fitness value. More formally,
a fitness landscape consists of three main components: a set of solutions, a fitness
function that evaluates these solutions, and a neighborhood relationship between
the solutions (i.e., the definition of distance between solutions). The fitness function
determines how good a solution is in terms of the objective value. It is the
driving force behind the search. Neighborhood relations are crucial in the analysis
of the fitness landscape. Typically, the shape of a fitness landscape depends on
the representation/encoding (see Sect. 4.2) of solutions and the chosen distance
measures between solutions for a given problem. That is, a particular problem can
have different fitness landscapes. For example, the landscape of real-valued and
binary encoding for a continuous optimization problem varies significantly.
More specially, a fitness landscape describes a surface of fitness values over the
solution space where optimization methods are searched. When an optimization
problem is a minimization problem, searching for an optimal solution to the
problem can be interpreted as walking on the surface of its fitness landscape toward
the lowest valleys, with the ability to overcome other valleys [3]. Therefore, the
10
20.0
8 15.0
10.0
6
5.0
4
0.0
0
2 2
4 0
6 2
2 8 6 4
1 4 7 10 13 16 8
10
10
1
Fig. 2.1 Fitness landscapes of one-dimensional (a) and two-dimensional (b) continuous optimiza-
tion problems
fitness landscape has been found to be effective for analyzing the behavior of the
optimization method in an optimization process and can help design an effective
optimization method. Figure 2.1 shows the fitness landscapes of two continuous
optimization problems in one dimension and two dimensions, respectively.
X = {x 1 , x 2 , · · · }
. (2.1)
Feasible region
can solve it in polynomial time. However, the continuous sphere function with
an uncountably infinite solution space can be easily solved by a gradient descent
method since it is convex and unimodal.
For a constrained optimization problem, only parts of the solution space are
feasible (see the feasible region in Fig. 2.2). A solution is feasible only if it satisfies
all constraints, i.e., it is within the feasible region of the solution space; otherwise,
the solution is infeasible, i.e., it is within the infeasible region in the solution space.
Infeasible solutions cannot be deployed. We need to find the feasible solution with
the best fitness value (see Chap. 10 for methods for finding feasible solutions).
F = {f (x 1 ), f (x 2 ), · · · }
. (2.2)
real world. For optimization problems with more than one objective function, the
comparison between solutions becomes complicated. Therefore, special methods
are needed to deal with the comparison difficulty (see Chap. 9). For constrained
optimization problems, the comparison between solutions is also complicated
since both objective values and constraint values need to be taken into account.
Constraint-handling techniques are normally needed to deal with constraints (see
Chap. 10) since solutions that violate any constraint are infeasible.
2.1.3 Neighbourhood
Global optimum
Fig. 2.3 The global optimum for a single objective minimization problem
(a) (b)
Fig. 2.4 Illustration for basin of attraction (a) and local optima (b)
Normally, the distribution and the number of optimal solutions of a problem are
strongly dependent on the type of the problem. In single-objective optimization
problems, there is one global optimum in general or multiple global optima in the
scenario of multimodal optimization (see Chap. 8). In multi-objective optimization
problems, there is a set of optimal solutions (widely known as Pareto-optimal
solutions; see Chap. 10), instead of a single optimal solution.
Before introducing the definition of the local optimum, we introduce the concept of
the basin of attraction. The basin of attraction of an optimum is the region of all its
neighborhood solutions whose objective values change monotonically as they move
closer to the optimum, as shown in Fig. 2.4a.
16 2 Fundamentals
A local optimum is a solution that has the best fitness value within its neighbor-
hood in the search space, as shown in Fig. 2.4b.
In some interesting cases, the local optimum is not a single point but a set of
points on ridges or plateaus with the same objective value. The shape of ridges and
plateaus can be linearly or non-linearly distributed in the fitness landscape.
To find the global optimum of a problem is not an easy task due to many factors,
such as a huge number of local optima, comparison difficulty between solutions in
constrained or multi-objective situations, dynamic environments, etc. To design an
efficient optimization algorithm, we should ensure that the algorithm can converge
to the region of global optimum, i.e., carry out exploitation for local optima. To
achieve the objective, the algorithm should be able to avoid trapping into local
optima and continue to explore promising areas that are not exploited, i.e., carry out
exploration for global optima (see the balance between exploration and exploitation
in Chap. 7).
The fitness landscape has played an important role in the theory of complex systems.
It has been found to be effective in optimization theory, shown to be powerful
for understanding the behavior of optimization algorithms, and to help improve
their performance. Therefore, by analyzing the structure of the underlying fitness
landscape, the behavior of optimization methods can be understood, and the design
of an effective algorithm becomes easier.
The structure of the fitness landscape significantly impacts the performance of
an optimization algorithm. Several properties associated with the structure of fitness
landscapes [8] will be discussed in detail as follows.
2.2.1 Modality
During landscape analysis in evolutionary theory, there has always been a tendency
to focus on the relationship between the problem difficulty and the landscape
modality. Generally speaking, more modality of the fitness landscape means more
difficulty of the problem. This is because modality biases search algorithms, as more
modality implies more candidate solutions with the same objective value, making it
more challenging for algorithms to determine the direction to be taken.
A landscape that has only one optimum is commonly called unimodal, as shown
in Fig. 2.5. On the contrary, a landscape with more than one optimum is known as
multimodal, as shown in Fig. 2.6. To some extent, the number of local optima in a
2.2 Properties of Fitness Landscape 17
140
1400
120
1200
100
1000
80
800
60
600
40
400
20
200
20
15
10 15 20
5 10
20 0 5
15 10 5 0
5 0 10 5
5 0 55 10 15 10
10 15 20 20 15
400
40000
300
30000
200
20000
100
10000
00
3
3
2
2
1
1
0
0
3 1
3 2 1
2 11 0 2
1 2
2 33
3
landscape significantly impacts the difficulty in finding the global optimum. A local
optimum is a point or region (a set of interconnected points with equal fitness) with
a fitness function value greater than those of all its nearest neighbors. Obviously, if
there are many local optima, it is difficult for a search algorithm to find all optima or
the global optimum since the algorithm needs to jump out many valleys, i.e., points,
which are in the basin of attraction of local optima, should cross all peaks along the
direction to the global optimum as shown in Fig. 2.4b. Therefore, it is a challenging
issue for search algorithms to address.
The number of optima does indeed have a significant impact on landscape
difficulty. However, for insightful analysis of landscape difficulty, it is crucial
to investigate the distribution, size, shape, and depth (or height) of the basins
of attraction. These factors may be more important in determining the level of
landscape difficulty. A landscape is deemed difficult when local optima possess
large basins of attraction, while the global optimum has a smaller basin of attraction.
An extreme scenario occurs when the global optimum is an isolated spike in a
plateau. Optima with basins of attraction of asymmetrical shapes are generally more
difficult to navigate than those with symmetrical shapes.
An objective-driven search algorithm inherently favors exploring local regions
that are easy to search. Solutions in such regions typically exhibit faster and larger
improvements in terms of the objective value compared to those in harder-to-search
regions. Consequently, solutions in the hard-search region are gradually removed
due to their poor fitness, leading the algorithm to converge in the easiest-to-search
region. However, if the easy-search region only contains local optima, the algorithm
is susceptible to premature convergence, wherein it becomes stuck in a local
optimum. Therefore, the modality of the fitness landscape presents fundamental
challenges for algorithms aiming to locate the global optimum when the search
space comprises local regions with varying levels of search difficulties.
2.2.2 Ruggedness
The search often takes place in a rugged landscape, which has a profound effect on
the problem difficulty and search speed of an algorithm. A rugged landscape can be
regarded as a multi-modality landscape with many sharp descents and ascends.
In simple cases, optimization methods can effectively track optimum solutions
using gradient information from fitness landscapes, particularly for objective func-
tions with continuous and low total variation. However, when objective functions
exhibit instability or frequent oscillations, as depicted in Fig. 2.7, it becomes more
challenging to determine the correct directions to proceed during the optimization
process. The ruggedness of these functions directly correlates with their difficulty
to optimize, with more rugged functions presenting greater challenges.
In short, ruggedness has to do with the level of variation in fitness values in
a fitness landscape, where the fitness values of neighborhood solutions have very
different fitness values. An extreme case is the needle-in-a-haystack problem, where
2.2 Properties of Fitness Landscape 19
3.0
3.0
2.5
2.5
2.0
2.0
1.5
1.5
1.0
1.0
0.5
0.5
40
40
20
20
40 0
40 20 0
20
00 20
20
20
20 40
40
40
40
Fig. 2.7 An example for ruggedness for a 2-D shifted Griewank function
each peak is like a needle without basin of attraction. The opposite of a very rugged
landscape would be a landscape with a single large basin/peak of attraction or a flat
landscape with no features. For the rugged landscape, a search algorithm is hardly
to get knowledge of where to go, causing the algorithm to be difficult to converge.
2.2.3 Deceptiveness
A typical fitness landscape should not only provide sufficient information to guide
the search for an algorithm, but the information should also guide the search in the
right direction, i.e., the direction toward the global optimum. In contrast, a fitness
landscape guides the optimization process far away from the optimum (and thus
also in the wrong direction), which is sometimes known as fitness deceptiveness
as shown in Fig. 2.8. The gradient of deceptive fitness landscapes usually leads the
search away from the global optimum.
Deceptiveness is intricately tied to the structure of optima within a landscape.
The relative positions of local optima to the global optimum and the presence of
isolation significantly influence the level of deceptiveness. For example, when an
algorithm discovers a region with superior fitness compared to others, it tends to
exploit this area under the assumption that it likely contains the global optimum.
However, in reality, the optimal solution may be distant from the currently searched
area [6]. Interestingly, certain optimization algorithms demonstrate an advantage
in solving problems with high levels of deceptiveness [2], whereas increased
20 2 Fundamentals
200
2000
150
1500
100
1000
50
500
6
4 6
2 4
0 2
44 22 2 0
0
0 2 4 2
2 44
6 4
6
Fig. 2.8 An example for deceptiveness for a 2-D rotated hybrid composition function
2.2.4 Neutrality
100
10000
80
8000
60
6000
40
4000
20
2000
20
1520
1015
5 10
05
20
20 15
50
15 10 55 10 5
10 00 55 1510
10 15 20 15
10 15 20
20 20
Fig. 2.9 An example for neutrality, 3-D map for 2-D shifted and rotated Rosenbrock function
20
20
15
15
10
10
55
30
2030
1020
30
30 0 10
20
20 10 0
10
10 10
00 10
10 20
20
20
20 30
30
30 30
Fig. 2.10 An example for neutrality, 3-D map for 2-D shifted Ackley function
2.2.5 Separability
2.2.6 Scalability
d tree [4]. In each subspace, the properties can be fine-tuned separately, such as
dimensionality, modality, the number of objectives, optima positions, shape and size
of basin of attraction, landscape structure (distribution of optima), inter-relationship
between decision variables, and domino convergence.
How to quantify the properties mentioned above is very important in fitness
landscape analysis. Normally, special techniques are needed to understand the
structure of the fitness landscape. In addition, how to visualize the fitness landscape
for discrete optimization problems or high-dimensional continuous optimization
problems while keeping the visual effect on these properties is also important yet a
challenging issue.
Efficient algorithms are preferred in terms of time and storage usage. Time and
storage are the two most commonly used measures to quantify the amount of
resources used. The efficiency of an algorithm depends on its time complexity and
space complexity.
that sorted by increasing order of magnitude, the commonly used time complexity
is 1, .lg n, n, .n lg n, .n2 , .n3 , .2n , and .n!.
When calculating the time complexity of an algorithm, one way is to look at how
many “for” loops there are, i.e., only one is .O(n), two is .O(n2 ), and so on. If the
upper bound of an algorithm is independent of the input size, it has a constant time,
denoted by .O(1). If the algorithm’s .T (n) = O(lg n), it has a logarithmic time, i.e.,
binary tree operations and binary search.
The best, worst, and average case complexity are the three common ways of mea-
suring the time complexity (also applied to other complexity measures) regarding
different inputs with the same size. Take the quick sort algorithm as an example.
The best-case complexity of quit sort is .O(n lg(n)) when each pivot divides the set
into half. The worst-case complexity occurs when there is no division at all during
the whole sorting process, which takes .O(n2 ). The average-case complexity is also
.O(n lg(n)) by taking all possible permutations of n elements with equal probability.
Sometimes, we can optimize an algorithm for reducing both the time complexity
and the space complexity. However, in practice, it is not always possible to achieve
the two objectives. A different algorithm may take less time but at the expense of
requiring more memory space and vice versa.
26 2 Fundamentals
Therefore, we should balance the trade-off between space and time according
to our constraints. That is to say, if we have memory constraints, then we can take
algorithms requiring less memory at the cost of more running time. On the contrary,
if we have running-time constraints, then we can take algorithms using less time at
the cost of more memory. For example, our task needs to implement a lookup table.
We can implement it by storing all entries in the memory, and then a result can be
quickly obtained for each enquiry. Other than that, we can recalculate each entry
as needed for each enquiry to save memory but at the cost of increasing computing
time.
The P versus NP problem is one of the seven Millennium Prize Problems selected
by the Clay Mathematics Institute in 2000. It is also one of the hardest problems in
the field of computer science concerning the speed of a computer to finish a task.
The P and NP problem is the question of whether all so-called NP problems are
equal to the P problems. Although this question is still an open question, the latest
investigation shows that 99% of experts believe that P ./= NP.1
A problem in class P represents the problem that could be addressed in
polynomial time with respect to the size of the problem. The problem is also called
deterministic polynomial problem, the P problem, for short. To be specific, a P
problem means that a solution can be obtained by an algorithm running in an amount
of time that is polynomial to its problem scale.
NP problems denote the non-deterministic polynomial problems. If the solution
of a problem can be verified in polynomial time, then it could be regarded as
NP. Verifying a solution in polynomial time does not mean finding a solution in
polynomial time. For example, finding priming factors of a large number will try
many different candidates, but verifying these candidates can be achieved by simply
multiplying these factors.
NP-hard problems refer to problems that are at least as hard as NP problems,
i.e., all NP problems can be reduced to NP-hard problems. Nevertheless, it is worth
noting that not all NP-hard problems can be classified as NP problems, which means
that verifying a solution may not take polynomial time.
The NP-complete problem is a problem that is both NP and NP-hard. Therefore,
finding an effective algorithm for any NP-complete problem suggests that an
efficient algorithm could be found for all NP problems as a solution for any problem
belonging to this type could be transformed into a solution for any other member
of the type. The TSP is a classical NP-complete problem, which widely exists in
logistic scheduling. It is unknown whether there are polynomial-time algorithms
for NP-complete problems, since it is still one of the most important problems in
References
1. Beaudoin, W., Verel, S., Collard, P., Escazut, C.: Deceptiveness and neutrality the ND family of
fitness landscapes. In: Proceedings of the 8th Annual Conference on Genetic and Evolutionary
Computation, pp. 507–514. ACM SIGEVO, Association for Computing Machinery, New York
(2006)
2. Goldberg, D.E.: Genetic algorithms and Walsh functions: Part II, deception and its analysis.
Complex Syst. 3(2), 153–171 (1989)
3. Jones, T.: Evolutionary algorithms, fitness landscapes and search. Ph.D. thesis, The University
of New Mexico, Albuquerque (1995)
4. Li, C., Nguyen, T.T., Zeng, S., Yang, M., Wu, M.: An open framework for constructing
continuous optimization problems. IEEE Trans. Cybern. 49(6), 2316–2330 (2018)
5. Li, X., Tang, K., Omidvar, M.N., Yang, Z., Qin, K.: Benchmark functions for the CEC’2013
special session and competition on large-scale global optimization. Technical report, RMIT
University, Cancun (2013)
6. Liepins, G.E., Vose, M.D.: Deceptiveness and genetic algorithm dynamics. In: Rawlins, G.J.
(ed.) Foundations of Genetic Algorithms, pp. 36–50. Elsevier, Amsterdam (1991)
7. Pitzer, E., Affenzeller, M.: A comprehensive survey on fitness landscape analysis. In: Fodor,
J., Klempous, R., Araujo, C.P.S. (eds.) Recent Advances in Intelligent Engineering Systems,
pp. 161–191. Springer, Berlin (2012)
8. Reeves, C., Rowe, J.E.: Genetic Algorithms: Principles and Perspectives: A Guide to GA
Theory. Springer Science & Business Media, Boston (2002)
9. Reidys, C.M., Stadler, P.F.: Neutrality in fitness landscapes. Appl. Math. Comput. 117(2–3),
321–350 (2001)
10. Rudnick, W.M.: Genetic algorithms and fitness variance with an application to the automated
design of artificial neural networks. Ph.D. thesis, Oregon Graduate Institute of Science &
Technology, Hillsboro (1992)
11. Shang, Y.W., Qiu, Y.H.: A note on the extended Rosenbrock function. Evol. Comput. 14(1),
119–126 (2006)
12. Smith, T., Husbands, P., O’Shea, M.: Fitness landscapes and evolvability. Evol. Comput. 10(1),
1–34 (2002)
13. Thierens, D., Goldberg, D.E., Pereira, A.G.: Domino convergence, drift, and the temporal-
salience structure of problems. In: 1998 IEEE International Conference on Evolutionary
Computation Proceedings. IEEE World Congress on Computational Intelligence, pp. 535–540.
IEEE, Anchorage (1998)
14. Weise, T., Niemczyk, S., Skubch, H., Reichle, R., Geihs, K.: A tunable model for multi-
objective, epistatic, rugged, and neutral fitness landscapes. In: Proceedings of the 10th Annual
Conference on Genetic and Evolutionary Computation, pp. 795–802. ACM SIGEVO, New
York (2008)
15. Wright, S.: The roles of mutation, inbreeding, crossbreeding, and selection in evolution. In:
Proceedings of the VI International Congress of Genetics, pp. 356–366. The International
Genetics Federation, The VI International Congress of Genetics, New York (1932)
16. Yang, M., Zhou, A., Li, C., Yao, X.: An efficient recursive differential grouping for large-scale
continuous problems. IEEE Trans. Evol. Comput. 25(1), 159–171 (2020)
Chapter 3
Canonical Optimization Algorithms
Numerical optimization algorithms use the above idea to solve the problem in a
recursive way. Suppose that .x k is the point obtained at the k-th iteration and .d k is
the direction along which the fitness function .f (x) declines at point .x k . .αk > 0 is
the step size along .d k . Then the next point .x k+1 is obtained by
x k+1 = x k + αk d k
. (3.1)
where .f (x k+1 ) < f (x k ) must hold. Such an algorithm is also called a line search.
Line search starts from an initial solution .x 0 . It finds a direction along which the
fitness value will decline and decides the step size to calculate the next solution.
The procedure is repeated until the termination is met. Algorithm 3.1 shows the
procedures of the algorithm.
‖x k+1 − x k ‖ < ϵx
. (3.3)
‖f (x k+1 ) − f (x k )‖ < ϵf
points, .x = −1 and .x = 1, where .x1 = 2. The key issue is that there is no sufficient
reduction in f .
There are some famous conditions to find such .α to obtain a sufficient reduction
in f :
(1) The Armijo condition
f (x k + αk d k ) ≤ f (x k ) + c1 αk d Tk ∇f (x k )
. (3.4)
where constant .c1 belonging to .(0, 1) generally is set to .c1 = 10−4 in practice.
.d k is the descent direction, so .d k ∇f (x k ) < 0 and the right-hand side of the
T
inequality is the subtractive function of .α. This condition guarantees that there
is some reduction in f as long as .α is not too small. There is an issue with
the Armijo condition. Any small .α that satisfies the Armijo condition cannot
make sufficient progress. To address this issue, we add another constraint to the
Armijo condition as follows:
(2) The Wolfe condition
f (x k + αk d k ) ≤ f (x k ) + c1 αk d Tk ∇f (x k )
. (3.5)
∇f (x k + αk d k )d Tk ≥ c2 d Tk ∇f (x k )
where .0 < c1 < c2 < 1, .c2 is a relatively large number compared with .c1 where
.c2 = 0.9 for Newton or quasi-Newton methods and .c2 = 0.1 for the nonlinear
conjugate gradient method. The second inequality guarantees that the slope of
Random documents with unrelated
content Scribd suggests to you:
Cassiterides yielded in abundance the ores of copper and tin, but
these had to be smelted and worked with all the accumulated results
of tentative skill before they yielded the copper or more useful
bronze. By whom or where this first knowledge was mastered is
unknown: the tendency is still to look to Asia, perhaps to Phœnicia,
or to the still undetermined cradle of the Aryans, wherever that may
prove to have been, for the birth of this phase of metallurgic art
which constituted so important a stage in early civilisation. Yet if the
ancient American missed it, it was not for want of opportunity.
Examples of the accidental fusion of copper by the sacrificial fires of
the Mound-Builders repeatedly occur in the mounds of the Ohio
valley. But no gifted native alchymist was prompt to read the lesson
and turn it to practical account.
Asia and Europe appear to have passed by a natural transition,
step by step, from their rudest stages of lithic art to polished stone,
and then to implements of metal. Some of the steps were doubtless
very slow. Worsaae believed that the use of bronze prevailed in
Denmark “five or six hundred years before the birth of Christ.”[76] In
Egypt it undoubtedly was known at a greatly earlier date. I still
incline to my early formed opinion, that gold was the first metal
worked. Found in nuggets, it could scarcely fail to attract attention.
It was easy to fashion into shape; and some of the small, highly
polished stone hammers seem fitter for this than any other work.[77]
The abundant gold ornaments of the New World at the time of the
discovery of Mexico and Peru accord with this idea. The like
attraction of the bright native copper, is proved by its employment
among the southern Indians for personal ornaments; and in this way
the economic use of the metals may have been first suggested.
From the working of gold nuggets, or of virgin copper, with the
hammer, to the smelting of the ores, was no trifling step; but that
knowledge once gained, the threshold of civilisation and true
progress had been reached. The history of the grand achievement is
embodied in the earliest myths both of the Old and the New World.
Tubal-cain, Dædalus, Hephæstus, Vulcan, Vœlund, Galant, the Luno
or the Celtic Fingal, and Wayland, the Saxon smith-god, are but
legendary variations of the first worker by whom the gift of
metallurgy was communicated to man; and so too the New World
has its Quetzalcoatl, or Vœlund of the Aztecs, the divine instructor of
their ancestors in the use of the metals. But whatever be the date of
this wise instructor, no share of the knowledge communicated by
him to that favoured race appears to have ever penetrated
northward of the Mexican Gulf.
It is vain to urge such dubious evidence as the fancied traces of a
mould-ridge, or the solitary example of a casting of uncertain age, in
proof of a knowledge of the furnace and the crucible among any
North American tribe. Everywhere in Europe the soil yields not only
its buried relics of gold, copper, and bronze, but also stone and
bronze moulds in which implements and personal ornaments were
cast. When the ingenious systematising of Danish archæologists had
familiarised the students of antiquity with the idea of a succession of
Stone, Bronze, and Iron periods in the history of Europe, the
question naturally followed, whether metallurgy did not begin, there,
as elsewhere, in the easy working of virgin copper. Dr. Latham
accordingly remarked, in his Ethnology of the British Islands, on the
supposition that no unalloyed copper relics had been found in
Britain: “Stone and bone first; then bronze, or copper and tin
combined; but no copper alone. I cannot get over this hiatus; cannot
imagine a metallurgic industry beginning with the use of alloys.” It
was a mistake, however, to assume that no copper relics had ever
been found. At first it had been taken for granted that all such
implements were of the familiar alloy. But so soon as the importance
of the distinction was recognised, examples of pure copper were
forthcoming. So early as 1822, Sir David Brewster described a large
axe of peculiar shape, and formed of copper, which was found in the
hard black till-clay at a depth of twenty feet under Ratho Bog, near
Edinburgh. This is no solitary example. The Scottish Museum of
Antiquities has other implements of pure copper; and Sir William
Wilde states in reference to the collections of the Royal Irish
Academy: “upon careful examination, it has been found that thirty of
the rudest, and apparently the very oldest celts, are of red, almost
unalloyed copper”; as is also the case with some other rudely
formed tools in the same collection.
It was a temporary advantage, doubtless, but a real loss, to the
Indian miners of Lake Superior that they found the native copper
there ready to hand, a pure ductile metal, probably regarded by
them as only a variety of stone which—unlike its rocky matrix,—they
could bend, or hammer into shape, without fracture. Its value as
such was widely appreciated. Copper tools, everywhere retaining the
specs, or larger crystals of silver, characteristic of the Lake Superior
veins, tell of the diffusion of the metal from that single source
throughout all the vast regions watered by the Mississippi and its
tributaries, and eastward by lake and river to the gulf of the St.
Lawrence and the mouth of the Hudson.
There was a time when this traffic must have been systematically
carried on; when the ancient miners of Lake Superior worked its rich
copper veins with industrious zeal; and when, probably as part of
the same aggressive energy, the valley of the Ohio was filling with a
settled population; its great earthworks were in process of
construction, and a native race entered on a course that gave
promise of social progress. But, from whatever cause, the work of
the old miners was abruptly terminated;[78] the race of the Mounds
vanished from the scenes of their ingenious toil; and rudest
barbarism resumed its sway over the whole northern continent. The
same Aryan race that, before the dawn of history, before the
Sanskrit-speaking people of India, or the Zends of Persia, entered on
their southern homes, spoke in its own cradleland the mother
tongue of Sanskrit, Greek, Celtic, and German, at length broke up
and went forth on its long wanderings. Whatever peoples it found
there; they were replaced by Celts, Romans, Greeks, Slavs, and
Teutons, who broke in upon the barbarism of prehistoric Europe;
displaced the older races, Allophylian, Neolithic, Iberian, Finnic, or by
whatever other name we may find it convenient to designate them;
but not without a considerable intermingling of the old blood with
that of the intruders. The sparsely settled continent gradually filled
up. Forests were cleared, swamps drained, rivers confined by
artificial banks and levées to their channels; and there grew up in
their new homes the Celtic, Classic, Slavic, and Teutonic tongues,
with all the varied culture and civilisation which they represent.
Agriculture, the special characteristic of the whole Aryan race,
flourished. They brought with them the cereals; and, with plenty, the
favoured race multiplied, till at length it has grown straitened within
the bounds of the old continent which it had made its own.
With the close of the fifteenth century one great cycle, that of
Europe’s mediæval era, came to an end; and then we trace the first
beginnings of that fresh scattering of the Aryan clan, and its new
western movement across the Ocean. It seems in a very striking
manner once more to repeat itself under our own eyes, as we look
abroad on the millions crowding from Europe and filling up the
western wilderness; hewing down the forests, cultivating the waste
prairies, and everywhere displacing the rude aborigines: yet here
also not without some interblending of the races; though the two
types, Aryan and pre-Aryan, meet under all the repellent influences
of high civilisation and the lowest barbarism. In the Canadian North-
West alone, the young province of Manitoba began its political
existence with a population of between 10,000 and 12,000 half-
breeds; in part at least, a hardy race of hunters and farmers; the
representatives of what is as certainly destined to constitute an
element in the new phases which the Aryan race already begins to
assume, under the diverse conditions of a new continent, as that
curious trace of Europe’s pre-Aryan people which attracted the
observant attention of Tacitus among the ancient Britons; and which
we are learning to recognise, with a new significance, as the
Melanochroi: the representatives of the old half-breed of Europe’s
prehistoric dawn.
[40] Gladstone, Juventus Mundi, pp. 474, 479.
[41] Indian Migrations, p. 24.
[42] Types of Mankind, p. 351.
[43] Early Man in Britain, p. 241.
[44] Ibid. p. 244.
[45] Alaska and its Resources, p. 237.
[46] Prehistoric Europe, p. 550.
[47] U.S. Geological Survey, 1872, p. 652. Report of
National Museum, 1887, p. 683, Fig. 11535.
[48] Report of National Museum, 1887, p. 678.
[49] Ancient Stone Implements of Great Britain, p. 140.
[50] Vide Prehistoric Man, 3d ed. vol. i. p. 180, Fig. 54.
[51] Report of the Peabody Museum, vol. ii., p. 38.
[52] Primitive Industry, p. 471.
[53] Ibid., p. 542.
[54] Primitive Industry, p. 547.
[55] Ibid., p. 545.
[56] Foster’s Prehistoric Races, p. 55.
[57] The Antiquity and Origin of the Trenton Gravel, p.
547.
[58] Primitive Industry, p. 481.
[59] Report of Washington National Museum, 1887-88,
pp. 677-702.
[60] Primitive Industry, p. 517.
[61] Indian Migrations as Evidence of Language (Horatio
Hale), p. 21.
[62] Indian Migrations, p. 22.
[63] Relations des Jésuites, 1660, p. 7. Quebec ed.
[64] Archæologia Americana, vol. ii.
[65] Brinton, Races and Peoples, p. 254.
[66] Bertram’s Travels through North and South Carolina,
Georgia, etc., 1791, p. 367.
[67] The League of the Iroquois, p. 2.
[68] Schoolcraft, History of the Indian Tribes, vol. ii., p.
78.
[69] Prehistoric Races of the United States, p. 259.
[70] Smithsonian Report, 1572, p. 353. The important
word not supplied here, it is obvious from the
context is absent by a mere typographical error.
[71] Ancient Monuments of the Mississippi Valley, vol. i.,
pp. 196-207.
[72] The North Americans of Antiquity, p. 95.
[73] Prehistoric Races of the United States, pp. 251-259.
[74] Antiquities of the Southern Indians, p. 225.
[75] Primitive Industry, pp. 411-422.
[76] Primæval Antiquities, p. 135.
[77] Prehistoric Annals of Scotland, 1st ed. 1851, p. 214;
2d ed. vol. i.
[78] Prehistoric Man, 3rd ed. vol. i., pp. 203-228.
V
ebookbell.com