Adaptive Learning Search A New Tool To Help Compre
Adaptive Learning Search A New Tool To Help Compre
net/publication/220160474
CITATIONS READS
19 144
4 authors, including:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Patrick Siarry on 01 June 2014.
JOHANN DRÉO∗
JEAN-PHILIPPE AUMASSON†
WALID TFAILI∗
PATRICK SIARRY∗
∗ Université Paris XII Val-de-Marne
Laboratoire Images, Signaux et Systèmes Intelligents (LISSI, EA 3956)
61, avenue du Général de Gaulle, 94010 Créteil, France
{dreo, siarry, tfaili}@univ-paris12.fr
† Security and Cryptography Laboratory, EPFL, Switzerland
The majority of the algorithms used to solve hard optimization problems today are pop-
ulation metaheuristics. These methods are often presented under a purely algorithmic
angle, while insisting on the metaphors which led to their design. We propose in this
article to regard population metaheuristics as methods making evolution a probabilistic
sampling of the objective function, either explicitly, implicitly, or directly, via processes
of learning, diversification, and intensification. We present a synthesis of some meta-
heuristics and their functioning seen under this angle, called Adaptive Learning Search.
We discuss how to design metaheuristics following this approach, and propose an imple-
mentation with our Open Metaheuristics framework, along with concrete examples.
1. Introduction
Optimization problems appear in many fields, as various as identification prob-
lems, supervised learning of neural networks, shortest path problems, etc.
Metaheuristics18 are a family of stochastic optimization algorithms, often applied
to hard combinatorial problems for which no more efficient method is known. They
have the advantage of being generic methods, thus do not require a complex tuning
for each problem, and can be used as a kind of “black boxes”. Recall that, gen-
erally, optimization algorithms search for a point into the search space, so as to
optimize (i.e., minimize or maximize) the objective function (also called fitness or
goal function). Metaheuristics are often divided into two sets:
1
2 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
2. Fundamental concepts
Metaheuristics share a certain number of properties. An essential one is that they
handle a sampling of the objective function, via common processes. The framework
of adaptive memory programming aims at stressing those processes.
The probabilistic sampling should ideally pick the best solutions with higher
probability. However, in an optimization problem, the effective goal is not to sam-
ple the objective function, but to find the distribution’s optimum. Thus, sampling
must concentrate on the areas of interest, while converging gradually towards the
optimum by means of algorithms. From the point of view of sampling, this conver-
gence is carried out by a progressive fall of dispersion in these areas.
IDEA uses a more general diversification than PSM, while not being limited to a
probability vector as model, but specifying that the search for the best probability
distribution forms an integral part of the algorithm. However, the fall of dispersion
is carried out by selecting the best individuals, no precision on the use of different
intensification principles is given.
mization process. We emphasize on the fact that the memorized data is not only a
raw input, but provides information on the distribution, and thus on the solutions.
Thereby, we propose to consider three terms to describe the main steps in a
population metaheuristic: learning, diversification and intensification, with respect
to a sampling either explicit, implicit, or direct. An ALS algorithm is thus organized
as presented in Algorithm 4.
3.2. Examples
We present several famous metaheuristics under the scope of ALS.
End
End
The main difficulty is how to estimate the distribution; the algorithms used for
this are based on an evaluation of the dependency of the variables, and can belong
to three different categories:
(1) Models without any dependency: the probability distribution is factorized from
univariant independent distributions, over each dimension. That choice has the
defect not to be realistic in case of hard optimization, where a dependency
between variables is often the rule.
(2) Models with bivariant dependency: the probability distribution is factorized
from bivariant distributions. In this case, the learning of distribution can be
extended to the notion of structure.
(3) Models with multiple dependencies: the factorization of the probability distri-
bution is obtained from statistics with an order higher than two.
8 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
For continuous problems, the distribution model is often based on a normal distri-
bution.
Some important variants were proposed, using for example “data clustering”
for multimodal optimization, parallel variants for discrete problems (see37 ). Con-
vergence theorems were also formulated, in particular with modeling by Markov
chains, or dynamic systems.
We model the EDA algorithms in the ALS scope in Algorithm 9.
Fig. 1. UML diagram of the template pattern, used to implement a metaheuristic as an ALS.
Algorithm 10 Implementation of the start method for the metaheuristic base class.
void itsMetaheuristic::start() {
// an initialization step before performing the optimization
initialization();
// while no stopping criterion reached
while( !isStoppingCriteria() ) {
learning(); // Learning phase
outputSample();
diversification(); // Diversification phase
outputSample();
intensification(); // Intensification phase
outputSample();
// one more iteration
iterationsCurrent++;
}
// an ending step, if necessary
end();
}
void itsRandom::diversification() {
// draw each point in an hyper cube
for( unsigned int i=0; i < getSampleSize(); i++) {
itsPoint p;
p.setSolution(
randomUniform(
getProblem()->boundsMinima(),
getProblem()->boundsMaxima()
)
);
setSample[i]( evaluate(p) ); // call for the problem
}
}
(2) At the diversification step, one draws a new sample according to the parameters.
(3) The intensification step selects the best points.
the same interface for all. One only has to define a subclass of the metaheuristic
factory for each metaheuristic that we program (see Figure 4). As a result, one
has not to specify the type of objects wanted, it even remains unknown, and the
programmer can easily change it by modifying the factory creating method.
Fig. 4. UML diagram of the abstract factory pattern, used to manipulate metaheuristics imple-
mentations.
14 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
iteration is thus crucial. In the ALS approach, this state is illustrated by the sample,
modified at each step of an iteration.
To achieve this goal, a structured output format is useful. We proposed an
XML (eXtensible Markup Language) file format, gathering all information of an
optimization session (see Figure 5). The advantages of such an approach is that one
can easily extract information from the files. As XML is extensible, one can also add
a new representation of information. For example, one can add a specific structure
to represent a solution (like a tree or a more complex representation) or add the
values of the metaheuristics parameters at each iteration (when the algorithm is
dynamic, for example). In a XML document, this unpredicted addition will not
disturb the possibility of extracting other information.
<optimization>
<iteration id="0">
<step class="start">
<sample>
<point><values>0.705915</values><solution>0.840188</solution></point>
<point><values>0.155538</values><solution>0.394383</solution></point>
</sample>
</step>
<evaluations>4</evaluations>
</iteration>
<iteration id="1">
<step class="learning">
<sample>
<point><values>0.705915</values><solution>0.840188</solution></point>
<point><values>0.155538</values><solution>0.394383</solution></point>
</sample>
</step>
<step class="diversification">
<sample>
<point><values>0.0771588</values><solution>0.277775</solution></point>
</sample>
</step>
<step class="intensification">
<sample>
<point><values>0.0348113</values><solution>0.186578</solution></point>
</sample>
</step>
<evaluations>7</evaluations>
</iteration>
<optimum>
<point><values>0.0348113</values><solution>0.186578</solution></point>
</optimum>
</optimization>
Fig. 5. Simple example of the oMetah output format for the optimization results. The algorithm
used is a simple EDA, using a sample of 2 points and optimizing the well known Sphere continuous
problem, in one dimension.
Open Metaheuristics also proposes a set of tags to include information about the
16 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
algorithm and the problem used to generate the results. This information, included
in the header of the XML file, are especially useful for verifying and reproducing
experiments.
Algorithm 14 Example of a test session for oMetah Lab. The script runs 10 times
an EDA and a random algorithm, on the Rosenbrock problem with 2 variables, and
outputs all the available graphics in a HTML report.
import ometahtest
import ometahstats
path = ’./ometah’
# Number of runs to calculate
runs = 10
u = ometahtest.Test(path,’-s 10 -i 10 -e 100 -p Rosenbrock -d 2 -m
CEDA’,runs)
u.start()
v = ometahtest.Test(path,’-s 10 -i 10 -e 100 -p Rosenbrock -d 2 -m
RA’,runs)
v.start()
# calculate all graphics and generate a HTML report
ometahstats.process(paths, ’all’,’html’)
20
10
Optima value
5
2
1
1 2 3 4 5 6 7 8 9 10 11
Iteration index
Fig. 6. Distribution of the sample on the values space for an Estimation of Distribution
Algorithm6 optimizing the Rosenbrock objective function13 with two dimensions. The distribu-
tions are obtained for each iteration, at the intensification step, through 50 runs, with a sample of
10 points, limited by 100 evaluations of the problem. Box plots are showing the median, quartiles,
minimum and maximum, the whiskers size is limited at 1.5 times the inter-quartiles distance.
4.5.4. Application
Open Metaheuristics has been successfully used to solve an optimization problem
in medical imaging17,16 : we employed optimization techniques for registration of
retinal angiograms (registration is an important tool for solving many medical image
analysis problems). Many common minimization strategies have been applied to
image registration problems48,32 , but metaheuristics have shown interesting results,
especially for high resolution difficult registration problems.
The key feature, in this application, was the ability to implement the problem in
a very simple manner, using the template pattern. Indeed, the separation between
metaheuristics, communication layer and problems permits to focus on the imple-
mentation of a single class, all the other processes being managed by the framework.
The implementation of the imaging part was tackled with the CImg54 template
library. Algorithm 15 shows the skeleton of the problem header. The main method
is the virtual objectiveFunction, which takes a point as an argument, and re-
turns the same point, with its value updated. The mandatory attributes, fixing the
problem bounds, are set up by the constructor.
Once the problem implemented, it is easy to optimize it with the available meta-
heuristics, and benefit of the automated data processing. Figure 9 shows examples
of graphics directly produced by Ometah. The source code is available online, in
the registration module of the Open Metaheuristic project15 , and the details of the
problem are presented in two publications17,16 .
18 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
Fig. 7. Distribution of the sample on the values space for an Estimation of Distribution Algorithm
(CEDA) and a Random Search (RA) optimizing the Rosenbrock objective function with two
dimensions. The distributions are obtained for each iteration, at the intensification step, through
50 runs, with a sample of 10 points, limited by 100 evaluations of the problem.
5. Conclusion
We have shown that population metaheuristics can be viewed as algorithms handling
a probabilistic sampling of a probability distribution, representing the objective
Adaptive learning search, a new tool to help comprehending metaheuristics 19
0.2
0.1
0.0
−0.1
densi
ty
−0.2
Y
d −0.3
−0.4
−0.5
Fig. 8. Distribution on the solutions space of the optima found by a simple Random Search on
the Rosenbrock problem with two dimensions. The optima are obtained through 50 runs, with a
sample of 10 points, limited by 100 evaluations of the problem. The density is estimated with a
two-dimensional kernel density method, with an axis-aligned bivariate normal kernel.
References
1. E. H. L. Aarts and P. J. M. Van Laarhoven. Statistical cooling : a general approach to
combinatorial optimisation problems. Philips Journal of Research, 40:193–226, 1985.
2. S. Baluja. Population-based Incremental Learning: A Method for Integrating Genetic
Search Based Function Optimization and Competitive Learning. Technical Report
CMU-CS-94-163, Carnegie Mellon University, 1994.
3. S. Baluja. Genetic Algorithms and Explicit Search Statistics. Advances in Neural
Information Processing Systems, 9:319–325, 1997.
4. S. Baluja and R. Caruana. Removing the Genetics from the Standard Genetic Algo-
20 Johann DRÉO, Jean-Philippe AUMASSON, Walid TFAILI, Patrick SIARRY
0
0 1 2 3 4 5 6 7 8
Image
Optima distribution
10
8
6
Frequency
4
2
0
Value
Fig. 9. Example of results obtained with the implementation of the registration problem opti-
mized by metaheuristics in the oMetah framework.
29. J. H. Holland. Outline for logical theory of adaptive systems. J. Assoc. Comput. Mach.,
3:297–314, 1962.
30. K. Hukushima and K Nemoto. Exchange Monte Carlo method and application to spin
glass simulations. Journal of the Physical Society of Japan, 65:1604–1608, 1996.
31. Y. Iba. Population Annealing: An approach to finite-temperature calculation. In Joint
Workshop of Hayashibara Foundation and SMAPIP. Hayashibara Forum, 2003.
32. M. Jenkinson and S. Smith. A global optimisation method for robust affine registration
of brain images. Medical Image Analysis, 5:143–156, 2001.
33. M. Jones. An Object-Oriented Framework for the Implementation of Search Tech-
niques. PhD thesis, University of East Anglia, 2000.
34. Maarten Keijzer, J. J. Merelo, G. Romero, and Marc Schoenauer. Evolving objects:
A general purpose evolutionary computation library. In Artificial Evolution, pages
231–244, 2001.
35. J. Kertesz and I. Kondor, editors. Advances in Computer Simulation, chapter Intro-
duction To Monte Carlo Algorithms. Springer-Verlag, 1998.
36. S. Kirkpatrick, C.D. Gelatt, and M.P. Vecchi. Optimization by simulated annealing.
Science, 220(4598):671–680, 1983.
37. P. Larrañaga and J.A. Lozano, editors. Estimation of Distribution Algorithms, A New
Tool for Evolutionary Computation. Genetic Algorithms and Evolutionary Computa-
tion. Kluwer Academic Publishers, 2002.
38. F. Liang and W. H. Wong. Evolutionary Monte Carlo: Application to Cp Model Sam-
pling and Change Point Theorem. Statistica Sinica, 10, 2000.
39. N. Metropolis, A. W. Rosenbluth, M. N. Rosenbluth, A. H. Teller, and E. Teller. Equa-
tion of state calculations by fast computing machines. Journal of Chemical Physics,
21:1087–1092, 1953.
40. N. Monmarché, E. Ramat, G. Dromel, M. Slimane, and G. Venturini. On the similar-
ities between AS, BSC and PBIL: toward the birth of a new meta-heuristics. E3i 215,
Université de Tours, 1999.
41. N. Monmarché, N. Ramat, L. Desbarat, and G. Venturini. Probabilistic search with
genetic algorithms and ant colonies. In A.S. Wu, editor, Proceedings of the 2000 Ge-
netic and Evolutionary Computation Conference Workshop, pages 209–211, 2000.
42. H. Mühlenbein and G. Paaß. From recombination of genes to the estimation of dis-
tributions I. Binary parameters. Lecture Notes in Computer Science 1411: Parallel
Problem Solving from Nature, PPSN IV:178–187, 1996.
43. J. A. Nelder and R. Mead. A simplex method for function minimization. Computer
Journal, 7:308–313, 1965.
44. M. E. J. Newman and R. G. Palmer. Error estimation in the histogram Monte Carlo
method. arxiv:cond-mat/98043006, 1998.
45. Y Okamoto and U. H. E. Hansmann. Thermodynamics of helix-coil transitions studied
by multicanonical algorithms. Journal Physical Chemistry, 99:11276–11287, 1995.
46. I. Rechenberg. Cybernetic Solution Path of an Experimental Problem. Royal Aircraft
Establishment Library Translation, 1965.
47. M.G.C. Resende. Greedy randomized adaptive search procedures (GRASP). Technical
Report TR 98.41.1, AT&T Labs-Research, 2000.
48. N. Ritter, R. Owens, J. Cooper, R. H. Eikelboom, and P. P. V. Saarloos. Registration
of stereo and temporal images of the retina. IEEE Trans. On Medical Imaging, 18:404–
418, 1999.
49. J. Rumbaugh, I. Jacobson, and G. Booch. The Unified Modeling Language Reference
Manual. Addison-Wesley, 1999.
50. G. Syswerda. Simulated Crossover in Genetic Algorithms. In L. D. Whitley, editor,
View publication stats