SlideShare a Scribd company logo
8
Most read
9
Most read
12
Most read
23-Jun-18
Genetic Algorithms
2
Evolution
 Oversimplified description of how evolution works in biology:
 Organisms (animals or plants) produce a number of offspring
which are almost, but not entirely, like themselves
 Variation may be due to mutation (random changes)
 Variation may be due to sexual reproduction (offspring have some
characteristics from each parent)
 Some of these offspring may survive to produce offspring of their
own—some won’t
 The “better adapted” offspring are more likely to survive
 Over time, later generations become better and better adapted
 Genetic algorithms use this same process to “evolve” better
programs
3
Genotypes and phenotypes
 Genes are the basic “instructions” for building an
organism
 A chromosome is a sequence of genes
 Biologists distinguish between an organism’s genotype
(the genes and chromosomes) and its phenotype (what
the organism actually is like)
 Example: You might have genes to be tall, but never grow to
be tall for other reasons (such as poor diet)
 Similarly, “genes” may describe a possible solution to a
problem, without actually being the solution
4
The basic genetic algorithm
 Start with a large “population” of randomly generated
“attempted solutions” to a problem
 Repeatedly do the following:
 Evaluate each of the attempted solutions
 Keep a subset of these solutions (the “best” ones)
 Use these solutions to generate a new population
 Quit when you have a satisfactory solution (or you run out of time)
Conceptual Algorithm
6
A really simple example
 Suppose your “organisms” are 32-bit computer words
 You want a string in which all the bits are ones
 Here’s how you can do it:
 Create 100 randomly generated computer words
 Repeatedly do the following:
 Count the 1 bits in each word
 Exit if any of the words have all 32 bits set to 1
 Keep the ten words that have the most 1s (discard the rest)
 From each word, generate 9 new words as follows:
 Pick a random bit in the word and toggle (change) it
 Note that this procedure does not guarantee that the next
“generation” will have more 1 bits, but it’s likely
7
A more realistic example, part I
 Suppose you have a large number of (x, y) data points
 For example, (1.0, 4.1), (3.1, 9.5), (-5.2, 8.6), ...
 You would like to fit a polynomial (of up to degree 5) through
these data points
 That is, you want a formula y = ax5 + bx4 + cx3 + dx2 +ex + f that gives
you a reasonably good fit to the actual data
 Here’s the usual way to compute goodness of fit:
 Compute the sum of (actual y – predicted y)2 for all the data points
 The lowest sum represents the best fit
 There are some standard curve fitting techniques, but let’s assume
you don’t know about them
 You can use a genetic algorithm to find a “pretty good” solution
8
A more realistic example, part II
 Your formula is y = ax5 + bx4 + cx3 + dx2 +ex + f
 Your “genes” are a, b, c, d, e, and f
 Your “chromosome” is the array [a, b, c, d, e, f]
 Your evaluation function for one array is:
 For every actual data point (x, y), ( red to mean “actual data”)
 Compute ý = ax5 + bx4 + cx3 + dx2 +ex + f
 Find the sum of (y – ý)2 over all x
 The sum is your measure of “badness” (larger numbers are worse)
 Example: For [0, 0, 0, 2, 3, 5] and the data points (1, 12) and (2, 22):
 ý = 0x5 + 0x4 + 0x3 + 2x2 +3x + 5 is 2 + 3 + 5 = 10 when x is 1
 ý = 0x5 + 0x4 + 0x3 + 2x2 +3x + 5 is 8 + 6 + 5 = 19 when x is 2
 (12 – 10)2 + (22 – 19)2 = 22 + 32 = 13
 If these are the only two data points, the “badness” of [0, 0, 0, 2, 3, 5] is 13
9
A more realistic example, part III
 Your algorithm might be as follows:
 Create 100 six-element arrays of random numbers
 Repeat 500 times (or any other number):
 For each of the 100 arrays, compute its badness (using all data
points)
 Keep the ten best arrays (discard the other 90)
 From each array you keep, generate nine new arrays as
follows:
 Pick a random element of the six
 Pick a random floating-point number between 0.0 and 2.0
 Multiply the random element of the array by the random
floating-point number
 After all 500 trials, pick the best array as your final answer
10
Asexual vs. sexual reproduction
 In the examples so far,
 Each “organism” (or “solution”) had only one parent
 Reproduction was asexual (without sex)
 The only way to introduce variation was through mutation
(random changes)
 In sexual reproduction,
 Each “organism” (or “solution”) has two parents
 Assuming that each organism has just one chromosome, new
offspring are produced by forming a new chromosome from
parts of the chromosomes of each parent
11
The really simple example again
 Suppose your “organisms” are 32-bit computer words,
and you want a string in which all the bits are ones
 Here’s how you can do it:
 Create 100 randomly generated computer words
 Repeatedly do the following:
 Count the 1 bits in each word
 Exit if any of the words have all 32 bits set to 1
 Keep the ten words that have the most 1s (discard the rest)
 From each word, generate 9 new words as follows:
 Choose one of the other words
 Take the first half of this word and combine it with the
second half of the other word
12
The example continued
 Half from one, half from the other:
0110 1001 0100 1110 1010 1101 1011 0101
1101 0100 0101 1010 1011 0100 1010 0101
0110 1001 0100 1110 1011 0100 1010 0101
 Or we might choose “genes” (bits) randomly:
0110 1001 0100 1110 1010 1101 1011 0101
1101 0100 0101 1010 1011 0100 1010 0101
0100 0101 0100 1010 1010 1100 1011 0101
 Or we might consider a “gene” to be a larger unit:
0110 1001 0100 1110 1010 1101 1011 0101
1101 0100 0101 1010 1011 0100 1010 0101
1101 1001 0101 1010 1010 1101 1010 0101
13
Comparison of simple examples
 In the simple example (trying to get all 1s):
 The sexual (two-parent, no mutation) approach, if it succeeds,
is likely to succeed much faster
 Because up to half of the bits change each time, not just one bit
 However, with no mutation, it may not succeed at all
 By pure bad luck, maybe none of the first (randomly generated) words
have (say) bit 17 set to 1
 Then there is no way a 1 could ever occur in this position
 Another problem is lack of genetic diversity
 Maybe some of the first generation did have bit 17 set to 1, but
none of them were selected for the second generation
 The best technique in general turns out to be sexual
reproduction with a small probability of mutation
14
Curve fitting with sexual reproduction
 Your formula is y = ax5 + bx4 + cx3 + dx2 +ex + f
 Your “genes” are a, b, c, d, e, and f
 Your “chromosome” is the array [a, b, c, d, e, f]
 What’s the best way to combine two chromosomes into
one?
 You could choose the first half of one and the second half of
the other: [a, b, c, d, e, f]
 You could choose genes randomly: [a, b, c, d, e, f]
 You could compute “gene averages:”
[(a+a)/2, (b+b)/2, (c+c)/2, (d+d)/2, (e+e)/2,(f+f)/2]
 The last may be the best, though it is difficult to know of a good
biological analogy for it
Three main types of rules
 The genetic algorithm uses three main types of
rules at each step to create the next generation
from the current population:
• Selection rules select the individuals, called parents, that
contribute to the population at the next generation.
• Crossover rules combine two parents to form children
for the next generation.
• Mutation rules apply random changes to individual
parents to form children.
15
16
Directed evolution
 Notice that, in the previous examples, we formed the
child organisms randomly
 We did not try to choose the “best” genes from each parent
 This is how natural (biological) evolution works
 Biological evolution is not directed—there is no “goal”
 Genetic algorithms use biology as inspiration, not as a set of
rules to be slavishly followed
 For trying to get a word of all 1s, there is an obvious
measure of a “good” gene
 But that’s mostly because it’s a silly example
 It’s much harder to detect a “good gene” in the curve-fitting
problem, harder still in almost any “real use” of a genetic
algorithm
17
Probabilistic matching
 In previous examples, we chose the N “best” organisms
as parents for the next generation
 A more common approach is to choose parents
randomly, based on their measure of goodness
 Thus, an organism that is twice as “good” as another is likely
to have twice as many offspring
 This has a couple of advantages:
 The best organisms will contribute the most to the next
generation
 Since every organism has some chance of being a parent, there
is somewhat less loss of genetic diversity
18
Genetic programming
 A string of bits could represent a program
 If you want a program to do something, you might try to
evolve one
 As a concrete example, suppose you want a program to
help you choose stocks in the stock market
 There is a huge amount of data, going back many years
 What data has the most predictive value?
 What’s the best way to combine this data?
 A genetic program is possible in theory, but it might
take millions of years to evolve into something useful
 How can we improve this?
19
Concluding remarks
 Genetic algorithms are—
 Fun! They are enjoyable to program and to work with
 This is probably why they are a subject of active research
 Mind-bogglingly slow—you don’t want to use them if you
have any alternatives
 Good for a very few types of problems
 Genetic algorithms can sometimes come up with a solution when you
can see no other way of tackling the problem
20
The End

More Related Content

What's hot (20)

PDF
INTRODUCTION TO ALGORITHMS Third Edition
PHI Learning Pvt. Ltd.
 
PDF
Wordnet Introduction
University of Technology
 
PDF
Knowledge representation
Amit Kumar Rathi
 
PPTX
Data Analysis: Evaluation Metrics for Supervised Learning Models of Machine L...
Md. Main Uddin Rony
 
PPTX
COMPILER DESIGN
Vetukurivenkatashiva
 
PPTX
AI Programming language (LISP)
Dr. SURBHI SAROHA
 
PPTX
Chapter 6 agent communications--agent communications
farshad33
 
PPTX
Truth management system
Mohammad Kamrul Hasan
 
PDF
Introduction to the Semantic Web
Marin Dimitrov
 
PPTX
Deep learning lecture - part 1 (basics, CNN)
SungminYou
 
PPT
First order logic
Rushdi Shams
 
PDF
Introduction to Statistical Machine Learning
mahutte
 
PPTX
LISP: Introduction to lisp
DataminingTools Inc
 
PPTX
Introduction to Genetic Algorithms
Dr. C.V. Suresh Babu
 
PPT
Biconnected components (13024116056)
Akshay soni
 
PPT
Query Decomposition and data localization
Hafiz faiz
 
PPTX
Lecture 2 data structures and algorithms
Aakash deep Singhal
 
PPTX
A brief introduction to mutual information and its application
Hyun-hwan Jeong
 
PDF
Introduction to Machine Learning with SciKit-Learn
Benjamin Bengfort
 
INTRODUCTION TO ALGORITHMS Third Edition
PHI Learning Pvt. Ltd.
 
Wordnet Introduction
University of Technology
 
Knowledge representation
Amit Kumar Rathi
 
Data Analysis: Evaluation Metrics for Supervised Learning Models of Machine L...
Md. Main Uddin Rony
 
COMPILER DESIGN
Vetukurivenkatashiva
 
AI Programming language (LISP)
Dr. SURBHI SAROHA
 
Chapter 6 agent communications--agent communications
farshad33
 
Truth management system
Mohammad Kamrul Hasan
 
Introduction to the Semantic Web
Marin Dimitrov
 
Deep learning lecture - part 1 (basics, CNN)
SungminYou
 
First order logic
Rushdi Shams
 
Introduction to Statistical Machine Learning
mahutte
 
LISP: Introduction to lisp
DataminingTools Inc
 
Introduction to Genetic Algorithms
Dr. C.V. Suresh Babu
 
Biconnected components (13024116056)
Akshay soni
 
Query Decomposition and data localization
Hafiz faiz
 
Lecture 2 data structures and algorithms
Aakash deep Singhal
 
A brief introduction to mutual information and its application
Hyun-hwan Jeong
 
Introduction to Machine Learning with SciKit-Learn
Benjamin Bengfort
 

Similar to Lec 7 genetic algorithms (20)

PPT
45 genetic-algorithms
grskrishna
 
PPTX
Genetic algorithm
Syed Muhammad Zeejah Hashmi
 
PPT
Soft computing06
university of sargodha
 
PPTX
Genetic Algorithm
Jagadish Mohanty
 
PDF
Soft Computing- Dr. H.s. Hota 28.08.14.pdf
forsatyam9451
 
PPT
Evolutionary algorithms
M S Prasad
 
PPTX
Genetic algorithm optimization technique.pptx
sridharece1
 
PPT
Genetic algorithm
garima931
 
PPTX
Genetic Algorithm
Fatemeh Karimi
 
PPTX
Genetic Algorithm by Example
Nobal Niraula
 
PPT
0101.genetic algorithm
Ahmad Almubarrok
 
PPT
Genetic algorithm
DurgeshPratapSIngh8
 
PPT
Ff meeting 25nov03
moniajit
 
PDF
Data Science - Part XIV - Genetic Algorithms
Derek Kane
 
PPT
Genetic algorithms
guest9938738
 
PDF
generic optimization techniques lecture slides
SardarHamidullah
 
PPT
GeneticAlgorithm
guestfbf1e1
 
PPTX
Lec 06
Nilt1234
 
PDF
3_GO_Olesya_Genetic_AlgorithmsOPTIMZTION.p.pdf
ssuser8d8cfc1
 
PPTX
GA of a Paper 2012.pptx
waqasjavaid26
 
45 genetic-algorithms
grskrishna
 
Genetic algorithm
Syed Muhammad Zeejah Hashmi
 
Soft computing06
university of sargodha
 
Genetic Algorithm
Jagadish Mohanty
 
Soft Computing- Dr. H.s. Hota 28.08.14.pdf
forsatyam9451
 
Evolutionary algorithms
M S Prasad
 
Genetic algorithm optimization technique.pptx
sridharece1
 
Genetic algorithm
garima931
 
Genetic Algorithm
Fatemeh Karimi
 
Genetic Algorithm by Example
Nobal Niraula
 
0101.genetic algorithm
Ahmad Almubarrok
 
Genetic algorithm
DurgeshPratapSIngh8
 
Ff meeting 25nov03
moniajit
 
Data Science - Part XIV - Genetic Algorithms
Derek Kane
 
Genetic algorithms
guest9938738
 
generic optimization techniques lecture slides
SardarHamidullah
 
GeneticAlgorithm
guestfbf1e1
 
Lec 06
Nilt1234
 
3_GO_Olesya_Genetic_AlgorithmsOPTIMZTION.p.pdf
ssuser8d8cfc1
 
GA of a Paper 2012.pptx
waqasjavaid26
 
Ad

More from Eyob Seyfu (17)

PDF
Chapter Five - Relational Query Languages.pdf
Eyob Seyfu
 
PDF
Chapter Three - Entity Relationship Diagram (ERD).pdf
Eyob Seyfu
 
PDF
Chapter Two: Database System Concepts and Architectures
Eyob Seyfu
 
PDF
Chapter One: Introduction to Database Systems
Eyob Seyfu
 
PDF
Lec 9 conclusion
Eyob Seyfu
 
PDF
Lec 8 ai implementation
Eyob Seyfu
 
PDF
Lec 6 learning
Eyob Seyfu
 
PDF
Lec 5 uncertainty
Eyob Seyfu
 
PDF
Lec 4 expert systems
Eyob Seyfu
 
PDF
Lec 3 1 reasoning
Eyob Seyfu
 
PDF
Lec 3 knowledge acquisition representation and inference
Eyob Seyfu
 
PDF
Lec 2 1 informed search
Eyob Seyfu
 
PDF
Lec 2 agents
Eyob Seyfu
 
PDF
Lec 0 about the course
Eyob Seyfu
 
PDF
Lec 1 introduction
Eyob Seyfu
 
PPTX
Security and Privacy Challenges in Cloud Computing Environments
Eyob Seyfu
 
PPTX
A Survey on Wireless Mesh Networks (WMN)
Eyob Seyfu
 
Chapter Five - Relational Query Languages.pdf
Eyob Seyfu
 
Chapter Three - Entity Relationship Diagram (ERD).pdf
Eyob Seyfu
 
Chapter Two: Database System Concepts and Architectures
Eyob Seyfu
 
Chapter One: Introduction to Database Systems
Eyob Seyfu
 
Lec 9 conclusion
Eyob Seyfu
 
Lec 8 ai implementation
Eyob Seyfu
 
Lec 6 learning
Eyob Seyfu
 
Lec 5 uncertainty
Eyob Seyfu
 
Lec 4 expert systems
Eyob Seyfu
 
Lec 3 1 reasoning
Eyob Seyfu
 
Lec 3 knowledge acquisition representation and inference
Eyob Seyfu
 
Lec 2 1 informed search
Eyob Seyfu
 
Lec 2 agents
Eyob Seyfu
 
Lec 0 about the course
Eyob Seyfu
 
Lec 1 introduction
Eyob Seyfu
 
Security and Privacy Challenges in Cloud Computing Environments
Eyob Seyfu
 
A Survey on Wireless Mesh Networks (WMN)
Eyob Seyfu
 
Ad

Recently uploaded (20)

PDF
Module - 5 Machine Learning-22ISE62.pdf
Dr. Shivashankar
 
PDF
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
PPTX
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
PPTX
Explore USA’s Best Structural And Non Structural Steel Detailing
Silicon Engineering Consultants LLC
 
PPTX
Unit_I Functional Units, Instruction Sets.pptx
logaprakash9
 
PDF
Bayesian Learning - Naive Bayes Algorithm
Sharmila Chidaravalli
 
PPTX
darshai cross section and river section analysis
muk7971
 
PDF
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
PPTX
Seminar Description: YOLO v1 (You Only Look Once).pptx
abhijithpramod20002
 
PDF
Submit Your Papers-International Journal on Cybernetics & Informatics ( IJCI)
IJCI JOURNAL
 
PDF
WD2(I)-RFQ-GW-1415_ Shifting and Filling of Sand in the Pond at the WD5 Area_...
ShahadathHossain23
 
PPTX
Distribution reservoir and service storage pptx
dhanashree78
 
PPTX
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
PDF
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
PPTX
Introduction to Internal Combustion Engines - Types, Working and Camparison.pptx
UtkarshPatil98
 
PDF
Clustering Algorithms - Kmeans,Min ALgorithm
Sharmila Chidaravalli
 
PPTX
MODULE 03 - CLOUD COMPUTING AND SECURITY.pptx
Alvas Institute of Engineering and technology, Moodabidri
 
PDF
Artificial Neural Network-Types,Perceptron,Problems
Sharmila Chidaravalli
 
PPTX
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
PDF
A Brief Introduction About Robert Paul Hardee
Robert Paul Hardee
 
Module - 5 Machine Learning-22ISE62.pdf
Dr. Shivashankar
 
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
Explore USA’s Best Structural And Non Structural Steel Detailing
Silicon Engineering Consultants LLC
 
Unit_I Functional Units, Instruction Sets.pptx
logaprakash9
 
Bayesian Learning - Naive Bayes Algorithm
Sharmila Chidaravalli
 
darshai cross section and river section analysis
muk7971
 
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
Seminar Description: YOLO v1 (You Only Look Once).pptx
abhijithpramod20002
 
Submit Your Papers-International Journal on Cybernetics & Informatics ( IJCI)
IJCI JOURNAL
 
WD2(I)-RFQ-GW-1415_ Shifting and Filling of Sand in the Pond at the WD5 Area_...
ShahadathHossain23
 
Distribution reservoir and service storage pptx
dhanashree78
 
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
Introduction to Internal Combustion Engines - Types, Working and Camparison.pptx
UtkarshPatil98
 
Clustering Algorithms - Kmeans,Min ALgorithm
Sharmila Chidaravalli
 
MODULE 03 - CLOUD COMPUTING AND SECURITY.pptx
Alvas Institute of Engineering and technology, Moodabidri
 
Artificial Neural Network-Types,Perceptron,Problems
Sharmila Chidaravalli
 
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
A Brief Introduction About Robert Paul Hardee
Robert Paul Hardee
 

Lec 7 genetic algorithms

  • 2. 2 Evolution  Oversimplified description of how evolution works in biology:  Organisms (animals or plants) produce a number of offspring which are almost, but not entirely, like themselves  Variation may be due to mutation (random changes)  Variation may be due to sexual reproduction (offspring have some characteristics from each parent)  Some of these offspring may survive to produce offspring of their own—some won’t  The “better adapted” offspring are more likely to survive  Over time, later generations become better and better adapted  Genetic algorithms use this same process to “evolve” better programs
  • 3. 3 Genotypes and phenotypes  Genes are the basic “instructions” for building an organism  A chromosome is a sequence of genes  Biologists distinguish between an organism’s genotype (the genes and chromosomes) and its phenotype (what the organism actually is like)  Example: You might have genes to be tall, but never grow to be tall for other reasons (such as poor diet)  Similarly, “genes” may describe a possible solution to a problem, without actually being the solution
  • 4. 4 The basic genetic algorithm  Start with a large “population” of randomly generated “attempted solutions” to a problem  Repeatedly do the following:  Evaluate each of the attempted solutions  Keep a subset of these solutions (the “best” ones)  Use these solutions to generate a new population  Quit when you have a satisfactory solution (or you run out of time)
  • 6. 6 A really simple example  Suppose your “organisms” are 32-bit computer words  You want a string in which all the bits are ones  Here’s how you can do it:  Create 100 randomly generated computer words  Repeatedly do the following:  Count the 1 bits in each word  Exit if any of the words have all 32 bits set to 1  Keep the ten words that have the most 1s (discard the rest)  From each word, generate 9 new words as follows:  Pick a random bit in the word and toggle (change) it  Note that this procedure does not guarantee that the next “generation” will have more 1 bits, but it’s likely
  • 7. 7 A more realistic example, part I  Suppose you have a large number of (x, y) data points  For example, (1.0, 4.1), (3.1, 9.5), (-5.2, 8.6), ...  You would like to fit a polynomial (of up to degree 5) through these data points  That is, you want a formula y = ax5 + bx4 + cx3 + dx2 +ex + f that gives you a reasonably good fit to the actual data  Here’s the usual way to compute goodness of fit:  Compute the sum of (actual y – predicted y)2 for all the data points  The lowest sum represents the best fit  There are some standard curve fitting techniques, but let’s assume you don’t know about them  You can use a genetic algorithm to find a “pretty good” solution
  • 8. 8 A more realistic example, part II  Your formula is y = ax5 + bx4 + cx3 + dx2 +ex + f  Your “genes” are a, b, c, d, e, and f  Your “chromosome” is the array [a, b, c, d, e, f]  Your evaluation function for one array is:  For every actual data point (x, y), ( red to mean “actual data”)  Compute ý = ax5 + bx4 + cx3 + dx2 +ex + f  Find the sum of (y – ý)2 over all x  The sum is your measure of “badness” (larger numbers are worse)  Example: For [0, 0, 0, 2, 3, 5] and the data points (1, 12) and (2, 22):  ý = 0x5 + 0x4 + 0x3 + 2x2 +3x + 5 is 2 + 3 + 5 = 10 when x is 1  ý = 0x5 + 0x4 + 0x3 + 2x2 +3x + 5 is 8 + 6 + 5 = 19 when x is 2  (12 – 10)2 + (22 – 19)2 = 22 + 32 = 13  If these are the only two data points, the “badness” of [0, 0, 0, 2, 3, 5] is 13
  • 9. 9 A more realistic example, part III  Your algorithm might be as follows:  Create 100 six-element arrays of random numbers  Repeat 500 times (or any other number):  For each of the 100 arrays, compute its badness (using all data points)  Keep the ten best arrays (discard the other 90)  From each array you keep, generate nine new arrays as follows:  Pick a random element of the six  Pick a random floating-point number between 0.0 and 2.0  Multiply the random element of the array by the random floating-point number  After all 500 trials, pick the best array as your final answer
  • 10. 10 Asexual vs. sexual reproduction  In the examples so far,  Each “organism” (or “solution”) had only one parent  Reproduction was asexual (without sex)  The only way to introduce variation was through mutation (random changes)  In sexual reproduction,  Each “organism” (or “solution”) has two parents  Assuming that each organism has just one chromosome, new offspring are produced by forming a new chromosome from parts of the chromosomes of each parent
  • 11. 11 The really simple example again  Suppose your “organisms” are 32-bit computer words, and you want a string in which all the bits are ones  Here’s how you can do it:  Create 100 randomly generated computer words  Repeatedly do the following:  Count the 1 bits in each word  Exit if any of the words have all 32 bits set to 1  Keep the ten words that have the most 1s (discard the rest)  From each word, generate 9 new words as follows:  Choose one of the other words  Take the first half of this word and combine it with the second half of the other word
  • 12. 12 The example continued  Half from one, half from the other: 0110 1001 0100 1110 1010 1101 1011 0101 1101 0100 0101 1010 1011 0100 1010 0101 0110 1001 0100 1110 1011 0100 1010 0101  Or we might choose “genes” (bits) randomly: 0110 1001 0100 1110 1010 1101 1011 0101 1101 0100 0101 1010 1011 0100 1010 0101 0100 0101 0100 1010 1010 1100 1011 0101  Or we might consider a “gene” to be a larger unit: 0110 1001 0100 1110 1010 1101 1011 0101 1101 0100 0101 1010 1011 0100 1010 0101 1101 1001 0101 1010 1010 1101 1010 0101
  • 13. 13 Comparison of simple examples  In the simple example (trying to get all 1s):  The sexual (two-parent, no mutation) approach, if it succeeds, is likely to succeed much faster  Because up to half of the bits change each time, not just one bit  However, with no mutation, it may not succeed at all  By pure bad luck, maybe none of the first (randomly generated) words have (say) bit 17 set to 1  Then there is no way a 1 could ever occur in this position  Another problem is lack of genetic diversity  Maybe some of the first generation did have bit 17 set to 1, but none of them were selected for the second generation  The best technique in general turns out to be sexual reproduction with a small probability of mutation
  • 14. 14 Curve fitting with sexual reproduction  Your formula is y = ax5 + bx4 + cx3 + dx2 +ex + f  Your “genes” are a, b, c, d, e, and f  Your “chromosome” is the array [a, b, c, d, e, f]  What’s the best way to combine two chromosomes into one?  You could choose the first half of one and the second half of the other: [a, b, c, d, e, f]  You could choose genes randomly: [a, b, c, d, e, f]  You could compute “gene averages:” [(a+a)/2, (b+b)/2, (c+c)/2, (d+d)/2, (e+e)/2,(f+f)/2]  The last may be the best, though it is difficult to know of a good biological analogy for it
  • 15. Three main types of rules  The genetic algorithm uses three main types of rules at each step to create the next generation from the current population: • Selection rules select the individuals, called parents, that contribute to the population at the next generation. • Crossover rules combine two parents to form children for the next generation. • Mutation rules apply random changes to individual parents to form children. 15
  • 16. 16 Directed evolution  Notice that, in the previous examples, we formed the child organisms randomly  We did not try to choose the “best” genes from each parent  This is how natural (biological) evolution works  Biological evolution is not directed—there is no “goal”  Genetic algorithms use biology as inspiration, not as a set of rules to be slavishly followed  For trying to get a word of all 1s, there is an obvious measure of a “good” gene  But that’s mostly because it’s a silly example  It’s much harder to detect a “good gene” in the curve-fitting problem, harder still in almost any “real use” of a genetic algorithm
  • 17. 17 Probabilistic matching  In previous examples, we chose the N “best” organisms as parents for the next generation  A more common approach is to choose parents randomly, based on their measure of goodness  Thus, an organism that is twice as “good” as another is likely to have twice as many offspring  This has a couple of advantages:  The best organisms will contribute the most to the next generation  Since every organism has some chance of being a parent, there is somewhat less loss of genetic diversity
  • 18. 18 Genetic programming  A string of bits could represent a program  If you want a program to do something, you might try to evolve one  As a concrete example, suppose you want a program to help you choose stocks in the stock market  There is a huge amount of data, going back many years  What data has the most predictive value?  What’s the best way to combine this data?  A genetic program is possible in theory, but it might take millions of years to evolve into something useful  How can we improve this?
  • 19. 19 Concluding remarks  Genetic algorithms are—  Fun! They are enjoyable to program and to work with  This is probably why they are a subject of active research  Mind-bogglingly slow—you don’t want to use them if you have any alternatives  Good for a very few types of problems  Genetic algorithms can sometimes come up with a solution when you can see no other way of tackling the problem