Lecture 6: The Building Block Hypothesis. Genetic Algorithms and Genetic Programming Lecture 6. The Schema Theorem Reminder

Similar documents
Genetic Algorithms and Genetic Programming Lecture 7

Automated Test Data Generation and Optimization Scheme Using Genetic Algorithm

Similarity Templates or Schemata. CS 571 Evolutionary Computation

Evolutionary Computation Part 2

When Will a Genetic Algorithm Outperform Hill-Climbing?

Chapter 14 Global Search Algorithms

Adaptive Crossover in Genetic Algorithms Using Statistics Mechanism

Non-deterministic Search techniques. Emma Hart

Genetic Algorithms Variations and Implementation Issues

Hybridization EVOLUTIONARY COMPUTING. Reasons for Hybridization - 1. Naming. Reasons for Hybridization - 3. Reasons for Hybridization - 2

Heuristic Optimisation

Single Candidate Methods

Artificial Intelligence Application (Genetic Algorithm)

Introduction to Genetic Algorithms. Based on Chapter 10 of Marsland Chapter 9 of Mitchell

Improved Leakage Model Based on Genetic Algorithm

Intelligent Reduction of Tire Noise

CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS

Genetic Algorithm Performance with Different Selection Methods in Solving Multi-Objective Network Design Problem

Outline of the module

MAXIMUM LIKELIHOOD ESTIMATION USING ACCELERATED GENETIC ALGORITHMS

Local Search. CS 486/686: Introduction to Artificial Intelligence Winter 2016

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

Machine Learning for Software Engineering

Chapter 9: Genetic Algorithms

CHAPTER 4 GENETIC ALGORITHM

Genetic Algorithms and Genetic Programming. Lecture 9: (23/10/09)

Local Search. CS 486/686: Introduction to Artificial Intelligence

Artificial Intelligence

Using Genetic Algorithms to optimize ACS-TSP

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

Citation for published version (APA): Lankhorst, M. M. (1996). Genetic algorithms in data analysis Groningen: s.n.

CS5401 FS2015 Exam 1 Key

CHAPTER 6 ORTHOGONAL PARTICLE SWARM OPTIMIZATION

Introduction to Artificial Intelligence 2 nd semester 2016/2017. Chapter 4: Beyond Classical Search

The Genetic Algorithm for finding the maxima of single-variable functions

Lecture: Iterative Search Methods

ARTIFICIAL INTELLIGENCE (CSCU9YE ) LECTURE 5: EVOLUTIONARY ALGORITHMS

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

A Performance Analysis of Compressed Compact Genetic Algorithm

Analysis of recombinative algorithms on a non-separable building-block problem

AI Programming CS S-08 Local Search / Genetic Algorithms

Binary Representations of Integers and the Performance of Selectorecombinative Genetic Algorithms

Heuristic Optimisation

Ar#ficial)Intelligence!!

Genetic Algorithms. Chapter 3

GENETIC ALGORITHM with Hands-On exercise

Suppose you have a problem You don t know how to solve it What can you do? Can you use a computer to somehow find a solution for you?

The Binary Genetic Algorithm. Universidad de los Andes-CODENSA

[Premalatha, 4(5): May, 2015] ISSN: (I2OR), Publication Impact Factor: (ISRA), Journal Impact Factor: 2.114

Introduction to Genetic Algorithms

Random Search Report An objective look at random search performance for 4 problem sets

A Parallel Evolutionary Algorithm for Discovery of Decision Rules

What is GOSET? GOSET stands for Genetic Optimization System Engineering Tool

An Empirical Investigation of Meta-heuristic and Heuristic Algorithms for a 2D Packing Problem

Local Search and Optimization Chapter 4. Mausam (Based on slides of Padhraic Smyth, Stuart Russell, Rao Kambhampati, Raj Rao, Dan Weld )

arxiv:cs/ v1 [cs.ne] 6 Apr 2001

REAL-CODED GENETIC ALGORITHMS CONSTRAINED OPTIMIZATION. Nedim TUTKUN

An Adaptive Hybrid Genetic-Annealing Approach

Hyperplane Ranking in. Simple Genetic Algorithms. D. Whitley, K. Mathias, and L. Pyeatt. Department of Computer Science. Colorado State University

Fall 09, Homework 5

Foundations of AI. 4. Informed Search Methods. Heuristics, Local Search Methods, Genetic Algorithms. Wolfram Burgard & Bernhard Nebel

DETERMINING MAXIMUM/MINIMUM VALUES FOR TWO- DIMENTIONAL MATHMATICLE FUNCTIONS USING RANDOM CREOSSOVER TECHNIQUES

Search Algorithms for Regression Test Suite Minimisation

Exploration vs. Exploitation in Differential Evolution

Foundations of AI. 4. Informed Search Methods. Heuristics, Local Search Methods, Genetic Algorithms

Escaping Local Optima: Genetic Algorithm

Outline. Motivation. Introduction of GAs. Genetic Algorithm 9/7/2017. Motivation Genetic algorithms An illustrative example Hypothesis space search

Aero-engine PID parameters Optimization based on Adaptive Genetic Algorithm. Yinling Wang, Huacong Li

Partitioning Sets with Genetic Algorithms

Introduction to Optimization

Introduction to Optimization

Artificial Intelligence

Artificial Intelligence

Evolving Genotype to Phenotype Mappings with a Multiple-Chromosome Genetic Algorithm

TDDC17. Intuitions behind heuristic search. Recall Uniform-Cost Search. Best-First Search. f(n) =... + h(n) g(n) = cost of path from root node to n

INF Biologically inspired computing Lecture 1: Marsland chapter 9.1, Optimization and Search Jim Tørresen

Mutation in Compressed Encoding in Estimation of Distribution Algorithm

Planning and Search. Genetic algorithms. Genetic algorithms 1

3.6.2 Generating admissible heuristics from relaxed problems

Evolutionary Algorithms. CS Evolutionary Algorithms 1

Job Shop Scheduling Problem (JSSP) Genetic Algorithms Critical Block and DG distance Neighbourhood Search

TDDC17. Intuitions behind heuristic search. Best-First Search. Recall Uniform-Cost Search. f(n) =... + h(n) g(n) = cost of path from root node to n

Genetic Algorithms for Real Parameter Optimization

Sparse Matrices Reordering using Evolutionary Algorithms: A Seeded Approach

Artificial Intelligence

CHAPTER 4 FEATURE SELECTION USING GENETIC ALGORITHM

Evolutionary Algorithms Selected Basic Topics and Terms

ACONM: A hybrid of Ant Colony Optimization and Nelder-Mead Simplex Search

A Genetic Algorithm for Graph Matching using Graph Node Characteristics 1 2

Elitist Schema Overlays: A Multi-Parent Genetic Operator

Time Complexity Analysis of the Genetic Algorithm Clustering Method

Memory-Based Immigrants for Genetic Algorithms in Dynamic Environments

Parallel Traveling Salesman. PhD Student: Viet Anh Trinh Advisor: Professor Feng Gu.

An Efficient Constraint Handling Method for Genetic Algorithms

Introduction to Design Optimization: Search Methods

Outline. Best-first search. Greedy best-first search A* search Heuristics Local search algorithms

Parameterless Genetic Algorithms: Review, Comparison and Improvement

5. Computational Geometry, Benchmarks and Algorithms for Rectangular and Irregular Packing. 6. Meta-heuristic Algorithms and Rectangular Packing

ANTICIPATORY VERSUS TRADITIONAL GENETIC ALGORITHM

Transcription:

Lecture 6: The Building Block Hypothesis 1 Genetic Algorithms and Genetic Programming Lecture 6 Gillian Hayes 9th October 2007 The Building Block Hypothesis Experimental evidence for the BBH The Royal Road functions Disruption and Hitchhiking GAs versus Hill-climbing The killer instinct Local search Highest when The Schema Theorem Reminder t) E(m(H, t + 1)) m(h, t)û(h, f(t) (1 p d(h) c l 1 )[(1 p m) o(h) ] û(h, t) is large fit d(h) is small short o(h) is small small number of defined bits Identifies building blocks of a good solution. 2 The Building Block Hypothesis 1 Schema Theorem: short, low-order, above average schemata receive exponentially increasing trials in subsequent generations of a genetic algorithm. During crossover, these building blocks become exchanged and combined. So the Schema Theorem identifies the building blocks of a good solution although it only addresses the disruptive effects of crossover (and the constructive effects of crossover are supposed to be a large part of why GAs work). How do we address these constructive effects? Building Block Hypothesis: a genetic algorithm seeks optimal performance through the juxtaposition of short, low-order, high-performance schemata, called the building blocks. 3

The Building Block Hypothesis 2 current winner best fitness 4 The Building Block Hypothesis 5 The Building Block Hypothesis is a hypothesis so we can do an experiment to test it. Experiment: use a problem which contains explicit building blocks and observe the population. Do the building blocks combine to give good solutions in the way the BBH predicts? small building blocks present t larger building blocks present BBH: crossover combines short, low-order schemata into increasingly fit candidate solutions. Requires: Mitchell, Forrest, Holland set up such a problem, using Royal Road functions. Details: Mitchell, Chapter 4, pp127 133 short, low-order, high-fitness schemata stepping stone solutions which combine H i and H j to create even higher fitness schemata The Royal Road Function(s) 1 Define fitness in terms of particular schemata: substrings that, if present in population, ought to be combinable into the optimal solution. They should lay out a Royal Road to the global optimum. The first RR function R 1 is defined using a list of schemata s i. Each s i has a fitness coefficient c i. The fitness R 1 (x) of some bit string x is given by: 6 The Royal Road Function(s) 1 7 Simple example using 16 bits. Suppose: s 1 = 1 1 1 1 * * * * * * * * * * * * s 2 = * * * * 1 1 1 1 * * * * * * * * s 3 = * * * * * * * * 1 1 1 1 * * * * s 4 = * * * * * * * * * * * * 1 1 1 1 and suppose c 1 = c 2 = c 3 = c 4 = 4 and s opt = 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 R 1 (x) = i c i δ i (x), δ i (x) = { 1 if x si 0 otherwise Then R 1 (opt) = 16 Take the string 1111 0100 1001 1111. It samples/matches s 1 and s 4. So δ 1 (x) = δ 4 (x) = 1 and δ 2 (x) = δ 3 (x) = 0. And R 1 (1111010010011111) = 8.

The Royal Road Function(s) 2 8 Hill Climbing 9 Several Royal Road functions defined in terms of different combinations of schemata, with building blocks at different levels, e.g. 4 contiguous 1s, 8 contiguous 1s, 16 contiguous 1s, etc. Try to evolve the string with all 1s and compare performance of GA against a number of hill-climbing schemes: Steepest-ascent hill climbing (SAHC) Next-ascent hill climbing (NAHC) Random-mutation hill climbing (RMHC) Steepest-ascent hill climbing (SAHC) 1. Let current-best be a random string. 2. From left to right flip each bit in the string. Record fitness of each one-bit mutant. 3. If any mutant is fitter than current-best, set current-best to fittest mutant and goto 2. 4. If no fitness increase, save current-best and goto 1. 5. After N evaluations return fittest current-best. Will the GA do better? Hill Climbing 10 Hill Climbing 11 Next-ascent hill climbing (NAHC) Random-mutation hill climbing (RMHC) 1. Let current-best be a random string. 2. From left, flip each bit i in string. If no fitness increase, flip bit back. If fitness increase, set current-best to new string and continue mutating new string from 1 bit after the bit at which fitness increase was found. 3. If no fitness increase, save current-best and goto 1. 4. After N evaluations return fittest current-best. 1. Let current-best be a random string. 2. Flip a random bit in current-best. If no fitness decrease, set current-best to mutated string. 3. Repeat 2. until optimal string found or N evaluations completed. 4. Return fittest current-best. See Mitchell p.129 for these algorithms.

Results 200 runs GA SAHC NAHC RMHC Mean 61,334 > max > max 6,179 No. evaluations to Median 54,208 > max > max 5,775 find optimal string max = 256000 Why did the GA do worse than RMHC? When do GAs perform well? By Markov chain analysis, RMHC s expected time is 6549 evaluations. OK. What s going wrong in the GA? Larger combinations of s i in the GA get broken up by crossover and disrupted by mutation. And the GA suffers from hitch-hiking: once an instance of a high-fitness schema is discovered, the unfit material, especially that just next to the fit parts, spreads along with the fit material. Slows discovery of good schemata in those positions. 12 Hitch-hiking Hitch-hikers prevent independent sampling particularly in those partitions falling between two closely-spaced others, e.g. hitch-hikers with s 2 and s 4 (0s near the ends of s 2 and s 4 ) drowned out instances of s 3. So sampling in s 3 region not independent of sampling in s 2 and s 4 regions. Early convergence to wrong schemata in a number of partitions of the string limits the effectiveness of crossover. Sampling of the different regions is not independent. See Mitchell Fig. 4.2 p.133 for hitch-hiking effect. 13 Analysis Easy problem, no local maxima (so hill-climbing works, RMHC is systematic). GA will out-perform HC on parallel machines (why?) GA is not sampling evenly; schema theorem does not hold. If partitions are sampled independently, schema theorem ought to hold. Mitchell proposes an idealised GA (IGA): sample a new string s i uniform-randomly 14 It works, and it s N times faster than HC IGA unusable in practice (why?) but gives us a lower bound on time GA needs to find optimal string In IGA each new string is an independent sample, whereas in RMHC each new sample differs from the previous by only one bit so RMHC takes longer to construct building blocks So we have some clues as to when GAs will do well... 15 if s i contains a new desired schema, keep it and cross it over with previous best string to incorporate new schema into the solution IGA aims to sample each partition independently and tend to keep best schemata in each partition static Building Block Hypothesis

When Do GAs Do Better Than Hill-climbing? To act like an ideal GA and outperform hill-climbing (at least in this sort of landscape) need Independent samples: big enough population, slow enough selection, high enough mutation, so that no bit-positions are fixed at same value in every chromosome Keeping desired schemata: strong enough selection to keep desired schemata but slow enough selection to avoid hitch-hiking We want crossover to cross over good schemata quickly when they re found to make better chromosomes (but we don t want crossover to disrupt solutions) Large N/long string so speedup over RMHC is worth it. Not possible to satisfy all constraints at once tailor to your problem 16 Where Now? Schema theorem starts to give us an idea of how GAs work but is flawed need better mathematical models of GA convergence...... but these better models don t make our GA go faster Standard GA finds good areas, but lacks the killer instinct to find the globally best solution Standard crossover often disrupts good solutions late in the run Binary representations of non-binary problems often slow the GA down rather than allowing it to sample more freely. The Hamming Cliff. 17 The Killer Instinct Want to get from good to best individuals. [De Jong] Say range of payoff values is [1,100]. Quickly get population with fitness say in [99,100]. Selective differential between best individuals and rest, e.g. 99.988 and 100.000, is very small. Why should GA prefer one over another? Dynamically scale fitness function as a function of generations or fitness range Use rank-proportional selection to maintain a constant selection differential. Slows down initial convergence but increases killer instinct in final stages. Elitism. Keep best individual found so far, or, selectively replace worst members of population 18 The Killer Instinct and Memetic Algorithms Hill-climbing local neighbourhood search is a fast single solution method which quickly gets stuck in local optima (cf. SAHC, NAHC) Genetic algorithms are a multi-solution technique which find good approximate solutions which are non-local optima Hence: try applying local search to each member of a population after crossover/mutation has been applied GA + LS = Memetic Algorithm 19 Aim is to shift balance from exploration at start to exploitation at end.

Making It Better 20 Start the GA from good initial positions: seeding. If you know roughly where a solution might lie, use this information. Use a representation close to the problem: does not have to be a fixed-length linear binary string Use operators that suit the representation chosen, e.g. specific positions crossover only in Run on parallel machines: island model GA (evolve isolated subpopulations, allow to migrate at intervals) Reading: Mitchell Chapter 4