LINEAR NEIGHBORHOOD EVOLUTION STRATEGY JOACHIM SPRAVE ABSTRACT. Approaches with non-panmictic populations have been made mostly in

Similar documents
Gíunter Rudolph. Informatik Centrum Dortmund e.v. Dí44221 Dortmund. Joachim Sprave. Universitíat Dortmund

Distributed Probabilistic Model-Building Genetic Algorithm

Selection Intensity in Asynchronous Cellular Evolutionary Algorithms

Dept. of Computer Science. The eld of time series analysis and forecasting methods has signicantly changed in the last

und Formale Beschreibungsverfahren D Karlsruhe, Germany fbranke, July 1996 Abstract

CS5401 FS2015 Exam 1 Key

EVOLUTIONARY ALGORITHMS FOR FUZZY LOGIC: A BRIEF OVERVIEW. Thomas B ACK, Frank KURSAWE. University of Dortmund, Department of Computer Science, LS XI

336 THE STATISTICAL SOFTWARE NEWSLETTER where z is one (randomly taken) pole of the simplex S, g the centroid of the remaining d poles of the simplex

Automata Construct with Genetic Algorithm

Algorithm Design (4) Metaheuristics

UNIT 9C Randomness in Computation: Cellular Automata Principles of Computing, Carnegie Mellon University

Drawdown Automata, Part 1: Basic Concepts

Mutations for Permutations

A Parameter Study for Differential Evolution

Computational Intelligence

where C is traversed in the clockwise direction, r 5 èuè =h, sin u; cos ui; u ë0; çè; è6è where C is traversed in the counterclockwise direction èhow

GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM

Escaping Local Optima: Genetic Algorithm

PARAMETER OPTIMIZATION FOR AUTOMATED SIGNAL ANALYSIS FOR CONDITION MONITORING OF AIRCRAFT SYSTEMS. Mike Gerdes 1, Dieter Scholz 1

Introduction to Optimization

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Introduction to Optimization

19.2 View Serializability. Recall our discussion in Section?? of how our true goal in the design of a

A Genetic Algorithm Framework

Heuristic Optimisation

A motivated definition of exploitation and exploration

CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS

Genetic Algorithms. Kang Zheng Karl Schober

A Novel Hybrid Self Organizing Migrating Algorithm with Mutation for Global Optimization

An Efficient Constraint Handling Method for Genetic Algorithms

Two-dimensional Totalistic Code 52

Suppose you have a problem You don t know how to solve it What can you do? Can you use a computer to somehow find a solution for you?

The Genetic Algorithm for finding the maxima of single-variable functions

Genetic Algorithm Performance with Different Selection Methods in Solving Multi-Objective Network Design Problem

An Introduction to Evolutionary Algorithms

Adaptive Crossover in Genetic Algorithms Using Statistics Mechanism

A new fine-grained evolutionary algorithm based on cellular learning automata

Variations on Genetic Cellular Automata

PARTICLE SWARM OPTIMIZATION (PSO) [1] is an

A COMPARATIVE STUDY OF FIVE PARALLEL GENETIC ALGORITHMS USING THE TRAVELING SALESMAN PROBLEM

Optimizing Flow Shop Sequencing Through Simulation Optimization Using Evolutionary Methods

OPTIMIZATION METHODS. For more information visit: or send an to:

Topological Machining Fixture Layout Synthesis Using Genetic Algorithms

Evolution Strategies, Network Random Keys, and the One-Max Tree Problem

Genetic Algorithm for Dynamic Capacitated Minimum Spanning Tree

A Learning Automata-based Memetic Algorithm

Using Genetic Algorithms to solve the Minimum Labeling Spanning Tree Problem

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

Exploring Complexity In Science and Technology. Logistics

1 Lab + Hwk 5: Particle Swarm Optimization

Adaptive Elitist-Population Based Genetic Algorithm for Multimodal Function Optimization

Exploration vs. Exploitation in Differential Evolution

Optimization of Function by using a New MATLAB based Genetic Algorithm Procedure

An Evolutionary Algorithm with Stochastic Hill-Climbing for the Edge-Biconnectivity Augmentation Problem

Evolutionary Computation Algorithms for Cryptanalysis: A Study

MINIMAL EDGE-ORDERED SPANNING TREES USING A SELF-ADAPTING GENETIC ALGORITHM WITH MULTIPLE GENOMIC REPRESENTATIONS

Parallel Evolutionary Optimisation with Constraint Propagation

entire search space constituting coefficient sets. The brute force approach performs three passes through the search space, with each run the se

Modeling Plant Succession with Markov Matrices

THE Multiconstrained 0 1 Knapsack Problem (MKP) is

Evolutionary Algorithms

MAXIMUM LIKELIHOOD ESTIMATION USING ACCELERATED GENETIC ALGORITHMS

The Simple Genetic Algorithm Performance: A Comparative Study on the Operators Combination

A Genetic Algorithm Applied to Graph Problems Involving Subsets of Vertices

GENETIC ALGORITHM-AIDED ROUTING ON 3D DYNAMIC NETWORKS

March 19, Heuristics for Optimization. Outline. Problem formulation. Genetic algorithms

Evolutionary Algorithms: Lecture 4. Department of Cybernetics, CTU Prague.

CELLULAR evolutionary algorithms (ceas) use populations

Evolutionary Computation. Chao Lan

1 Lab + Hwk 5: Particle Swarm Optimization

Mechanical Component Design for Multiple Objectives Using Elitist Non-Dominated Sorting GA

A Hybrid Approach for Misbehavior Detection in Wireless Ad-Hoc Networks

Genetic Algorithms for Real Parameter Optimization

Use of Genetic Algorithms for Resource Mapping in Distributed Systems

Genetic Algorithm for Seismic Velocity Picking

Improving resiliency using graph based evolutionary algorithms

Chapter 9: Genetic Algorithms

METAHEURISTICS Genetic Algorithm

Integration of Declarative Approaches (System Demonstration)

5. Computational Geometry, Benchmarks and Algorithms for Rectangular and Irregular Packing. 6. Meta-heuristic Algorithms and Rectangular Packing

Non-deterministic Search techniques. Emma Hart

Improving GA Search Reliability Using Maximal Hyper-Rectangle Analysis

Polyhedron Evolver Evolution of 3D Shapes with Evolvica

sequential evolutions

A Multi-Level Method for the Steady State Solution of Markov Chains

V.Petridis, S. Kazarlis and A. Papaikonomou

Cellular Automata. Cellular Automata contains three modes: 1. One Dimensional, 2. Two Dimensional, and 3. Life

Dynamic Representations and Escaping Local Optima: Improving Genetic Algorithms and Local Search

Genetic Algorithm Based Template Optimization for a Vision System: Obstacle Detection

Evolutionary Algorithms. CS Evolutionary Algorithms 1

Comparative Analysis of Various Evolutionary and Memetic Algorithms

Genetic Algorithm for Dynamic Capacitated Minimum Spanning Tree

NICHING EVOLUTIONARY ALGORITHMS FOR MULTIMODAL AND DYNAMIC OPTIMIZATION

CHAPTER 6 ORTHOGONAL PARTICLE SWARM OPTIMIZATION

Automated Test Data Generation and Optimization Scheme Using Genetic Algorithm

Genetic algorithms and finite element coupling for mechanical optimization

Solving Sudoku Puzzles with Node Based Coincidence Algorithm

Abstract. 1 Introduction

Review: Final Exam CPSC Artificial Intelligence Michael M. Richter

PARALLEL GENETIC ALGORITHMS IMPLEMENTED ON TRANSPUTERS

Transcription:

LINEAR NEIGHBORHOOD EVOLUTION STRATEGY JOACHIM SPRAVE Department of Computer Science, University of Dortmund 44221 Dortmund, Germany E-Mail: sprave@ls11.informatik.uni-dortmund.de ABSTRACT Global selection in panmictic evolutionary algorithms often causes premature convergence. Approaches with non-panmictic populations have been made mostly in parallel implementations to eliminate global control. This results in new types of selection, which may also improve the reliability of sequential evolutionary algorithms. Most previous approaches in the past were based on genetic algorithms ègaè, but this paper focuses evolution strategies èesè. Three ES-based approaches with a ring topology are introduced, one with a GA-like kind of mating selection, and two which are very close to the standard èç; çè-es and èç + çè-es. They have been compared to traditional ES in numerical experiments. The results show the dependence of convergence reliability and neighborhood sizes, and that local selection can improve the exploitation facility of ES with only a linear slow-down in exploration. 1. Parallel Evolutionary Algorithms 1.1. Interaction Models There are at least two diæerent approaches to parallel evolutionary algorithms èeaè. The ærst is called the migration model and results in coarse grain distributed applications. A couple of instances of a standard EA are running on diæerent processors, providing a migration facility to exchange genetic material between populations. The travelers can be chosen in many waysíyou can allow the best to travel, urge the worst, choose travelers randomly or anything else. Communication can be accomplished asynchronously to prevent the processors from being idle. By tuning the local population sizes and migration frequencies these algorithms can be scaled to a balanced usage of processing and communication resources. The subpopulations are still panmictic: for each individual, any other individual in the same population is a potential mating partner. The second approach is known as diæusion model, plant pollination model 5 or neighborhood model. The individuals are placed in an arbitrary space, e.g., on a two-dimensional grid. On top of this topology it is easy to deæne some kind of neigh-

borhood. By constraining evolutionary processes to this neighborhood, the need for any kind of global knowledge is eliminated. Mapping this to a grid of processors results in a æne grain approach for parallel EAs. In contrast to migration, these algorithms lack the feature of scalability. Nevertheless, because of their regular structure, they are well-suited for data parallel computers, such as the Connection Machine. The neighborhood model is not only an approach for parallelism, it is as well a better model of nature. The existence of a global selection authority is a very strong simpliæcation in classic EA. To implement the two-dimensional neighborhood model on MIMD-machines, the processors themselves are connected to a torus, and a subgrid of individuals is placed on every processor. Now the borders must be exchanged between adjacent subgrids after every generation. Because EAs are very robust, there is no need for synchronization of the tasks placed on diæerent processors. An implementation like this, based on a genetic algorithm ègaè, is DIOGENES 11. It has shown that the size of the subgrids is the parameter to balance computation and communication, thus the parallelism is scalable as in the migration model. Numerical results from DIOGENES have shown that local selection can improve the ability of EAs to ænd the global optimum of multi-modal problems. 1.2. Local Selection in Genetic Algorithms There are many parallel EAs using the neighborhood model, and most are GAbased. These approaches have in common that a new selection operator had to be introduced. To generate the successor of an individual for the next generation, Collins and Jeæerson 4 took two series of samples from the neighborhood and chose the best of each series as parents for the new individual. Sprave 11 made experiments with diæerent selection strategies. The best results were achieved by a deterministic selection: each individual was replaced by the oæspring of itself and its ættest neighbor. Spiessens and Manderick 10 took a selection scheme as close as possible to sequential GA, a local proportionate selection: ærst, each individual is replaced by one of its neighbors, selected according to the local ætness distribution. After that, each individual selects a mating partner from its neighborhood èwhich itself has been selected by local proportionate selection in the step beforeè, to produce a successor for this position in next generation. All neighborhood GAs have shown that the diversity in a population can be preserved for many generations. Thus, exploration of the search space is improved, and

the exploitation is not necessarily worse than in a panmictic GA. 2. Linear Neighborhood Evolution Strategy Using local selection also in sequential EAs should lead to more diversity during the optimization process, and thus increase the probability of discovering the global optimum of multimodal objective functions. This approach looks very much like a Cellular Automaton ~ ècaè. There are two violations of the rules which Wolfram 12 deæned a CA to possess: the transitions are not deterministic, and, in case of an ES, the states are not discrete. In the theory of cellular automata, most analysis concerns one-dimensional or linear cellular automata, because they are simple but have nearly all the properties of multi-dimensional CA. So, the linear topology seems to be a good choice for analyzing cellular EA, too. 2.1. Local Selection Schemes The next three sections describe three local selection schemes for ES. Although they are described for a ring topology in this paper, they are well-suited for any kind of topology as long as a neighborhood of more than two individuals is deæned for each place in the population. The Linear-Neighborhood-ES èlnesè is an ES with a neighborhood deæned by using the index of each individual. The neighborhood with radius ç of the individual a i consists of NB ç èiè :=fa èi,çè ; ::; a èi+çè g è1è where the indices are calculated modulo population size so that the result is a ring topology. 2.2. Local Mating Selection The diæerence of selection schemes of GA-like proportionate selection, ranking and so on, and the ES's èç; çè-selection can be reduced to diæerent acceptance probability functions. Nevertheless, selection in GA is more or less intended as a mating selection, wherein the best individuals have most oæspring. In ES, a surplus of oæspring is produced and only the best of them survive and each has the same chance to be chosen as a parent. While it is easy to constrain the GA's mating selection to a neighborhood, the extinctive selection used in ES is inherently global. Of course, it is possible to transfer the GA's selection scheme to ES by choosing two parents from the neighborhood by

proportionate selection, and replace the central individual by their oæspring in the next generation. The neighborhood size is typically small to maximize local eæects. Therefore, the simple mating selection described below should produce similar results like ranking or proportionate selection, except for the fact that, as long as the neighborhood size is not greater than the population size, one individual cannot be chosen twice to produce one oæspring. For each individual a i, do synchronously èin parallel or pseudo-parallelè: æ Choose the two best from its neighborhood. æ Recombine them to one oæspring named a 0 i. æ Mutate a 0 i. æ Replace a i by a 0 i. Recombination and mutation operators are the same as in traditional ES èfor details see Schwefel and Bíack 2 è. Note that if the neighborhood is the whole population, this selection becomes a è2;pè-es èwhere P is the size of the ring topologyè and, if ç is at least P,aè1;Pè. 2.3. Local èç; çè-selection To make comparison to standard ES easy, a èç; çè-selection was combined with the neighborhood model. Assuming that ç is a multiple of the population size ç, the calculation of a new generation in a èç; çè-lnes works as follows: For each individual a i, do synchronously èin parallel or pseudo-parallelè: æ Choose ç=ç times two parents randomly from NB ç èiè and recombine them to a 0 i;j ; 1 ç j ç ç=ç. æ Replace a i by a 0 i which ischosen as the best of the a 0 i;j. æ Mutate a 0 i as in a panmictic ES. Now there are two extreme values for ç:

æ ç = bç=2c This case is very close to a èç; çè-es, but not the ç best are the parents of the next generation. Instead, there are ç sets of ç=ç individuals created, and the best of each set is taken to the next generation. The selection pressure is weaker, but because the parents are taken randomly from the whole population, it should be still suæcient to generate an eæcient search. æ ç =0 This is nothing but a è1; ç=çè-es, started ç times in parallel. Another interesting case is ç = 1, the smallest symmetrical neighborhood which is possible. It is expected to be slower in unimodal problems, and better in multimodal functions. 2.4. Local èç + çè-selection If the oæspring is only accepted when it is better than its local predecessor, the algorithm behaves very much like a èç + çè-es. The quality function value of each position in the population is monotonic decreasing èin case of minimizationè. If there is a lower boundary for the step sizes, the proof of global convergence for the panmictic è+è-es 3 is valid for this kind of ES, too. 3. Results All test functions are minimization problems taken from a comparison of genetic algorithms and evolution strategies by Hoæmeister and Bíack 6.For the sphere model, all evaluations are averaged over 50 runs with diæerent seed values for the random number generator, and n diæerent step sizes ç i were used èwhere n is the dimension of the test functionè. For all test functions discrete recombination was taken for x i, and intermediate recombination for the step sizes ç i. All step sizes are varied self-adaptive, correlated mutations have not been used at all. The main objective of the experiments was to observe the eæects produced by local selection. A given parameter setting was tested with neighborhood radius ç varying from 1 to ç=2. 3.1. Sphere Model The ærst test function is the sphere model

f 1 è~xè = nx i=1 x 2 i ; ~x 2 R n è2è n = 30;,5:12 ç x i ç 5:12 This function can show the ability of the algorithms to proceed straight to the optimum in unimodal environments. Figure 1 shows the best function value found after 500; 000 function evaluations varying the neighborhood radius from 1 to 50. Population size ç was set to 100 for LNES. The è+è and è; è strategies produced six children per position each generation, while the mating-lnes does not need to produce a surplus. For comparison, a è100; 600è-ES and a è10; 60è-ES were run with the same initialization. 1 1e-10 1e-20 1e-30 best f(x) 1e-40 1e-50 1e-60 1e-70 1e-80 (100,600)-ES (10,60)-ES (100,600)-LNES mating-lnes (100+600)-LNES 1e-90 0 10 20 30 40 50 neighborhood radius Figure 1: Evaluation of f 1 The local èç + çè-strategy fails to make any progress towards the optimum èthe worst initial value which is possible is 768:2è. It is diæcult already for a panmictic èç + çè strategy to learn n step sizes, but it is obviously impossible for the local variant. Mating selection works well with a neighborhood radius of six or more. This is a strong hint that this kind of selection also needs a ratio of at least six children per

parent to learn individual step sizes, as suggested by Schwefel 8. Beyond that ratio, it has a strong emphasis on local search. Although the èç; çè-lnes is much slower than traditional ES, it has the same kind of linear convergence in quadratic environments. In contrast to the former strategy, it performs much better with very small neighborhood sizes, whichisvery advantageous for parallel implementations: the smaller the neighborhoods, the less communication is required. 3.2. Generalized Rastrigin Function The generalized Rastrigin function is a very diæcult problem for panmictic ES, as shown by Hoæmeister and Bíack 6, and it was used by Rudolph 7 to show the ability of the migration model ES to ænd the global optimum in multimodal environments. f 7 è~xè =na + nx i=1 x 2 i, A cosè!x iè ; ~x 2 R n è3è n = 20; A = 10;! =2ç;,5:12 ç x i ç 5:12 percentage of successfull runs 100 80 60 40 20 (100,600)-ES (100+600)-ES (100,600)-LNES (100+600)-LNES 0 0 5 10 15 20 25 30 35 40 45 50 neighborhood radius Figure 2: Evaluation of f 7

Figure 2 shows the percentage of successful runs èfrom 250 runs in totalè dependent on the neighborhood radius. Because all variants with multiple step sizes failed to ænd the global optimum within 100 runs, they are not drawn. The most successful runs, about 65è, were performed by the è100; 600è-LNES with one step size for all x i and small neighborhood sizes. Larger neighborhood sizes are worse, but nevertheless nearly as good as the panmictic è100; 600è-ES. A è10; 60è-ES did not ænd the optimum even once in 150 runs. The mating-lnes suæers from premature convergence: its emphasis on local search is too strong even with only one step size: very few runs were successful. 3.3. Shekel's Foxholes Shekel's Foxholes 9 is known as a hard problem for ES. It consists of a large plateau with a few holes of diæerent depths. On the plateau, an ES gets no hint which direction to go to. Once the population falls into one of the holes, it is nearly impossible for the procedure to leave it and search for another. 1 25X f 5 è~xè = 1 + K j=1 P 1 ; ~x 2 R 2 è4è 2 c j + i=1èx i, a ij è 6 K = 500; f 5 èa 1j ;a 2j è ç c j = j a 1j =16èj mod 5, 2è; a 2j =16èbj=5c,2è To avoid that some initial points are already placed in holes, the start population is not sampled from an interval including the holes as in the comparison of Hoæmeister and Bíack 6, but initialized far away from the holes: x i =10; 000 for all individuals. A è100; 500è-LNES with neighborhood size 1 is still able to ænd the global optimum with a high probability. To understand the internal process, a visualization method has been applied to LNES which isvery often used for Cellular Automata: each objective function value is mapped to a color, so each generation can be displayed as a line of dots. Drawing these lines one below the other shows the ætness distribution inside the population over the time. Figure 3 shows atypical run of LNES on Shekel's Foxholes. In the beginning, the entire population is white, which indicates individuals searching on the plateau. Every hole is mapped to a greyscale, the darker the deeper. It can be seen that, while some points of the population are already in holes, the rest of the population keeps on creeping around on the plateau. The global optimum, painted in black, has been found several times and is expanding slowly but steadily over the population. Eventually, after a number of generations depending on neighborhood sizes, the optimal solution would take over the whole population.

Figure 3: Typical ætness distribution for Shekel's Foxholes 4. Summary Linear neighborhood ES have properties that are similar to traditional ES. Except for the mating selection, the emphasis on local search ismuch weaker than in traditional ES, but the exploration facility ismuch better instead, because they are able to explore a set of attractive areas in parallel. LNES are free of global evolutionary operators, and are therefore well-suited for parallel implementations. The granularity of parallel implementations can be tuned by the local population size and the neighborhood radius. As already shown 11, the neighborhood model for evolutionary algorithms also works with asynchronous communication to exchange the overlapping neighborhoods. The local mating selection introduced here is an alternative approach toprovide the selection pressure which is necessary for self adaption of step sizes, but it is not suited for global optimization. 1. Richard K. Belew and Lashon B. Booker, editors. Proceedings of the Fourth International Conference on Genetic Algorithms and their Applications, University of California, San Diego, USA, 1991. Morgan Kaufmann Publishers. 2. Thomas Bíack and Hans-Paul Schwefel. An overview of evolutionary algorithms for parameter optimization. Evolutionary Computation, 1è1è:1í23, 1993. 3. Joachim Born. Evolutionsstrategien zur numerischen Líosung von Adaptationsaufgaben. Doctoral dissertation A, Humboldt-University, Berlin, 1978.

4. Robert J. Collins and David R. Jeæerson. Selection in massively parallel genetic algorithms. In Belew and Booker 1, pages 249í256. 5. David E. Goldberg. Genetic algorithms in search, optimization and machine learning. Addison Wesley, 1989. 6. Frank Hoæmeister and Thomas Bíack. Genetic algorithms and evolution strategies similarities and diæerences. Papers on Economics & Evolution 9103, European Study Group for Evolutionary Economics èesgeeè, November 1991. 7. Gíunter Rudolph. Globale Optimierung mit Parallelen Evolutionsstrategien. Diploma thesis, University of Dortmund, Department of Computer Science, Dortmund, Germany, July 1990. 8. Hans-Paul Schwefel. Collective phenomena in evolutionary systems. In Preprints of the 31st Annual Meeting of the International Society for General System Research, Budapest, volume 2, pages 1025í1033, June 1987. 9. J. Shekel. Test functions for multimodal search techniques. In Fifth Annual Princeton Conference on Information Schience and Systems, 1971. 10. Piet Spiessens and Bernard Manderick. A massively parallel genetic algorithm: Implementation and ærst analysis. In Belew and Booker 1, pages 279í286. 11. Joachim Sprave. Parallelisierung Genetischer Algorithmen zur Suche und Optimierung. Diploma thesis, University of Dortmund, December 1990. 12. Stephen Wolfram. Statistical mechanics of cellular automata. In S. Wolfram, editor, Theory and Applications of Cellular Automata, chapter 1, pages 7 í 50. World Scientiæc Publishing, Singapore, 1986.