Multi-objective Optimization

Similar documents
Incorporation of Scalarizing Fitness Functions into Evolutionary Multiobjective Optimization Algorithms

Evolutionary multi-objective algorithm design issues

Recombination of Similar Parents in EMO Algorithms

Multi-objective Optimization

A Similarity-Based Mating Scheme for Evolutionary Multiobjective Optimization

Evolutionary Algorithms: Lecture 4. Department of Cybernetics, CTU Prague.

Comparison of Evolutionary Multiobjective Optimization with Reference Solution-Based Single-Objective Approach

Lamarckian Repair and Darwinian Repair in EMO Algorithms for Multiobjective 0/1 Knapsack Problems

MULTI-OBJECTIVE OPTIMIZATION

Multi-Objective Optimization using Evolutionary Algorithms

Multi-Objective Optimization using Evolutionary Algorithms

Mechanical Component Design for Multiple Objectives Using Elitist Non-Dominated Sorting GA

Finding Sets of Non-Dominated Solutions with High Spread and Well-Balanced Distribution using Generalized Strength Pareto Evolutionary Algorithm

Multi-Objective Evolutionary Algorithms

Multiobjective Prototype Optimization with Evolved Improvement Steps

An Evolutionary Multi-Objective Crowding Algorithm (EMOCA): Benchmark Test Function Results

SPEA2+: Improving the Performance of the Strength Pareto Evolutionary Algorithm 2

Multi-objective Optimization Algorithm based on Magnetotactic Bacterium

International Journal of Computer Techniques - Volume 3 Issue 2, Mar-Apr 2016

Towards Understanding Evolutionary Bilevel Multi-Objective Optimization Algorithm

STUDY OF MULTI-OBJECTIVE OPTIMIZATION AND ITS IMPLEMENTATION USING NSGA-II

Investigating the Effect of Parallelism in Decomposition Based Evolutionary Many-Objective Optimization Algorithms

Mechanical Component Design for Multiple Objectives Using Elitist Non-Dominated Sorting GA

Multi-Objective Optimization Using Genetic Algorithms

NCGA : Neighborhood Cultivation Genetic Algorithm for Multi-Objective Optimization Problems

Compromise Based Evolutionary Multiobjective Optimization Algorithm for Multidisciplinary Optimization

An Evolutionary Algorithm for the Multi-objective Shortest Path Problem

Improved S-CDAS using Crossover Controlling the Number of Crossed Genes for Many-objective Optimization

Using ɛ-dominance for Hidden and Degenerated Pareto-Fronts

Part II. Computational Intelligence Algorithms

Performance Assessment of DMOEA-DD with CEC 2009 MOEA Competition Test Instances

Late Parallelization and Feedback Approaches for Distributed Computation of Evolutionary Multiobjective Optimization Algorithms

Solving Bilevel Multi-Objective Optimization Problems Using Evolutionary Algorithms

Approximation Model Guided Selection for Evolutionary Multiobjective Optimization

GECCO 2007 Tutorial / Evolutionary Multiobjective Optimization. Eckart Zitzler ETH Zürich. weight = 750g profit = 5.

division 1 division 2 division 3 Pareto Optimum Solution f 2 (x) Min Max (x) f 1

Multi-objective Ranking based Non-Dominant Module Clustering

ENERGY OPTIMIZATION IN WIRELESS SENSOR NETWORK USING NSGA-II

Hybrid Genetic Algorithms for Multi-objective Optimisation of Water Distribution Networks

DCMOGADES: Distributed Cooperation model of Multi-Objective Genetic Algorithm with Distributed Scheme

Genetic Algorithm Performance with Different Selection Methods in Solving Multi-Objective Network Design Problem

Decomposition of Multi-Objective Evolutionary Algorithm based on Estimation of Distribution

Multi-Objective Pipe Smoothing Genetic Algorithm For Water Distribution Network Design

The Multi-Objective Genetic Algorithm Based Techniques for Intrusion Detection

EVOLUTIONARY algorithms (EAs) are a class of

DEMO: Differential Evolution for Multiobjective Optimization

Finding a preferred diverse set of Pareto-optimal solutions for a limited number of function calls

Development of Evolutionary Multi-Objective Optimization

Evolutionary Computation

Evolutionary Multi-Objective Optimization Without Additional Parameters

Unsupervised Feature Selection Using Multi-Objective Genetic Algorithms for Handwritten Word Recognition

Deconstructing Multi-objective Evolutionary Algorithms: An Iterative Analysis on the Permutation Flow-Shop Problem

Optimizing Delivery Time in Multi-Objective Vehicle Routing Problems with Time Windows

Approximation-Guided Evolutionary Multi-Objective Optimization

A GENETIC ALGORITHM APPROACH FOR TECHNOLOGY CHARACTERIZATION. A Thesis EDGAR GALVAN

CHAPTER 2 MULTI-OBJECTIVE REACTIVE POWER OPTIMIZATION

A Predictive Pareto Dominance Based Algorithm for Many-Objective Problems

Multiobjective Formulations of Fuzzy Rule-Based Classification System Design

R2-IBEA: R2 Indicator Based Evolutionary Algorithm for Multiobjective Optimization

Two Heuristic Operations to Improve the Diversity of Two-objective Pareto Solutions

Revision of a Floating-Point Genetic Algorithm GENOCOP V for Nonlinear Programming Problems

A Fast Elitist Non-Dominated Sorting Genetic Algorithm for Multi-Objective Optimization: NSGA-II

Combining Convergence and Diversity in Evolutionary Multi-Objective Optimization

A MULTIOBJECTIVE GENETIC ALGORITHM APPLIED TO MULTIVARIABLE CONTROL OPTIMIZATION

Ajay Sharma Gaurav Kapur VK Kaushik LC Mangal RC Agarwal. Defence Electronics Applications Laboratory, Dehradun DRDO, India

Handling Constraints in Multi-Objective GA for Embedded System Design

Overview of NSGA-II for Optimizing Machining Process Parameters

GT HEURISTIC FOR SOLVING MULTI OBJECTIVE JOB SHOP SCHEDULING PROBLEMS

An Evolutionary Algorithm with Advanced Goal and Priority Specification for Multi-objective Optimization

Multiobjective hboa, Clustering, and Scalability. Martin Pelikan Kumara Sastry David E. Goldberg. IlliGAL Report No February 2005

Bio-inspired Optimization and Design

Design of Curves and Surfaces Using Multi-Objective Optimization

Chapter 2 Some Single- and Multiobjective Optimization Techniques 2.1 Introduction

Reference Point Based Evolutionary Approach for Workflow Grid Scheduling

Discovering Knowledge Rules with Multi-Objective Evolutionary Computing

International Conference on Computer Applications in Shipbuilding (ICCAS-2009) Shanghai, China Vol.2, pp

Procesamiento Paralelo para Problemas Multiobjetivo en Entornos Dinámicos

CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS

X/$ IEEE

Multiobjective Optimisation. Why? Panorama. General Formulation. Decision Space and Objective Space. 1 of 7 02/03/15 09:49.

Global Optimization of a Magnetic Lattice using Genetic Algorithms

Evolutionary Algorithms. CS Evolutionary Algorithms 1

Bi-Objective Optimization for Scheduling in Heterogeneous Computing Systems

Suppose you have a problem You don t know how to solve it What can you do? Can you use a computer to somehow find a solution for you?

Solving a hybrid flowshop scheduling problem with a decomposition technique and a fuzzy logic based method

The Genetic Algorithm for finding the maxima of single-variable functions

A Distance Metric for Evolutionary Many-Objective Optimization Algorithms Using User-Preferences

Efficient Hybrid Multi-Objective Evolutionary Algorithm

CHAPTER 5 STRUCTURAL OPTIMIZATION OF SWITCHED RELUCTANCE MACHINE

Engineering design using genetic algorithms

Evolutionary Algorithm for Embedded System Topology Optimization. Supervisor: Prof. Dr. Martin Radetzki Author: Haowei Wang

Double Archive Pareto Local Search

An Improved Multi-Objective Evolutionary Algorithm with Adaptable Parameters

Dynamic Uniform Scaling for Multiobjective Genetic Algorithms

Solving Multi-objective Optimisation Problems Using the Potential Pareto Regions Evolutionary Algorithm

ROBUST MULTI-OBJECTIVE OPTIMIZATION OF WATER DISTRIBUTION NETWORKS

Improved Pruning of Non-Dominated Solutions Based on Crowding Distance for Bi-Objective Optimization Problems

Parallel Multi-objective Optimization using Master-Slave Model on Heterogeneous Resources

Indicator-Based Selection in Multiobjective Search

Design of an Optimal Nearest Neighbor Classifier Using an Intelligent Genetic Algorithm

Transcription:

Jugal K. Kalita

Single vs. Single vs. Single Objective Optimization: When an optimization problem involves only one objective function, the task of finding the optimal solution is called single-objective optimization. Example: Find a CAR for me with Minimum cost. : When an optimization problem involves more than one objective function, the task of finding one or more optimal solutions is known as multi-objective optimization. Example: Find a CAR for me with minimum cost and maximum comfort.

: Introduction Multi-objective Problems Problems with more than one objective function, typically conflicting objectives Cars Example: Maximize luxury vs. Minimize price Another example: Maximiize f 1 (x) = x 2 and f 2 (x) = (x 2) 2 when x is in the range [ 10 3, 10 3 ]

: Introduction Multi-objective Problems A more complex example: Maximize the following two objective functions when the variables are within the range [ π, π]: f 1 (x) = [1 + (A 1 B 1 ) 2 + (A 1 B 2 ) 2] f 2 (x) = [(x 1 + 3) 2 + (x 2 + 1) 2] A 1 = 0.5sin(1) 2cos(1) + sin(2) 1.5cos(2) A 2 = 1.5sin(1) cos(1) + 2sin(2) 0.5cos(2) B 1 = 0.5sin(x 1 ) 2cos(x 1 ) + sin(x 2 ) 1.5cos(x 2 ) B 2 = 1.5sin(x 1 ) cos(x 1 ) + 2sin(x 2 ) 0.5cos(x 2 )

: Introduction General Formulation of a Multi-objective Problem Minimiize F(X) where F(X) = {f i : i = 1, M} X = {X j : j = 1, N} subject to C(X) 0 where C = {C k : k = 1, P} H(X) = 0 where H = {H k : k = 1, Q}

: Introduction Multi-objective optimization (MOO) is the optimization of conflicting objectives. In some problems, it is possible to find a way of combining the objectives into a single objective. But, in some other problems, it is not possible to do so. Sometimes the differences are qualitative and the relative importance of these objectives cannot be numerically quantified. Frequently, there is no unique solutions, rather there is a trade-off among optimal solutions.

: Introduction Conceptual Example Suppose you need to fly on a long trip: Should you choose the cheapest ticket (more connections) or shortest flying time (more expensive)? It is impossible to put a value on time, so these two objectives cannot be linked. Also, the relative importance will vary. There may be a business emergency you need to go fix quickly. Or, maybe you are on a very tight budget.

Pareto-optimal Solutions Pareto-Optimal Solutions A MOO problem with constraints will have many solutions in the feasible region. Even though we may not be able to assign numerical relative importance to the multiple objectives, we can still classify some possible solutions as better than others.

Pareto-optimal Solutions Example of Pareto-Optimal Solutions Suppose in our airplane-trip example from earlier, we find the following tickets. Ticket Travel Time (hrs) Price ($) A 10 1700 B 9 2000 C 8 1800 D 7.5 2300 E 6 2200

Pareto-optimal Solutions Comparison of Solutions If we compare tickets A and B, we cannot say that either is superior without knowing the relative importance of Travel Time vs. Price. However, comparing tickets B and C shows that C is better than B in both objectives, so we can say that C dominates B. So, as long as C is a feasible option, there is no reason we would choose B.

Pareto-optimal Solutions Comparison of Solutions If we finish the comparisons, we also see that D is dominated by E. The rest of the options (A, C, and E) have a trade-off associated with Time vs. Price, so none is clearly superior to the others. We call {A, C, E} the non-dominated set of solutions because none of the solutions are dominated.

Pareto-optimal Solutions Graph of Solutions Usually, solutions of this type form a typical shape, shown in the graph below.

Pareto-optimal Solutions Types of Solutions Solutions that lie along the line are non-dominated solutions while those that lie inside the line are dominated because there is always another solution on the line that has at least one objective that is better.

Pareto-optimal Solutions Pareto-optimal Solutions The line is called the Pareto front and solutions on it are called Pareto-optimal. All Pareto-optimal solutions are non-dominated. Thus, it is important in MOO to find the solutions as close as possible to the Pareto front and as far along it as possible.

Pareto-optimal Solutions Finding the Pareto Front For the following feasible region with objectives f 1 and f 2 where both f 1 and f 2 are minimized.

Pareto-optimal Solutions Graphical Example One way to imagine finding points on the Pareto front is by using a combination of numerical weights for the two objectives. Here, we have a weight of w 1 on f 1 and weight w 2 on f 2. The equation of a slanted line is w 1 f 1 + w 2 f 2 for various values of w 1 and w 2 such that w 1 + w 2 = 1.

Pareto-optimal Solutions Finding the Pareto Front If this is done for a 90 span of lines, all the points on the Pareto front will be found.

Pareto-optimal Solutions Realistic Procedures There are different methods used in practice, but one is to use a genetic algorithm to enumerate points along the Pareto front over several iterations, then use some method to rank the quality of the trade-offs based on the particular application being modeled.

Pareto-optimal Solutions Practicality of this Procedure Actually, this is not the procedure that is used in practice, but it is a good illustration of the concept. This procedure would require finding all possible points in the feasible region and then using many combinations of weights. For more than two objectives, the complexities and the number of combinations make this impractical.

Pareto-optimal Solutions Finding a Point on the Pareto Front It should be remembered that each point on the Pareto front is found by solving an optimization problem. If the problem is nonlinear or very complex, the simple step of getting just one solution may not be trivial. This is the case whether we a genetic algorithm or some other method. We have to find points on the Pareto Front.

Mutli-objective GA References Development of Multiobjective GAs The 1st generation: Vector Evaluation approach Vector Evaluated Genetic Algorithm (VEGA) J. D. Schaffer: Multiple objective optimization with vector evaluated genetic algorithms, Proc. of 1st Inter. Conf. on GAs and Their Applic., 1985. The 2nd generation: Pareto ranking + Diversity Multiobjective Genetic Algorithm (MOGA) Fonseca, C. M. and Fleming, P. J: Genetic algorithms for multiobjective optimization: formulation, discussion and generalization, Proc. of the 5th Inter. Conf. on GAs, 1993. Non-dominated Sorting Genetic Algorithm (NSGA) Srinivas, N. and Deb, K.: Multiobjective function optimization using nondominated sorting genetic algorithms, Evolutionary Computation, 1995.

Mutli-objective GA References Development of Multiobjective GAs The 3rd generation: Weighted Sum + Elitist Preserve Random-weight Approach (RWA): Ishibuchi, H and T. Murata: A multiobjective genetic local search algorithm and its application to flowshop scheduling, IEEE Trans. on Sys., Man & Cyber., 1998. Strength Pareto Evolutionary Algorithm (SPEA): E. Zitzler and L. Thiele: Multiobjective Evolutionary Algorithms: A Comparative Case Study and the Strength Pareto Approach, IEEE Trans. on Evolutionary Comput., 1999. Adaptive Weight Approach (AWA): Gen M,. and Cheng, R. : Genetic Algorithms and Engineering Optimization, John Wiley & Sons, New York, 2000. Non-dominated Sorting Genetic Algorithm II (NSGAII): Deb, K., A. Pratap, S. Agarwal & T. Meyarivan: A Fast and Elitist Multiobjective Genetic Algorithm: NSGA-II IEEE Trans. on Evolutionary Comput., 2002

VEGA Vector Evaluated Genetic Algorithms (Schaffer, 1985) VEGA is the first notable work to solve MOO problems in which it uses a vector fitness measure to create the next generation. The selection step in each generation becomes a loop. Each time through the loop the appropriate fraction of the next generation, or subpopulation, is selected on the basis of each objective. The entire population is shuffled thoroughly to apply crossover and mutation operators. This is performed to achieve the mating of individuals of different subpopulations.

VEGA VEGA 1985 Divide population into M equal blocks for M objectives at every generation. Reproduce each block with one objective function. Complete population participates in crossover and mutation. Use proportionate selection operator. To reduce the positional bias in the population, shuffle the population before partition.

VEGA VEGA 1985

Pareto Ranking Pareto Ranking Approach Pareto ranking approach consists of assigning fitness ranks to chromosomes based on the level of non-domination. Best ranked chromosomes have better chance at reproduction. A Pareto ranking-based fitness assignment method was first suggested by Goldberg, D.: Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley, Reading, MA, 1989.

Goldberg s Method Pareto Ranking (Goldberg 1989) Step 1: Assign rank 1 to the nondominated chromosomes Step 2: Remove rank 1 chromosomes from contention Step 3: Find the next nondominated chromosomes, remove them from contention, assign rank 2 to them. Step 4: Continue until the entire population is ranked. Figure: Simple case with two objectives to be minimized using Goldberg s Ranking

MOGA Multi-objective Genetic Algorithm (Fonseca & Fleming 1993) Step 1: Assign rank 1 to the nondominated chromosomes Step 2: Remove rank 1 chromosomes from contention Step 3: Find the next nondominated chromosomes, remove them from contention, assign rank equal to the number of a chromosome s dominating individuals +1. Step 4: Continue until the entire population is ranked. Figure: Simple case with two objectives to be minimized using MOGA

NSGA Non-dominated Sorting Genetic Algorithm (Deb 1995) Step 1: Identify nondominated individuals, and assign a large dummy fitness value. Step 2: To maintain diversity in the population, share these individuals with their dummy fitness values. Step 3: After sharing, ignore these nondominated individuals temporarily. Step 4: Identify the second nondominated front in the rest of the population and assign a dummy fitness value smaller than the minimum shared dummy fitness of the previous front. Step 5: Continue this process until the entire population is classified into several fronts.

NSGA Salient Points about NSGA Assign fitness first to the non-dominated set and successively to dominated sets. A solution i of the first (or best) non-dominated set is assigned a fitness value F i = N where N is the population size. Assigning higher fitness to solutions belonging to a better non-dominated set ensures selection pressure toward the Pareto-optimal front. Maintain diversity of solutions in a specific front by reducing the assigned fitness based on the crowdedness (the number of neighboring solutions) of a solution. This is done using a sharing function. Degrading the fitness of solutions in a crowded area in a certain front emphasizes solutions in less crowded regions.

NSGA NSGA... Figure: Simple case with two objectives to be minimized using NSGA Figure: Flow chart of NSGA

NSGA NSGA Analysis It is relatively easy to implement. It needs a parameter called the sharing factor to decide how the fitnesses of crowded solutions in a specific front should be shared. The solutions it obtains to a problem seem to be very sensitive to the sharing factor.

NSGA-II Non-Dominated Sorting Genetic Algorithm NSGA-II Deb et al. (2000, 2002) proposed a new version of NSGA called NSGA-II NSGA-II is more efficient computationally speaking. It uses elitism and doesn t need a sharing factor or parameter.

NSGA-II Innovations in NSGA-II There are three innovations. A Fast non-dominated sorting algorithm for solutions A way to efficiently compute what s called crowding distance. A simple crowded distance comparison We will look at all three of these. First, let s look at the main algorithm.

NSGA-II Non-Dominated Sorting Genetic Algorithm NSGA-II Steps before main loop begins Initialize population P0 Sort P0 on the basis of nondomination Fitness of an individual is equal to an individual s nondomination level Binary Tournament Selection Mutation and recombination create an offspring population Q 0

NSGA-II Non-Dominated Sorting Genetic Algorithm NSGA-II Primary loop R t = P t + Q t (t= generation) Sort R t on the basis of nondomination Create Pt+1 by adding individuals from each level of R t until P t+1 is of size N Create Qt+1 by applying Binary Tournament Selection, Mutation, and Recombination to P t+1 Increment t

NSGA-II NSGA Flowchart Figure: Flow chart of NSGA II

NSGA-II NSGA Flow Diagram Figure: P t is the parent population and Q t is the offspring population for generation t. F 1 are the best solutions from the combined population, parents and offspring. F 2 are the next best solutions, etc.

NSGA-II Non-dominated Soring For each solution, calculate two entities: 1) domination count n p, the number of solutions that dominate the solution p, and 2) S p, a set of solutions that the solution p dominates. All solutions in the first non-dominated front have their domination count as 0. For each solution p with n p = 0, we visit each member q of its set S p and reduce its domination count by 1. In doing so, if for any member q, the domination count becomes 0, we put it in a separate list Q. Q is the second non-dominated front. Continue the above procedure with each member of Q and the third front is identified. Continue till all fronts are identified.

NSGA-II Non-dominated Sorting Figure: Result of non-dominated sorting

NSGA-II Non-dominated sorting 1: k=0 2: repeat 3: k k + 1 4: R.k non-dominated solutions in P 5: P P R.k 6: until P = Φ Here R 1 R k denote partitions of rank i = 1 k

NSGA-II Diversity Preservation It is desired that there is a good spread of solutions. This is done by estimating the density of the solutions and by defining a crowded-comparison operator. To estimate the density of solutions surrounding a particular solution in the population, we sort all solutions in each objective dimension and calculate the average distance of two points on either side of this point along each of the objectives. See algorithm at the bottom of page 185 to see how each solution is given a crowding distance along each objective dimension.

NSGA-II Crowding Distance Metric Figure: Crowding distance calculation. For solution i find the average distance from the two solutions immediately before and immediately after, along each objective dimension.

NSGA-II Crowded-Comparsion Operator The crowded-comparison operator ( n ) is used to select solutions at various stages of the algorithm toward a uniformly spread-out Pareto-optimal front. Assume that every individual i in the population has two attributes: 1. Non-domination rank i rank 2. crowding distance i distance We now define a partial order n as i n if (I rank < j rank ) or ((i rank = j rank ) and (i distance > j distance ))