Benchmark Functions for the CEC 2008 Special Session and Competition on Large Scale Global Optimization

Similar documents
Large Scale Global Optimization Challenge

Unidimensional Search for solving continuous high-dimensional optimization problems

The 100-Digit Challenge:

Cooperative Co-evolution with Delta Grouping for Large Scale Non-separable Function Optimization


Performance Assessment of DMOEA-DD with CEC 2009 MOEA Competition Test Instances

An Improved Tree Seed Algorithm for Optimization Problems

Modified Shuffled Frog-leaping Algorithm with Dimension by Dimension Improvement

A Hybrid Fireworks Optimization Method with Differential Evolution Operators

A Novel Hybrid Self Organizing Migrating Algorithm with Mutation for Global Optimization

Particle Swarm Optimization Artificial Bee Colony Chain (PSOABCC): A Hybrid Meteahuristic Algorithm

Université Libre de Bruxelles

Small World Particle Swarm Optimizer for Global Optimization Problems

A Hybrid Metaheuristic Based on Differential Evolution and Local Search with Quadratic Interpolation

IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE

Particle Swarm Optimization

Cooperative Coevolution using The Brain Storm Optimization Algorithm

Benchmark Functions for the CEC 2010 Special Session and Competition on Large-Scale Global Optimization

Modified Particle Swarm Optimization

An Empirical Study on Influence of Approximation Approaches on Enhancing Fireworks Algorithm

ACONM: A hybrid of Ant Colony Optimization and Nelder-Mead Simplex Search

Benchmark Functions for CEC 2013 Special Session and Competition on Niching Methods for Multimodal Function Optimization

Center-Based Initialization of Cooperative Co-evolutionary Algorithm for Large-scale Optimization

Theoretical Foundations of SBSE. Xin Yao CERCIA, School of Computer Science University of Birmingham

Exploration vs. Exploitation in Differential Evolution

Comparative Study of Meta-heuristics Optimization Algorithm using Benchmark Function

A Comparative Analysis on the Performance of Particle Swarm Optimization and Artificial Immune Systems for Mathematical Test Functions.

An Incremental Ant Colony Algorithm with Local Search for Continuous Optimization

Smoothing and Auxiliary Functions Based Cooperative Coevolution for Global Optimization

QUANTUM BASED PSO TECHNIQUE FOR IMAGE SEGMENTATION

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques

Optimization of Benchmark Functions Using Artificial Bee Colony (ABC) Algorithm

This article can be cited as I. H. Khan, A Comparative Study of Evolutionary Algorithms, International Journal of Artificial Intelligence, vol.

The Simple Genetic Algorithm Performance: A Comparative Study on the Operators Combination

متلب سایت MatlabSite.com

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

An Island Based Hybrid Evolutionary Algorithm for Optimization

An Incremental ACO R with Local Search for Continuous Optimization Problems

A Simulated Annealing algorithm for GPU clusters

EE 553 Term Project Report Particle Swarm Optimization (PSO) and PSO with Cross-over

PARTICLE SWARM OPTIMIZATION (PSO) [1] is an

DS-PSO: Particle Swarm Optimization with Dynamic and Static Topologies

Committee-based Active Learning for Surrogate-Assisted Particle Swarm Optimization of Expensive Problems

A Cultivated Differential Evolution Algorithm using modified Mutation and Selection Strategy

ABC Optimization: A Co-Operative Learning Approach to Complex Routing Problems

Combining Two Local Searches with Crossover: An Efficient Hybrid Algorithm for the Traveling Salesman Problem

Thesis Submitted for the degree of. Doctor of Philosophy. at the University of Leicester. Changhe Li. Department of Computer Science

A New Cooperative Co-evolution Algorithm Based on Variable Grouping and Local Search for Large Scale Global Optimization

Small World Network Based Dynamic Topology for Particle Swarm Optimization

SHADE with Iterative Local Search for Large-Scale Global Optimization

Binary Differential Evolution Strategies

Opportunistic Self Organizing Migrating Algorithm for Real-Time Dynamic Traveling Salesman Problem

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini

Non-Uniform Mapping in Real-Coded Genetic Algorithms

A THREAD BUILDING BLOCKS BASED PARALLEL GENETIC ALGORITHM

Cuckoo search with varied scaling factor

Ying TAN. Peking University, China. Fireworks Algorithm (FWA) for Optimization. Ying TAN. Introduction. Conventional FWA.

Optimal Design of a Parallel Beam System with Elastic Supports to Minimize Flexural Response to Harmonic Loading

Partial Opposition-based Learning Using Current Best Candidate Solution

Feeding the Fish Weight Update Strategies for the Fish School Search Algorithm. Andreas Janecek

Simulated Raindrop Algorithm for Global Optimization

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization

The Modified IWO Algorithm for Optimization of Numerical Functions

Research Article Differential Evolution Algorithm with Self-Adaptive Population Resizing Mechanism

A hybrid constrained optimization approach coupling PSO and adaptive constraint-handling technique

Modified Particle Swarm Optimization with Novel Modulated Inertia for Velocity Update

SwarmOps for Java. Numeric & Heuristic Optimization Source-Code Library for Java The Manual Revision 1.0

A Comparison of Several Heuristic Algorithms for Solving High Dimensional Optimization Problems

Self-Adaptive Differential Evolution Algorithm in Constrained Real-Parameter Optimization

Using Interval Scanning to Improve the Performance of the Enhanced Unidimensional Search Method

SwarmOps for Matlab. Numeric & Heuristic Optimization Source-Code Library for Matlab The Manual Revision 1.0

Chaos Genetic Algorithm Instead Genetic Algorithm

Adaptative Clustering Particle Swarm Optimization

DIFFERENTIAL evolution (DE) is a simple yet powerful

NICHING EVOLUTIONARY ALGORITHMS FOR MULTIMODAL AND DYNAMIC OPTIMIZATION

Framework for Bat Algorithm Optimization Metaheuristic

A Comparative Study on Nature Inspired Algorithms with Firefly Algorithm

Optimized Algorithm for Particle Swarm Optimization

With the fast growing size and complexity of modern

Individual Parameter Selection Strategy for Particle Swarm Optimization

Feeding the Fish Weight Update Strategies for the Fish School Search Algorithm

Experiments with Firefly Algorithm

An evolutionary annealing-simplex algorithm for global optimisation of water resource systems

SINCE PARTICLE swarm optimization (PSO) was introduced

Assessing Particle Swarm Optimizers Using Network Science Metrics

Optimistic Path using Artificial Bee Colony Approach

A Polar Coordinate Particle Swarm Optimiser

Pre-requisite Material for Course Heuristics and Approximation Algorithms

Scalability of generalized adaptive differential evolution for large-scale continuous optimization

A novel meta-heuristic algorithm for numerical function optimization: Blind, naked mole-rats (BNMR) algorithm

THE DEVELOPMENT OF THE POTENTIAL AND ACADMIC PROGRAMMES OF WROCLAW UNIVERISTY OF TECHNOLOGY METAHEURISTICS

PARTICLE Swarm Optimization (PSO), an algorithm by

Automatic differentiation based for particle swarm optimization steepest descent direction

FIREWORKS algorithm (FWA) is a newly proposed swarm

Imperialist Competitive Algorithm using Chaos Theory for Optimization (CICA)

Modified K-Means Algorithm for Genetic Clustering

Offspring Generation Method using Delaunay Triangulation for Real-Coded Genetic Algorithms

FUZZY FA: A MODIFIED FIREFLY ALGORITHM

The Coral Reefs Optimization Algorithm: An Efficient Meta-heuristic for Solving Hard Optimization Problems

Transcription:

Benchmark Functions for the CEC 2008 Special Session and Competition on Large Scale Global Optimization K. Tang 1, X. Yao 1, 2, P. N. Suganthan 3, C. MacNish 4, Y. P. Chen 5, C. M. Chen 5, Z. Yang 1 1 Nature Inspired Computation and Applications Laboratory (NICAL), Department of Computer Science and Technology, the University of Science and Technology of China, Hefei, Anhui, China 2 The Center of Excellence for Research in Computational Intelligence and Applications (CERCIA), School of Computer Science, the University of Birminham, Edgbaston, Birmingham B15 2TT, U.K. 3 School of EEE, Nanyang Technological University, Singapore, 639798 4 School of Computer Science & Software Engineering, the University of Western Australia, M002, 35 Stirling Highway, Crawley, Western Australia, 6009 5 Natural Computing Laboratory, Department of Computer Science, National Chiao Tung University, Taiwan ketang@ustc.edu.cn, x.yao@cs.bham.ac.uk, epnsugan@ntu.edu.sg, cara@csse.uwa.edu.au, ypchen@nclab.tw, ccming@nclab.tw, zhyuyang@mail.ustc.edu.cn 1

In the past two decades, different kinds of nature-inspired optimization algorithms have been designed and applied to solve optimization problems, e.g., simulated annealing (SA), evolutionary algorithms (EAs), differential evolution (DE), particle swarm optimization (PSO), Ant Colony Optimisation (ACO), Estimation of Distribution Algorithms (EDA), etc. Although these approaches have shown excellent search abilities when applying to some 30-100 dimensional problems, many of them suffer from the "curse of dimensionality", which implies that their performance deteriorates quickly as the dimensionality of search space increases. The reasons appear to be two-fold. First, complexity of the problem usually increases with the size of problem, and a previously successful search strategy may no longer be capable of finding the optimal solution. Second, the solution space of the problem increases exponentially with the problem size, and a more efficient search strategy is required to explore all the promising regions in a given time budget. Historically, scaling EAs to large size problems have attracted much interest, including both theoretical and practical studies. The earliest practical approach might be the parallelism of an existing EA. Later, cooperative coevolution appears to be another promising method. However, existing work on this topic are often limited to the test problems used in individual studies, and a systematic evaluation platform is not available in the literature for comparing the scalability of different EAs. In this report, 6 benchmark functions are given based on [1] and [2] for high-dimensional optimization. All of them are scalable for any size of dimension. The codes in Matlab and C for them are available at http://nical.ustc.edu.cn/cec08ss.php. The other benchmark function (Function 7 - FastFractal "DoubleDip") is generated based on [3] [4]. The C code for function 7 has been contributed by Ales Zamuda from the University of Maribor, Slovenia. It uses the GJC / CNI interface to run the Java code from C++. In the package, C code is provided in a separate zip file, named cec08-f7-cpp.zip. The mathematical formulas and properties of these functions are described in Section 2, and the evaluation criteria are given in Section 3. 1. Summary of the 7 CEC 08 Test Functions Unimodal Functions (2): F 1 : Shifted Sphere Function F 2 : Shifted Schwefel s Problem 2.21 Multimodal Functions (5): F 3 : Shifted Rosenbrock s Function F 4 : Shifted Rastrigin s Function F 5 : Shifted Griewank s Function F 6 : Shifted Ackley s Function F 7 : FastFractal DoubleDip Function 2

2. Definitions of the 7 CEC 08 Test Functions 2.1 Unimodal Functions: 2.1.1. F 1 : Shifted Sphere Function D 2 1( ) = i + _ 1 i= 1 F x z f bias, z = x o, x = [ x1, x2,..., x D ] D: dimensions. o = [ o1, o2,..., o D ] : the shifted global optimum. Figure 2-1 3-D map for 2-D function Properties: Unimodal Shifted Separable Scalable Dimension D as 100, 500 and 1000 x [ 100,100] D, Global optimum: x* = o, F (*) 1 x = f_bias1 = - 450 Associated Data files: Name: sphere shift_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 3

2.1.2. F 2 : Schwefel s Problem 2.21 F ( x ) = max{ z,1 i D} + f _ bias, z = x o, x = [ x1, x2,..., x D ] 2 i 2 i D: dimensions. o = [ o1, o2,..., o D ] : the shifted global optimum. Figure 2-2 3-D map for 2-D function Properties: Unimodal Shifted Non-separable Scalable Dimension D as 100, 500 and 1000 x [ 100,100] D, Global optimum: x* = o, F (*) 1 x = f_bias1 = - 450 Associated Data files: Name: schwefel_shift_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 4

2.2 Multimodal Functions 2.2.1. F 3 : Shifted Rosenbrock s Function D 1 2 2 2 3( ) = (100( i i+ 1) + ( i 1) ) + _ 3 i= 1 F x z z z f bias, z = x o + 1, x = [ x1, x2,..., x D ] D: dimensions o = [ o1, o2,..., o D ] : the shifted global optimum 10 x 10 10 8 6 4 2 0 100 50 0 50 Figure 2-3 3-D map for 2-D function Properties: Multi-modal Shifted Non-separable Scalable Having a very narrow valley from local optimum to global optimum Dimension D as 100, 500 and 1000 x [ 100,100] D, Global optimum * * x = o, F ( ) 3 x = f_bias3 = 390 100 100 50 0 50 100 Associated Data file: Name: rosenbrock_shift_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 5

2.2.2. F 4 : Shifted Rastrigin s Function D 2 4( ) = ( i 10cos(2 π i) + 10) + _ 4 i= 1 F x z z f bias, z = x o, x = [ x1, x2,..., x D ] D: dimensions o = [ o1, o2,..., o D ] : the shifted global optimum Figure 2-4 3-D map for 2-D function Properties: Multi-modal Shifted Separable Scalable Local optima s number is huge Dimension D as 100, 500 and 1000 x [ 5,5] D, Global optimum * * x = o, F ( ) 4 x = f_bias4 = - 330 Associated Data file: Name: rastrigin_shifit_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 6

2.2.3. F 5 : Shifted Griewank s Function z z F x f bias, z = ( x o ), x = [ x1, x2,..., x D ] D 2 D i i 5( ) = cos( ) + 1 + _ 5 i= 1 4000 i= 1 i D: dimensions o = [ o1, o2,..., o D ] : the shifted global optimum Figure 2-5 3-D map for 2-D function Properties: Multi-modal Shifted Non-separable Scalable Dimension D as 100, 500 and 1000 x [ 600,600] D, Global optimum * * x = o, F ( ) 5 x = f_bias5 = -180 Associated Data file: Name: griewank_shfit_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 7

2.2.4. F 6 : Shifted Ackley s Function 1 1 F x z z e f bias, z = ( x o ), D D 2 6( ) = 20 exp( 0.2 i ) exp( cos(2 π i)) + 20 + + _ 6 D i= 1 D i= 1 = [ x1, x2,..., x D ], D: dimensions = [ o1, o2,..., o D ] x o : the shifted global optimum; Figure 2-6 3-D map for 2-D function Properties: Multi-modal Shifted Separable Scalable Dimension D as 100, 500 and 1000 x [ 32,32] D, Global optimum * * x = o, F ( ) 6 x = f_bias6 = - 140 Associated Data file: Name: ackley_shift_func_data.mat Variable: o 1*1000 vector the shifted global optimum When using, cut o=o(1:d) for D=100, 500 8

2.2.5. F 7 : FastFractal DoubleDip Function Note: To make use of this function in a.m file, please include the following line at the beginning of your file that calls the function: javaclasspath('fractalfunctions.jar') D F ( x ) = fractal1d( x + twist( x + )) 7 i ( imod D) 1 i= 1 ( ) = 4( 4-2 3 + 2 ) twist y y y y k 1 3 2 ran2( o) 1 fractal1d( x) doubledip( x, ran1( o), ) k 1 2 (2 ran1( o)) k = 1 1 1 x c + x c x c + s < x< doubledip( x, c, s) = 0, otherwise 6 4 2 ( 6144( ) 3088( ) 392( ) 1), 0.5 0.5 x = [,,..., ], D: dimensions x1 x2 x D ran1(o): double, pseudorandomly chosen, with seed o, with equal probability from the interval [0,1] ran2(o): integer, pseudorandomly chosen, with seed o, with equal probability from the set {0,1,2} fractal1d(x) is an approximation to a recursive algorithm, it does not take account of wrapping at the boundaries, or local re-seeding of the random generators - please use the executable provided Figure 2-7 3-D map for 2-D function 9

Properties: Multi-modal Non-separable Scalable Dimension D as 100, 500 and 1000 x [ 1,1] D, Global optimum unknown, F ( x * ) unknown 7 Associated Data file: Name: fastfractal_doubledip_data.mat Variable: o integer seeds the random generators 10

3. Evaluation Criteria 3.1 Description of the Evaluation Criteria Problems: 7 minimization problems Dimensions: D=100, 500, 1000 Runs / problem: 25 (Do not run many 25 runs to pick the best run) Max_FES: 5000*D (Max_FES_100D= 500000; for 500D=2500000; for 1000D=5000000) Initialization: Uniform random initialization within the search space Global Optimum: All problems have the global optimum within the given bounds and there is no need to perform search outside of the given bounds for these problems. Termination: Terminate when reaching Max_FES. 1) Record function error value (f(x)-f(x*)) after 1 100 *FES, 1 *FES and FES at 10 termination (due to Max_FES) for each run. For each function, sort the error values in 25 runs from the smallest (best) to the largest (worst) Present the following: 1 st (best), 7 th, 13 th (median), 19 th, 25 th (worst) function values Mean and STD for the 25 runs NOTE: The value of f(x*) is not available for function 7 (FastFractal DoubleDip Function). In this case, please record the function value f(x) directly (i.e., we regard the f(x*) is 0). 2) Convergence Graphs (or Run-length distribution graphs) Convergence Graphs for each problem for D=1000. The graph would show the median performance of the total runs with termination by the Max_FES. The semi-log graphs should show log10( f(x)- f(x*)) vs FES for the first 6 functions (as shown in Figure 3-1). NOTE: The function 7 always takes negative value. In this case, please plot ( f(x)- f(x*) vs FES directly (as shown in Figure 3-2). 11

3) Parameters We discourage participants searching for a distinct set of parameters for each problem/dimension/etc. Please provide details on the following whenever applicable: a) All parameters to be adjusted b) Actual parameter values used. c) Estimated cost of parameter tuning in terms of number of FEs d) Corresponding dynamic ranges e) Guidelines on how to adjust the parameters 4) Encoding If the algorithm requires encoding, then the encoding scheme should be independent of the specific problems and governed by generic factors such as the search ranges. FYI: Estimated runtime for the test suite Dimension: 1000D Problems: Functions 1-7 Algorithm: Differential Evolution Runs: Only one run Max_FES: 5000000 PC: CPU-P4 2.8G, RAM-512M Runtime: 15 hours 12

3.2 Example System: Windows XP (SP1) CPU: Pentium(R) 4 3.00GHz RAM: 1 G Language: Matlab 7.1 Algorithm: Particle Swarm Optimizer (PSO) Results, D=100, Max_FES=500000 Table 3-1 Error Values Achieved for Problems 1-7, with D=100 FES 5.00e+3 5.00e+4 5.00e+5 Prob 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 2 3 4 5 6 7 13

Results, D=500, Max_FES=2500000 Table 3-2 Error Values Achieved for Problems 1-7, with D=500 FES 2.50e+4 2.50e+5 2.50e+6 Prob 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 2 3 4 5 6 7 14

Results, D=1000, Max_FES=5000000 Table 3-3 Error Values Achieved for Problems 1-7, with D=1000 FES 5.00e+4 5.00e+5 5.00e+6 Prob 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 st (Best) 7 th 13 th (Median) 19 th 25 th (Worst) Mean Std 1 2 3 4 5 6 7 15

Convergence Graphs (Function 1-6, 1000D) 10 15 f 1 10 10 f 2 f 3 log(f(x) f(x*)) 10 5 10 0 10 5 f 4 f 5 f 6 10 10 10 15 0 1 2 3 4 5 FEs x 10 6 Figure 3-1 Convergence Graph for Functions 1-6 16

Convergence Graphs (Function 7, 1000D) 0.6 x 104 f 7 0.7 0.8 0.9 f(x) f(x*) 1 1.1 1.2 1.3 1.4 1.5 0 1 2 3 4 5 FEs x 10 6 Figure 3-2 Convergence Graph for Function 7 17

References: [1] P. N. Suganthan, N. Hansen, J. J. Liang, K. Deb, Y. P. Chen, A. Auger, and S. Tiwari, Problem Definitions and Evaluation Criteria for the CEC 2005 Special Session on Real- Parameter Optimization, Technical Report, Nanyang Technological University, Singapore, http://www.ntu.edu.sg/home/epnsugan, 2005. [2] X. Yao, Y. Liu, and G. Lin, Evolutionary Programming Made Faster, IEEE Transactions on Evolutionary Computation, vol. 3, no. 2, pp. 82 102, 1999. [3] MacNish, C.,Towards Unbiased Benchmarking of Evolutionary and Hybrid Algorithms for Real-valued Optimisation, Connection Science, Vol.19, No. 4, December 2007. To appear. [4] MacNish, C., Benchmarking Evolutionary and Hybrid Algorithms using Randomized Self-Similar Landscapes, Proc. 6th International Conference on Simulated Evolution and Learning (SEAL'06), LNCS 4247, pp. 361-368, Springer, 2006. 18