1 Lab + Hwk 5: Particle Swarm Optimization

Size: px
Start display at page:

Download "1 Lab + Hwk 5: Particle Swarm Optimization"

Transcription

1 1 Lab + Hwk 5: Particle Swarm Optimization This laboratory requires the following equipment: C programming tools (gcc, make). Webots simulation software. Webots User Guide Webots Reference Manual. The laboratory duration is about 3 hours. Homework will be due on the 7 th day after your lab session, at 12 noon. Please format your solution as a PDF file with the name [name]_lab[#].pdf, where [name] is your account user name and [#] is the number of the lab+homework assignment, and upload it as the Report submission onto the student section of the course webpage. All additional material (movies, code, and any other relevant files) should be archived as a zip file with the name [name]_lab[#].tgz (hint: tar cvzf [name]_lab[#].tgz [directory]) and uploaded as the Additional Material submission onto the student section of the course webpage. Every misnamed file will result in a 1 point penalty to your assignment grade. Keep in mind that no late solutions will be accepted unless supported by health reasons (and a note from the doctor). If there are extreme circumstances which will prevent you from attending a particular lab session, it may be possible to make arrangements with the TAs before the lab in question takes place. The more advance notice you can give in these situations, the more likely it is that we will be able to work something out. 1.1 Grading In the following text you will find several exercises and questions. The notation S x means that the question can be solved using only additional simulation; you are not required to write or produce anything for these questions. The notation Q x means that the question can be answered theoretically, without any simulation; your answers to these questions must be submitted in your Report. The length of answers should be approximately two sentences unless otherwise noted. The notation I x means that the problem has to be solved by implementing a piece of code and performing a simulation; the code written for these questions must be submitted in your Additional Material. Please use this notation in your answer file! The number of points for each exercise is given between parentheses. The combined total number of points for the laboratory and homework exercises is Optimization In many instances, we want to find either the minimum or maximum value of some function. If the function in question is complex or even unknown, it is impossible to JP; TL, Lab+Hwk 5: Particle Swarm Optimization 1

2 accomplish this task analytically. Furthermore, if the function parameter space is very large, a systematic search of this space is often too computationally expensive. Therefore, a number of techniques have been developed to find near-optimal solutions (i.e. local minima and maxima), such as genetic algorithms and particle swarm optimization. 1.3 The Particle Swarm To find minima/maxima, repeated evaluations of the function in question must be done at different points. A good optimization algorithm will use the information obtained from these evaluations to choose the locations of future evaluations, and eventually to decide where the minimum/maximum is. In Particle Swarm Optimization, a set of particles is initialized with a random position for each particle. Each particle is also assigned a random velocity in the search space. Particles are then flown through the search space, moving at their respective velocities. Optimization is achieved by each particle remembering at what position they achieved the best evaluation of the function, or the particle s personal best. Particles also remember the best achieving position of their neighborhood. The neighborhood for some particle A is a group of particles to which A belongs. This group can be topological (whatever particles are closest) or fixed throughout the algorithm. It can consist of either a subset of the particle swarm (local neighborhood), or the entire swarm (global neighborhood). The neighborhood best of a neighborhood is the position that yielded the best evaluation by any particle in the neighborhood. At each generation of the algorithm, the velocity of each particle is updated, using a randomized attraction to both the particle best and the neighborhood best. This allows particles to move towards areas of the search space that have yielded good results, and in doing so, discover nearby better results. In this way, the particles will eventually converge on possibly global optima. The idea for this type of behavior took inspiration from the ways birds act while flying in a flock. 2 Lab: Using PSO 2.1 PSO on the Sphere Function The Sphere function is defined as follows: f ( x) = n 2 x i i= 1 We want to find the set of x i that give the value closest to 0 (given by x i = 0 for all x i ). The values for x i are initially randomly distributed between and This function is often used as a simple test for optimization functions; although it may seem trivial to a human observer, it can be difficult for an optimization algorithm to figure out that the best solution is all zeroes. We will use PSO with a swarm size of 30 and the neighborhood of a particle given by the three nearest particles on each side (e.g. particle 8 has neighborhood {5, 6, 7, 8, 9, 10, 11}) with particle 29 being next to particle 0. We will optimize the Sphere Comment [AM1]: I.e. mention neighborhood index = 7 JP; TL, Lab+Hwk 5: Particle Swarm Optimization 2

3 function with n = 10, and, therefore, ten components in the vector representing each particle. In other words, a particle is a set of 10 numbers, and the closer to 0 all the numbers become, the better the performance. Copy the pso directory from the student section of the course webpage into your home directory. Compile the PSO program using make. You can execute the program with./main <number of iterations>. S: Run the program with 10 iterations, 100 iterations, 1000 iterations, and iterations. At the end of each trial, both the achieved value and the found solution for 10 runs will be printed along with the average value. Lower values mean better the solutions. Q 1 (5): What do you observe? S: Try the simulation with iterations. Q 2 (5): How many iterations do you think will be necessary to get the exact answer? Why? S: The definition VMAX in main.c controls the maximum velocity that particles can achieve in the simulation. Using 100 iterations, try varying VMAX from 0.5 to 4.0. Q 3 (5): What do you observe? S: Try varying VMAX from to Q 4 (5): What happens now? Why? It can be very inconvenient to have to tune VMAX to give the optimal performance for a simulation. A feature that was developed for PSO soon after its invention was an inertia coefficient. This was a coefficient typically less than 1 which is multiplied with the velocity at each iteration, in order to naturally slow particles down without imposing a hard velocity threshold. S: Uncomment line 77 of pso.c (it will be marked in the code), set VMAX to 4.0, and run the program again. Try increasing the number of iterations. Q 5 (5): What do you observe now? 2.2 PSO for Unsupervised Learning We ve established that PSO can do a good job of optimizing simple mathematical functions. The next step is to test it in a more demanding environment, in particular in the presence of noisy functions. In Evolution of Homing Navigation in a Real Mobile Robot (Floreano and Mondada, 1996), a genetic algorithm was used to train a robot to move through a maze and avoid obstacles. We want to use Webots to evolve robotic controllers for the mobile robots, comparing the performance of GA and PSO. For the experiments in this lab, we use the e-puck robot instead of the Khepera that was used in the paper. The behavior we wish to achieve on the robots is obstacle avoidance, the same as in the Floreano and Mondada paper. Copy the ga_obs_sup, pso_obs_sup and obs_con controllers and the ga_obs and pso_obs worlds from the student section of the course webpage. Compile all controllers. These worlds are simple walled arenas with no obstacles and twenty e-puck robots. We use a two neuron, single-layer, neural networks to control each robot, with proximity sensors and the JP; TL, Lab+Hwk 5: Particle Swarm Optimization 3

4 relative position center of mass as inputs and the motor speed as outputs. There are recursive and lateral connections from the outputs of both neurons back into each other. The controller we evolve is the set of weights for these neurons (8 proximity sensors + 2 recursive/lateral connections + 1 threshold = 11 weights for each neuron => 22 weights total). The fitness of a controller is measured with the equation: F = V ( 1 v)(1 i), V, v, i [0,1] where V is a measure of the average rotation speed of the two wheels over the trial, v is a measure of the average difference in wheel speed of the two wheels over the trial, and i is the average activation value of the proximity sensor with the highest activity. Each of these terms in the fitness function encourage the robot to go fast, go straight, and not stay close to walls, respectively. The default time for a single evaluation of fitness is ~30 s. The genetic algorithm we re using is similar to that used in 1996 by Floreano and Mondada. It selects the best performing half of the population as potential parents. Out of these, parents are selected using a Roulette Wheel scheme. Crossover occurs with probability 0.2 and mutation occurs with probability 0.15 to create children to replenish the missing half of the population Fast Heterogeneous Learning Using PSO or GA on a single robot, it can take a long time to evolve highperforming controllers. This is because the robot will have to test every member of the population/swarm one after another for each iteration of the algorithm. For example, doing 100 iterations of PSO with 20 particles with each evaluation taking 1 minute means we need 2000 minutes or over 36 hours to complete! One way to speed up this process is to use multiple robots doing heterogeneous learning. In heterogeneous learning, each robot is using a different controller than the others. By using multiple robots, we can distribute the particles among the robots and evaluate their performance in parallel, thus saving a lot of time. Going back to our previous example, if we have 20 robots each evaluating a different particle, it will only take 1 minute to evaluate 20 particles, and the total evolution time goes down to about 1.7 hours. We will now test evolution of obstacle avoidance behavior using heterogeneous learning with GA and PSO. In the follow questions, you will be asked to run simulations that may take several minutes to complete. In order to save time, you may want to look ahead to the next question to see if it is something which you can work on while you wait. S: Load the ga_obs world. Run the world to simulate the evolution. You will need to use fast mode to simulate quickly enough, but you may want to observe how the robots are behaving occasionally (pay attention to whether they are near the beginning or end of an evolutionary run). The simulation will do 10 separate evolutionary runs and print the final fitness for each, as well as the average fitness at the end. When the evolution is finished, observe the performance of the robots (they will be running the best found controller). JP; TL, Lab+Hwk 5: Particle Swarm Optimization 4

5 Q 6 (5): What kind of fitness values do you get? S: Now load the pso_obs world, and run the simulations again. Q 7 (5): How does the performance compare? We now switch to using the modified version of our genetic algorithm specifically designed for noisy fitness function evaluation (see Zhang, 2003). At each iteration, the performance of the potential parent set is reevaluated, and the new fitness value is combined with previous values to get a combined estimate. Q 8 (5): Why might this modification help to improve performance? S: In the ga.c file, set the NOISY definition from 0 to 1. This enables the modification. Run the simulations again. Q 9 (5): Record the final fitness values. How do the new values compare? The modification for GA for handling noisy fitness function evaluations was adapted for PSO in Particle Swarm Optimization for Unsupervised Robotic Learning (Pugh, 2005). In this, at each iteration, the previous best positions of the particles are re-evaluated, and the new fitness value combined with previous ones. S: In the pso.c file, set the NOISY definition from 0 to 1. This enables the modification. Run the simulations again. Q 10 (5): Record the final fitness values. How do the new fitness values compare? Q 11 (5): In both modifications, the number of iterations run by the algorithm is divided by two. Why does this make sense? JP; TL, Lab+Hwk 5: Particle Swarm Optimization 5

6 3 Hwk: Analyzing PSO Depending on the function being optimized, different parameter settings will give better results in PSO. We will now explore the influence of parameters on a more complex numerical function, the Rastrigin function, given by: n 2 [ x i 10cos( 2 x i ) + 10] f ( x) = π i= 1 This function contains many local optima which could cause premature convergence of particles. S: Go back to the pso directory. Rerun optimization of the Sphere function (with the inertia coefficient) setting the number of particles to 3, 5, 10, 15, 30, and 60. For each number of particles n, set the number of iterations to 3000/n (i.e. 1000, 600, 300, 200, 100, 50; this will keep the computational time approximately constant). You can specify the number of particles from the command line with./main <number of iterations> <number of particles>. Q 12 (5): What performances did you get for each number of particles? Which number is best? S: In pso.c, change the definition of FUNCTION from 1 to 2. This selects the Rastrigin function. Try running the simulation again while varying the number of particles. Q 13 (5): What performances did you get now for each number of particles? Which number is best? Q 14 (10): How do the results for the Sphere and Rastrigin functions compare? Why is this happening? You may need up to five lines to fully explain this. Next, we will explore when it s useful to use a global neighborhood, where all the particles are attracted to the single best found position, and when it s useful to use a local neighborhood, when different groups of particles are attracted to different positions. S: On the Sphere function, rerun the optimization setting the definition of NB (number of neighbors on each side) to 1 (local neighborhood) and 15 (global neighborhood) with 15 particles and 100 iterations. Q 15 (5): What performances did you get? Which neighborhood type is best? S: Now set the number of iterations to 1000, and run the same experiments on the Rastrigin function. Q 16 (5): What performances did you get now? Which neighborhood type is best? Q 17 (5): For evolving aggregation in Webots, robots can cluster well either by moving forwards towards the group or by moving backwards towards the group. However, robots are better able to navigate to a more tight cluster going forwards, as they have more proximity sensors to see what they are approaching, and will therefore give a slightly better fitness overall. Would a global or local neighborhood be preferable here? Q 18 (10): What if you want to optimize the Sphere function, but with Gaussian noise added. Would this affect your neighborhood choice? Why? JP; TL, Lab+Hwk 5: Particle Swarm Optimization 6

7 4 References Floreano D. and Mondada F., "Evolution of Homing Navigation in a Real Mobile Robot", IEEE Trans. on System, Man, and Cybernetics: Part B, 26(3): , Zhang Y., Martinoli A., and Antonsson E. K. Evolutionary Design of a Collective Sensory System, In H. Lipson, E. K. Antonsson, and J. R. Koza, editors, Proc. of the 2003 AAAI Spring Symposium on Computational Synthesis, March 2003, Stanford, CA, pp AAAI Press Pugh, J., Zhang, Y., and Martinoli, A. Particle Swarm Optimization for Unsupervised Robotic Learning, In Proc. of the IEEE Swarm Intelligence Symposium 2005, Pasadena, CA. JP; TL, Lab+Hwk 5: Particle Swarm Optimization 7

1 Lab + Hwk 5: Particle Swarm Optimization

1 Lab + Hwk 5: Particle Swarm Optimization 1 Lab + Hwk 5: Particle Swarm Optimization This laboratory requires the following equipment: C programming tools (gcc, make), already installed in GR B001 Webots simulation software Webots User Guide Webots

More information

1 Lab 5: Particle Swarm Optimization

1 Lab 5: Particle Swarm Optimization 1 Lab 5: Particle Swarm Optimization This laboratory requires the following: (The development tools are installed in GR B0 01 already): C development tools (gcc, make, etc.) Webots simulation software

More information

Discrete Multi-Valued Particle Swarm Optimization

Discrete Multi-Valued Particle Swarm Optimization Discrete Multi-d Particle Swarm Optimization Jim Pugh and Alcherio Martinoli Swarm-Intelligent Systems Group École Polytechnique Fédérale de Lausanne 05 Lausanne, Switzerland Email: {jim.pugh,alcherio.martinoli}@epfl.ch

More information

1 Lab+Hwk 3: Introduction to the Webots Robotics Simulation Software

1 Lab+Hwk 3: Introduction to the Webots Robotics Simulation Software 1 Lab+Hwk 3: Introduction to the Webots Robotics Simulation Software This laboratory requires the following software (pre-installed in BC07/08): C development tools (gcc, make). Webots simulation software.

More information

An Introduction to Distributed Robotic Systems

An Introduction to Distributed Robotic Systems An Introduction to Distributed Robotic Systems Alcherio Martinoli School of Architecture, Civil and Environmental Engineering Distributed Intelligent Systems and Algorithms Laboratory École Polytechnique

More information

Modified Particle Swarm Optimization

Modified Particle Swarm Optimization Modified Particle Swarm Optimization Swati Agrawal 1, R.P. Shimpi 2 1 Aerospace Engineering Department, IIT Bombay, Mumbai, India, swati.agrawal@iitb.ac.in 2 Aerospace Engineering Department, IIT Bombay,

More information

LECTURE 16: SWARM INTELLIGENCE 2 / PARTICLE SWARM OPTIMIZATION 2

LECTURE 16: SWARM INTELLIGENCE 2 / PARTICLE SWARM OPTIMIZATION 2 15-382 COLLECTIVE INTELLIGENCE - S18 LECTURE 16: SWARM INTELLIGENCE 2 / PARTICLE SWARM OPTIMIZATION 2 INSTRUCTOR: GIANNI A. DI CARO BACKGROUND: REYNOLDS BOIDS Reynolds created a model of coordinated animal

More information

Kyrre Glette INF3490 Evolvable Hardware Cartesian Genetic Programming

Kyrre Glette INF3490 Evolvable Hardware Cartesian Genetic Programming Kyrre Glette kyrrehg@ifi INF3490 Evolvable Hardware Cartesian Genetic Programming Overview Introduction to Evolvable Hardware (EHW) Cartesian Genetic Programming Applications of EHW 3 Evolvable Hardware

More information

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES 6.1 INTRODUCTION The exploration of applications of ANN for image classification has yielded satisfactory results. But, the scope for improving

More information

Tracking Changing Extrema with Particle Swarm Optimizer

Tracking Changing Extrema with Particle Swarm Optimizer Tracking Changing Extrema with Particle Swarm Optimizer Anthony Carlisle Department of Mathematical and Computer Sciences, Huntingdon College antho@huntingdon.edu Abstract The modification of the Particle

More information

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les Chaotic Crossover Operator on Genetic Algorithm Hüseyin Demirci Computer Engineering, Sakarya University, Sakarya, 54187, Turkey Ahmet Turan Özcerit Computer Engineering, Sakarya University, Sakarya, 54187,

More information

EE 553 Term Project Report Particle Swarm Optimization (PSO) and PSO with Cross-over

EE 553 Term Project Report Particle Swarm Optimization (PSO) and PSO with Cross-over EE Term Project Report Particle Swarm Optimization (PSO) and PSO with Cross-over Emre Uğur February, 00 Abstract In this work, Particle Swarm Optimization (PSO) method is implemented and applied to various

More information

Chapter 14 Global Search Algorithms

Chapter 14 Global Search Algorithms Chapter 14 Global Search Algorithms An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Introduction We discuss various search methods that attempts to search throughout the entire feasible set.

More information

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India. Volume 3, Issue 3, March 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Training Artificial

More information

PARTICLE SWARM OPTIMIZATION (PSO)

PARTICLE SWARM OPTIMIZATION (PSO) PARTICLE SWARM OPTIMIZATION (PSO) J. Kennedy and R. Eberhart, Particle Swarm Optimization. Proceedings of the Fourth IEEE Int. Conference on Neural Networks, 1995. A population based optimization technique

More information

GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM

GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM Journal of Al-Nahrain University Vol.10(2), December, 2007, pp.172-177 Science GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM * Azhar W. Hammad, ** Dr. Ban N. Thannoon Al-Nahrain

More information

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-&3 -(' ( +-   % '.+ % ' -0(+$, The structure is a very important aspect in neural network design, it is not only impossible to determine an optimal structure for a given problem, it is even impossible to prove that a given structure

More information

Speculative Evaluation in Particle Swarm Optimization

Speculative Evaluation in Particle Swarm Optimization Speculative Evaluation in Particle Swarm Optimization Matthew Gardner, Andrew McNabb, and Kevin Seppi Department of Computer Science, Brigham Young University Abstract. Particle swarm optimization (PSO)

More information

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding e Scientific World Journal, Article ID 746260, 8 pages http://dx.doi.org/10.1155/2014/746260 Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding Ming-Yi

More information

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM BAHAREH NAKISA, MOHAMMAD NAIM RASTGOO, MOHAMMAD FAIDZUL NASRUDIN, MOHD ZAKREE AHMAD NAZRI Department of Computer

More information

Evolutionary Algorithms. CS Evolutionary Algorithms 1

Evolutionary Algorithms. CS Evolutionary Algorithms 1 Evolutionary Algorithms CS 478 - Evolutionary Algorithms 1 Evolutionary Computation/Algorithms Genetic Algorithms l Simulate natural evolution of structures via selection and reproduction, based on performance

More information

Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization

Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization J.Venkatesh 1, B.Chiranjeevulu 2 1 PG Student, Dept. of ECE, Viswanadha Institute of Technology And Management,

More information

Particle Swarm Optimization

Particle Swarm Optimization Particle Swarm Optimization Gonçalo Pereira INESC-ID and Instituto Superior Técnico Porto Salvo, Portugal gpereira@gaips.inesc-id.pt April 15, 2011 1 What is it? Particle Swarm Optimization is an algorithm

More information

PARTICLE SWARM OPTIMIZATION (PSO) [1] is an

PARTICLE SWARM OPTIMIZATION (PSO) [1] is an Proceedings of International Joint Conference on Neural Netorks, Atlanta, Georgia, USA, June -9, 9 Netork-Structured Particle Sarm Optimizer Considering Neighborhood Relationships Haruna Matsushita and

More information

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques Nasser Sadati Abstract Particle Swarm Optimization (PSO) algorithms recently invented as intelligent optimizers with several highly

More information

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization 2017 2 nd International Electrical Engineering Conference (IEEC 2017) May. 19 th -20 th, 2017 at IEP Centre, Karachi, Pakistan Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic

More information

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 20 CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 2.1 CLASSIFICATION OF CONVENTIONAL TECHNIQUES Classical optimization methods can be classified into two distinct groups:

More information

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm A. Lari, A. Khosravi and A. Alfi Faculty of Electrical and Computer Engineering, Noushirvani

More information

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra Pattern Recall Analysis of the Hopfield Neural Network with a Genetic Algorithm Susmita Mohapatra Department of Computer Science, Utkal University, India Abstract: This paper is focused on the implementation

More information

Mobile Robot Path Planning in Static Environment

Mobile Robot Path Planning in Static Environment Mobile Robot Path Planning in Static Environment A Thesis Submitted in Partial Fulfilment of the Requirements for the Degree of Bachelor of Technology in Computer Science & Engineering Submitted by: Raman

More information

Fall 09, Homework 5

Fall 09, Homework 5 5-38 Fall 09, Homework 5 Due: Wednesday, November 8th, beginning of the class You can work in a group of up to two people. This group does not need to be the same group as for the other homeworks. You

More information

Machine Evolution. Machine Evolution. Let s look at. Machine Evolution. Machine Evolution. Machine Evolution. Machine Evolution

Machine Evolution. Machine Evolution. Let s look at. Machine Evolution. Machine Evolution. Machine Evolution. Machine Evolution Let s look at As you will see later in this course, neural networks can learn, that is, adapt to given constraints. For example, NNs can approximate a given function. In biology, such learning corresponds

More information

Generation of Ultra Side lobe levels in Circular Array Antennas using Evolutionary Algorithms

Generation of Ultra Side lobe levels in Circular Array Antennas using Evolutionary Algorithms Generation of Ultra Side lobe levels in Circular Array Antennas using Evolutionary Algorithms D. Prabhakar Associate Professor, Dept of ECE DVR & Dr. HS MIC College of Technology Kanchikacherla, AP, India.

More information

FOREST PLANNING USING PSO WITH A PRIORITY REPRESENTATION

FOREST PLANNING USING PSO WITH A PRIORITY REPRESENTATION FOREST PLANNING USING PSO WITH A PRIORITY REPRESENTATION P.W. Brooks and W.D. Potter Institute for Artificial Intelligence, University of Georgia, USA Overview Background: (NIO Project 1 ) PSO -- GA --

More information

CS 135 Lab Assignments Week 1

CS 135 Lab Assignments Week 1 CS 135 Lab Assignments Week 1 Professor: Matt B. Pedersen This handout is the assignment that you must finish for the lab portion of the course in week 1. You must finish the assignments yourself; if you

More information

ARTIFICIAL INTELLIGENCE (CSCU9YE ) LECTURE 5: EVOLUTIONARY ALGORITHMS

ARTIFICIAL INTELLIGENCE (CSCU9YE ) LECTURE 5: EVOLUTIONARY ALGORITHMS ARTIFICIAL INTELLIGENCE (CSCU9YE ) LECTURE 5: EVOLUTIONARY ALGORITHMS Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Optimisation problems Optimisation & search Two Examples The knapsack problem

More information

Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization

Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization Richa Agnihotri #1, Dr. Shikha Agrawal #1, Dr. Rajeev Pandey #1 # Department of Computer Science Engineering, UIT,

More information

Particle Swarm Optimization

Particle Swarm Optimization Dario Schor, M.Sc., EIT schor@ieee.org Space Systems Department Magellan Aerospace Winnipeg Winnipeg, Manitoba 1 of 34 Optimization Techniques Motivation Optimization: Where, min x F(x), subject to g(x)

More information

Luo, W., and Li, Y. (2016) Benchmarking Heuristic Search and Optimisation Algorithms in Matlab. In: 22nd International Conference on Automation and Computing (ICAC), 2016, University of Essex, Colchester,

More information

Homework Assignment #3

Homework Assignment #3 CS 540-2: Introduction to Artificial Intelligence Homework Assignment #3 Assigned: Monday, February 20 Due: Saturday, March 4 Hand-In Instructions This assignment includes written problems and programming

More information

ATI Material Do Not Duplicate ATI Material. www. ATIcourses.com. www. ATIcourses.com

ATI Material Do Not Duplicate ATI Material. www. ATIcourses.com. www. ATIcourses.com ATI Material Material Do Not Duplicate ATI Material Boost Your Skills with On-Site Courses Tailored to Your Needs www.aticourses.com The Applied Technology Institute specializes in training programs for

More information

Small World Network Based Dynamic Topology for Particle Swarm Optimization

Small World Network Based Dynamic Topology for Particle Swarm Optimization Small World Network Based Dynamic Topology for Particle Swarm Optimization Qingxue Liu 1,2, Barend Jacobus van Wyk 1 1 Department of Electrical Engineering Tshwane University of Technology Pretoria, South

More information

Evolutionary Design of a Collective Sensory System

Evolutionary Design of a Collective Sensory System From: AAAI Technical Report SS-03-02. Compilation copyright 2003, AAAI (www.aaai.org). All rights reserved. Evolutionary Design of a Collective Sensory System Yizhen Zhang 1,2, Alcherio Martinoli 1, and

More information

HPSOM: A HYBRID PARTICLE SWARM OPTIMIZATION ALGORITHM WITH GENETIC MUTATION. Received February 2012; revised June 2012

HPSOM: A HYBRID PARTICLE SWARM OPTIMIZATION ALGORITHM WITH GENETIC MUTATION. Received February 2012; revised June 2012 International Journal of Innovative Computing, Information and Control ICIC International c 2013 ISSN 1349-4198 Volume 9, Number 5, May 2013 pp. 1919 1934 HPSOM: A HYBRID PARTICLE SWARM OPTIMIZATION ALGORITHM

More information

Model Parameter Estimation

Model Parameter Estimation Model Parameter Estimation Shan He School for Computational Science University of Birmingham Module 06-23836: Computational Modelling with MATLAB Outline Outline of Topics Concepts about model parameter

More information

Self-learning Mobile Robot Navigation in Unknown Environment Using Evolutionary Learning

Self-learning Mobile Robot Navigation in Unknown Environment Using Evolutionary Learning Journal of Universal Computer Science, vol. 20, no. 10 (2014), 1459-1468 submitted: 30/10/13, accepted: 20/6/14, appeared: 1/10/14 J.UCS Self-learning Mobile Robot Navigation in Unknown Environment Using

More information

Optimization of Noisy Fitness Functions by means of Genetic Algorithms using History of Search with Test of Estimation

Optimization of Noisy Fitness Functions by means of Genetic Algorithms using History of Search with Test of Estimation Optimization of Noisy Fitness Functions by means of Genetic Algorithms using History of Search with Test of Estimation Yasuhito Sano and Hajime Kita 2 Interdisciplinary Graduate School of Science and Engineering,

More information

Genetic.io. Genetic Algorithms in all their shapes and forms! Genetic.io Make something of your big data

Genetic.io. Genetic Algorithms in all their shapes and forms! Genetic.io Make something of your big data Genetic Algorithms in all their shapes and forms! Julien Sebrien Self-taught, passion for development. Java, Cassandra, Spark, JPPF. @jsebrien, julien.sebrien@genetic.io Distribution of IT solutions (SaaS,

More information

Swarm Intelligence Particle Swarm Optimization. Erick Luerken 13.Feb.2006 CS 790R, University of Nevada, Reno

Swarm Intelligence Particle Swarm Optimization. Erick Luerken 13.Feb.2006 CS 790R, University of Nevada, Reno Swarm Intelligence Particle Swarm Optimization Erick Luerken 13.Feb.2006 CS 790R, University of Nevada, Reno Motivation Discuss assigned literature in terms of complexity leading to actual applications

More information

Neural Network Weight Selection Using Genetic Algorithms

Neural Network Weight Selection Using Genetic Algorithms Neural Network Weight Selection Using Genetic Algorithms David Montana presented by: Carl Fink, Hongyi Chen, Jack Cheng, Xinglong Li, Bruce Lin, Chongjie Zhang April 12, 2005 1 Neural Networks Neural networks

More information

Constrained Functions of N Variables: Non-Gradient Based Methods

Constrained Functions of N Variables: Non-Gradient Based Methods onstrained Functions of N Variables: Non-Gradient Based Methods Gerhard Venter Stellenbosch University Outline Outline onstrained Optimization Non-gradient based methods Genetic Algorithms (GA) Particle

More information

THE ROLE OF SENSORY-MOTOR COORDINATION Identifying Environmental Motion Dynamics with Dynamic Neural Networks

THE ROLE OF SENSORY-MOTOR COORDINATION Identifying Environmental Motion Dynamics with Dynamic Neural Networks THE ROLE OF SENSORY-MOTOR COORDINATION Identifying Environmental Motion Dynamics with Dynamic Neural Networks Stephen Paul McKibbin, Bala Amavasai, Arul N. Selvan, Fabio Caparrelli and W. A. F. W. Othman

More information

Review: Final Exam CPSC Artificial Intelligence Michael M. Richter

Review: Final Exam CPSC Artificial Intelligence Michael M. Richter Review: Final Exam Model for a Learning Step Learner initially Environm ent Teacher Compare s pe c ia l Information Control Correct Learning criteria Feedback changed Learner after Learning Learning by

More information

A Modified PSO Technique for the Coordination Problem in Presence of DG

A Modified PSO Technique for the Coordination Problem in Presence of DG A Modified PSO Technique for the Coordination Problem in Presence of DG M. El-Saadawi A. Hassan M. Saeed Dept. of Electrical Engineering, Faculty of Engineering, Mansoura University, Egypt saadawi1@gmail.com-

More information

Small World Particle Swarm Optimizer for Global Optimization Problems

Small World Particle Swarm Optimizer for Global Optimization Problems Small World Particle Swarm Optimizer for Global Optimization Problems Megha Vora and T.T. Mirnalinee Department of Computer Science and Engineering S.S.N College of Engineering, Anna University, Chennai,

More information

A Method Based Genetic Algorithm for Pipe Routing Design

A Method Based Genetic Algorithm for Pipe Routing Design 5th International Conference on Advanced Engineering Materials and Technology (AEMT 2015) A Method Based Genetic Algorithm for Pipe Routing Design Changtao Wang 1, a, Xiaotong Sun 2,b,Tiancheng Yuan 3,c

More information

Experimental Study on Bound Handling Techniques for Multi-Objective Particle Swarm Optimization

Experimental Study on Bound Handling Techniques for Multi-Objective Particle Swarm Optimization Experimental Study on Bound Handling Techniques for Multi-Objective Particle Swarm Optimization adfa, p. 1, 2011. Springer-Verlag Berlin Heidelberg 2011 Devang Agarwal and Deepak Sharma Department of Mechanical

More information

Particle Filter Localization

Particle Filter Localization E190Q Autonomous Mobile Robots Lab 4 Particle Filter Localization INTRODUCTION Determining a robots position in a global coordinate frame is one of the most important and difficult problems to overcome

More information

EVOLVING ENGINEERING DESIGN TRADE-OFFS. Yizhen Zhang Collective Robotics Group

EVOLVING ENGINEERING DESIGN TRADE-OFFS. Yizhen Zhang Collective Robotics Group Proceedings of DETC 3 ASME 23 Design Engineering Technical Conferences and Computers and Information in Engineering Conference Chicago, Illinois USA, September 2-6, 23 DETC23/DTM-48676 EVOLVING ENGINEERING

More information

Introduction to Genetic Algorithms

Introduction to Genetic Algorithms Advanced Topics in Image Analysis and Machine Learning Introduction to Genetic Algorithms Week 3 Faculty of Information Science and Engineering Ritsumeikan University Today s class outline Genetic Algorithms

More information

Hwk 1: UNIX, C Programming, and Memory Management

Hwk 1: UNIX, C Programming, and Memory Management Hwk : UNIX, C Programming, and Memory Management This homework may require the following equipment: C compiler Note that the homework is graded. For any questions, please contact us at sista@groupes.epfl.ch..

More information

Genetic Algorithms: Setting Parmeters and Incorporating Constraints OUTLINE OF TOPICS: 1. Setting GA parameters. 2. Constraint Handling (two methods)

Genetic Algorithms: Setting Parmeters and Incorporating Constraints OUTLINE OF TOPICS: 1. Setting GA parameters. 2. Constraint Handling (two methods) Genetic Algorithms: Setting Parmeters and Incorporating Constraints OUTLINE OF TOPICS: 1. Setting GA parameters general guidelines for binary coded GA (some can be extended to real valued GA) estimating

More information

Section 1.5 Transformation of Functions

Section 1.5 Transformation of Functions Section.5 Transformation of Functions 6 Section.5 Transformation of Functions Often when given a problem, we try to model the scenario using mathematics in the form of words, tables, graphs and equations

More information

Genetic programming. Lecture Genetic Programming. LISP as a GP language. LISP structure. S-expressions

Genetic programming. Lecture Genetic Programming. LISP as a GP language. LISP structure. S-expressions Genetic programming Lecture Genetic Programming CIS 412 Artificial Intelligence Umass, Dartmouth One of the central problems in computer science is how to make computers solve problems without being explicitly

More information

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization Cell-to-switch assignment in cellular networks using barebones particle swarm optimization Sotirios K. Goudos a), Konstantinos B. Baltzis, Christos Bachtsevanidis, and John N. Sahalos RadioCommunications

More information

Dr. Mustafa Jarrar. Chapter 4 Informed Searching. Sina Institute, University of Birzeit

Dr. Mustafa Jarrar. Chapter 4 Informed Searching. Sina Institute, University of Birzeit Lecture Notes, Advanced Artificial Intelligence (SCOM7341) Sina Institute, University of Birzeit 2 nd Semester, 2012 Advanced Artificial Intelligence (SCOM7341) Chapter 4 Informed Searching Dr. Mustafa

More information

A Structural Optimization Method of Genetic Network Programming. for Enhancing Generalization Ability

A Structural Optimization Method of Genetic Network Programming. for Enhancing Generalization Ability International Journal of Engineering Innovation and Management Vol.6, No.2, 2016 A Structural Optimization Method of Genetic Network Programming for Enhancing Generalization Ability Shingo Mabu, Yamaguchi

More information

Offspring Generation Method using Delaunay Triangulation for Real-Coded Genetic Algorithms

Offspring Generation Method using Delaunay Triangulation for Real-Coded Genetic Algorithms Offspring Generation Method using Delaunay Triangulation for Real-Coded Genetic Algorithms Hisashi Shimosaka 1, Tomoyuki Hiroyasu 2, and Mitsunori Miki 2 1 Graduate School of Engineering, Doshisha University,

More information

Approach Using Genetic Algorithm for Intrusion Detection System

Approach Using Genetic Algorithm for Intrusion Detection System Approach Using Genetic Algorithm for Intrusion Detection System 544 Abhijeet Karve Government College of Engineering, Aurangabad, Dr. Babasaheb Ambedkar Marathwada University, Aurangabad, Maharashtra-

More information

IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE

IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE Fang Wang, and Yuhui Qiu Intelligent Software and Software Engineering Laboratory, Southwest-China Normal University,

More information

A NEW APPROACH TO SOLVE ECONOMIC LOAD DISPATCH USING PARTICLE SWARM OPTIMIZATION

A NEW APPROACH TO SOLVE ECONOMIC LOAD DISPATCH USING PARTICLE SWARM OPTIMIZATION A NEW APPROACH TO SOLVE ECONOMIC LOAD DISPATCH USING PARTICLE SWARM OPTIMIZATION Manjeet Singh 1, Divesh Thareja 2 1 Department of Electrical and Electronics Engineering, Assistant Professor, HCTM Technical

More information

CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS

CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS CHAPTER 6 REAL-VALUED GENETIC ALGORITHMS 6.1 Introduction Gradient-based algorithms have some weaknesses relative to engineering optimization. Specifically, it is difficult to use gradient-based algorithms

More information

Predicting Diabetes using Neural Networks and Randomized Optimization

Predicting Diabetes using Neural Networks and Randomized Optimization Predicting Diabetes using Neural Networks and Randomized Optimization Kunal Sharma GTID: ksharma74 CS 4641 Machine Learning Abstract This paper analysis the following randomized optimization techniques

More information

Exploration vs. Exploitation in Differential Evolution

Exploration vs. Exploitation in Differential Evolution Exploration vs. Exploitation in Differential Evolution Ângela A. R. Sá 1, Adriano O. Andrade 1, Alcimar B. Soares 1 and Slawomir J. Nasuto 2 Abstract. Differential Evolution (DE) is a tool for efficient

More information

CT79 SOFT COMPUTING ALCCS-FEB 2014

CT79 SOFT COMPUTING ALCCS-FEB 2014 Q.1 a. Define Union, Intersection and complement operations of Fuzzy sets. For fuzzy sets A and B Figure Fuzzy sets A & B The union of two fuzzy sets A and B is a fuzzy set C, written as C=AUB or C=A OR

More information

Inertia Weight. v i = ωv i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) The new velocity update equation:

Inertia Weight. v i = ωv i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) The new velocity update equation: Convergence of PSO The velocity update equation: v i = v i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) for some values of φ 1 and φ 2 the velocity grows without bound can bound velocity to range [ V max,v

More information

Topological Machining Fixture Layout Synthesis Using Genetic Algorithms

Topological Machining Fixture Layout Synthesis Using Genetic Algorithms Topological Machining Fixture Layout Synthesis Using Genetic Algorithms Necmettin Kaya Uludag University, Mechanical Eng. Department, Bursa, Turkey Ferruh Öztürk Uludag University, Mechanical Eng. Department,

More information

International Journal of Digital Application & Contemporary research Website: (Volume 1, Issue 7, February 2013)

International Journal of Digital Application & Contemporary research Website:   (Volume 1, Issue 7, February 2013) Performance Analysis of GA and PSO over Economic Load Dispatch Problem Sakshi Rajpoot sakshirajpoot1988@gmail.com Dr. Sandeep Bhongade sandeepbhongade@rediffmail.com Abstract Economic Load dispatch problem

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION 1 CHAPTER 1 INTRODUCTION 1.1 OPTIMIZATION OF MACHINING PROCESS AND MACHINING ECONOMICS In a manufacturing industry, machining process is to shape the metal parts by removing unwanted material. During the

More information

GREEN-PSO: Conserving Function Evaluations in Particle Swarm Optimization

GREEN-PSO: Conserving Function Evaluations in Particle Swarm Optimization GREEN-PSO: Conserving Function Evaluations in Particle Swarm Optimization Stephen M. Majercik 1 1 Department of Computer Science, Bowdoin College, Brunswick, Maine, USA smajerci@bowdoin.edu Keywords: Abstract:

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Informed Search and Exploration Chapter 4 (4.3 4.6) Searching: So Far We ve discussed how to build goal-based and utility-based agents that search to solve problems We ve also presented

More information

Section 1.5 Transformation of Functions

Section 1.5 Transformation of Functions Section 1.5 Transformation of Functions 61 Section 1.5 Transformation of Functions Often when given a problem, we try to model the scenario using mathematics in the form of words, tables, graphs and equations

More information

Particle Swarm Optimization applied to Pattern Recognition

Particle Swarm Optimization applied to Pattern Recognition Particle Swarm Optimization applied to Pattern Recognition by Abel Mengistu Advisor: Dr. Raheel Ahmad CS Senior Research 2011 Manchester College May, 2011-1 - Table of Contents Introduction... - 3 - Objectives...

More information

Crew Scheduling Problem: A Column Generation Approach Improved by a Genetic Algorithm. Santos and Mateus (2007)

Crew Scheduling Problem: A Column Generation Approach Improved by a Genetic Algorithm. Santos and Mateus (2007) In the name of God Crew Scheduling Problem: A Column Generation Approach Improved by a Genetic Algorithm Spring 2009 Instructor: Dr. Masoud Yaghini Outlines Problem Definition Modeling As A Set Partitioning

More information

AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS

AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS AN IMPROVED ITERATIVE METHOD FOR SOLVING GENERAL SYSTEM OF EQUATIONS VIA GENETIC ALGORITHMS Seyed Abolfazl Shahzadehfazeli 1, Zainab Haji Abootorabi,3 1 Parallel Processing Laboratory, Yazd University,

More information

Particle swarm optimization of artificial neural networks for autonomous robots. Mahmood Rahmani

Particle swarm optimization of artificial neural networks for autonomous robots. Mahmood Rahmani Thesis for the degree of Master of Science in Complex Adaptive Systems Particle swarm optimization of artificial neural networks for autonomous robots Mahmood Rahmani Department of Applied Physics Chalmers

More information

Section 1.5 Transformation of Functions

Section 1.5 Transformation of Functions 6 Chapter 1 Section 1.5 Transformation of Functions Often when given a problem, we try to model the scenario using mathematics in the form of words, tables, graphs and equations in order to explain or

More information

Solving Optimization Problems with MATLAB Loren Shure

Solving Optimization Problems with MATLAB Loren Shure Solving Optimization Problems with MATLAB Loren Shure 6 The MathWorks, Inc. Topics Introduction Least-squares minimization Nonlinear optimization Mied-integer programming Global optimization Optimization

More information

Objectives. Part 1: forward kinematics. Physical Dimension

Objectives. Part 1: forward kinematics. Physical Dimension ME 446 Laboratory #1 Kinematic Transformations Report is due at the beginning of your lab time the week of February 20 th. One report per group. Lab sessions will be held the weeks of January 23 rd, January

More information

Automatic differentiation based for particle swarm optimization steepest descent direction

Automatic differentiation based for particle swarm optimization steepest descent direction International Journal of Advances in Intelligent Informatics ISSN: 2442-6571 Vol 1, No 2, July 2015, pp. 90-97 90 Automatic differentiation based for particle swarm optimization steepest descent direction

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS

A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS A GENETIC ALGORITHM FOR CLUSTERING ON VERY LARGE DATA SETS Jim Gasvoda and Qin Ding Department of Computer Science, Pennsylvania State University at Harrisburg, Middletown, PA 17057, USA {jmg289, qding}@psu.edu

More information

Homework /681: Artificial Intelligence (Fall 2017) Out: October 18, 2017 Due: October 29, 2017 at 11:59PM

Homework /681: Artificial Intelligence (Fall 2017) Out: October 18, 2017 Due: October 29, 2017 at 11:59PM Homework 2 15-381/681: Artificial Intelligence (Fall 2017) Out: October 18, 2017 Due: October 29, 2017 at 11:59PM Homework Policies Homework is due on Autolab by the posted deadline. Assignments submitted

More information

The movement of the dimmer firefly i towards the brighter firefly j in terms of the dimmer one s updated location is determined by the following equat

The movement of the dimmer firefly i towards the brighter firefly j in terms of the dimmer one s updated location is determined by the following equat An Improved Firefly Algorithm for Optimization Problems Amarita Ritthipakdee 1, Arit Thammano, Nol Premasathian 3, and Bunyarit Uyyanonvara 4 Abstract Optimization problem is one of the most difficult

More information

Assignment 3 ITCS-6010/8010: Cloud Computing for Data Analysis

Assignment 3 ITCS-6010/8010: Cloud Computing for Data Analysis Assignment 3 ITCS-6010/8010: Cloud Computing for Data Analysis Due by 11:59:59pm on Tuesday, March 16, 2010 This assignment is based on a similar assignment developed at the University of Washington. Running

More information

Unified Engineering Fall 2004

Unified Engineering Fall 2004 Massachusetts Institute of Technology Department of Aeronautics and Astronautics Cambridge, MA 02139 Unified Engineering Fall 2004 Problem Set #3 Due Date: Tuesday, Sept. 28, 2004 at 5pm M4 M5 M6 C4,C5,C6,

More information

PARTICLE Swarm Optimization (PSO), an algorithm by

PARTICLE Swarm Optimization (PSO), an algorithm by , March 12-14, 2014, Hong Kong Cluster-based Particle Swarm Algorithm for Solving the Mastermind Problem Dan Partynski Abstract In this paper we present a metaheuristic algorithm that is inspired by Particle

More information

Proportional Relationships: Connections

Proportional Relationships: Connections Proportional Relationships: Connections Henri Picciotto A set of activities using multiple mathematical tools (and some real world applications) to help middle school students make connections between

More information

Genetic Algorithms. Kang Zheng Karl Schober

Genetic Algorithms. Kang Zheng Karl Schober Genetic Algorithms Kang Zheng Karl Schober Genetic algorithm What is Genetic algorithm? A genetic algorithm (or GA) is a search technique used in computing to find true or approximate solutions to optimization

More information

Escaping Local Optima: Genetic Algorithm

Escaping Local Optima: Genetic Algorithm Artificial Intelligence Escaping Local Optima: Genetic Algorithm Dae-Won Kim School of Computer Science & Engineering Chung-Ang University We re trying to escape local optima To achieve this, we have learned

More information