Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Similar documents
A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

GENETIC ALGORITHM VERSUS PARTICLE SWARM OPTIMIZATION IN N-QUEEN PROBLEM

Simulation of Back Propagation Neural Network for Iris Flower Classification

Traffic Signal Control Based On Fuzzy Artificial Neural Networks With Particle Swarm Optimization

Meta- Heuristic based Optimization Algorithms: A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Feature weighting using particle swarm optimization for learning vector quantization classifier

PARTICLE SWARM OPTIMIZATION (PSO)

International Journal of Digital Application & Contemporary research Website: (Volume 1, Issue 7, February 2013)

CHAPTER 6 HYBRID AI BASED IMAGE CLASSIFICATION TECHNIQUES

Handling Multi Objectives of with Multi Objective Dynamic Particle Swarm Optimization

THREE PHASE FAULT DIAGNOSIS BASED ON RBF NEURAL NETWORK OPTIMIZED BY PSO ALGORITHM

Mobile Robot Path Planning in Static Environments using Particle Swarm Optimization

Modified Particle Swarm Optimization

Reconfiguration Optimization for Loss Reduction in Distribution Networks using Hybrid PSO algorithm and Fuzzy logic

A NEW APPROACH TO SOLVE ECONOMIC LOAD DISPATCH USING PARTICLE SWARM OPTIMIZATION

Hybrid Particle Swarm and Neural Network Approach for Streamflow Forecasting

LECTURE 16: SWARM INTELLIGENCE 2 / PARTICLE SWARM OPTIMIZATION 2

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

Comparison of Some Evolutionary Algorithms for Approximate Solutions of Optimal Control Problems

SIMULTANEOUS COMPUTATION OF MODEL ORDER AND PARAMETER ESTIMATION FOR ARX MODEL BASED ON MULTI- SWARM PARTICLE SWARM OPTIMIZATION

Particle Swarm Optimization

Neural Networks Laboratory EE 329 A

Particle Swarm Optimization applied to Pattern Recognition

Simplifying Handwritten Characters Recognition Using a Particle Swarm Optimization Approach

Index Terms PSO, parallel computing, clustering, multiprocessor.

Hybrid Particle Swarm-Based-Simulated Annealing Optimization Techniques

Particle Swarm Optimization Artificial Bee Colony Chain (PSOABCC): A Hybrid Meteahuristic Algorithm

Evolutionary Algorithms For Neural Networks Binary And Real Data Classification

Inertia Weight. v i = ωv i +φ 1 R(0,1)(p i x i )+φ 2 R(0,1)(p g x i ) The new velocity update equation:

Research Article Path Planning Using a Hybrid Evolutionary Algorithm Based on Tree Structure Encoding

PARALLEL PARTICLE SWARM OPTIMIZATION IN DATA CLUSTERING

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Kyrre Glette INF3490 Evolvable Hardware Cartesian Genetic Programming

An improved PID neural network controller for long time delay systems using particle swarm optimization algorithm

Particle swarm optimization for mobile network design

Generation of Ultra Side lobe levels in Circular Array Antennas using Evolutionary Algorithms

Robust Descriptive Statistics Based PSO Algorithm for Image Segmentation

1 Lab 5: Particle Swarm Optimization

Particle Swarm Optimization Approach for Scheduling of Flexible Job Shops

A PSO-based Generic Classifier Design and Weka Implementation Study

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

Image Compression: An Artificial Neural Network Approach

A RANDOM SYNCHRONOUS-ASYNCHRONOUS PARTICLE SWARM OPTIMIZATION ALGORITHM WITH A NEW ITERATION STRATEGY

Weight Optimization for a Neural Network using Particle Swarm Optimization (PSO)

International Journal of Advance Research in Computer Science and Management Studies

Step Size Optimization of LMS Algorithm Using Particle Swarm Optimization Algorithm in System Identification

QUANTUM BASED PSO TECHNIQUE FOR IMAGE SEGMENTATION

A Comparative Study of Genetic Algorithm and Particle Swarm Optimization

Comparative Analysis of Swarm Intelligence Techniques for Data Classification

Particle Swarm Optimization

PARTICLE SWARM OPTIMIZATION APPLICATION IN OPTIMIZATION

Particle Swarm Optimization Applied to Job Shop Scheduling

PSO-BP ALGORITHM IMPLEMENTATION FOR MATERIAL SURFACE IMAGE IDENTIFICATION

Computer-Aided Diagnosis for Lung Diseases based on Artificial Intelligence: A Review to Comparison of Two- Ways: BP Training and PSO Optimization

A Naïve Soft Computing based Approach for Gene Expression Data Analysis

Three-Dimensional Off-Line Path Planning for Unmanned Aerial Vehicle Using Modified Particle Swarm Optimization

Improving Tree-Based Classification Rules Using a Particle Swarm Optimization

Overlapping Swarm Intelligence for Training Artificial Neural Networks

ARTIFICIAL INTELLIGENCE (CSCU9YE ) LECTURE 5: EVOLUTIONARY ALGORITHMS

An Improved Fuzzy Algorithm for Image Segmentation

KEYWORDS: Mobile Ad hoc Networks (MANETs), Swarm Intelligence, Particle Swarm Optimization (PSO), Multi Point Relay (MPR), Throughput.

Identification Of Iris Flower Species Using Machine Learning

ARMA MODEL SELECTION USING PARTICLE SWARM OPTIMIZATION AND AIC CRITERIA. Mark S. Voss a b. and Xin Feng.

Dr. Ramesh Kumar, Nayan Kumar (Department of Electrical Engineering,NIT Patna, India, (Department of Electrical Engineering,NIT Uttarakhand, India,

Designing of Optimized Combinational Circuits Using Particle Swarm Optimization Algorithm

Comparing Classification Performances between Neural Networks and Particle Swarm Optimization for Traffic Sign Recognition

A MULTI-SWARM PARTICLE SWARM OPTIMIZATION WITH LOCAL SEARCH ON MULTI-ROBOT SEARCH SYSTEM

A Classifier with the Function-based Decision Tree

Convolutional Code Optimization for Various Constraint Lengths using PSO

A Modified PSO Technique for the Coordination Problem in Presence of DG

Effect of the PSO Topologies on the Performance of the PSO-ELM

Cell-to-switch assignment in. cellular networks. barebones particle swarm optimization

A hybrid particle swarm optimization back-propagation algorithm for feedforward neural network training

IN recent years, neural networks have attracted considerable attention

Hybrid PSO-SA algorithm for training a Neural Network for Classification

A Comparative Analysis on the Performance of Particle Swarm Optimization and Artificial Immune Systems for Mathematical Test Functions.

A Novel Particle Swarm Optimization-based Algorithm for the Optimal Centralized Wireless Access Network

Optimization of Benchmark Functions Using Artificial Bee Colony (ABC) Algorithm

Artificial Neural Network Training by using Regrouping Particle Swarm Optimization

1 Lab + Hwk 5: Particle Swarm Optimization

Performance Analysis of Data Mining Classification Techniques

GA is the most popular population based heuristic algorithm since it was developed by Holland in 1975 [1]. This algorithm runs faster and requires les

Neuro-fuzzy, GA-Fuzzy, Neural-Fuzzy-GA: A Data Mining Technique for Optimization

Particle Swarm Optimization For N-Queens Problem

Parameter Selection of a Support Vector Machine, Based on a Chaotic Particle Swarm Optimization Algorithm

A Data Classification Algorithm of Internet of Things Based on Neural Network

Particle Swarm Optimization Based Approach for Location Area Planning in Cellular Networks

Witold Pedrycz. University of Alberta Edmonton, Alberta, Canada

A Comparative Study of the Application of Swarm Intelligence in Kruppa-Based Camera Auto- Calibration

Modified Particle Swarm Optimization with Novel Modulated Inertia for Velocity Update

IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY A REVIEW: OPTICAL CHARACTER RECOGNITION Swati Tomar *1 & Amit Kishore 2 *1

Artificial Bee Colony (ABC) Optimization Algorithm for Solving Constrained Optimization Problems

Using CODEQ to Train Feed-forward Neural Networks

KTH ROYAL INSTITUTE OF TECHNOLOGY. Lecture 14 Machine Learning. K-means, knn

Machine Learning: Algorithms and Applications Mockup Examination

Optimization design of a RC slab under impact load using DPSO-based multiple optimal solutions

Optimal Power Flow Using Particle Swarm Optimization

Constrained Classification of Large Imbalanced Data

A Study on Optimization Algorithms for Clustering Gene Expression Data

1 Lab + Hwk 5: Particle Swarm Optimization

Transcription:

Volume 3, Issue 3, March 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Training Artificial Neural Network using Particle Swarm Optimization Algorithm Diptam Dutta Heritage Institute of Technology Argha Roy* Netaji Subhash Engg. College Kaustav Choudhury Heritage Institute of Technology Abstract - In this paper, the adaptation of network weights using Particle Swarm Optimization (PSO) was proposed as a mechanism to improve the performance of Artificial Neural Network (ANN) in classification of IRIS dataset. Classification is a machine learning technique used to predict group membership for data instances. To simplify the problem of classification neural networks are being introduced. This paper focuses on IRIS plant classification using Neural Network. The problem concerns the identification of IRIS plant species on the basis of plant attribute measurements. Classification of IRIS data set would be discovering patterns from examining petal and sepal size of the IRIS plant and how the prediction was made from analyzing the pattern to form the class of IRIS plant. By using this pattern and classification, in future upcoming years the unknown data can be predicted more precisely. Artificial neural networks have been successfully applied to problems in pattern classification, function approximations, optimization, and associative memories. In this work, Multilayer feed- forward networks are trained using back propagation learning algorithm. Keywords - Artificial neural network, particle swarm optimization, machine learning, back-propagation, IRIS. I. INTRODUCTION We view particle swarm optimization as a mid-level form of A-life or biologically derived algorithm, occupying the space in nature between evolutionary searches, which requires neural processing, which occurs on the order of milliseconds. Social optimization occurs in the time frame of ordinary experience - in fact, it is ordinary experience. In addition to its ties with A-life, particle swarm optimization has obvious ties with evolutionary computation. Conceptually, it seems to lie somewhere between genetic algorithms and evolutionary programming. Here we describe the use of back propagation neural networks (BPNN) towards the identification of iris plants on the basis of the following measurements: sepal length, sepal width, petal length, and petal width. There is a comparison of the fitness of neural networks with input data normalized by column, row, sigmoid, and column constrained sigmoid normalization. Also contained within the paper is an analysis of the performance results of back propagation neural networks with various numbers of hidden layer neurons, and differing number of cycles (epochs). The analysis of the performance of the neural networks is based on several criteria: incorrectly identified plants by training set (recall) and testing set (accuracy), specific error within incorrectly identified plants, overall data set error as tested, and class identification precision. II. LITERATURE REVIEW The most widely used method of training for feed-forward ANNs is back-propagation (BP) algorithm [10]. Feed-forward ANNs are commonly used for function approximation and pattern classifications. Back-propagation algorithm and its variations such as QuickProp [11] and RProp [12] are likely to reach local minima especially in case that the error surface is rugged. In addition, the efficiency of BP methods depends on the selection of appropriate learning parameters. The other training methods for feed-forward ANNs include those that are based on evolutionary computation and heuristic principles such as Genetic Algorithm (GA), and PSO. A. Artificial Intelligence : A precise definition of intelligence is unavailable. It is probably explained best by discussing some of the aspects. In general, intelligence has something to do with the process of knowledge and thinking, also called cognition. These mental processes are needed for, i.e., solving a mathematical problem or playing a game of chess. One needs to possess a certain intelligence to be able to do these tasks. Not only the deliberate thought processes are part of cognition, also the unconscious processes like perceiving and recognizing an object belong to it. B. Particle swarm optimization (PSO): Particle swarm optimization (PSO) [1] [2] is a stochastically global optimization method that belongs to the family of Swarm Intelligence and Artificial Life. Similar to artificial neural network (ANN) and Genetic Algorithms (GA) [7][8] 2013, IJARCSSE All Rights Reserved Page 430

which is the simplified models of the neural system & the natural selections of the evolutionary theory, PSO is based on the principles that flock of birds, school of fish, or swarm of bee s searches for food sources where at the beginning the perfect location is not known. However, they eventually they reach the best location of food source by means of communicating with each other. C. Artificial Neural Network (ANN): An Artificial Neural Network, often just called a neural network, is a mathematical model inspired by biological neural networks. A neural network consists of an interconnected group of artificial neurons, and it processes information using a connectionist approach to computation. In most cases a neural network is an adaptive system that changes its structure during a learning phase. Neural networks are used to model complex relationships between inputs and outputs or to find patterns in data. III. PSO-BACK PROPAGATION (BP) ALGORITHM The PSO BP is an optimization algorithm combining the PSO with the BP. Similar to the GA, the PSO algorithm is a global algorithm, which has a strong ability to find global optimistic result, this PSO algorithm, The BP algorithm, on the contrary, has a strong ability to find local optimistic result, but its ability to find the global optimistic result is weak. By combining the PSO with the BP, The fundamental idea for this hybrid algorithm is that at the beginning stage of searching for the optimum, the PSO is employed to accelerate the training speed. When the fitness function value has not changed for some generations, or value changed is smaller than a predefined number, the searching process is switched to gradient descending searching according to this heuristic knowledge. Similar to the APSO algorithm [7], the PSO BP algorithm s searching process is also started from initializing a group of random particles. First, all the particles are updated according to the Equations. Until a new generation set of particles are generated, and then those new particles are used to search the global best position in the solution space. Finally the BP algorithm is used to search around the global optimum. In this way, this hybrid algorithm may find an optimum more quickly. A. Pseudo Code for the Algorithm: Initialize particle END DO Calculate fitness value If the fitness value is better than the best fitness value (pbest) in history Set current value as the new pbest End Choose the particle with the best fitness value of all the particles as gbest Calculate particle velocity according equation (a) Update particle position according equation (b) End While maximum iterations or minimum error criteria is not attained B. Flow Chart: 2013, IJARCSSE All Rights Reserved Page 431

IV. PROPOSED WORK The proposed optimization algorithm combines the PSO with the back-propagation (BP). Similar to the GA, the PSO algorithm is a global algorithm, which has a strong ability to find global optimistic result, this PSO algorithm, The BP algorithm, on the contrary, has a strong ability to find local optimistic result, but its ability to find the global optimistic result is weak. By combining the PSO with the BP, The fundamental idea for this hybrid algorithm is that at the beginning stage of searching for the optimum, the PSO is employed to accelerate the training speed. When the fitness function value has not changed for some generations, or value changed is smaller than a predefined number, the searching process is switched to gradient descending searching according to this heuristic knowledge. The algorithm s searching process is also started from initializing a group of random particles. First, all the particles are updated according to the Equations. Until a new generation set of particles are generated, and then those new particles are used to search the global best position in the solution space. Finally the BP algorithm is used to search around the global optimum. In this way, this hybrid algorithm may find an optimum more quickly. Fig 1: Proposed two layered feed forward neural network structure V. RESULTS AND DISCUSSIONS Different ranges of values are taken for x and y. for some specific ranges of x and y, we are analyzing different runs over each iterations. And by using MATLAB, we can easily find the difference between the particles. It is the language for technical computing. MATLAB is the easiest language for solving mathematical equations or these type of functions as compared to C programming, by which we can easily implement different functions. MATLAB is very time consuming. The fittest network architecture identified used column normalization, 54 cycles, 1 hidden layer with 6 hidden layer neurons, a step width of 0.15, a maximum non-propagated error of 0.1, and a value of 1 for the number of update steps. We analyze the data using specific value given in IRIS dataset (sample provided 5 dataset shown) X1 5.1 4.9 4.7 4.6 5 X2 3.5 3 3.2 3.1 3.6 X3 1.4 1.4 1.3 1.5 1.4 X4 0.2 0.2 0.2 0.2 0.2 Table 1: Sample IRIS Dataset To get the output in a binary pattern we need to normalize the output value. H1 H2 F i =100 H3 Fig 2: Process Normalization 2013, IJARCSSE All Rights Reserved Page 432

A. Output: Fi-Xai Fi= final weighted average Xai (threshold function) = 0.5 (defined) Now if, Output =1, (Fi-Xai) >= 0 =0, (Fi-Xai) < 0 Thus the final output result takes the shape of Setosa 010 Versicolor 100 Virginnica 001 Table2: Output Pattern B. Weight calculation: The Constant factor here taken as C1 =1, to calculate [6] [10] the weighted average value: H[i] = Hij *X[i] 0<=i<=150 0<=j<=5 F[i] = Wij * H[j] 0<=i<=150 0<=j<=5 C. Classification performance: As shown in this plot at the Epoch 46 the validation performance returns less Mean square Error. Mean square error is the average square between output & target. The projected result for 54 Epoch we get the test data matrix with the accuracy rate of classified pattern of 97.3% Fig 3: Plot of error per iteration VI. CONCLUSION Particle swarm optimization is an extremely simple algorithm that seems to be effective for optimizing a wide range of functions. The adjustment p i toward and p g by the particle swarm optimizer is conceptually similar to the crossover operation utilized by genetic algorithms. It uses the concept of fitness, as do all evolutionary computation paradigms. Unique to the concept of particle swarm optimization is flying potential solutions through hyperspace, accelerating toward "better" solutions. In this simulation, we demonstrated the efficiency that this method possesses. Lastly, this method can be employed in training in various ANNs with different topologies. 2013, IJARCSSE All Rights Reserved Page 433

REFERENCES [1] Kennedy, J.; Eberhart, R. (1995). "Particle Swarm Optimization". Proceedings of IEEE International Conference on Neural Networks. [2] Kennedy, J.; Eberhart, R.C. (2001). Swarm Intelligence. Morgan Kaufmann. [3] Poli, R. (2008). "Analysis of the publications on the applications of particle swarm optimisation". Journal of Artificial Evolution and Applications 2008. [4] Shi, Y.; Eberhart, R.C. (1998). "Parameter selection in particle swarm optimization". Proceedings of Evolutionary Programming. [5] Pedersen, M.E.H. (2010). "Good parameters for particle swarm optimization". [6] Trelea, I.C. (2003). "The Particle Swarm Optimization Algorithm: convergence analysis and parameter selection". Information Processing Letters. [7] Zhang, G.P., 2000. Neural networks for classification: a survey. IEEE Transactions on Systems Man and Cybernetics. [8] Rudolph, G., 1997. Local convergence rates of simple evolutionary algorithms with cauchy mutations. [9] Liu, H.B., Tang, Y.Y., Meng, J., Ji, Y., 2004. Neural networks learning using vbest model particle swarm optimization. [10] HechtNelso R., 1989, back propagation neural network [11] Scott E. Fahlman: An Empirical Study of Learning Speed in Back-Propagation Networks, September 1988 [12] Rprop Description and Implementation Details Martin Riedmiller, 1994. Technical report 2013, IJARCSSE All Rights Reserved Page 434