Simulation of Back Propagation Neural Network for Iris Flower Classification

Similar documents
Neural Networks Laboratory EE 329 A

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Image Compression: An Artificial Neural Network Approach

IN recent years, neural networks have attracted considerable attention

Performance Analysis of Data Mining Classification Techniques

Comparative Analysis of Swarm Intelligence Techniques for Data Classification

A Systematic Overview of Data Mining Algorithms. Sargur Srihari University at Buffalo The State University of New York

Supervised Learning in Neural Networks (Part 2)

Performance Measure of Hard c-means,fuzzy c-means and Alternative c-means Algorithms

Week 3: Perceptron and Multi-layer Perceptron

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

Techniques for Dealing with Missing Values in Feedforward Networks

Artificial Neural Network-Based Prediction of Human Posture

Optimum size of feed forward neural network for Iris data set.

Computer-Aided Diagnosis for Lung Diseases based on Artificial Intelligence: A Review to Comparison of Two- Ways: BP Training and PSO Optimization

TechTalk on Artificial Intelligence

Back propagation Algorithm:

Using CODEQ to Train Feed-forward Neural Networks

AMOL MUKUND LONDHE, DR.CHELPA LINGAM

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Introduction to Artificial Intelligence

Experimental Approach for the Evaluation of Neural Network Classifier Algorithms

International Journal of Advanced Research in Computer Science and Software Engineering

Machine Learning 13. week

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Neural Network Classifier for Isolated Character Recognition

A Novel Technique for Optimizing the Hidden Layer Architecture in Artificial Neural Networks N. M. Wagarachchi 1, A. S.

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

Comparision between Quad tree based K-Means and EM Algorithm for Fault Prediction

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

Lecture #11: The Perceptron

ANN exercise session

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Climate Precipitation Prediction by Neural Network

Identification Of Iris Flower Species Using Machine Learning

Weight Optimization for a Neural Network using Particle Swarm Optimization (PSO)

OMBP: Optic Modified BackPropagation training algorithm for fast convergence of Feedforward Neural Network

Neural Network Approach for Automatic Landuse Classification of Satellite Images: One-Against-Rest and Multi-Class Classifiers

Hybrid PSO-SA algorithm for training a Neural Network for Classification

Cursive Handwriting Recognition System Using Feature Extraction and Artificial Neural Network

2. Neural network basics

COMBINING NEURAL NETWORKS FOR SKIN DETECTION

THE discrete multi-valued neuron was presented by N.

Input: Concepts, Instances, Attributes

Ripple Down Rule learner (RIDOR) Classifier for IRIS Dataset

Center for Automation and Autonomous Complex Systems. Computer Science Department, Tulane University. New Orleans, LA June 5, 1991.

Hands on Datamining & Machine Learning with Weka

International Journal of Electrical and Computer Engineering 4: Application of Neural Network in User Authentication for Smart Home System

Liquefaction Analysis in 3D based on Neural Network Algorithm

Cluster Analysis and Visualization. Workshop on Statistics and Machine Learning 2004/2/6

Design and Performance Analysis of and Gate using Synaptic Inputs for Neural Network Application

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

CS6220: DATA MINING TECHNIQUES

Evolutionary Algorithms For Neural Networks Binary And Real Data Classification

International Journal of Advance Engineering and Research Development OPTICAL CHARACTER RECOGNITION USING ARTIFICIAL NEURAL NETWORK

KTH ROYAL INSTITUTE OF TECHNOLOGY. Lecture 14 Machine Learning. K-means, knn

Neuron Selectivity as a Biologically Plausible Alternative to Backpropagation

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Yuki Osada Andrew Cannon

Channel Performance Improvement through FF and RBF Neural Network based Equalization

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

A NEW ALGORITHM FOR OPTIMIZING THE SELF- ORGANIZING MAP

Cluster Analysis using Spherical SOM

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6)

Basic Concepts Weka Workbench and its terminology

Notes on Multilayer, Feedforward Neural Networks

Neural Network Neurons

Khmer Character Recognition using Artificial Neural Network

Volume 1, Issue 3 (2013) ISSN International Journal of Advance Research and Innovation

Department of Electronics and Telecommunication Engineering 1 PG Student, JSPM s Imperial College of Engineering and Research, Pune (M.H.

Logical Rhythm - Class 3. August 27, 2018

DATA MINING INTRODUCTION TO CLASSIFICATION USING LINEAR CLASSIFIERS

Classification with Diffuse or Incomplete Information

Proceedings of the 2016 International Conference on Industrial Engineering and Operations Management Detroit, Michigan, USA, September 23-25, 2016

Extreme Learning Machines. Tony Oakden ANU AI Masters Project (early Presentation) 4/8/2014

CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS

BL5229: Data Analysis with Matlab Lab: Learning: Clustering

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Function Approximation and Feature Selection Tool

Model Selection Introduction to Machine Learning. Matt Gormley Lecture 4 January 29, 2018

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

An Intelligent Technique for Image Compression

An Introduction to Cluster Analysis. Zhaoxia Yu Department of Statistics Vice Chair of Undergraduate Affairs

An Efficient Method for Extracting Fuzzy Classification Rules from High Dimensional Data

Data Mining. Neural Networks

CROSS-CORRELATION NEURAL NETWORK: A NEW NEURAL NETWORK CLASSIFIER

ARCHITECTURE OPTIMIZATION MODEL FOR THE MULTILAYER PERCEPTRON AND CLUSTERING

Statistical Methods in AI

Non-linear gating network for the large scale classification model CombNET-II

Radial Basis Function (RBF) Neural Networks Based on the Triple Modular Redundancy Technology (TMR)

CMPT 882 Week 3 Summary

Global Journal of Engineering Science and Research Management

Data mining with Support Vector Machine

Transcription:

American Journal of Engineering Research (AJER) e-issn: 2320-0847 p-issn : 2320-0936 Volume-6, Issue-1, pp-200-205 www.ajer.org Research Paper Open Access Simulation of Back Propagation Neural Network for Iris Flower Classification R.A.Abdulkadir 1,Khalipha A. Imam 2 M.B. Jibril 3 1,3 Electrical Engineering Department, Kano University of Science and TechnologyWudil 2 Electrical Engineering Department, Kano State Polytechnic ABSTRACT:One of the most dynamic research and application areas of neural networks is classification. In this paper, the use of matlab coding for simulation of backpropagation neural network for classification of Iris dataset is demonstrated. Fisher s Iris data base collected from uci repository is used. The data set contains 3 classes of 50 instances each, where each class refers to a type of iris plant. Sepal length, sepal width, petal length and petal width are the four features used to classify each flower to its category. The three classes of the flower are Iris Setosa, Iris Versicolor and Iris Verginica. The network is trained for different epochs with different number of neurons in the hidden layer. The performance of the network is evaluated by plotting the error versus the number of iterations, furthermore by testing the network with different samples of the iris flower data. The successfully trained network classified the testing data correctly; indicating 100% recognition. Keywords:Classification, back propagation, artificial neural network, iris flower I. INTRODUCTION Classification is one of the most frequently used decision making tasks of human activity. A classification problem arises when an object needs to be assigned into a predefined group or class based on a number of observed attributes related to that object. Many problems in business, science, industry, and medicine can be treated as classification problems. Examples include deformed banknote identification [1], bankruptcy prediction [2], credit scoring [3], quality management [4], speech recognition [5] and handwritten character recognition [6]. Artificial Neural Networks (ANN) are widely used for pattern recognition and classification owing to the fact that they can approximate complex systems that are difficult to model using conventional modelling techniques such as mathematical modelling. Although significant advancement has been made in classification related areas of neural networks, various issues in applying neural networks still remain and have not been totally addressed. In this research todemonstrate how some of these issues can be tackle, back propagation neural network is simulated for iris flower dataset classification, by writing a matlab code. II. PREVIOUS WORKS There are so many experts research on iris flower dataset. R.A. Fisher first introduced this dataset in his famous paper The use of multiple measurements in taxonomic problems [7]. Later, Patrick S. Hoey analysed the Iris dataset via two distinct statistical methods; He first plotted the dataset onto scatterplots to determine patterns in the data in relation to the Iris classifications. Furthermore, develop an application in Java that will run a series of methods on the dataset to extract relevant statistical information from the dataset. With these two methods, with these two methods he made a concrete prediction about the dataset [8]. In a seminar VitalyBorovinskiy demonstrated the solution of iris flower classification problem by 3 types of neural networks; multilayer perceptron, radial basis function network, and probabilistic neural network. The result indicates that multilayer perceptron performed better both during the training and testing [9]. The adaptation of network weights using Particle Swarm Optimization (PSO) as a mechanism to improve the performance of Artificial Neural Network (ANN) in the classification of IRIS dataset was proposed by Dutta D., Roy A., Reddy k [10]. An Artificial Neural Network (ANN) is an information processing tool that is inspired by the biological nervous system such as the brain. Artificial Neural Networks (ANN) are widely used to approximate complex systems that are difficult to model using conventional modelling techniques such as mathematical modelling. The most common applications are function approximation (feature extraction), and pattern recognition and w w w. a j e r. o r g Page 200

classification. There is no exact available formula to decide what architecture of ANN and which training algorithm will solve a given problem. The best solution is obtained by trial and error. One can get an idea by looking at a problem and decide to start with simple networks; going on to complex ones till the solution is within the acceptable limits of error. There are different neural network architectures. The basic architectures include multi-layered feed-forward networks (Figure 2.0) that are trained using back-propagation training algorithms. Figure 1: Multi-layered feed-forward neural network III. METHOD AND METHODOLOGY: 3.1 Problem description Fisher s Iris data base (Fisher, 1936) is perhaps the best known database to be found in the pattern recognition literature. The data set contains 3 classes of 50 instances each, where each class refers to a type of iris plant. One class is linearly separable from the other two; the latter are not linearly separable from each other. The data base contains the following attributes: 1). sepal length in cm 2). sepal width in cm 3). petal length in cm 4). petal width in cm 5). class: - Iris Setosa - Iris Versicolour - Iris Virginica Fisher s Iris data base is available in Matlab (load fisheriris) and in Internet (for example, on http://archive.ics.uci.edu/ml/datasets/iris). The goal of this work is to use an Artificial Neural Network to classify a specimen into its category using the above mentioned four features. This is done by using Multilayer Perceptron (MLP) feedforward network using backpropagation training algorithm. 3.2 Software and Data Pre-processing: One can develop ANN architectures and write training algorithms in any known higher level language (e.g C or Matlab). MATLAB has a proven tool box that helps one apply, already developed training algorithms, to solve the problem. MATLAB also has a GUI based system. This windows-based platform provides a step-bystep visual procedure to build, train and simulate the ANN. For this work a matlab code is written to simulate the network. Matlab R2016b is used. The features obtained from many specimens are actually considered as raw data. Some amount of preprocessing is always carried out on the input and output data in order to make it suitable for the network. These procedures help obtain faster and efficient training. Since there are three classes of flowers, a three-state numerical code must be assigned. The network must be trained to reproduce the code at the outputs. The output can be left in decimal form where the decimal numbers 1, 2, and 3 represent the classes or the classes can be coded in binary form. For this problem the outputs are coded as: 1. I. Setosa : Code [0.1 0.1 0.9] 2. I. Versicolor: Code [0.1 0.9 0.1] 3. I. Verginica: Code [0.9 0.1 0.1] w w w. a j e r. o r g Page 201

3.3 Data Setup: The data set file is saved in the matlab folder as irisdata.mat. The file is loaded using the command loadirisdata; in the code. This command uploads the file and includes all the variables setup in the current workspace. The data is divided as follows: Training inputs (4 120): 4 rows by 120 columns. The number 4 represents the features which will be presented, one-by-one, in parallel fashion. Each pattern of 4 features is called a vector and there are 120 such vectors for training. Target outputs (3 120): There are three outputs for each feature vector. There are 120 such vectors for training and will be the targets for the input vectors. Test inputs (4 30): Once the network has been trained, it has to be tested on data that it has not seen before. 30 vectors are used for testing the network. Test outputs (3 30): These are the expected classes for the test inputs. The simulated outputs of the network will be compared to these actual classes in order to determine the performance of the trained network. 3.4 Network Architecture Details: Following network parameters are considered: Input Layer: A single layer of none processing neurons is used. Since there are four features in the input data four neurons are used in the input layer. Output Layer: Single layer of processing neurons, since the output of the network is in 3 elements vector form, the output layer will have three neurons. Hidden Layer: One hidden layer is used. The number of neurons in hidden layers is experimented. With each combination of number of neurons in the hidden layer the network is trained and the error function (sum squared error) is plotted. The number of neurons that ensures the best generalization is chosen. Activation functions: At each processing neuron an activation function is required to determine how that neuron is activated. Sigmoid activation function is used in both the hidden and output layers. Synaptic weights and Bias: The synaptic weights between input and hidden layer neurons, and hidden to output layer neurons and the hidden layer bias are randomly initialized using rand command which is later adjusted and updated according to backpropagation training algorithm. Epochs: The number of iterations is also experimented. The error function is plotted versus the number of iterations. The number of iterations that ensures the best generalization is chosen. IV. RESULT AND DISCUSSION A backpropagation neural network is simulated to classify iris flower dataset; by writing a program code using matlab. 120 out of the total 150 instances in the data set with their corresponding target outputs were used for training the network. The simulation was carried out using Intel core i5-4200u CPU @ 2.29 GHz PC with 4GB RAM, windows 10, 64-bit OS, and MATLAB 2016b software. The network was trained for 1000 epochs first with three neurons in the hidden layer and then with four neurons in the hidden layer. The network was retrained for 10000 epochs with three neurons hidden layer and later with four neurons in the hidden layer. The performance of the developed network was assessed by plotting the error versus the number of iterations, furthermore by testing the network with a different 30 samples of the iris plant data (10 from every iris plant class). The error graph and the testing data classification result are presented below. Figure 2: Error square plot for 1000 Epochs with 3 neurons in the hidden layer w w w. a j e r. o r g Page 202

Table I: Test data classification for 1000 Epochs with 3 neurons in the hidden layer Test Iris Setosa 10 9 1 Iris Versicolor 10 0 10 Iris Virginnica 10 0 10 Figure 3: Error square plot for 1000 Epochs with 4 neurons in the hidden layer Table II: Test data classification for 1000 Epochs with 4 neurons in the hidden layer Test Iris Setosa 10 10 0 Iris Versicolor 10 0 10 Iris Virginnica 10 0 10 Figure 4: Error square plot for 10000 Epochs with 3 neurons in the hidden layer w w w. a j e r. o r g Page 203

Table III: Test data classification for 1000 Epochs with 4 neurons in the hidden layer Test Iris Setosa 10 9 1 Iris Versicolor 10 0 10 Iris Virginnica 10 0 10 Figure 5: Error square plot for 10000 Epochs with 4 neurons in the hidden layer Table IV: Test data classification for 10000 Epochs with 4 neurons in the hidden layer Test Iris Setosa 10 10 0 Iris Versicolor 10 10 0 Iris Virginnica 10 10 0 Final parameters of the successfully trained neural network are presented in Table 4.5. The successfully trained neural network classified the testing data correctly; hence, 100% recognition is achieved. Table V: Neural Network Final Training Parameters Input Nodes 4 Hidden Nodes 4 Output Nodes 3 Iterations 10000 w w w. a j e r. o r g Page 204

V. CONCLUSION A backpropagation neural network has been simulated for iris flower dataset classification. The simulation is by writing a matlab code that uses matlab commands. Fisher s Iris data base collected from uci repository is used which is available on http://archive.ics.uci.edu/ml/datasets/iris. The three classes of the flower were coded as; I. Setosa : Code [0.1 0.1 0.9], I. Verginica : Code [0.1 0.9 0.1] and I. Versicolor : Code [0.9 0.1 0.1]. The network was trained for 1000 epochs with different number of neurons in the hidden layer. The network was retrained for 10000 epochs. The performance of the developed network was evaluated by plotting the error versus the number of iterations, furthermore by testing the network with different samples of the iris flower data. The successfully trained neural network classified the testing data correctly; indicating 100% recognition. The research served as a means for understanding backpropagation neural network for pattern recognition and classification. It also gave opportunity to get more familiar with coding in matlab. REFERENCE [1] A. KHASHMAN, B. SEKEROGLU and K. DIMILILER, "Deformed Banknote Identification Using Pattern Averaging and Neural Networks," in 4th WSEAS Int. Conf. on COMPUTATIONAL INTELLIGENCE, MAN-MACHINE SYSTEMS and CYBERNETICS, Florida, 2005. [2] X. Brédart, "Bankruptcy Prediction Model Using Neural Networks," Accounting and Finance Research, vol. 3, no. 2, pp. 124-128, 2014. [3] M. A. Doori and B. Beyrouti, "Credit Scoring Model Based on Back Propagation Neural Network Using Various Activation and Error Function," IJCSNS International Journal of Computer Science and Network Security, vol. 14, no. 3, pp. 16-24, 2014. [4] J. Combaz, J.-C. Fernandez and M. Jaber, "Using Neural Networks for Quality Management," Verimag, Centre Equation, France, n.d. [5] P. Rani, S. Kakkar and S. Rani, "Speech Recognition using Neural Network," International Journal of Computer Applications, pp. 11-14, 2015. [6] A. Pal and D. Singh, "Handwritten English Character Recognition Using Neural Network," International Journal of Computer Science & Communication, vol. 1, no. 2, pp. 141-144, 2010. [7] F. R. A., "The use of multiple measurements in taxonomic problems," Annual Eugenics, vol. 7, no. II, pp. 179-188, 1936. [8] P. S. Hoey, "Statistical Analysis of the Iris Flower Dataset," University of Massachusetts At Lowell, Massachusetts, 2004. [9] V. Borovinskiy, "Classification of Iris data set," University of Ljubljana, Ljubljana, 2009. [10] D. Dutta, A. Roy and K. Choudhury, "Training Artificial Neural Network using Particle Swarm Optimization Algorithm," International Journal of Advanced Research in Computer Science and Software Engineering, vol. 3, no. 3, pp. 430-434, 2013. [11] N. M. A., Neural Network and Deep Learning, Australia: Determination press, 2015. [12] B. Krose and P. v. d. Smagt, An Introduction to Neural Network, Amsterdam: University of Amsterdam, 1996. [13] Wikipedia, "Kosaciecszczecinkowaty Iris setosa.jpg," 2 December 2014. [Online]. Available: https://en.wikipedia.org/wiki/file:kosaciec_szczecinkowaty_iris_setosa.jpg. [Accessed 6 December 2016]. [14] Wikipedia, "Blue Flag, Ottawa.jpg," 29 June 2005. [Online]. Available: https://en.wikipedia.org/wiki/file:blue_flag,_ottawa.jpg. [Accessed 8 December 2016]. [15] Wikipedia, "Iris virginica.jpg," 22 March 2010. [Online]. Available: https://en.wikipedia.org/wiki/file:iris_virginica.jpg. [Accessed 11 December 2016]. [16] D. Kriesel, "A Brief Introduction to Neural Networks," 2007. [Online]. Available: http://www.dkriesel.com. [Accessed 20 11 2016]. [17] D. Anderson and G. McNeill, Artificial Neural Networks Technology, New York: Kaman Sciences Corporation, 1992. [18] D. Graupe, Principles of Artificial Neural Networks, Singapore: World Scientific Publishing Co. Pte. Ltd., 2007. w w w. a j e r. o r g Page 205