Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

Similar documents
Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

International Journal of Advanced Research in Computer Science and Software Engineering

Supervised Learning in Neural Networks (Part 2)

Neural Network Neurons

Ensembles of Neural Networks for Forecasting of Time Series of Spacecraft Telemetry

Neural Networks Laboratory EE 329 A

Climate Precipitation Prediction by Neural Network

Cursive Handwriting Recognition System Using Feature Extraction and Artificial Neural Network

COMBINING NEURAL NETWORKS FOR SKIN DETECTION

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

ISSN: (Online) Volume 3, Issue 9, September 2015 International Journal of Advance Research in Computer Science and Management Studies

Simulation of Back Propagation Neural Network for Iris Flower Classification

OMBP: Optic Modified BackPropagation training algorithm for fast convergence of Feedforward Neural Network

Computer-Aided Diagnosis for Lung Diseases based on Artificial Intelligence: A Review to Comparison of Two- Ways: BP Training and PSO Optimization

Neural Nets. CSCI 5582, Fall 2007

Liquefaction Analysis in 3D based on Neural Network Algorithm

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Face Recognition Using K-Means and RBFN

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

Fast Learning for Big Data Using Dynamic Function

Notes on Multilayer, Feedforward Neural Networks

Image Compression: An Artificial Neural Network Approach

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

Human Identification at a Distance Using Body Shape Information

DETERMINATION OF OVER-LEARNING AND OVER-FITTING PROBLEM IN BACK PROPAGATION NEURAL NETWORK

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Iris recognition using SVM and BP algorithms

Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification

Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Comparison of Leaf Recognition using Multi-layer Perceptron and Support Vector Machine

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

A neural network that classifies glass either as window or non-window depending on the glass chemistry.

A Comparative Study of Selected Classification Algorithms of Data Mining

Neural Network Approach for Automatic Landuse Classification of Satellite Images: One-Against-Rest and Multi-Class Classifiers

CP365 Artificial Intelligence

Perceptrons and Backpropagation. Fabio Zachert Cognitive Modelling WiSe 2014/15

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

A Novel Technique for Optimizing the Hidden Layer Architecture in Artificial Neural Networks N. M. Wagarachchi 1, A. S.

Artificial Neuron Modelling Based on Wave Shape

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Correlation Based Feature Selection with Irrelevant Feature Removal

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

Neural Network Classifier for Isolated Character Recognition

Data Mining Techniques Methods Algorithms and Tools

APPLICATION OF A MULTI- LAYER PERCEPTRON FOR MASS VALUATION OF REAL ESTATES

Performance Analysis of Data Mining Classification Techniques

Perceptron: This is convolution!

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6)

International Research Journal of Computer Science (IRJCS) ISSN: Issue 09, Volume 4 (September 2017)

Implementation of FPGA-Based General Purpose Artificial Neural Network

Research Article Optimization of Process Parameters in Injection Moulding of FR Lever Using GRA and DFA and Validated by Ann

Evolutionary Algorithms For Neural Networks Binary And Real Data Classification

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

Hand Written Character Recognition using VNP based Segmentation and Artificial Neural Network

A Matlab based Face Recognition GUI system Using Principal Component Analysis and Artificial Neural Network

IN recent years, neural networks have attracted considerable attention

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

A New Technique to Optimize User s Browsing Session using Data Mining

AMOL MUKUND LONDHE, DR.CHELPA LINGAM

Data Mining. Neural Networks

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Keras: Handwritten Digit Recognition using MNIST Dataset

A Survey on Edge Detection Techniques using Different Types of Digital Images

The Edge Domination in Prime Square Dominating Graphs

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Mobile Application with Optical Character Recognition Using Neural Network

Advanced Spam Detection Methodology by the Neural Network Classifier

Inversion of Fracture Parameters by Using the Artificial Neural Network

Neural network based Numerical digits Recognization using NNT in Matlab

HANDWRITTEN GURMUKHI CHARACTER RECOGNITION USING WAVELET TRANSFORMS

Neural Networks. Neural Network. Neural Network. Neural Network 2/21/2008. Andrew Kusiak. Intelligent Systems Laboratory Seamans Center

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK IN GAIT RECOGNITION

A Study on the Neural Network Model for Finger Print Recognition

A Deeper Insight of Neural Network Spatial Interaction Model s Performance Trained with Different Training Algorithms

11/14/2010 Intelligent Systems and Soft Computing 1

Computational Intelligence Meets the NetFlix Prize

CS 4510/9010 Applied Machine Learning

Analysis of Modified Rule Extraction Algorithm and Internal Representation of Neural Network

Multilayer Feed-forward networks

Data Mining on Agriculture Data using Neural Networks

Opening the Black Box Data Driven Visualizaion of Neural N

Week 3: Perceptron and Multi-layer Perceptron

Code Mania Artificial Intelligence: a. Module - 1: Introduction to Artificial intelligence and Python:

THE NEURAL NETWORKS: APPLICATION AND OPTIMIZATION APPLICATION OF LEVENBERG-MARQUARDT ALGORITHM FOR TIFINAGH CHARACTER RECOGNITION

REDUCTION OF NOISE TO IMPROVE QUALITY OF VOICE SIGNAL DATA USING NEURAL NETWORKS

Recognition of Handwritten Digits using Machine Learning Techniques

Gender Classification Technique Based on Facial Features using Neural Network

Keywords: Extraction, Training, Classification 1. INTRODUCTION 2. EXISTING SYSTEMS

PERFORMANCE ANALYSIS AND VALIDATION OF CLUSTERING ALGORITHMS USING SOFT COMPUTING TECHNIQUES

CMPT 882 Week 3 Summary

Proceedings of the 2016 International Conference on Industrial Engineering and Operations Management Detroit, Michigan, USA, September 23-25, 2016

COMPUTATIONAL INTELLIGENCE

Backpropagation in Neural Nets, and an Introduction to Vision. CSCI 5582, Fall 2007

Identification of Multisensor Conversion Characteristic Using Neural Networks

NOVATEUR PUBLICATIONS INTERNATIONAL JOURNAL OF INNOVATIONS IN ENGINEERING RESEARCH AND TECHNOLOGY [IJIERT] ISSN: VOLUME 5, ISSUE

FUFM-High Utility Itemsets in Transactional Database

Imputation for missing data through artificial intelligence 1

Transcription:

Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 5, May 2015, pg.358 364 RESEARCH ARTICLE ISSN 2320 088X Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach Foram Panchal 1, Mahesh Panchal 2 1 Department of Computer Science & Engineering, Kalol Institute of Technology, Gujarat, India 2 Deputy Director at GTU, Gujarat, India 1 forampanchal87@gmail.com, 2 mkhpanchal@gmail.com Abstract Artificial Neural Network is appropriate and most effective for pattern recognition, signal processing, and classification problems. For getting the proper result, ANN requires accurate data preprocessing, architecture selection, network training. The major problem in ANN is to select the number of hidden nodes. The random selection of number of hidden nodes leads towards the problem of Overfitting or Underfitting. The Competitive Learning Approach works on real world data to train the ANN and finds the optimized number of hidden layer. The proposed method calculates the number of hidden layer on the basis of input data also numbers of hidden nodes are dynamic as they are generated while training the ANN. Keyword Artificial Neural Network (ANN), Competitive learning approach, Hidden nodes I. INTRODUCTION Artificial Neural Network is appropriate and most effective for pattern recognition, signal processing, and classification problems. For proper result ANN requires correct preprocessing of data, selection of architecture and training for the network. But the major problem is to select the number of hidden nodes, because the random selection of number of hidden nodes may cause the problem of Overfitting and Underfitting.Sometimes the network matches the data narrowly and loses its generalization ability over the test data which give rise to overtraining. This paper gives information regarding the competitive learning approach of finding hidden nodes for artificial neural network. The advantage of proposed method is that it is not approximately calculating number of hidden nodes but based on similarity between input data. In proposed method number of hidden nodes are not predefined but generated during the 2015, IJCSMC All Rights Reserved 358

training time. The organization of this paper is, in section II various methods for finding number of hidden nodes are explain. Section III contains proposed method to finding number of hidden nodes. Section IV is about experiments and its result. Last section contains the conclusion about the study that is done so far and so forth future work. II. VARIOUS METHODS TO FIND NUMBER OF HIDDEN NODES A. Structured trial and error method [1] The characteristics of this method is to do repeated attempts until it successes or until the last attempt of agent. This method is dividing in to two approaches. 1. Forward Approach This method works with bottom to top approach. Initially we take lower number of hidden neurons to train and test the NN. Then we train and test NN by increasing number of hidden neurons step by step. This process is repeated until it gives improved result for training and testing the NN. 2. Backward Approach This method works with top to bottom approach. Initially we take big number of hidden neurons to train and test the NN. Then we train and test NN by decreasing number of hidden neurons step by step. This process is repeated until it gives improved result for training and testing the NN. B. Two phase method [1][5] The characteristic of this method is to do repeated attempts until it successes or until the last attempt of agent, which is similar to above method but the approach is different. Total 4 groups are created from the dataset. In the first phase two of the groups are used to train the NN. During the second phase the third group is used to test the NN.Output values from the trained NN are determined by using the remaining fourth group. Leastpossible error term regarding number of neurons in the hidden layer is obtained by repeating the process. C. Rule of Thumb Methods [2] The number of neurons to be used in hidden layer is determined by this method as described below: The number of hidden neurons should not be greater than the size of the input layer and as well as be less than the output layer. 2015, IJCSMC All Rights Reserved 359

Suppose the size of the input layer is 8 and the size of the output layer is 4 the hidden neurons are in range between 4 and 8. The number of hidden neurons should be summation of 2/3 of the input layer size and the size of the output layer. If we have input neurons 9 and output neurons 2 then we have 8 hidden neurons. The number of hidden neurons should not be more than double the input layer size. If we have input neurons 6 then we have hidden neurons less than 12. D. Method based on some random criteria [3] Data collection : Data Normalization : Designing the network : Real time data was collected from Suzlon Energy India Ltd., India wind farm for a period from April 2011 to Dec. 2012. Normalization of data is essential as the variables of different units. Data are scaled within 0 to 1. Selection of proposed criteria : Selection of proposed criteria : Scaling will result in improved accuracy. Set-up parameter includes epochs and dimensions. The dimensions like number of input, hidden, and output neuron are to be designed. The three input parameters are temperature, wind direction and wind speed. For the proposed model, 101 various criteria were examined to estimate training process and errors For the proposed model, 101 various criteria were examined to estimate training process and errors. 2015, IJCSMC All Rights Reserved 360

Table 1 Criteria for fixing number of hidden nodes Divide the data into training and testing : 7000 data was used for training, and 3000 data was used for testing data. III. METHODOLOGY In this section, methods used for proposed work is discussed in detail. Expected outcomes after applying proposed methods are included along with implementation strategies. A. Competitive approach Competitive approach results in win-lose consequences as it works similar to the process of distributive negotiation. B. Flow of Work 1. Initialize number of input nodes. 2. Initialize similarity threshold. 3. Now read a sample from training data. 4. If it is first sample then add one hidden node and output node. 5. X = sqrt (sum (square(weights between input-hidden nodes))) Y = sqrt (sum (square (values of new sample))) 6. Now find the difference of X and Y. 7. If difference is greater than threshold value there is no need to create hidden node otherwise create hidden node. 8. Now assign weight to new hidden node from input-hidden nodes as input values. 9. Now check class of new sample if it is similar to network class then no need to create new output node otherwise create output node. 2015, IJCSMC All Rights Reserved 361

10. Now assign weight to new output node from hidden-output nodes as output values. 11. Repeat steps 3 to 10. 12. As a final result we get the number of hidden nodes necessary to train the NN. Figure 1 Flow Chart Figure 1 Flow chart of Competitive Approach 2015, IJCSMC All Rights Reserved 362

IV. EXPERIMENTS AND RESULT For the experimental purpose Sales Forecasting data set with 6 inputs and 2 outputs has been taken. Proposed method (Competitive Learning Approach) has performed on this dataset also Backpropagation with momentum and Conjugate Gradient Method perform on the dataset for comparison purpose. 7 6 5 4 3 2 1 0 1 2 3 4 5 6 7 8 Similarity Threshold 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 No.of Hidden nodes 6 6 3 1 1 1 1 1 MSE 2.4076 2.4076 0.25143 0.03051 0.03051 0.03051 0.03051 0.03051 Figure - 2 Sales forecasting using Competitive Learning Approach Figure - 3Sales forecasting using Backpropagation and Conjugate Gradient [4] From the figure 2 as we increased the similarity threshold value, number of hidden nodes decreased also the MSE is decreased according to the number of hidden nodes decreased and similarity threshold increased. 2015, IJCSMC All Rights Reserved 363

V. CONCLUSION From the above experiment we conclude that as we increased similarity threshold, no of hidden node are also decreased and we get better MSE (min square error). We also conclude that all the previous methods to find number of hidden nodes are based on approximation, also we have to decide the number of hidden nodes before training the neural network. But proposed method is not based on approximation rather it gives exact solution. Also there is no need to decide number of hidden node before training of ANN. In proposed method number of hidden nodes is generated automatically during the training of ANN, based on the value of similarity threshold. REFERENCES [1] Saurabh Karsoliya, Approximating Number of Hidden layer neurons in Multiple Hidden Layer BPNN Architecture, International Journal of Engineering Trends and Technology- Volume 3 Issue 6-2012 [2] Gaurang Panchal, Amit Ganatra, Y P Kosta and Devyani Panchal, Behaviour Analysis of Multilayer Perceptrons with Multiple Hidden Neurons and Hidden Layers, International Journal of Computer Theory and Engineering, Vol. 3, No. 2, April 2011, ISSN: 1793-8201 [3] K. Ghana Sheila and S. N. Deepak, Review on Methods to Fix Number of Hidden Neurons in Neural networks, Mathematical Problems in Engineering, vol. 2013, Article ID 425740, 11 pages, 2013. doi:10.1155/2013/425740 [4] Foram Panchal and Mahesh Panchal, Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network, International Journal of Computer Science and Mobile Computing, Vol.3, Issue.11, Nov 2014, pages 455-464 [5] Blum, A., 1992, Neural Networks in C++,NY:Wiley 2015, IJCSMC All Rights Reserved 364