Effect of Hidden Layer Neurons on the Classification of Optical Character Recognition Typed Arabic Numerals

Size: px
Start display at page:

Download "Effect of Hidden Layer Neurons on the Classification of Optical Character Recognition Typed Arabic Numerals"

Transcription

1 Journal of Computer Science (7): , 008 ISSN Science Publications Effect of Hidden Layer Neurons on the Classification of Optical Character Recognition Typed Arabic Numerals Nidal F. Shilbayeh and Mahmoud Z. Isandarani Faculty of Information Technology for Graduate Studies, Middle East University for Graduate Studies, P.O. Box, Post Code: 60 Faculty of Science and Information Technology, Al-Zaytoonah Private University of Jordan, P.O. Box 9597, Post Code: 9, Amman-Jordan Abstract: Problem statement: The effect of varying the number of nodes in the hidden layer and number of iterations are important factors in the recognition rate. In this paper, a novel and effective criterion based on Cross Pruning (CP) algorithm is proposed to optimize number of hidden neurons and number of iterations in Multi Layer Perceptron (MLP) neural based recognition system. Our technique uses rule-based and neural networ pattern recognition methods in an integrated system in order to perform learning and recognition of dynamically printed numerals Approach: The study investigates the effect of varying the size of the networ hidden layers (pruning) and number of iterations (epochs) on the classification and performance of the used MLP. The optimum number of hidden neurons and epochs is experimentally established through the use of our novel Cross Pruning (CP) algorithm and via designing special neural based software. The designed software implements sigmoid as its shaping function. Results: Experimental results are presented on the classification accuracy and recognition. Significant recognition rate improvement is achieved using 000 epochs and 5 hidden neurons in our MLP OCR numeral recognition system. Conclusions/Recommendations: Our approach has a significant improvement in learning and classification of any numeral, character MLP based recognition system. Key words: Neural networ, MLP, pruning, pattern recognition, classification, OCR INTRODUCTION A maor problem in applying neural networs is specifying the size of the networ. Even for moderately size networs the number of parameters may become large compared to the number of data [-]. Artificial neural networs have been successfully applied to problems in pattern classification, function approximation, optimization, pattern matching and associative memories [0]. Multilayer feed forward networs trained using the bac propagation-learning algorithm is limited to searching for a suitable set of weights. This initiates the problem of selecting appropriate topology (number of hidden layers) and weights to solve the learning problem in hand. Two small networs are unable to adequately learn the problem while excessively large networs tend to over fit the training data and consequently result in poor generalization and wea performance. Most practical learning problems are nown to be computationally complex and hard to optimize. Networ pruning offers an excellent approach for dynamically determining an appropriate networ topology. Pruning techniques involves training different networ sizes and compare the performance in terms of classification and error. The process comprises methodical and consistence elimination or addition of neurons which subsequently removes or adds weights to the hidden layers of the system. Most of available systems capture inputs as a sequence of coordinate points, taing into account character blending and merging and the problem of characters that have close similarity to each other. This is solved via pre-processing of the characters prior to recognition, hence, performing a shape recognition process. The post-processing recognition process can be achieved using zones that define directions of travel, where characters are recognized as connected zones using a looup table for matching and classification. The primary tas of OCR recognition is to tae an input and correctly assign it as one of the possible put classes. This process can be divided into two Corresponding Author:Nidal F. Shilbayeh, Faculty of Information Technology for Graduate Studies, Middle East University for Graduate Studies, P.O. Box, Post Code:

2 J. Computer Sci., (7): , 008 general stages: feature selection and classification. Feature selection is critical to the whole process since the classifier will not be able to recognize from poorly selected features []. Some requirements for character recognition system design suggest themselves such as: To create a system which can recognize nonperfect numerals To create a system which can use both static and dynamic information components To perform recognition with image data presented at field rate or as fast as possible To recognize as quicly as possible, even before the full image sampling is completed To use a recognition method which requires a small amount of computational time and memory To create an expandable system which can recognize additional types of characters High recognition rate In the event of an error, non-recognition is implemented instead of miss-recognition, especially in critical conditions Recognizes inputs quicly Require minimal training In this study networ performance is examined as a function of both networ size and number of iterations. This is carried using a feed forward multi layer perceptron neural model. This system captures and intelligently recognizes numerals. There are many proposed methods for on-line recognition, which use a wide variety of pattern recognition techniques. Neural networs have been proposed as good candidates for character recognition. Studies have also been carried for OCR recognition, comparing techniques such as dynamic programming, Hidden Marov Models and recurrent neural networs [-6]. MATERIALS AND METHODS System design: Figure shows the used OCR MLP. In designing our OCR system the following factors are considered: Initial weights: Since we used a gradient decent learning algorithm that is nown to dynamically treat all weights in similar manner and to avoid a situation of no learning, random function is included in our algorithm. Such a function is specifically designed to guarantee that the networ will not end in global minima which will adversely affect the system performance and will result in massive errors and miss classifications. Learning rate: The learning rate is bounded due to the following: Two small a learning rate will cause the system to tae a very long time to converge To large a learning rate will cause the system to oscillate indefinitely and diverge Based on the above two mentioned factors our learning rate was chosen to be equal 0.5. Transfer function: Even though some literature indicates that anti symmetric functions will cause the system to learn faster, this research used the standard sigmoid function which is nown to be a stable function due to its range from 0-. All the above is considered in our research when hidden layer neurons and number of epochs are closely examined. Mathematical modeling: Figure shows our designed and implemented MLP neural engine used in our OCR system and characterized by the following three main equations: ( 0) in The put of the input layer i Input = () i Neural networ Adust Weight Fig. Used OCR MLP system Preparation Stage Preprocessing ( ) w i 6 50 ( ) w Fig. : OCR neural classification system Target Compare + Exp?input ( ) Recognition Process Postprocessing 579

3 J. Computer Sci., (7): , 008 ( ) ( 0) ( ) f = i i The put of the hidden layer () ( ) ( ) ( ) f = The put of the put layer () Using Eq. - we can rewrite Eq. to become: ( ) ( ) ( ) ( ) ( ) = f = f f iniw i.w i () Learning in an MLP is nown to occur via modification of weights as indicated in the following equations: ( ) (5) w = η t arg p ( ) (6) w = η t arg 0 i i p Where η the learning rate and p is is the pattern on which the equation is applied. It is clear that for the put layer the hidden layer appears as an input layer and for the input layer the hidden layer appears as an put layer. Equation 5 and 6 are obtained using the sigmoid function as the shaping function, which taes the following form: f ( input ) = (7) + exp input ( ) Applying Eq. 7 to Eq., and we obtain the shaped puts for the three layers used in our MLP. ( ) ( 0) layer 0 : i = ini (8) Applying Eq. 9, 0 and will produce all layers puts Using Eq. 5 and 6, the necessary weight modification values are obtained Adding the results of 5 and 6 to the original weights according to the following equations will update the weight matrix w t + = w t + w () i i i w t + = w t + w () Each time Eq. and are applied, a single cycle is completed (epoch). However there is a need to compute the sum squared error for each two layers and for the overall networ using the following equation E( w ) ( ) i = t arg () p RESULTS Table - show the effect of varying both number of hidden neurons and number of epochs on the recognition of Arabic numerals, while Table and 5 clearly shows classification accuracy and sum squared Table Recognition results using 00 epochs Reference Hidden Neurons ( ) ( ) layer : = ( 0) ( ) + exp i i i ( ) layer ( ) : = ( ) ( ) + exp i (9) (0) Calculation of weight adustments in our used MLP follows the following steps: 580 Table Recognition results using 000 epochs Reference Hidden Neurons

4 J. Computer Sci., (7): , 008 errors. Table 6 and 7 show the extracted weight matrices for both input-hiddenand hidden-put, which is made available via the specially Designed Brain Database (DBD) that automatically adds the new trained engine weights to its contents while eeping trac of all previously used weights. Table Recognition results using 0000 epochs Reference Hidden Neurons Table Percentage classification Percentage Percentage Percentage Hidden Classification classification classification Neurons 00 Epochs 000 Epochs 0000 Epochs DISCUSSION Traditionally, pruning is carried on hidden layers and hidden layer neurons however we introduce in our wor the principle of Cross Pruning (CP) or what we can call the pruning matrix whereby we carried pruning on both hidden layer neurons and number of epochs which serves as a two dimensional controlling matrix function for the design and implementation of MLP based recognition systems [5-,7-9]. Figure shows the effect of pruning and learning process on the classification accuracy and recognition rate of our MLP based OCR system. It is clear from the graphs that best accuracy and classification is obtained when using 000 epochs and 5 hidden neurons. It observed that using 00 epochs resulted in under learning and poor recognition of tested numerals with slow oscillations due to slow recognition time while at the other extreme of 0000 epochs instability and fast oscillations which indicates shorter recall time and over learning with slightly better recognition than 00 epochs case. Table 5 Classification and sum squared errors Sum squared Sum squared Sum squared Hidden error error error Neurons (00 Epochs) (00 Epochs) (00 Epochs) Table 6: Hidden-put weight matrix W :0 W :0 W :60 W 6:80 W 8:00 W 0:0 W :0 W :60 W 6:80 W 8:

5 J. Computer Sci., (7): , 008 Table 7: Input-hidden weight matrix W :80 W 8:60 W 6:0 W :0 W :00 W 0:80 W 8:560 W 56:60 W 6:70 W 7:

6 J. Computer Sci., (7): , 008 Percentage classification Character classification 00 Epochs 000 Epochs 0000 Epochs No. of hidden neurons Fig.: Effect of cross pruning on character classification Number of Incorrect classifications Numerals miss-classifications 8 Typed 00 7 Typed Typed Reference values Fig. : Effect of cross pruning on classification errors Number of Incorrect classifications Numerals miss-classifications Reference values Typed 00 Typed 000 Typed 0000 Fig. 5: Effect of cross pruning on miss-classification Figure demonstrates application of Eq. in conunction with our cross pruning approach to obtain a much more reliable and reflective representation for the errors that have occurred as a result of miss classification. It is clear from the diagram that the greatest percentage of miss classification occurs when using a 00 epochs with larger oscillation period compared to a 000 epochs the number of tested hidden neurons with the difference of amplitude reduction in the sum squared errors at earlier stage due to the large number of cycles but still unstable due to over teaching and over fitting compared to the 00 epochs case which suffers instability due to under learning and under fitting. It is clear from the curve that the optimum number of epochs that produce a stable classification system with minimum number of errors as a function of hidden layer neurons is a 000 epochs. Figure 5 deals with an important aspect of designing our MLP OCR system and in general contributes to designing a much more accurate classification system through the plotting and analyzing of what we can call Statistical Classification Reach (SCR) whereby the figure indicates through the use of our Cross Pruning (CP) technique the effect of varying both number of epochs and number of hidden layer neurons on the absolute statistical difference of numeral miss classification which can be generalized to cover character and word miss classification. It is clear from Fig. 5 and by applying our SCR technique that our system will be most unstable when using 00 and 0000 epochs with low number of hidden layer neurons and it also shows that the numeral the system had most difficulty in learning and classifying is the numeral as it shares common properties with numerals, 8. The figure also indicates that epoch numbers and hidden neuron numbers had no effect on the classification of the numeral 7 as its composition is simple and does not share many features with other numerals; hence it is difficult to be confused or misclassified. CONCLUSION It is nown that the choice of hidden units depend on many factors such as the amount of noise in the training data, the type of the shaping or activation function, the training algorithm, the number of input and put units and number of training patterns. Overall, deciding how many hidden layer neurons to use is a complex tas which this research tried to clarify and quantify. In doing so and to have a meaningful characterization for the system performance in terms of classification and to include and analyze the dependency of all the mentioned factors the concept of cross pruning is introduced together with the Statistical Classification Reach (SCR) and weight matrix auto update. These two techniques helped in deciding the optimal number of hidden layer neurons and number of epochs and pinpointed to the numerals that the system found difficult to classify. Our approach is certainly novel in terms of reducing time and effort and simplifying approaches to system design and modification which will assist in eliminating any wea points found in numeral based system in terms of numeral, MLP based recognition system. 58

7 REFERENCES. Thivierge, J.P., F. Rivestand and T.R. Shultz, 00. A dual-phase technique for pruning constructive networs. Proceedings of the IEEE International Joint Conference on Neural Networs, July 0-, IEEE Xplore Press, USA., pp: ber=07.. Pareh, R., J. Yang and V. Honavar, 000. Constructive neural-networ learning algorithms for pattern classification. IEEE Trans. Neural Networs, : xpl/freeabs_all.sp?tp=&arnumber=890&isnum ber=800.. Engelbrecht, A., 00. A new pruning heuristic based on variance analysis of sensitivity information. IEEE Trans. Neural Networs, : DOI: 0.09/ Pareh, R., J. Yang and V. Honavar, 000. Constructive neural networ learning algorithms for multi-category pattern classification. IEEE Trans. Neural Networs, : 6-5. Digital Obect Identifier 0.09/ Yang, J., R. Pareh and V. Honavar, 000. Comparison of performance of variants of singlelayer perceptron algorithms on non-separable data. Neural, Parallel Sci. Comput., 8: onid=b960d8e989607cccca67b? doi= Chen, C.H. and V. Honavar, 999. A neural networ architecture for syntax analysis. IEEE Trans. Neural Networs, 0: Yang, J., R. Pareh and V. Honavar, 998. DistAl: An inter-pattern distance based constructive learning algorithm. Technical Report TR /0/50/index.html. 8. Yao, X., 99. A review of evolutionary artificial neural networs. Int. J. Intell. Syst., : Cantú-Paz, E., 00. Pruning neural networs with distribution estimation algorithms. Lecture Notes Comput. Sci., 7: ist.psu.edu/viewdoc/summary?doi= Amin, H., K.M. Curtis and B.R. Hayes Gill, 997. Dynamically pruning put weights in an expanding multilayer perceptron neural networ. Proceedings of the th International Conference on Digital Signal Processing, July -, Santorini, Greece, pp: DOI: 0.09/ICDSP J. Computer Sci., (7): , Chen, H.H., M.T. Manry and H. Chandrasearan, 999. A neural networ training algorithm utilizing multiple sets of linear equations. Neurocomputing, 5: viewdoc/summary?doi=0.... Ponnapalli, 999. A formal selection and pruning algorithm for feedforward artificial networ optimization. IEEE Trans. Neural Networs, 0: DOI: 0.09/ Poliar, R., L. Udpa, S. Udpa and V. Honavar, 00. Learn++: An incremental learning algorithm for supervised neural networs. IEEE Trans. Syst., Manand Cybernetics-Part C: Appl. Rev., : DOI: 0.09/ Costa, M., A. Braga and B. Menezes, 00. Constructive and pruning methods for neural networ design. Proceeding of the 7th Brazilian Symposium on Neural Networs, Nov. -, IEEE Computer Society, USA., pp: Grippo, L., 000. Convergent on-line algorithms for supervised learning in neural networs. IEEE Trans. Neural Networ, : DOI: 0.09/ Burgess, N., 99. A constructive algorithm that converges for realvalued input patterns. Int. J. Neural Syst., 5: DOI: 0.09/ Reed, R., 99. Pruning algorithms: A survey. IEEE Trans. Neural Networ, : DOI: 0.09/ Hancoc, P.J.B., 99. Pruning neural nets by genetic algorithm. Proceedings of the International Conference on Artificial Neural Networs, 99, Elsevier Science, Amsterdam, Netherlands, pp: summary?doi= Lim, T.J., W.Y. Loh and Y.S. Shih, 000. A comparison of prediction accuracy, complexity and training time of thirty-three old and new classification algorithms. Mach. Learn., 0: summary?doi= Setiono, R. and A. Gaweda, 000. Neural networ pruning for function approximation. Proceeding of the International Joint Conference on Neural Networs, July -7, IEEE Computer Society, Los Alamitos CA., pp: amodele=affichen&cpsidt=795.. de Kruif, B.J. and T.J.A. de Vries, 00. Pruning error minimization in least squares support vector machines. IEEE Trans. Neural Networs, : DOI: 0.09/TNN

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions ENEE 739Q: STATISTICAL AND NEURAL PATTERN RECOGNITION Spring 2002 Assignment 2 Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions Aravind Sundaresan

More information

11/14/2010 Intelligent Systems and Soft Computing 1

11/14/2010 Intelligent Systems and Soft Computing 1 Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works The neuron as a simple computing element The perceptron Multilayer neural networks Accelerated learning in

More information

Image Compression: An Artificial Neural Network Approach

Image Compression: An Artificial Neural Network Approach Image Compression: An Artificial Neural Network Approach Anjana B 1, Mrs Shreeja R 2 1 Department of Computer Science and Engineering, Calicut University, Kuttippuram 2 Department of Computer Science and

More information

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Neural Networks CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Biological and artificial neural networks Feed-forward neural networks Single layer

More information

Supervised Learning in Neural Networks (Part 2)

Supervised Learning in Neural Networks (Part 2) Supervised Learning in Neural Networks (Part 2) Multilayer neural networks (back-propagation training algorithm) The input signals are propagated in a forward direction on a layer-bylayer basis. Learning

More information

In this assignment, we investigated the use of neural networks for supervised classification

In this assignment, we investigated the use of neural networks for supervised classification Paul Couchman Fabien Imbault Ronan Tigreat Gorka Urchegui Tellechea Classification assignment (group 6) Image processing MSc Embedded Systems March 2003 Classification includes a broad range of decision-theoric

More information

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting.

An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting. An Algorithm For Training Multilayer Perceptron (MLP) For Image Reconstruction Using Neural Network Without Overfitting. Mohammad Mahmudul Alam Mia, Shovasis Kumar Biswas, Monalisa Chowdhury Urmi, Abubakar

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

INTELLIGENT PROCESS SELECTION FOR NTM - A NEURAL NETWORK APPROACH

INTELLIGENT PROCESS SELECTION FOR NTM - A NEURAL NETWORK APPROACH International Journal of Industrial Engineering Research and Development (IJIERD), ISSN 0976 6979(Print), ISSN 0976 6987(Online) Volume 1, Number 1, July - Aug (2010), pp. 87-96 IAEME, http://www.iaeme.com/iierd.html

More information

Dept. of Computing Science & Math

Dept. of Computing Science & Math Lecture 4: Multi-Laer Perceptrons 1 Revie of Gradient Descent Learning 1. The purpose of neural netor training is to minimize the output errors on a particular set of training data b adusting the netor

More information

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used. 1 4.12 Generalization In back-propagation learning, as many training examples as possible are typically used. It is hoped that the network so designed generalizes well. A network generalizes well when

More information

Fast Learning for Big Data Using Dynamic Function

Fast Learning for Big Data Using Dynamic Function IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Fast Learning for Big Data Using Dynamic Function To cite this article: T Alwajeeh et al 2017 IOP Conf. Ser.: Mater. Sci. Eng.

More information

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India.

Argha Roy* Dept. of CSE Netaji Subhash Engg. College West Bengal, India. Volume 3, Issue 3, March 2013 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Training Artificial

More information

Channel Performance Improvement through FF and RBF Neural Network based Equalization

Channel Performance Improvement through FF and RBF Neural Network based Equalization Channel Performance Improvement through FF and RBF Neural Network based Equalization Manish Mahajan 1, Deepak Pancholi 2, A.C. Tiwari 3 Research Scholar 1, Asst. Professor 2, Professor 3 Lakshmi Narain

More information

A neural network that classifies glass either as window or non-window depending on the glass chemistry.

A neural network that classifies glass either as window or non-window depending on the glass chemistry. A neural network that classifies glass either as window or non-window depending on the glass chemistry. Djaber Maouche Department of Electrical Electronic Engineering Cukurova University Adana, Turkey

More information

Why MultiLayer Perceptron/Neural Network? Objective: Attributes:

Why MultiLayer Perceptron/Neural Network? Objective: Attributes: Why MultiLayer Perceptron/Neural Network? Neural networks, with their remarkable ability to derive meaning from complicated or imprecise data, can be used to extract patterns and detect trends that are

More information

Efficient Object Tracking Using K means and Radial Basis Function

Efficient Object Tracking Using K means and Radial Basis Function Efficient Object Tracing Using K means and Radial Basis Function Mr. Pradeep K. Deshmuh, Ms. Yogini Gholap University of Pune Department of Post Graduate Computer Engineering, JSPM S Rajarshi Shahu College

More information

Role of Hidden Neurons in an Elman Recurrent Neural Network in Classification of Cavitation Signals

Role of Hidden Neurons in an Elman Recurrent Neural Network in Classification of Cavitation Signals Role of Hidden Neurons in an Elman Recurrent Neural Network in Classification of Cavitation Signals Ramadevi R Sathyabama University, Jeppiaar Nagar, Rajiv Gandhi Road, Chennai 600119, India Sheela Rani

More information

Accelerating the convergence speed of neural networks learning methods using least squares

Accelerating the convergence speed of neural networks learning methods using least squares Bruges (Belgium), 23-25 April 2003, d-side publi, ISBN 2-930307-03-X, pp 255-260 Accelerating the convergence speed of neural networks learning methods using least squares Oscar Fontenla-Romero 1, Deniz

More information

A Novel Technique for Optimizing the Hidden Layer Architecture in Artificial Neural Networks N. M. Wagarachchi 1, A. S.

A Novel Technique for Optimizing the Hidden Layer Architecture in Artificial Neural Networks N. M. Wagarachchi 1, A. S. American International Journal of Research in Science, Technology, Engineering & Mathematics Available online at http://www.iasir.net ISSN (Print): 2328-3491, ISSN (Online): 2328-3580, ISSN (CD-ROM): 2328-3629

More information

Classification and Regression using Linear Networks, Multilayer Perceptrons and Radial Basis Functions

Classification and Regression using Linear Networks, Multilayer Perceptrons and Radial Basis Functions ENEE 739Q SPRING 2002 COURSE ASSIGNMENT 2 REPORT 1 Classification and Regression using Linear Networks, Multilayer Perceptrons and Radial Basis Functions Vikas Chandrakant Raykar Abstract The aim of the

More information

A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems

A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems Chapter 5 A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems 5.1 Introduction Many researchers have proposed pruning algorithms in numerous ways to optimize

More information

International Journal of Electrical and Computer Engineering 4: Application of Neural Network in User Authentication for Smart Home System

International Journal of Electrical and Computer Engineering 4: Application of Neural Network in User Authentication for Smart Home System Application of Neural Network in User Authentication for Smart Home System A. Joseph, D.B.L. Bong, and D.A.A. Mat Abstract Security has been an important issue and concern in the smart home systems. Smart

More information

Notes on Multilayer, Feedforward Neural Networks

Notes on Multilayer, Feedforward Neural Networks Notes on Multilayer, Feedforward Neural Networks CS425/528: Machine Learning Fall 2012 Prepared by: Lynne E. Parker [Material in these notes was gleaned from various sources, including E. Alpaydin s book

More information

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS Neural Networks Classifier Introduction INPUT: classification data, i.e. it contains an classification (class) attribute. WE also say that the class

More information

Seismic regionalization based on an artificial neural network

Seismic regionalization based on an artificial neural network Seismic regionalization based on an artificial neural network *Jaime García-Pérez 1) and René Riaño 2) 1), 2) Instituto de Ingeniería, UNAM, CU, Coyoacán, México D.F., 014510, Mexico 1) jgap@pumas.ii.unam.mx

More information

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5 Artificial Neural Networks Lecture Notes Part 5 About this file: If you have trouble reading the contents of this file, or in case of transcription errors, email gi0062@bcmail.brooklyn.cuny.edu Acknowledgments:

More information

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska

Classification Lecture Notes cse352. Neural Networks. Professor Anita Wasilewska Classification Lecture Notes cse352 Neural Networks Professor Anita Wasilewska Neural Networks Classification Introduction INPUT: classification data, i.e. it contains an classification (class) attribute

More information

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks

Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks Simulation of Zhang Suen Algorithm using Feed- Forward Neural Networks Ritika Luthra Research Scholar Chandigarh University Gulshan Goyal Associate Professor Chandigarh University ABSTRACT Image Skeletonization

More information

Artificial Neural Network based Curve Prediction

Artificial Neural Network based Curve Prediction Artificial Neural Network based Curve Prediction LECTURE COURSE: AUSGEWÄHLTE OPTIMIERUNGSVERFAHREN FÜR INGENIEURE SUPERVISOR: PROF. CHRISTIAN HAFNER STUDENTS: ANTHONY HSIAO, MICHAEL BOESCH Abstract We

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Image Data: Classification via Neural Networks Instructor: Yizhou Sun yzsun@ccs.neu.edu November 19, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining

More information

IN recent years, neural networks have attracted considerable attention

IN recent years, neural networks have attracted considerable attention Multilayer Perceptron: Architecture Optimization and Training Hassan Ramchoun, Mohammed Amine Janati Idrissi, Youssef Ghanou, Mohamed Ettaouil Modeling and Scientific Computing Laboratory, Faculty of Science

More information

Data Mining. Neural Networks

Data Mining. Neural Networks Data Mining Neural Networks Goals for this Unit Basic understanding of Neural Networks and how they work Ability to use Neural Networks to solve real problems Understand when neural networks may be most

More information

Conditional Random Field for tracking user behavior based on his eye s movements 1

Conditional Random Field for tracking user behavior based on his eye s movements 1 Conditional Random Field for tracing user behavior based on his eye s movements 1 Trinh Minh Tri Do Thierry Artières LIP6, Université Paris 6 LIP6, Université Paris 6 8 rue du capitaine Scott 8 rue du

More information

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation K. Roy, U. Pal and B. B. Chaudhuri CVPR Unit; Indian Statistical Institute, Kolkata-108; India umapada@isical.ac.in

More information

MATLAB representation of neural network Outline Neural network with single-layer of neurons. Neural network with multiple-layer of neurons.

MATLAB representation of neural network Outline Neural network with single-layer of neurons. Neural network with multiple-layer of neurons. MATLAB representation of neural network Outline Neural network with single-layer of neurons. Neural network with multiple-layer of neurons. Introduction: Neural Network topologies (Typical Architectures)

More information

Constructively Learning a Near-Minimal Neural Network Architecture

Constructively Learning a Near-Minimal Neural Network Architecture Constructively Learning a Near-Minimal Neural Network Architecture Justin Fletcher and Zoran ObradoviC Abetract- Rather than iteratively manually examining a variety of pre-specified architectures, a constructive

More information

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6)

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6) International Journals of Advanced Research in Computer Science and Software Engineering Research Article June 17 Artificial Neural Network in Classification A Comparison Dr. J. Jegathesh Amalraj * Assistant

More information

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM Annals of the University of Petroşani, Economics, 12(4), 2012, 185-192 185 IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM MIRCEA PETRINI * ABSTACT: This paper presents some simple techniques to improve

More information

CMPT 882 Week 3 Summary

CMPT 882 Week 3 Summary CMPT 882 Week 3 Summary! Artificial Neural Networks (ANNs) are networks of interconnected simple units that are based on a greatly simplified model of the brain. ANNs are useful learning tools by being

More information

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013 Application of Neural Network for Different Learning Parameter in Classification of Local Feature Image Annie anak Joseph, Chong Yung Fook Universiti Malaysia Sarawak, Faculty of Engineering, 94300, Kota

More information

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-&3 -(' ( +-   % '.+ % ' -0(+$, The structure is a very important aspect in neural network design, it is not only impossible to determine an optimal structure for a given problem, it is even impossible to prove that a given structure

More information

Neural Networks Based Time-Delay Estimation using DCT Coefficients

Neural Networks Based Time-Delay Estimation using DCT Coefficients American Journal of Applied Sciences 6 (4): 73-78, 9 ISSN 1546-939 9 Science Publications Neural Networks Based Time-Delay Estimation using DCT Coefficients Samir J. Shaltaf and Ahmad A. Mohammad Department

More information

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 11, November 2014,

More information

Visual object classification by sparse convolutional neural networks

Visual object classification by sparse convolutional neural networks Visual object classification by sparse convolutional neural networks Alexander Gepperth 1 1- Ruhr-Universität Bochum - Institute for Neural Dynamics Universitätsstraße 150, 44801 Bochum - Germany Abstract.

More information

Week 3: Perceptron and Multi-layer Perceptron

Week 3: Perceptron and Multi-layer Perceptron Week 3: Perceptron and Multi-layer Perceptron Phong Le, Willem Zuidema November 12, 2013 Last week we studied two famous biological neuron models, Fitzhugh-Nagumo model and Izhikevich model. This week,

More information

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers A. Salhi, B. Minaoui, M. Fakir, H. Chakib, H. Grimech Faculty of science and Technology Sultan Moulay Slimane

More information

Deep Learning. Volker Tresp Summer 2014

Deep Learning. Volker Tresp Summer 2014 Deep Learning Volker Tresp Summer 2014 1 Neural Network Winter and Revival While Machine Learning was flourishing, there was a Neural Network winter (late 1990 s until late 2000 s) Around 2010 there

More information

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu Natural Language Processing CS 6320 Lecture 6 Neural Language Models Instructor: Sanda Harabagiu In this lecture We shall cover: Deep Neural Models for Natural Language Processing Introduce Feed Forward

More information

Pattern Classification Algorithms for Face Recognition

Pattern Classification Algorithms for Face Recognition Chapter 7 Pattern Classification Algorithms for Face Recognition 7.1 Introduction The best pattern recognizers in most instances are human beings. Yet we do not completely understand how the brain recognize

More information

Character Recognition Using Convolutional Neural Networks

Character Recognition Using Convolutional Neural Networks Character Recognition Using Convolutional Neural Networks David Bouchain Seminar Statistical Learning Theory University of Ulm, Germany Institute for Neural Information Processing Winter 2006/2007 Abstract

More information

Research on Evaluation Method of Product Style Semantics Based on Neural Network

Research on Evaluation Method of Product Style Semantics Based on Neural Network Research Journal of Applied Sciences, Engineering and Technology 6(23): 4330-4335, 2013 ISSN: 2040-7459; e-issn: 2040-7467 Maxwell Scientific Organization, 2013 Submitted: September 28, 2012 Accepted:

More information

Today. Gradient descent for minimization of functions of real variables. Multi-dimensional scaling. Self-organizing maps

Today. Gradient descent for minimization of functions of real variables. Multi-dimensional scaling. Self-organizing maps Today Gradient descent for minimization of functions of real variables. Multi-dimensional scaling Self-organizing maps Gradient Descent Derivatives Consider function f(x) : R R. The derivative w.r.t. x

More information

International Journal of Advance Engineering and Research Development OPTICAL CHARACTER RECOGNITION USING ARTIFICIAL NEURAL NETWORK

International Journal of Advance Engineering and Research Development OPTICAL CHARACTER RECOGNITION USING ARTIFICIAL NEURAL NETWORK Scientific Journal of Impact Factor(SJIF): 4.14 International Journal of Advance Engineering and Research Development Special Issue for ICPCDECT 2016,Volume 3 Issue 1 e-issn(o): 2348-4470 p-issn(p): 2348-6406

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Perceptron as a graph

Perceptron as a graph Neural Networks Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University October 10 th, 2007 2005-2007 Carlos Guestrin 1 Perceptron as a graph 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0-6 -4-2

More information

Machine Learning 13. week

Machine Learning 13. week Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of

More information

THE NEURAL NETWORKS: APPLICATION AND OPTIMIZATION APPLICATION OF LEVENBERG-MARQUARDT ALGORITHM FOR TIFINAGH CHARACTER RECOGNITION

THE NEURAL NETWORKS: APPLICATION AND OPTIMIZATION APPLICATION OF LEVENBERG-MARQUARDT ALGORITHM FOR TIFINAGH CHARACTER RECOGNITION International Journal of Science, Environment and Technology, Vol. 2, No 5, 2013, 779 786 ISSN 2278-3687 (O) THE NEURAL NETWORKS: APPLICATION AND OPTIMIZATION APPLICATION OF LEVENBERG-MARQUARDT ALGORITHM

More information

ESTIMATION OF SUBSURFACE QANATS DEPTH BY MULTI LAYER PERCEPTRON NEURAL NETWORK VIA MICROGRAVITY DATA

ESTIMATION OF SUBSURFACE QANATS DEPTH BY MULTI LAYER PERCEPTRON NEURAL NETWORK VIA MICROGRAVITY DATA Advances in Geosciences Vol. 20: Solid Earth (2008) Ed. Kenji Satake c World Scientific Publishing Company ESTIMATION OF SUBSURFACE QANATS DEPTH BY MULTI LAYER PERCEPTRON NEURAL NETWORK VIA MICROGRAVITY

More information

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R.

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R. Lecture 24: Learning 3 Victor R. Lesser CMPSCI 683 Fall 2010 Today s Lecture Continuation of Neural Networks Artificial Neural Networks Compose of nodes/units connected by links Each link has a numeric

More information

COMBINING NEURAL NETWORKS FOR SKIN DETECTION

COMBINING NEURAL NETWORKS FOR SKIN DETECTION COMBINING NEURAL NETWORKS FOR SKIN DETECTION Chelsia Amy Doukim 1, Jamal Ahmad Dargham 1, Ali Chekima 1 and Sigeru Omatu 2 1 School of Engineering and Information Technology, Universiti Malaysia Sabah,

More information

Neural Network Weight Selection Using Genetic Algorithms

Neural Network Weight Selection Using Genetic Algorithms Neural Network Weight Selection Using Genetic Algorithms David Montana presented by: Carl Fink, Hongyi Chen, Jack Cheng, Xinglong Li, Bruce Lin, Chongjie Zhang April 12, 2005 1 Neural Networks Neural networks

More information

Website: HOPEFIELD NETWORK. Inderjeet Singh Behl, Ankush Saini, Jaideep Verma. ID-

Website:   HOPEFIELD NETWORK. Inderjeet Singh Behl, Ankush Saini, Jaideep Verma.  ID- International Journal Of Scientific Research And Education Volume 1 Issue 7 Pages 154-162 2013 ISSN (e): 2321-7545 Website: http://ijsae.in HOPEFIELD NETWORK Inderjeet Singh Behl, Ankush Saini, Jaideep

More information

Implementation of a Library for Artificial Neural Networks in C

Implementation of a Library for Artificial Neural Networks in C Implementation of a Library for Artificial Neural Networks in C Jack Breese TJHSST Computer Systems Lab 2007-2008 June 10, 2008 1 Abstract In modern computing, there are several approaches to pattern recognition

More information

Simulation of Back Propagation Neural Network for Iris Flower Classification

Simulation of Back Propagation Neural Network for Iris Flower Classification American Journal of Engineering Research (AJER) e-issn: 2320-0847 p-issn : 2320-0936 Volume-6, Issue-1, pp-200-205 www.ajer.org Research Paper Open Access Simulation of Back Propagation Neural Network

More information

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism) Artificial Neural Networks Analogy to biological neural systems, the most robust learning systems we know. Attempt to: Understand natural biological systems through computational modeling. Model intelligent

More information

Neuron Selectivity as a Biologically Plausible Alternative to Backpropagation

Neuron Selectivity as a Biologically Plausible Alternative to Backpropagation Neuron Selectivity as a Biologically Plausible Alternative to Backpropagation C.J. Norsigian Department of Bioengineering cnorsigi@eng.ucsd.edu Vishwajith Ramesh Department of Bioengineering vramesh@eng.ucsd.edu

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION CHAPTER 1 INTRODUCTION 1.1 Introduction Pattern recognition is a set of mathematical, statistical and heuristic techniques used in executing `man-like' tasks on computers. Pattern recognition plays an

More information

Machine Learning Classifiers and Boosting

Machine Learning Classifiers and Boosting Machine Learning Classifiers and Boosting Reading Ch 18.6-18.12, 20.1-20.3.2 Outline Different types of learning problems Different types of learning algorithms Supervised learning Decision trees Naïve

More information

THE CLASSICAL method for training a multilayer feedforward

THE CLASSICAL method for training a multilayer feedforward 930 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 4, JULY 1999 A Fast U-D Factorization-Based Learning Algorithm with Applications to Nonlinear System Modeling and Identification Youmin Zhang and

More information

COMPUTATIONAL INTELLIGENCE

COMPUTATIONAL INTELLIGENCE COMPUTATIONAL INTELLIGENCE Fundamentals Adrian Horzyk Preface Before we can proceed to discuss specific complex methods we have to introduce basic concepts, principles, and models of computational intelligence

More information

Neural Networks (Overview) Prof. Richard Zanibbi

Neural Networks (Overview) Prof. Richard Zanibbi Neural Networks (Overview) Prof. Richard Zanibbi Inspired by Biology Introduction But as used in pattern recognition research, have little relation with real neural systems (studied in neurology and neuroscience)

More information

Optimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem

Optimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS Optimization Model of K-Means Clustering Using Artificial Neural Networks to Handle Class Imbalance Problem To cite this article:

More information

STEREO-DISPARITY ESTIMATION USING A SUPERVISED NEURAL NETWORK

STEREO-DISPARITY ESTIMATION USING A SUPERVISED NEURAL NETWORK 2004 IEEE Workshop on Machine Learning for Signal Processing STEREO-DISPARITY ESTIMATION USING A SUPERVISED NEURAL NETWORK Y. V. Venkatesh, B. S. Venhtesh and A. Jaya Kumar Department of Electrical Engineering

More information

CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS

CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS 8.1 Introduction The recognition systems developed so far were for simple characters comprising of consonants and vowels. But there is one

More information

Artificial Neural Network-Based Prediction of Human Posture

Artificial Neural Network-Based Prediction of Human Posture Artificial Neural Network-Based Prediction of Human Posture Abstract The use of an artificial neural network (ANN) in many practical complicated problems encourages its implementation in the digital human

More information

Performance Analysis of Data Mining Classification Techniques

Performance Analysis of Data Mining Classification Techniques Performance Analysis of Data Mining Classification Techniques Tejas Mehta 1, Dr. Dhaval Kathiriya 2 Ph.D. Student, School of Computer Science, Dr. Babasaheb Ambedkar Open University, Gujarat, India 1 Principal

More information

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa Instructors: Parth Shah, Riju Pahwa Lecture 2 Notes Outline 1. Neural Networks The Big Idea Architecture SGD and Backpropagation 2. Convolutional Neural Networks Intuition Architecture 3. Recurrent Neural

More information

Supervised Learning with Neural Networks. We now look at how an agent might learn to solve a general problem by seeing examples.

Supervised Learning with Neural Networks. We now look at how an agent might learn to solve a general problem by seeing examples. Supervised Learning with Neural Networks We now look at how an agent might learn to solve a general problem by seeing examples. Aims: to present an outline of supervised learning as part of AI; to introduce

More information

Approximating a Policy Can be Easier Than Approximating a Value Function

Approximating a Policy Can be Easier Than Approximating a Value Function Computer Science Technical Report Approximating a Policy Can be Easier Than Approximating a Value Function Charles W. Anderson www.cs.colo.edu/ anderson February, 2 Technical Report CS-- Computer Science

More information

This leads to our algorithm which is outlined in Section III, along with a tabular summary of it's performance on several benchmarks. The last section

This leads to our algorithm which is outlined in Section III, along with a tabular summary of it's performance on several benchmarks. The last section An Algorithm for Incremental Construction of Feedforward Networks of Threshold Units with Real Valued Inputs Dhananjay S. Phatak Electrical Engineering Department State University of New York, Binghamton,

More information

A Population-Based Learning Algorithm Which Learns Both. Architectures and Weights of Neural Networks y. Yong Liu and Xin Yao

A Population-Based Learning Algorithm Which Learns Both. Architectures and Weights of Neural Networks y. Yong Liu and Xin Yao A Population-Based Learning Algorithm Which Learns Both Architectures and Weights of Neural Networks y Yong Liu and Xin Yao Computational Intelligence Group Department of Computer Science University College,

More information

Classification of Printed Chinese Characters by Using Neural Network

Classification of Printed Chinese Characters by Using Neural Network Classification of Printed Chinese Characters by Using Neural Network ATTAULLAH KHAWAJA Ph.D. Student, Department of Electronics engineering, Beijing Institute of Technology, 100081 Beijing, P.R.CHINA ABDUL

More information

Iris Recognition Using Discrete Cosine Transform and Artificial Neural Networks

Iris Recognition Using Discrete Cosine Transform and Artificial Neural Networks Journal of Computer Science 5 (5): 369-373, 2009 ISSN 1549-3636 2009 Science Publications Iris Recognition Using Discrete Cosine Transform and Artificial Neural Networks Ahmad M. Sarhan Department of Computer

More information

Mobile Application with Optical Character Recognition Using Neural Network

Mobile Application with Optical Character Recognition Using Neural Network Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 1, January 2015,

More information

Neural Network Neurons

Neural Network Neurons Neural Networks Neural Network Neurons 1 Receives n inputs (plus a bias term) Multiplies each input by its weight Applies activation function to the sum of results Outputs result Activation Functions Given

More information

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK IN GAIT RECOGNITION

CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK IN GAIT RECOGNITION 75 CHAPTER 6 COUNTER PROPAGATION NEURAL NETWORK IN GAIT RECOGNITION 6.1 INTRODUCTION Counter propagation network (CPN) was developed by Robert Hecht-Nielsen as a means to combine an unsupervised Kohonen

More information

Neural Network Classifier for Isolated Character Recognition

Neural Network Classifier for Isolated Character Recognition Neural Network Classifier for Isolated Character Recognition 1 Ruby Mehta, 2 Ravneet Kaur 1 M.Tech (CSE), Guru Nanak Dev University, Amritsar (Punjab), India 2 M.Tech Scholar, Computer Science & Engineering

More information

SUPPORT VECTOR MACHINES AND PROBABILISTIC NEURAL NETWORKS IN THE ASSESSMENT OF THE RISK OF DAMAGE TO WATER SUPPLY SYSTEMS IN MINING AREAS

SUPPORT VECTOR MACHINES AND PROBABILISTIC NEURAL NETWORKS IN THE ASSESSMENT OF THE RISK OF DAMAGE TO WATER SUPPLY SYSTEMS IN MINING AREAS SUPPORT VECTOR MACHINES AND PROBABILISTIC NEURAL NETWORKS IN THE ASSESSMENT OF THE RISK OF DAMAGE TO WATER SUPPLY SYSTEMS IN MINING AREAS Janusz RUSEK * Department of Engineering Surveying and Civil Engineering

More information

Learning and Generalization in Single Layer Perceptrons

Learning and Generalization in Single Layer Perceptrons Learning and Generalization in Single Layer Perceptrons Neural Computation : Lecture 4 John A. Bullinaria, 2015 1. What Can Perceptrons do? 2. Decision Boundaries The Two Dimensional Case 3. Decision Boundaries

More information

A Compensatory Wavelet Neuron Model

A Compensatory Wavelet Neuron Model A Compensatory Wavelet Neuron Model Sinha, M., Gupta, M. M. and Nikiforuk, P.N Intelligent Systems Research Laboratory College of Engineering, University of Saskatchewan Saskatoon, SK, S7N 5A9, CANADA

More information

OPTIMIZATION OF TURNING PROCESS USING A NEURO-FUZZY CONTROLLER

OPTIMIZATION OF TURNING PROCESS USING A NEURO-FUZZY CONTROLLER Sixteenth National Convention of Mechanical Engineers and All India Seminar on Future Trends in Mechanical Engineering, Research and Development, Deptt. Of Mech. & Ind. Engg., U.O.R., Roorkee, Sept. 29-30,

More information

RIMT IET, Mandi Gobindgarh Abstract - In this paper, analysis the speed of sending message in Healthcare standard 7 with the use of back

RIMT IET, Mandi Gobindgarh Abstract - In this paper, analysis the speed of sending message in Healthcare standard 7 with the use of back Global Journal of Computer Science and Technology Neural & Artificial Intelligence Volume 13 Issue 3 Version 1.0 Year 2013 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global

More information

Ensembles of Neural Networks for Forecasting of Time Series of Spacecraft Telemetry

Ensembles of Neural Networks for Forecasting of Time Series of Spacecraft Telemetry ISSN 1060-992X, Optical Memory and Neural Networks, 2017, Vol. 26, No. 1, pp. 47 54. Allerton Press, Inc., 2017. Ensembles of Neural Networks for Forecasting of Time Series of Spacecraft Telemetry E. E.

More information

Hybrid Method with ANN application for Linear Programming Problem

Hybrid Method with ANN application for Linear Programming Problem Hybrid Method with ANN application for Linear Programming Problem L. R. Aravind Babu* Dept. of Information Technology, CSE Wing, DDE Annamalai University, Annamalainagar, Tamilnadu, India ABSTRACT: This

More information

MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS

MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS MODIFIED KALMAN FILTER BASED METHOD FOR TRAINING STATE-RECURRENT MULTILAYER PERCEPTRONS Deniz Erdogmus, Justin C. Sanchez 2, Jose C. Principe Computational NeuroEngineering Laboratory, Electrical & Computer

More information

Neural Networks and Deep Learning

Neural Networks and Deep Learning Neural Networks and Deep Learning Example Learning Problem Example Learning Problem Celebrity Faces in the Wild Machine Learning Pipeline Raw data Feature extract. Feature computation Inference: prediction,

More information

Multi Layer Perceptron trained by Quasi Newton learning rule

Multi Layer Perceptron trained by Quasi Newton learning rule Multi Layer Perceptron trained by Quasi Newton learning rule Feed-forward neural networks provide a general framework for representing nonlinear functional mappings between a set of input variables and

More information

Extending reservoir computing with random static projections: a hybrid between extreme learning and RC

Extending reservoir computing with random static projections: a hybrid between extreme learning and RC Extending reservoir computing with random static projections: a hybrid between extreme learning and RC John Butcher 1, David Verstraeten 2, Benjamin Schrauwen 2,CharlesDay 1 and Peter Haycock 1 1- Institute

More information

ECE 285 Class Project Report

ECE 285 Class Project Report ECE 285 Class Project Report Based on Source localization in an ocean waveguide using supervised machine learning Yiwen Gong ( yig122@eng.ucsd.edu), Yu Chai( yuc385@eng.ucsd.edu ), Yifeng Bu( ybu@eng.ucsd.edu

More information