Neural Networks CMSC475/675

Similar documents
Data Mining. Neural Networks

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Introduction to Neural Networks

For Monday. Read chapter 18, sections Homework:

11/14/2010 Intelligent Systems and Soft Computing 1

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Multilayer Feed-forward networks

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

COMPUTATIONAL INTELLIGENCE

NEURAL NETWORKS. Typeset by FoilTEX 1

Artificial Neural Networks

2. Neural network basics

Artificial Neural Networks. Introduction to Computational Neuroscience Ardi Tampuu

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

Character Recognition Using Convolutional Neural Networks

CMPT 882 Week 3 Summary

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Dr. Qadri Hamarsheh Supervised Learning in Neural Networks (Part 1) learning algorithm Δwkj wkj Theoretically practically

Neural Networks. Neural Network. Neural Network. Neural Network 2/21/2008. Andrew Kusiak. Intelligent Systems Laboratory Seamans Center

Website: HOPEFIELD NETWORK. Inderjeet Singh Behl, Ankush Saini, Jaideep Verma. ID-

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

Lecture 20: Neural Networks for NLP. Zubin Pahuja

11/14/2010 Intelligent Systems and Soft Computing 1

Logical Rhythm - Class 3. August 27, 2018

Perceptrons and Backpropagation. Fabio Zachert Cognitive Modelling WiSe 2014/15

464 Index. Associative memory paradigms

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty)

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Artificial Neural Networks Lecture Notes Part 5. Stephen Lucci, PhD. Part 5

IMPLEMENTING DEEP LEARNING USING CUDNN 이예하 VUNO INC.

Supervised Learning in Neural Networks (Part 2)

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

A Dendrogram. Bioinformatics (Lec 17)

Neural Networks: What can a network represent. Deep Learning, Fall 2018

Neural Networks: What can a network represent. Deep Learning, Spring 2018

Neural Networks. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Yuki Osada Andrew Cannon

! References: ! Computer eyesight gets a lot more accurate, NY Times. ! Stanford CS 231n. ! Christopher Olah s blog. ! Take ECS 174!

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

Neural Nets. General Model Building

Feedback Alignment Algorithms. Lisa Zhang, Tingwu Wang, Mengye Ren

Machine Learning 13. week

Exercise 2: Hopeld Networks

Opening the Black Box Data Driven Visualizaion of Neural N

Chapter 5 Neural Network Concepts and Paradigms

Neural Networks (Overview) Prof. Richard Zanibbi

Pattern Classification Algorithms for Face Recognition

Machine Learning in Biology

Correlation Matrix Memories: Improving Performance for Capacity and Generalisation

Practical Tips for using Backpropagation

Optimization Methods for Machine Learning (OMML)

Instantaneously trained neural networks with complex inputs

Deep (1) Matthieu Cord LIP6 / UPMC Paris 6

Administrative. Assignment 1 due Wednesday April 18, 11:59pm

Back propagation Algorithm:

Reservoir Computing for Neural Networks

Neural Networks (pp )

Hardware Neuronale Netzwerke - Lernen durch künstliche Evolution (?)

Neural Network Neurons

Biologically inspired object categorization in cluttered scenes

Connectionist Learning Architecture Based on an Optical Thin-Film Multilayer Model

Introduction AL Neuronale Netzwerke. VL Algorithmisches Lernen, Teil 2b. Norman Hendrich

Ensemble methods in machine learning. Example. Neural networks. Neural networks

Using neural nets to recognize hand-written digits. Srikumar Ramalingam School of Computing University of Utah

Climate Precipitation Prediction by Neural Network

II. ARTIFICIAL NEURAL NETWORK

Multi Layer Perceptron with Back Propagation. User Manual

Morphogenesis. Simulation Results

Deep Learning. Volker Tresp Summer 2014

ECG782: Multidimensional Digital Signal Processing

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Deep Learning. Architecture Design for. Sargur N. Srihari

POLITECNICO DI MILANO Faculty of Industrial Engineering and Informatics Department of Electronics, Informatics and Bioengineering

Hidden Units. Sargur N. Srihari

CSC 578 Neural Networks and Deep Learning

Alex Waibel

Neural Network and Deep Learning. Donglin Zeng, Department of Biostatistics, University of North Carolina

Lecture 17: Neural Networks and Deep Learning. Instructor: Saravanan Thirumuruganathan

Multi-layer Perceptron Forward Pass Backpropagation. Lecture 11: Aykut Erdem November 2016 Hacettepe University

Future Image Prediction using Artificial Neural Networks

Machine Learning Classifiers and Boosting

Backpropagation and Neural Networks. Lecture 4-1

Neural Network Approach for Automatic Landuse Classification of Satellite Images: One-Against-Rest and Multi-Class Classifiers

DEEP LEARNING REVIEW. Yann LeCun, Yoshua Bengio & Geoffrey Hinton Nature Presented by Divya Chitimalla

IMPROVEMENTS TO THE BACKPROPAGATION ALGORITHM

Introduction to Multilayer Perceptrons

Instructor: Jessica Wu Harvey Mudd College

COMPUTER SIMULATION OF COMPLEX SYSTEMS USING AUTOMATA NETWORKS K. Ming Leung

arxiv: v1 [cs.lg] 25 Jan 2018

Reservoir Computing with Emphasis on Liquid State Machines

CS311 - Neural Nets Lab Thursday, April 11

CP365 Artificial Intelligence

Cryptography Algorithms using Artificial Neural Network Arijit Ghosh 1 Department of Computer Science St. Xavier s College (Autonomous) Kolkata India

1. Introduction. 2. Motivation and Problem Definition. Volume 8 Issue 2, February Susmita Mohapatra

INTRODUCTION TO DEEP LEARNING

ANFIS: ADAPTIVE-NETWORK-BASED FUZZY INFERENCE SYSTEMS (J.S.R. Jang 1993,1995) bell x; a, b, c = 1 a

Image Compression: An Artificial Neural Network Approach

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Transcription:

Introduction to Neural Networks CMSC475/675

Chapter 1 Introduction

Why ANN Introduction Some tasks can be done easily (effortlessly) by humans but are hard by conventional paradigms on Von Neumann machine with algorithmic approach Pattern recognition (old friends, hand-written characters) Content addressable recall Approximate, common sense reasoning (driving, playing piano, baseball player) These tasks are often ill-defined, experience based, hard to apply logic

Introduction Von Neumann machine -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- One or a few high speed (ns) processors with considerable computing power One or a few shared high speed buses for communication Sequential memory access by address Problem-solving knowledge is separated from the computing component Hard to be adaptive Human Brain -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Large # (10 11 ) of low speed processors (ms) with limited computing power Large # (10 15 ) of low speed connections Content addressable recall (CAM) Problem-solving knowledge resides in the connectivity of neurons Adaptation by changing the connectivity ii

Biological neural activity Each neuron has a body, an axon, and many dendrites Can be in one of the two states: firing ing and rest. Neuron fires if the total incoming stimulus exceeds the threshold Synapse: thin gap gpbetween axon of one neuron and dendrite of another. Signal exchange Synaptic strength/efficiency

Introduction What is an (artificial) neural network A set of nodes (units, neurons, processing elements) Each node has input and output Each node performs a simple computation by its node function Weighted connections between nodes Connectivity gives the structure/architecture of the net What can be computed by a NN is primarily determined by the connections and their weights A very much simplified version of networks of neurons in animal nerve systems

NN -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Nodes input output node function Connections connection strength Introduction Bio NN -------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Cell body signals from other neurons firing frequency firing mechanism Synapses synaptic strength Highly parallel, simple local computation (at neuron level) achieves global results as emerging property of the interaction (at network level) Pattern directed (meaning of individual nodes only in the context of a pattern) Fault-tolerant/graceful degrading g Learning/adaptation plays important role.

Pitts & McCulloch (1943) History of NN First mathematical ti model of biological i lneurons All Boolean operations can be implemented by these neuron-like nodes (with different threshold and excitatory/inhibitory connections). Competitor to Von Neumann model for general purpose computing device in 1940 s and early 50 s Origin of automata theory. Hebb (1949) Hebbian rule of learning: increase the connection strength between neurons i and j whenever both i and j are activated. Or increase the connection strength between nodes i and j whenever both nodes are simultaneously ON or OFF.

History of NN Early booming (50 s early 60 s) Rosenblatt (1958) Perceptron: network of threshold nodes for pattern classification Perceptron learning rule Percenptron convergence theorem: x 1 x 2 everything that can be represented by a perceptron can be learned Widrow and Hoff (1960, 1962) Learning rule based on gradient descent (with differentiable unit) Minsky s attempt to build a general purpose machine with Pitts/McCullock units x n

History of NN The setback (mid 60 s late 70 s) Serious problems with perceptron model (Minsky s book 1969) Single layer perceonptrons cannot represent (learn) simple functions such as XOR Multi-layer of non-linear units may have greater power but there is no learning rule for such nets Scaling problem: connection weights may grow infinitely The first two problems overcame by latter effort in 80 s, but the scaling gproblem persists Death of Rosenblatt (1964) Striving of Von Neumann machine and AI

History of NN Renewed enthusiasm and flourish (80 s present) New techniques Backpropagation learning for multi-layer feed forward nets (with non-linear, differentiable node functions) Thermodynamic models (Hopfield net, Boltzmann machine, etc.) Unsupervised learning Impressive application (character recognition, speech recognition, text-to-speech transformation,,p process control, associative memory, etc.) Traditional approaches face difficult challenges Caution: Don t underestimate difficulties and limitations Poses more problems than solutions

ANN Neuron Models Each node has one or more inputs from other nodes, and one output to other nodes Input/output values can be Binary {0, 1} Bipolar {-1, 1} Continuous (bounded or not) All inputs to one node come in at tthe same time and remain activated until the output is produced Weights associated with links f ( net) is the most popular node function where net i 1 w i x i n General neuron model Weighted input summation

Node Function Identity function : f ( net) net. Constant function : f ( net) c. Step (threshold) function where c is called dthe thresholdh Ramp function Step function Ramp function

Sigmoid function S-shaped Continuous and everywhere differentiable Rotationally symmetric about some point (net = c) Asymptotically approaches saturation points Examples: Node Function Sigmoid function When y = 0 and z = 0: a = 0,b= b 1,c= c 0. When y = 0 and z = -0.5 a = -0.5, b = 0.5, c = 0. Larger x gives steeper curve where tanh( u ) e u e e u e u u

Gaussian function Bell-shaped (radial basis) Node Function Continuous f(net) asymptotically approaches 0 (or some constant) when net is large Single maximum (when net = ) Example: Gaussian function

Network Architecture (Asymmetric) Fully Connected Networks Every node is connected to every other node Connection may be excitatory (positive), inhibitory (negative), or irrelevant ( 0). Most general Symmetric fully connected nets: weights are symmetric (w ij = w ji ) Input nodes: receive input from the environment Output nodes: send signals to the environment Hidden nodes: no direct interaction to the environment

Layered Networks Network Architecture Nodes are partitioned into subsets, called layers. No connections that lead from nodes in layer j to those in layer k if j > k. Inputs from the environment are applied to nodes in layer 0 (input layer). ) Nodes in input layer are place holders with no computation occurring (i.e., their node functions are identity function)

Feedforward Networks Network Architecture A connection is allowed from a node in layer i only to nodes in layer i + 1. Most widely used architecture. Conceptually, nodes at higher levels successively abstract features from preceding layers

Acyclic Networks Network Architecture Connections do not form directed cycles Multi-layered feedforward nets are acyclic Recurrent Networks Nets with directed cycles Much harder to analyze than acyclic nets. Modular nets Consists of several modules, each of which is itself a neural net for a particular sub-problem Sparse connections between modules