Opening the Black Box Data Driven Visualizaion of Neural N

Similar documents
Supervised Learning in Neural Networks (Part 2)

Data Mining. Neural Networks

11/14/2010 Intelligent Systems and Soft Computing 1

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Multilayer Feed-forward networks

CHAPTER 7 MASS LOSS PREDICTION USING ARTIFICIAL NEURAL NETWORK (ANN)

CS 4510/9010 Applied Machine Learning. Neural Nets. Paula Matuszek Fall copyright Paula Matuszek 2016

CS6220: DATA MINING TECHNIQUES

Neural Network Weight Selection Using Genetic Algorithms

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS

WHAT TYPE OF NEURAL NETWORK IS IDEAL FOR PREDICTIONS OF SOLAR FLARES?

CMPT 882 Week 3 Summary

5 Learning hypothesis classes (16 points)

Neural Networks (pp )

CP365 Artificial Intelligence

Neural Network Learning. Today s Lecture. Continuation of Neural Networks. Artificial Neural Networks. Lecture 24: Learning 3. Victor R.

Machine Learning 13. week

Motivation. Problem: With our linear methods, we can train the weights but not the basis functions: Activator Trainable weight. Fixed basis function

Assignment # 5. Farrukh Jabeen Due Date: November 2, Neural Networks: Backpropation

Supervised Learning (contd) Linear Separation. Mausam (based on slides by UW-AI faculty)

Neural Networks. By Laurence Squires

6. NEURAL NETWORK BASED PATH PLANNING ALGORITHM 6.1 INTRODUCTION

Traffic Signs Recognition using HP and HOG Descriptors Combined to MLP and SVM Classifiers

In this assignment, we investigated the use of neural networks for supervised classification

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

Ensemble methods in machine learning. Example. Neural networks. Neural networks

turning data into dollars

Neural Networks. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Lecture #11: The Perceptron

Perceptron: This is convolution!

Machine Learning in Biology

Keywords: ANN; network topology; bathymetric model; representability.

.. Spring 2017 CSC 566 Advanced Data Mining Alexander Dekhtyar..

Seismic regionalization based on an artificial neural network

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

COMPUTATIONAL INTELLIGENCE

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Classification Algorithms for Face Recognition

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES

International Journal of Advanced Research in Computer Science and Software Engineering

Yuki Osada Andrew Cannon

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

Neural Networks CMSC475/675

Learning. Learning agents Inductive learning. Neural Networks. Different Learning Scenarios Evaluation

Graph Neural Network. learning algorithm and applications. Shujia Zhang

Neural Nets for Adaptive Filter and Adaptive Pattern Recognition

Back propagation Algorithm:

Logical Rhythm - Class 3. August 27, 2018

Learning from Data: Adaptive Basis Functions

A Neural Network for Real-Time Signal Processing

Introduction to Neural Networks

Instructor: Jessica Wu Harvey Mudd College

Week 3: Perceptron and Multi-layer Perceptron

11/14/2010 Intelligent Systems and Soft Computing 1

Character Recognition Using Convolutional Neural Networks

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

For Monday. Read chapter 18, sections Homework:

Analytical model A structure and process for analyzing a dataset. For example, a decision tree is a model for the classification of a dataset.

Image Compression: An Artificial Neural Network Approach

Artificial Intelligence Introduction Handwriting Recognition Kadir Eren Unal ( ), Jakob Heyder ( )

Administrative. Assignment 1 due Wednesday April 18, 11:59pm

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Random Search Report An objective look at random search performance for 4 problem sets

Artificial neural networks are the paradigm of connectionist systems (connectionism vs. symbolism)

A System for Joining and Recognition of Broken Bangla Numerals for Indian Postal Automation

Neural Network Classifier for Isolated Character Recognition

Support Vector Machines

Volume Illumination, Contouring

Artificial Neural Networks MLP, RBF & GMDH

Global Journal of Engineering Science and Research Management

Deep neural networks II

1-Nearest Neighbor Boundary

ECG782: Multidimensional Digital Signal Processing

Neural Network Approach for Automatic Landuse Classification of Satellite Images: One-Against-Rest and Multi-Class Classifiers

The exam is closed book, closed notes except your one-page cheat sheet.

Object-Based Classification & ecognition. Zutao Ouyang 11/17/2015

Neural Network Application Design. Supervised Function Approximation. Supervised Function Approximation. Supervised Function Approximation

A Class of Instantaneously Trained Neural Networks

A Systematic Overview of Data Mining Algorithms

The exam is closed book, closed notes except your one-page (two-sided) cheat sheet.

Neural Network Neurons

Technical University of Munich. Exercise 7: Neural Network Basics

Deep Learning with Tensorflow AlexNet

Extreme Learning Machines. Tony Oakden ANU AI Masters Project (early Presentation) 4/8/2014

Unit V. Neural Fuzzy System

Lecture on Modeling Tools for Clustering & Regression

An introduction to neural networks for beginners. By Dr Andy Thomas Adventures in Machine Learning

Deep Learning. Deep Learning provided breakthrough results in speech recognition and image classification. Why?

Matrix-Vector Multiplication by MapReduce. From Rajaraman / Ullman- Ch.2 Part 1

Backpropagation and Neural Networks. Lecture 4-1

27: Hybrid Graphical Models and Neural Networks

Inversion of Fracture Parameters by Using the Artificial Neural Network

Image Registration Lecture 4: First Examples

Computational Intelligence Meets the NetFlix Prize

Use of Artificial Neural Networks to Investigate the Surface Roughness in CNC Milling Machine

Noisy iris recognition: a comparison of classifiers and feature extractors

Multi Layer Perceptron with Back Propagation. User Manual

Transcription:

Opening the Black Box Data Driven Visualizaion of Neural Networks September 20, 2006

Aritificial Neural Networks Limitations of ANNs Use of Visualization (ANNs) mimic the processes found in biological neural networks. The brain is a massively parallel information processing system formed by about ten billion nerve cells (neurons) and their synapses. Used for predicting and learning from a given set of data. Based on the combination of neurons, connections, transfer functions, learning algorithms, and neuron layout methods.

Limitations of ANNs Aritificial Neural Networks Limitations of ANNs Use of Visualization ANNs act as a black box. Information stored is a set of weights and connections that provides no insight as to how a task is performed. The relationship between the input and output is not clear. ANNs do not provide a solution that is easily understood or validated. Parameter selection is tedious, and often done in a trial-and-error process.

Use of Visualization Aritificial Neural Networks Limitations of ANNs Use of Visualization Visualizations of ANNs give insight to how the NN works. Viz also used in network structuring and parameter selection, saving time and cost. Previous ANN viz did not focus on data used by network. This paper takes a data driven approach. In addition to ANN viz, the user is allowed to interact with data and see the effect on the network.

In General ANN Basics Specifics for this paper Each node (neuron) is connected to all other nodes in the adjacent layer. Each connection between nodes has a weight, with the weights modulating the value across the connection. Notation: Nodes in input layer: I 1, I 2, I 3,..., I m Nodes in Hidden layer: H 1, H 2, H 3,..., H n Weight of connection between I i and H j : W ij The value of a node in the hidden layer is given by: H j = TF ( n i=1 W ij I i )

(Cont.) ANN Basics Specifics for this paper The value of a node in the output layer (O k ): O k = TF ( m i=1 W jk H j ) TF (x) is a non-linear transfer function. Transfer function is what converts NN to a non-linear system. Non-linear needed since classifying NNs make terminal desicions. When calculating the value of an output node, the same transfer function used on the previous layer is applied to the summed results from that (previous) layer.

Training ANN Basics Specifics for this paper At the beginning... We need a set of training inputs and desired outputs. Weights are set at random. Weights are then iteratively modified to obtain minimum error. The criterion for minimum is some arbitrary threshold (maybe a small one). After training, the network may be used for data similar to the training set.

Specifics ANN Basics Specifics for this paper NN is feed-forward. Input always moves from left to right, no loops. Back-propogation training algorithm used. Mean squared error calculated first calculated at output level for a given input. Error is then minimized, by changing weights, in a back-to-front manner at the next level. Error minimization iteratively continues until threshold reached. Repeat steps 1-3 on the other test data. Start all over again.

(Cont.) ANN Basics Specifics for this paper The standard sigmoid function is used for the transfer function. 1 f (x) = (1+e x ) A continuous differentiable version of a stair-stepping function. Is the most common transfer fucntion used for classification.

Volume Classification Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty Trained by a small set of input data including the corresponding class IDs. Input vector includes a voxel s scalar value, gradient magnitude, its neighbors scalar values, and its position.

Spam Classification Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty Trained by a set of pre-classified spam and non-spam email messages (117, 283 respectively). Words and phrases from emails form input vectors. This NN used because of its large size (82 input nodes and 100 hidden nodes).

Single Data Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty Probe is provided to select from data domain. NN is then shown as data are passing through the network. Input and output nodes are colored based on the selected voxel s value. A connection s width is based on its importance. The focus is on the input-hidden layer, so results can be mapped to input data domain. Importance based on weight. The layer s influence is propogated by multiplying each weight between both layers that connect to the same hidden node.

Set of Data Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty A region of data is now selected. Now, input node size is also set, also based on importance, or contribution of the node to the result. For large NNs, the nodes can be arranged by order of importance. Relative importance is calculated by a ratio of weights associated with that node to that of all nodes (in the same layer). The user can (maybe) set a threshold that will hide less important connections.

Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty (Cont.) The assumption is made that selected data are similar (values are close to the mean of values), so weights between input-hidden layer are divided by those means. The standard deviation is also calculated, and nodes are assigned color based on the similarity of mean and standard deviation.

Visualizing Uncertainty Datasets Visualizing Weights With Volume Classification NN Visualizing Uncertainty NN outputs a value representing uncertainty. High and low values indicate low uncertainty. Middle values indicate hight uncertainty. Parallel coordinates are used to show the inputs and outputs when training or classifying. Colors correspond to how well the NN classified certain sets of data. This can be used to understand what type of test data is needed. Boundaries between materials, due to interpolation will appear (by color) to have a high level of uncertainty. The fix for this is to assign a voxel s color based on the value of the neighboring voxel along the opposite normal direction.

, Uses, Rendering Times, etc. None listed! The authors do note that based on their rendered node/connection weights for a particular NN, 4 nodes were removed from the hidden layer. Cost of classification was reduced by 15% Result 0.5% different from original result.