Machine learning in fmri

Similar documents
Evaluating Classifiers

Evaluating Classifiers

CS4491/CS 7265 BIG DATA ANALYTICS

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

Evaluation Metrics. (Classifiers) CS229 Section Anand Avati

Evaluating Machine Learning Methods: Part 1

List of Exercises: Data Mining 1 December 12th, 2015

Cross- Valida+on & ROC curve. Anna Helena Reali Costa PCS 5024

Evaluating Machine-Learning Methods. Goals for the lecture

INTRODUCTION TO MACHINE LEARNING. Measuring model performance or error

MEASURING CLASSIFIER PERFORMANCE

Classification Part 4

Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates?

Machine Learning and Bioinformatics 機器學習與生物資訊學

Classification. Instructor: Wei Ding

ECLT 5810 Evaluation of Classification Quality

Use of Synthetic Data in Testing Administrative Records Systems

Data Mining Classification: Alternative Techniques. Imbalanced Class Problem

Data Mining Classification: Bayesian Decision Theory

A DISCRETIZATION METHOD BASED ON MAXIMIZING THE AREA UNDER RECEIVER OPERATING CHARACTERISTIC CURVE

Network Traffic Measurements and Analysis

CS 584 Data Mining. Classification 3

CS145: INTRODUCTION TO DATA MINING

Probabilistic Classifiers DWML, /27

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Machine Learning nearest neighbors classification. Luigi Cerulo Department of Science and Technology University of Sannio

Best First and Greedy Search Based CFS and Naïve Bayes Algorithms for Hepatitis Diagnosis

INTRODUCTION TO DATA MINING. Daniel Rodríguez, University of Alcalá

Business Club. Decision Trees

Model s Performance Measures

CS435 Introduction to Big Data Spring 2018 Colorado State University. 3/21/2018 Week 10-B Sangmi Lee Pallickara. FAQs. Collaborative filtering

10 Classification: Evaluation

Evaluating Classifiers

Lecture 6 K- Nearest Neighbors(KNN) And Predictive Accuracy

[Programming Assignment] (1)

CLASSIFICATION JELENA JOVANOVIĆ. Web:

Detecting DGA Malware Traffic Through Behavioral Models. Erquiaga, María José Catania, Carlos García, Sebastían

Data Mining: Classifier Evaluation. CSCI-B490 Seminar in Computer Science (Data Mining)

A Comparative Study of Locality Preserving Projection and Principle Component Analysis on Classification Performance Using Logistic Regression

Data Mining and Knowledge Discovery Practice notes 2

Chuck Cartledge, PhD. 23 September 2017

Machine Learning for. Artem Lind & Aleskandr Tkachenko

DATA MINING OVERFITTING AND EVALUATION

Weka ( )

Data Mining and Knowledge Discovery: Practice Notes

Decision trees. For this lab, we will use the Carseats data set from the ISLR package. (install and) load the package with the data set

DATA MINING LECTURE 9. Classification Decision Trees Evaluation

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane

CSE Data Mining Concepts and Techniques STATISTICAL METHODS (REGRESSION) Professor- Anita Wasilewska. Team 13

A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995)

Time series representations: a state-of-the-art

Data Mining and Knowledge Discovery: Practice Notes

Title of Projects: - Implementing a simple Recommender System on user based rating and testing the same.

EE795: Computer Vision and Intelligent Systems

Lecture 25: Review I

Machine Learning: Symbolische Ansätze

Evaluation. Evaluate what? For really large amounts of data... A: Use a validation set.

A Novel Approach to Compute Confusion Matrix for Classification of n-class Attributes with Feature Selection

CS249: ADVANCED DATA MINING

Package hmeasure. February 20, 2015

Nächste Woche. Dienstag, : Vortrag Ian Witten (statt Vorlesung) Donnerstag, 4.12.: Übung (keine Vorlesung) IGD, 10h. 1 J.

ML4Bio Lecture #1: Introduc3on. February 24 th, 2016 Quaid Morris

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

Introduction to Automated Text Analysis. bit.ly/poir599

Stages of (Batch) Machine Learning

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

Feature Subset Selection using Clusters & Informed Search. Team 3

Evaluation of different biological data and computational classification methods for use in protein interaction prediction.

Partitioning Data. IRDS: Evaluation, Debugging, and Diagnostics. Cross-Validation. Cross-Validation for parameter tuning

Large Scale Data Analysis Using Deep Learning

Introduction to Machine Learning

CHAPTER 3 MACHINE LEARNING MODEL FOR PREDICTION OF PERFORMANCE

Global Probability of Boundary

Part II: A broader view

2. On classification and related tasks

I211: Information infrastructure II

Discrimination of resting-state fmri for Schizophrenia patients with Lattice Computing based features

K- Nearest Neighbors(KNN) And Predictive Accuracy

Privacy Preserving Probabilistic Record Linkage

Naïve Bayes Classification. Material borrowed from Jonathan Huang and I. H. Witten s and E. Frank s Data Mining and Jeremy Wyatt and others

Metrics Overfitting Model Evaluation Research directions. Classification. Practical Issues. Huiping Cao. lassification-issues, Slide 1/57

CSE 190, Fall 2015: Midterm

Adaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented wh

Classification of Protein Crystallization Imagery

WP1: Video Data Analysis

Basic Tokenizing, Indexing, and Implementation of Vector-Space Retrieval

Data Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 5 of Data Mining by I. H. Witten, E. Frank and M. A.

Scalable Selective Traffic Congestion Notification

SNS College of Technology, Coimbatore, India

.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. for each element of the dataset we are given its class label.

Clustering will not be satisfactory if:

I How does the formulation (5) serve the purpose of the composite parameterization

Binary Diagnostic Tests Clustered Samples

Assignment No: 2. Assessment as per Schedule. Specifications Readability Assignments

Data Clustering. Danushka Bollegala

Cross-validation and the Bootstrap

Classification. Slide sources:

Pattern recognition (4)

Information Management course

Transcription:

Machine learning in fmri Validation Alexandre Savio, Maite Termenón, Manuel Graña 1 Computational Intelligence Group, University of the Basque Country December, 2010 1/18

Outline 1 Motivation The validation problem 2 Types of validation methods Hold-out Cross-validation The confusion matrix Measures of t 2/18

Outline 1 Motivation The validation problem 2 Types of validation methods Hold-out Cross-validation The confusion matrix Measures of t 3/18

The purpose of validation I We have: a model with one or more unknown parameters a data set to which the model can be t (the training data set). The tting process optimizes the model parameters to make the model t the training data as well as possible. 4/18

Over tting I If we then take an independent sample of validation data from the same population as the training data, it will generally turn out that the model does not t the validation data as well as it ts the training data. This is particularly likely to happen when: the size of the training data set is small, or when the number of parameters in the model is large. 5/18

Outline 1 Motivation The validation problem 2 Types of validation methods Hold-out Cross-validation The confusion matrix Measures of t 6/18

Hold-Out Split dataset into two groups: Training set: used to train the classi er Test set: used to estimate the error rate of the trained classi er 7/18

Cross-validation Cross-validation is: a way to predict the t of a model to a hypothetical validation set when an explicit validation set is not available, using computation in place of mathematical analysis. 8/18

Repeated random sub-sampling Each split randomly selects a ( xed) number of examples without replacement. For each data split we retrain the classi er from scratch with the training examples and estimate Ei with the test examples. 9/18

K-fold For each of K experiments, use K-1 folds for training and the remaining one for testing. The advantage of K-Fold cross-validation is that all the examples in the dataset are eventually used for both training and testing. Common choice for K-Fold cross-validation is K=10. 10/18

Leave-one-out Leave-one-out is the degenerate case of K-Fold cross Validation, where K is chosen as the total number of examples. For each experiment use N-1 examples for training and the remaining example for testing, then N experiments will be performed. 11/18

How many folds are needed? With a large number of folds [+] The bias of the true error rate estimator will be small (the estimator will be very accurate). [-] The variance of the true error rate estimator will be large. [-] The computational time will be very large as well (many experiments). With a small number of folds [+] The number of experiments and, therefore, computation time are reduced. [+] The variance of the estimator will be small. [-] The bias of the estimator will be large (conservative or higher than the true error rate). In practice, the choice of the number of folds depends on the size of the dataset. 12/18

Three-way data splits 13/18

Confusion matrix 14/18

Measures of t I TP + FP + FN + TN = N TN Accuracy: Acc = TP + N True Positive Rate or Sensitivity or Recall: TPR = TPTP +FN = sensitivity False Positive Rate: FPR = FPFP +TN = 1 speci city 15/18

Measures of t II Precision The proportion of the predicted positive cases that were correct, as calculated using the equation: Precision = TPTP +FP F-score: F score = Precision Recall FN Misclassi cation Ratio: MCR = FP + N Jaccard Similarity: Jaccard = TP +TP FN +FP Dice's Coe cient: DC = 2 TP2+ TP FN +FP 16/18

Measures of t III Accurate Segmentation Ratio (ASR) The ASR is only meaningful when we are dealing with multiple class problems. Then we will have speci c contigency matrices, one for each class of the problem. c (TPc ) ASR = sumsum where sumc means the sum over all c (TPc +FPc ) di erent classes. 17/18

Summary Cross-validation is a way to predict the t of a model to a hypothetical validation set when an explicit validation set is not available. K-fold cross-validation and leave-one-out are the most used approaches in fmri machine learning applications. There are many coe cients of t measure which can be chosen. Accuracy, Sensitivity and Speci city are the most used in medical literature. 18/18