Minimum Redundancy and Maximum Relevance Feature Selec4on. Hang Xiao
|
|
- Benedict Mills
- 5 years ago
- Views:
Transcription
1 Minimum Redundancy and Maximum Relevance Feature Selec4on Hang Xiao
2 Background Feature a feature is an individual measurable heuris4c property of a phenomenon being observed In character recogni4on: horizontal and ver4cal profiles, number of internal holes, stroke detec4on In speech recogni4on: noise ra4os, length of sounds, rela4ve power, filter matches In microarray : genes expression
3 Background Relevance between features Correla4on F- sta4s4c Mutual informa4on I(x, y) = Z Z p(x, y)log p(x, y) p(x)p(y) dxdy p(x,y) : joint distribu4on func4on of X and Y p(x), p(y) : marginal probability distribu4on func4ons Independent : p(x,y) = p(x)p(y) è I(x,y) = 0
4 Feature Selec4on Problem Maximal relevance selec4ng the features with the highest relevance to the target class c, based on mutual info., F- test, etc. without considering rela4onships among features Minimal Redundancy Selected features are correlated Selected features cover narrow regions in space
5 mrmr: Discrete Variables Maximize Relevance: max V I,V I = 1 S S is the set of features I(i,j) is mutual informa4on between feature i and j Minimal Redundancy: min W I,W I = 1 S 2 X I(h, i) i2s X i,j2s I(i, j)
6 mrmr: Con4nuous Variables Maximum relevance: F- sta4s4c F(i,h) max V F,V F = 1 S Minimum redundancy : Correla4on cor(i,j) min W c,w c = 1 S 2 X X F (i, h) i2s i,j cor(i, j)
7 Combine Relevance and Redundancy Addi4ve combina4on max(v W ) Mul4plica4ve combina4on max(v/w )
8 Most Related Methods Most used feature selec4on methods: top- ranking features without considering rela4onships among features. Yu & Liu, 2003/2004. informa4on gain, essen4ally similar approach Wrapper: not filter approach, classifier- involved and thus features do not generalize well PCA and ICA: Feature are orthogonal or independent, but not in the original feature space
9 Class Predic4on Methods Naive Bayes (NB) classifier p(h k s) / i2s p(g i h k ) {g 1, g 2,, g m } gene expression level p(g i h k ) is condi4onal table (density) Support Vector Machine SVM Draw an op4mal hyperplane in the feature vector space
10 Class Predic4on Methods Linear Discriminant Analysis (LDA) Find a linear combina4on of feature ANOVA, regression analysis Logis4c Regression (LR) a linear combina4on of the feature variables transformed into probabili4es by a logis4c func4on
11 Microarray Gene Expression Data Sets for Cancer Classifica4on
12 LOOCV : Leave- One- Out Cross Valida4on Baseline feature : based solely on maximum relevance
13 The role of redundancy reduc4on (a) Relevance V I, and (b) Redundancy for MRMR features on discre4zed NCI dataset. (c) The respec4ve LOOCV errors obtained using the Naive Bayes classifier
14 Do mrmr Features Generalize Well on Unseen Data? Child Leukemia data (7 classes, 215 training samples, 112 tes4ng samples) tes4ng errors. M is the number of features used in classifica4on
15 What is the Rela4onship of mrmr Features and Various Data Discre4za4on Schemes? LOOCV tes4ng results classifier(#error) for binarized NCI and Lymphoma data using SVM classifier.
16 Comparison with other work
17 Theore4cal basis of mrmr Maximum Dependency Criterion Sta4s4c associa4on Defini4on : mutual informa4on I(S m,h) Mutual Informa4on For two variables x and y I(x, y) = Z Z p(x, y)log p(x, y) p(x)p(y) dxdy For mul4variate variable S m and the target h I(S m ; h) = Z Z p(s m,h)log p(s m,h) p(s m )p(h) ds mdh
18 High- Dimensional Mutual Informa4on For mul4variate variable Sm and the target h I(S m ; h) = = = Z Z p(s m,h)log p(s m,h) p(s m )p(h) ds mdh Z Z p(s m 1,x m,h)log p(s m 1,x m,h) p(s m 1,x m )p(h) ds mdh Z Z... p(x 1,...,x m,h)log p(x 1,...,x m,h) p(x 1,...,x m )p(h) dx 1...dx m dh Es4mate high- dimensional I(S m,h) is so difficult An ill- posed problem to find inverse of large co- variance matrix Insufficient number of samples Combinatorial 4me complex O(C( Ω, S ))
19 Factorize the Mutual Informa4on Mutual informa4on for mul4variate variable S m and the target h I(S m ; h) = Z Z p(s m,h)log p(s m,h) p(s m )p(h) ds mdh Define: Z Z J(x 1,x 2,...,x m )=... It can be proved: p(x 1,...,x m )log p(x 1,x 2,...,x m ) p(x 1 )...p(x m ) dx 1...dx m I(S m,h)=j(h, S m ) J(S m )
20 Factorize I(S m,h) Relevance of S={x 1,x 2, } and h, or RL(S,h) Redundancy among variables {x 1,x 2,...}, or RD(S) RL = 1 S X x i 2S I(x i,h) RD = 1 S 2 I(S m,h)=j(h, S m ) J(S m ) X x i x j 2S I(x i,x j ) For incremental search, max I(S,h) is equivalent to max [RL(S,h) RD(S)], so called min- Redundancy- Max- Relevance(mRMR)
21 Advantages of mrmr Both relevance and redundancy es4ma4on are low- dimensional problems (i.e. involving only 2 variables). This is much easier than directly es4ma4ng mul4variate density or mutual informa4on in the high- dimensional space! Fast speed More reliable es4ma4on mrmr is an op4mal first- order approxima4on of I(.) maximiza4on Relevance- only ranking only maximizes J(.)!
22 Search Algorithm of mrmr Greedy search algorithm In the pool Ω, find the variable x 1 that has the largest I(x 1,h). Exclude x 1 from Ω Search x 2 so that it maximizes I(x 2,h) - I(.,x 2 )/ Ω Iterate this process un4l an expected number of variables have been obtained, or other constraints are sa4sfied Complexity O( S * Ω )
23 Comparing Max- Dep and mrmr: Complexity of Feature Selec4on
24 Comparing Max- Dep and mrmr: Accuracy of Feature Selected in Classifica4on Leave- One- Out cross valida4on of feature classifica4on accuracies of mrmr and MaxDep
25 Use Wrappers to Refine Features mrmr is a filter approach Fast Features might be redundant Independent of the classifier Wrappers seek to minimize the number of errors directly Slow Features are less robust Dependent on classifier Beuer predic4on accuracy Use mrmr first to generate a short feature pool and use wrappers to get a least redundant feature set with beuer accuracy
26 Use Wrappers to Refine Features Forward wrappers (incremental selec4on) Backward wrappers (decremental selec4on) NCI Data
27 Conclusions The Max- Dependency feature selec4on can be efficiently implemented as the mrmr algorithm Significantly outperforms the widely used max- relevance selec4on method: mrmr features cover a broader feature space with less features mrmr is very efficient and useful for gene selec4on and many other applica4ons. The programs are ready!
Machine Learning Crash Course: Part I
Machine Learning Crash Course: Part I Ariel Kleiner August 21, 2012 Machine learning exists at the intersec
More informationLogis&c Regression. Aar$ Singh & Barnabas Poczos. Machine Learning / Jan 28, 2014
Logis&c Regression Aar$ Singh & Barnabas Poczos Machine Learning 10-701/15-781 Jan 28, 2014 Linear Regression & Linear Classifica&on Weight Height Linear fit Linear decision boundary 2 Naïve Bayes Recap
More informationCS 6140: Machine Learning Spring Final Exams. What we learned Final Exams 2/26/16
Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Assignment
More informationCS 6140: Machine Learning Spring 2016
CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa?on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Logis?cs Assignment
More informationFeature Selec+on. Machine Learning Fall 2018 Kasthuri Kannan
Feature Selec+on Machine Learning Fall 2018 Kasthuri Kannan Interpretability vs. Predic+on Types of feature selec+on Subset selec+on/forward/backward Shrinkage (Lasso/Ridge) Best model (CV) Feature selec+on
More information! What is mrmr feature selection. ! Applications in cancer classification. ! Applications in image pattern recognition. ! Theoretical basis of mrmr
Miniu Redundancy and Maxiu Relevance Feature election and Its Applications Hanchuan Peng Janelia Far Research Capus, Howard Hughes Medical Institute What is RMR feature selection Applications in cancer
More informationIntroduc)on to Probabilis)c Latent Seman)c Analysis. NYP Predic)ve Analy)cs Meetup June 10, 2010
Introduc)on to Probabilis)c Latent Seman)c Analysis NYP Predic)ve Analy)cs Meetup June 10, 2010 PLSA A type of latent variable model with observed count data and nominal latent variable(s). Despite the
More informationCSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates
CSCI 599 Class Presenta/on Zach Levine Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates April 26 th, 2012 Topics Covered in this Presenta2on A (Brief) Review of HMMs HMM Parameter Learning Expecta2on-
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 3 Parametric Distribu>ons We want model the probability
More informationFeature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani
Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate
More informationAbout the Course. Reading List. Assignments and Examina5on
Uppsala University Department of Linguis5cs and Philology About the Course Introduc5on to machine learning Focus on methods used in NLP Decision trees and nearest neighbor methods Linear models for classifica5on
More informationCS 6140: Machine Learning Spring 2017
CS 6140: Machine Learning Spring 2017 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Logis@cs Grades
More informationDecision Trees, Random Forests and Random Ferns. Peter Kovesi
Decision Trees, Random Forests and Random Ferns Peter Kovesi What do I want to do? Take an image. Iden9fy the dis9nct regions of stuff in the image. Mark the boundaries of these regions. Recognize and
More informationStages of (Batch) Machine Learning
Evalua&on Stages of (Batch) Machine Learning Given: labeled training data X, Y = {hx i,y i i} n i=1 Assumes each x i D(X ) with y i = f target (x i ) Train the model: model ß classifier.train(x, Y ) x
More informationSearch Engines. Informa1on Retrieval in Prac1ce. Annotations by Michael L. Nelson
Search Engines Informa1on Retrieval in Prac1ce Annotations by Michael L. Nelson All slides Addison Wesley, 2008 Retrieval Models Provide a mathema1cal framework for defining the search process includes
More informationMachine Learning. CS 232: Ar)ficial Intelligence Naïve Bayes Oct 26, 2015
1 CS 232: Ar)ficial Intelligence Naïve Bayes Oct 26, 2015 Machine Learning Part 1 of course: how use a model to make op)mal decisions (state space, MDPs) Machine learning: how to acquire a model from data
More informationDeformable Part Models
Deformable Part Models References: Felzenszwalb, Girshick, McAllester and Ramanan, Object Detec@on with Discrimina@vely Trained Part Based Models, PAMI 2010 Code available at hkp://www.cs.berkeley.edu/~rbg/latent/
More informationPa#ern Recogni-on for Neuroimaging Toolbox
Pa#ern Recogni-on for Neuroimaging Toolbox Pa#ern Recogni-on Methods: Basics João M. Monteiro Based on slides from Jessica Schrouff and Janaina Mourão-Miranda PRoNTo course UCL, London, UK 2017 Outline
More information10703 Deep Reinforcement Learning and Control
10703 Deep Reinforcement Learning and Control Russ Salakhutdinov Machine Learning Department rsalakhu@cs.cmu.edu Policy Gradient II Used Materials Disclaimer: Much of the material and slides for this lecture
More informationCSE 473: Ar+ficial Intelligence Machine Learning: Naïve Bayes and Perceptron
CSE 473: Ar+ficial Intelligence Machine Learning: Naïve Bayes and Perceptron Luke ZeFlemoyer --- University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI
More informationCollective classification in network data
1 / 50 Collective classification in network data Seminar on graphs, UCSB 2009 Outline 2 / 50 1 Problem 2 Methods Local methods Global methods 3 Experiments Outline 3 / 50 1 Problem 2 Methods Local methods
More informationML4Bio Lecture #1: Introduc3on. February 24 th, 2016 Quaid Morris
ML4Bio Lecture #1: Introduc3on February 24 th, 216 Quaid Morris Course goals Prac3cal introduc3on to ML Having a basic grounding in the terminology and important concepts in ML; to permit self- study,
More informationUnsupervised Learning
Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover
More information3 Feature Selection & Feature Extraction
3 Feature Selection & Feature Extraction Overview: 3.1 Introduction 3.2 Feature Extraction 3.3 Feature Selection 3.3.1 Max-Dependency, Max-Relevance, Min-Redundancy 3.3.2 Relevance Filter 3.3.3 Redundancy
More informationStatistical dependence measure for feature selection in microarray datasets
Statistical dependence measure for feature selection in microarray datasets Verónica Bolón-Canedo 1, Sohan Seth 2, Noelia Sánchez-Maroño 1, Amparo Alonso-Betanzos 1 and José C. Príncipe 2 1- Department
More informationA New Maximum-Relevance Criterion for Significant Gene Selection
A New Maximum-Relevance Criterion for Significant Gene Selection Young Bun Kim 1,JeanGao 1, and Pawel Michalak 2 1 Department of Computer Science and Engineering 2 Department of Biology The University
More informationQuick review: Data Mining Tasks... Classifica(on [Predic(ve] Regression [Predic(ve] Clustering [Descrip(ve] Associa(on Rule Discovery [Descrip(ve]
Evaluation Quick review: Data Mining Tasks... Classifica(on [Predic(ve] Regression [Predic(ve] Clustering [Descrip(ve] Associa(on Rule Discovery [Descrip(ve] Classification: Definition Given a collec(on
More informationCOMP 9517 Computer Vision
COMP 9517 Computer Vision Pa6ern Recogni:on (1) 1 Introduc:on Pa#ern recogni,on is the scien:fic discipline whose goal is the classifica:on of objects into a number of categories or classes Pa6ern recogni:on
More informationPredic'ng ALS Progression with Bayesian Addi've Regression Trees
Predic'ng ALS Progression with Bayesian Addi've Regression Trees Lilly Fang and Lester Mackey November 13, 2012 RECOMB Conference on Regulatory and Systems Genomics The ALS Predic'on Prize Challenge: Predict
More informationSearch Engines. Informa1on Retrieval in Prac1ce. Annota1ons by Michael L. Nelson
Search Engines Informa1on Retrieval in Prac1ce Annota1ons by Michael L. Nelson All slides Addison Wesley, 2008 Evalua1on Evalua1on is key to building effec$ve and efficient search engines measurement usually
More informationCross- Valida+on & ROC curve. Anna Helena Reali Costa PCS 5024
Cross- Valida+on & ROC curve Anna Helena Reali Costa PCS 5024 Resampling Methods Involve repeatedly drawing samples from a training set and refibng a model on each sample. Used in model assessment (evalua+ng
More informationMul$variate classifica$on. Astro 585 Spring 2013
Mul$variate classifica$on Astro 585 Spring 2013 Concepts of classifica$on Here we consider situa9ons where the mul9variate dataset under study represents a new test set that is a mixture of classes that
More informationInformation theory methods for feature selection
Information theory methods for feature selection Zuzana Reitermanová Department of Computer Science Faculty of Mathematics and Physics Charles University in Prague, Czech Republic Diplomový a doktorandský
More informationSupervised vs unsupervised clustering
Classification Supervised vs unsupervised clustering Cluster analysis: Classes are not known a- priori. Classification: Classes are defined a-priori Sometimes called supervised clustering Extract useful
More informationCS6200 Informa.on Retrieval. David Smith College of Computer and Informa.on Science Northeastern University
CS6200 Informa.on Retrieval David Smith College of Computer and Informa.on Science Northeastern University Query Process Retrieval Models Provide a mathema.cal framework for defining the search process
More informationCAP5415-Computer Vision Lecture 13-Support Vector Machines for Computer Vision Applica=ons
CAP5415-Computer Vision Lecture 13-Support Vector Machines for Computer Vision Applica=ons Guest Lecturer: Dr. Boqing Gong Dr. Ulas Bagci bagci@ucf.edu 1 October 14 Reminders Choose your mini-projects
More informationECE 634: Digital Video Systems Mo7on models: 1/19/17
ECE 634: Digital Video Systems Mo7on models: 1/19/17 Professor Amy Reibman MSEE 356 reibman@purdue.edu hip://engineering.purdue.edu/~reibman/ece634/index.html Outline Today: Mo7on models for mo7on es7ma7on
More information10601 Machine Learning. Model and feature selection
10601 Machine Learning Model and feature selection Model selection issues We have seen some of this before Selecting features (or basis functions) Logistic regression SVMs Selecting parameter value Prior
More informationTangible Visualiza.on. Andy Wu Synaesthe.c Media Lab GVU Center Georgia Ins.tute of Technology
Tangible Visualiza.on Andy Wu Synaesthe.c Media Lab GVU Center Georgia Ins.tute of Technology Introduc.on Informa.on Visualiza.on (Infovis) is the study of the visual representa.on of complex informa.on,
More informationA novel feature selection method based on normalized mutual information
Appl Intell (2012) 37:100 120 DOI 10.1007/s10489-011-0315-y A novel feature selection method based on normalized mutual information La The Vinh Sungyoung Lee Young-Tack Park Brian J. d Auriol Published
More informationExtending Heuris.c Search
Extending Heuris.c Search Talk at Hebrew University, Cri.cal MAS group Roni Stern Department of Informa.on System Engineering, Ben Gurion University, Israel 1 Heuris.c search 2 Outline Combining lookahead
More informationSVM Classification in -Arrays
SVM Classification in -Arrays SVM classification and validation of cancer tissue samples using microarray expression data Furey et al, 2000 Special Topics in Bioinformatics, SS10 A. Regl, 7055213 What
More informationIntroduc)on to Informa)on Visualiza)on
Introduc)on to Informa)on Visualiza)on Seeing the Science with Visualiza)on Raw Data 01001101011001 11001010010101 00101010100110 11101101011011 00110010111010 Visualiza(on Applica(on Visualiza)on on
More informationFeatures: representation, normalization, selection. Chapter e-9
Features: representation, normalization, selection Chapter e-9 1 Features Distinguish between instances (e.g. an image that you need to classify), and the features you create for an instance. Features
More informationEnsemble- Based Characteriza4on of Uncertain Features Dennis McLaughlin, Rafal Wojcik
Ensemble- Based Characteriza4on of Uncertain Features Dennis McLaughlin, Rafal Wojcik Hydrology TRMM TMI/PR satellite rainfall Neuroscience - - MRI Medicine - - CAT Geophysics Seismic Material tes4ng Laser
More informationof Manchester The University COMP14112 Markov Chains, HMMs and Speech Revision
COMP14112 Lecture 11 Markov Chains, HMMs and Speech Revision 1 What have we covered in the speech lectures? Extracting features from raw speech data Classification and the naive Bayes classifier Training
More informationTerraSwarm. A Machine Learning and Op0miza0on Toolkit for the Swarm. Ilge Akkaya, Shuhei Emoto, Edward A. Lee. University of California, Berkeley
TerraSwarm A Machine Learning and Op0miza0on Toolkit for the Swarm Ilge Akkaya, Shuhei Emoto, Edward A. Lee University of California, Berkeley TerraSwarm Tools Telecon 17 November 2014 Sponsored by the
More informationSearch Engines. Informa1on Retrieval in Prac1ce. Annotations by Michael L. Nelson
Search Engines Informa1on Retrieval in Prac1ce Annotations by Michael L. Nelson All slides Addison Wesley, 2008 Classifica1on and Clustering Classifica1on and clustering are classical padern recogni1on
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationCS395T Visual Recogni5on and Search. Gautam S. Muralidhar
CS395T Visual Recogni5on and Search Gautam S. Muralidhar Today s Theme Unsupervised discovery of images Main mo5va5on behind unsupervised discovery is that supervision is expensive Common tasks include
More informationDecision Trees: Representa:on
Decision Trees: Representa:on Machine Learning Fall 2017 Some slides from Tom Mitchell, Dan Roth and others 1 Key issues in machine learning Modeling How to formulate your problem as a machine learning
More informationBoW model. Textual data: Bag of Words model
BoW model Textual data: Bag of Words model With text, categoriza9on is the task of assigning a document to one or more categories based on its content. It is appropriate for: Detec9ng and indexing similar
More information10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors
Dejan Sarka Anomaly Detection Sponsors About me SQL Server MVP (17 years) and MCT (20 years) 25 years working with SQL Server Authoring 16 th book Authoring many courses, articles Agenda Introduction Simple
More informationClassification: Decision Trees
Classification: Decision Trees IST557 Data Mining: Techniques and Applications Jessie Li, Penn State University 1 Decision Tree Example Will a pa)ent have high-risk based on the ini)al 24-hour observa)on?
More informationAnalyzing Vocal Patterns to Determine Emotion Maisy Wieman, Andy Sun
Analyzing Vocal Patterns to Determine Emotion Maisy Wieman, Andy Sun 1. Introduction The human voice is very versatile and carries a multitude of emotions. Emotion in speech carries extra insight about
More informationSampling Strategies for Object Classifica6on. Gautam Muralidhar
Sampling Strategies for Object Classifica6on Gautam Muralidhar Reference papers The Pyramid Match Kernel Grauman and Darrell Approximated Correspondences in High Dimensions Grauman and Darrell Video Google
More informationFeature Selection. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 262
Feature Selection Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester 2016 239 / 262 What is Feature Selection? Department Biosysteme Karsten Borgwardt Data Mining Course Basel
More informationBENCHMARKING ATTRIBUTE SELECTION TECHNIQUES FOR MICROARRAY DATA
BENCHMARKING ATTRIBUTE SELECTION TECHNIQUES FOR MICROARRAY DATA S. DeepaLakshmi 1 and T. Velmurugan 2 1 Bharathiar University, Coimbatore, India 2 Department of Computer Science, D. G. Vaishnav College,
More informationPreface to the Second Edition. Preface to the First Edition. 1 Introduction 1
Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches
More informationAlignment and Image Comparison
Alignment and Image Comparison Erik Learned- Miller University of Massachuse>s, Amherst Alignment and Image Comparison Erik Learned- Miller University of Massachuse>s, Amherst Alignment and Image Comparison
More informationClassification. Slide sources:
Classification Slide sources: Gideon Dror, Academic College of TA Yaffo Nathan Ifill, Leicester MA4102 Data Mining and Neural Networks Andrew Moore, CMU : http://www.cs.cmu.edu/~awm/tutorials 1 Outline
More informationMachine Learning for Speaker Recogni2on and Bioinforma2cs
Machine Learning for Speaker Recogni2on and Bioinforma2cs Man-Wai MAK Dept of Electronic and Informa8on Engineering, The Hong Kong Polytechnic University http://wwweiepolyueduhk/~mwmak/ UTS/PolyU Workshop
More informationCSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo
CSE 6242 A / CS 4803 DVA Feb 12, 2013 Dimension Reduction Guest Lecturer: Jaegul Choo CSE 6242 A / CS 4803 DVA Feb 12, 2013 Dimension Reduction Guest Lecturer: Jaegul Choo Data is Too Big To Do Something..
More informationSta$s$cs & Experimental Design with R. Barbara Kitchenham Keele University
Sta$s$cs & Experimental Design with R Barbara Kitchenham Keele University 1 Analysis of Variance Mul$ple groups with Normally distributed data 2 Experimental Design LIST Factors you may be able to control
More informationChunking: An Empirical Evalua3on of So7ware Architecture (?)
Chunking: An Empirical Evalua3on of So7ware Architecture (?) Rachana Koneru David M. Weiss Iowa State University weiss@iastate.edu rachana.koneru@gmail.com With participation by Audris Mockus, Jeff St.
More informationLecture 2. White- box Tes2ng and Structural Coverage (see Amman and Offut, Chapter 2)
Lecture 2 White- box Tes2ng and Structural Coverage (see Amman and Offut, Chapter 2) White- box Tes2ng (aka. Glass- box or structural tes2ng) An error may exist at one (or more) loca2on(s) Line numbers
More informationToday s Class. High Dimensional Data & Dimensionality Reduc8on. Readings for This Week: Today s Class. Scien8fic Data. Misc. Personal Data 2/22/12
High Dimensional Data & Dimensionality Reduc8on Readings for This Week: Graphical Histories for Visualiza8on: Suppor8ng Analysis, Communica8on, and Evalua8on, Jeffrey Heer, Jock D. Mackinlay, Chris Stolte,
More informationVector Semantics. Dense Vectors
Vector Semantics Dense Vectors Sparse versus dense vectors PPMI vectors are long (length V = 20,000 to 50,000) sparse (most elements are zero) Alterna>ve: learn vectors which are short (length 200-1000)
More informationAlignment and Image Comparison. Erik Learned- Miller University of Massachuse>s, Amherst
Alignment and Image Comparison Erik Learned- Miller University of Massachuse>s, Amherst Alignment and Image Comparison Erik Learned- Miller University of Massachuse>s, Amherst Alignment and Image Comparison
More informationOrigin- des*na*on Flow Measurement in High- Speed Networks
IEEE INFOCOM, 2012 Origin- des*na*on Flow Measurement in High- Speed Networks Tao Li Shigang Chen Yan Qiao Introduc*on (Defini*ons) Origin- des+na+on flow between two routers is the set of packets that
More informationApply. A. Michelle Lawing Ecosystem Science and Management Texas A&M University College Sta,on, TX
Apply A. Michelle Lawing Ecosystem Science and Management Texas A&M University College Sta,on, TX 77843 alawing@tamu.edu Schedule for today My presenta,on Review New stuff Mixed, Fixed, and Random Models
More informationMTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen
MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen Lecture 2: Feature selection Feature Selection feature selection (also called variable selection): choosing k < d important
More informationSta$s$cs & Experimental Design with R. Barbara Kitchenham Keele University
Sta$s$cs & Experimental Design with R Barbara Kitchenham Keele University 1 Comparing two or more groups Part 5 2 Aim To cover standard approaches for independent and dependent groups For two groups Student
More informationMulti-voxel pattern analysis: Decoding Mental States from fmri Activity Patterns
Multi-voxel pattern analysis: Decoding Mental States from fmri Activity Patterns Artwork by Leon Zernitsky Jesse Rissman NITP Summer Program 2012 Part 1 of 2 Goals of Multi-voxel Pattern Analysis Decoding
More informationLouis Fourrier Fabien Gaie Thomas Rolf
CS 229 Stay Alert! The Ford Challenge Louis Fourrier Fabien Gaie Thomas Rolf Louis Fourrier Fabien Gaie Thomas Rolf 1. Problem description a. Goal Our final project is a recent Kaggle competition submitted
More informationUse JSL to Scrape Data from the Web and Predict Football Wins! William Baum Graduate Sta/s/cs Student University of New Hampshire
Use JSL to Scrape Data from the Web and Predict Football Wins! William Baum Graduate Sta/s/cs Student University of New Hampshire Just for Fun! I m an avid American football fan Sports sta/s/cs are easily
More informationQuiz Section Week 8 May 17, Machine learning and Support Vector Machines
Quiz Section Week 8 May 17, 2016 Machine learning and Support Vector Machines Another definition of supervised machine learning Given N training examples (objects) {(x 1,y 1 ), (x 2,y 2 ),, (x N,y N )}
More informationMachine Learning Techniques for Data Mining
Machine Learning Techniques for Data Mining Eibe Frank University of Waikato New Zealand 10/25/2000 1 PART VII Moving on: Engineering the input and output 10/25/2000 2 Applying a learner is not all Already
More informationCOMP61011 Foundations of Machine Learning. Feature Selection
OMP61011 Foundations of Machine Learning Feature Selection Pattern Recognition: The Early Days Only 200 papers in the world! I wish! Pattern Recognition: The Early Days Using eight very simple measurements
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2013 http://ce.sharif.edu/courses/91-92/2/ce725-1/ Agenda Features and Patterns The Curse of Size and
More informationCS 5614: (Big) Data Management Systems. B. Aditya Prakash Lecture #19: Machine Learning 1
CS 5614: (Big) Data Management Systems B. Aditya Prakash Lecture #19: Machine Learning 1 Supervised Learning Would like to do predicbon: esbmate a func3on f(x) so that y = f(x) Where y can be: Real number:
More informationSIFT (Scale Invariant Feature Transform) descriptor
Local descriptors SIFT (Scale Invariant Feature Transform) descriptor SIFT keypoints at loca;on xy and scale σ have been obtained according to a procedure that guarantees illumina;on and scale invance.
More informationAdvancing Feature Selection Research
Advancing Feature Selection Research ASU Feature Selection Repository Zheng Zhao zhaozheng@asu.edu Computer Science & Engineering Arizona State University Shashvata Sharma shashvata.sharma@asu.edu Computer
More informationClustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford
Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically
More informationExploratory data analysis for microarrays
Exploratory data analysis for microarrays Jörg Rahnenführer Computational Biology and Applied Algorithmics Max Planck Institute for Informatics D-66123 Saarbrücken Germany NGFN - Courses in Practical DNA
More informationCS6375: Machine Learning Gautam Kunapuli. Mid-Term Review
Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes
More informationSlides for Data Mining by I. H. Witten and E. Frank
Slides for Data Mining by I. H. Witten and E. Frank 7 Engineering the input and output Attribute selection Scheme-independent, scheme-specific Attribute discretization Unsupervised, supervised, error-
More informationUsing Classical Planners for Tasks with Con5nuous Ac5ons in Robo5cs
Using Classical Planners for Tasks with Con5nuous Ac5ons in Robo5cs Stuart Russell Joint work with Siddharth Srivastava, Lorenzo Riano, Pieter Abbeel Using Classical Planners for Tasks with Con5nuous Ac5ons
More informationComparison of different preprocessing techniques and feature selection algorithms in cancer datasets
Comparison of different preprocessing techniques and feature selection algorithms in cancer datasets Konstantinos Sechidis School of Computer Science University of Manchester sechidik@cs.man.ac.uk Abstract
More informationStable and Accurate Feature Selection
Stable and Accurate Feature Selection Gokhan Gulgezen, Zehra Cataltepe,andLeiYu Istanbul Technical University, Computer Engineering Department, Istanbul, Turkey Binghamton University, Computer Science
More informationEvalua&ng methods for represen&ng model error using ensemble data assimila&on
Evalua&ng methods for represen&ng model error using ensemble data assimila&on Jeff Whitaker NOAA Earth System Research Lab, Boulder, CO, USA jeffrey.s.whitaker@noaa.gov 1 Evalua&ng model error schemes
More information7 Techniques for Data Dimensionality Reduction
7 Techniques for Data Dimensionality Reduction Rosaria Silipo KNIME.com The 2009 KDD Challenge Prediction Targets: Churn (contract renewals), Appetency (likelihood to buy specific product), Upselling (likelihood
More informationFEATURE SELECTION TECHNIQUES
CHAPTER-2 FEATURE SELECTION TECHNIQUES 2.1. INTRODUCTION Dimensionality reduction through the choice of an appropriate feature subset selection, results in multiple uses including performance upgrading,
More informationRegression Tes+ng. Midterm Wednesday Oct. 26, 7:00-8:30pm Room 142
Regression Tes+ng Computer Science 521-621 Fall 2011 Prof. L. J. Osterweil Material adapted from slides originally prepared by Prof. L. A. Clarke Midterm Wednesday Oct. 26, 7:00-8:30pm Room 142 Closed
More informationContent-based image and video analysis. Machine learning
Content-based image and video analysis Machine learning for multimedia retrieval 04.05.2009 What is machine learning? Some problems are very hard to solve by writing a computer program by hand Almost all
More informationDiscriminate Analysis
Discriminate Analysis Outline Introduction Linear Discriminant Analysis Examples 1 Introduction What is Discriminant Analysis? Statistical technique to classify objects into mutually exclusive and exhaustive
More informationRapid Extraction and Updating Road Network from LIDAR Data
Rapid Extraction and Updating Road Network from LIDAR Data Jiaping Zhao, Suya You, Jing Huang Computer Science Department University of Southern California October, 2011 Research Objec+ve Road extrac+on
More informationCITS4009 Introduc0on to Data Science
School of Computer Science and Software Engineering CITS4009 Introduc0on to Data Science SEMESTER 2, 2017: CHAPTER 3 EXPLORING DATA 1 Chapter Objec0ves Using summary sta.s.cs to explore data Exploring
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationRegularization and model selection
CS229 Lecture notes Andrew Ng Part VI Regularization and model selection Suppose we are trying select among several different models for a learning problem. For instance, we might be using a polynomial
More information