Stability of Feature Selection Algorithms

Similar documents
Stability of Feature Selection Algorithms

Model Mining for Robust Feature Selection

Weighted Heuristic Ensemble of Filters

Stable Feature Selection via Dense Feature Groups

ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A COMPARATIVE ANALYSIS

Exploratory data analysis for microarrays

Statistical Pattern Recognition

Statistical Pattern Recognition

Gene selection through Switched Neural Networks

Statistical Pattern Recognition

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Statistical dependence measure for feature selection in microarray datasets

A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection (Kohavi, 1995)

Dimensionality Reduction, including by Feature Selection.

Stable Feature Selection with Minimal Independent Dominating Sets

BENCHMARKING ATTRIBUTE SELECTION TECHNIQUES FOR MICROARRAY DATA

Filter methods for feature selection. A comparative study

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Accelerometer Gesture Recognition

Comparison of different preprocessing techniques and feature selection algorithms in cancer datasets

STABILITY OF FEATURE SELECTION ALGORITHMS AND ENSEMBLE FEATURE SELECTION METHODS IN BIOINFORMATICS

Feature Selection with Decision Tree Criterion

SVM Classification in -Arrays

Feature Selection Using Modified-MCA Based Scoring Metric for Classification

1 More configuration model

CS 229 Final Project Report Learning to Decode Cognitive States of Rat using Functional Magnetic Resonance Imaging Time Series

ECLT 5810 Clustering

Understanding Clustering Supervising the unsupervised

Gene signature selection to predict survival benefits from adjuvant chemotherapy in NSCLC patients

ECLT 5810 Clustering

BIOINF 585: Machine Learning for Systems Biology & Clinical Informatics

Redundancy Based Feature Selection for Microarray Data

Features: representation, normalization, selection. Chapter e-9

Big Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1

Predict Outcomes and Reveal Relationships in Categorical Data

STA 4273H: Statistical Machine Learning

9/29/13. Outline Data mining tasks. Clustering algorithms. Applications of clustering in biology

CPSC 340: Machine Learning and Data Mining. Kernel Trick Fall 2017

International Journal of Scientific & Engineering Research, Volume 5, Issue 7, July ISSN

Machine Learning Techniques for Data Mining

Gene Clustering & Classification

Slides for Data Mining by I. H. Witten and E. Frank

Noise-based Feature Perturbation as a Selection Method for Microarray Data

Technical Report. Feature Selection for High-Dimensional Data with RapidMiner. Sangkyun Lee, *Benjamin Schowe, *Viswanath Sivakumar, Katharina Morik

A FAST CLUSTERING-BASED FEATURE SUBSET SELECTION ALGORITHM

Individualized Error Estimation for Classification and Regression Models

Stable and Accurate Feature Selection

FEATURE EXTRACTION TECHNIQUES USING SUPPORT VECTOR MACHINES IN DISEASE PREDICTION

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

EECS730: Introduction to Bioinformatics

Towards robust feature selection for high-dimensional, small sample settings

April 3, 2012 T.C. Havens

Multicollinearity and Validation CIVL 7012/8012

A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis

Feature Selection and Classification for Small Gene Sets

Lecture 3: Linear Classification

REMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD

Data Mining Final Project on NIPS Competition 2003

Fast or furious? - User analysis of SF Express Inc

International Journal of Research in Advent Technology, Vol.7, No.3, March 2019 E-ISSN: Available online at

Outlier Ensembles. Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY Keynote, Outlier Detection and Description Workshop, 2013

FEATURE SELECTION TECHNIQUES

6. Dicretization methods 6.1 The purpose of discretization

AN ENSEMBLE OF FILTERS AND WRAPPERS FOR MICROARRAY DATA CLASSIFICATION

IBM SPSS Categories. Predict outcomes and reveal relationships in categorical data. Highlights. With IBM SPSS Categories you can:

Nearest Neighbor Predictors

Hybrid Feature Selection for Modeling Intrusion Detection Systems

Missing Data Estimation in Microarrays Using Multi-Organism Approach

Machine Learning: An Applied Econometric Approach Online Appendix

Fall 09, Homework 5

SSV Criterion Based Discretization for Naive Bayes Classifiers

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1

An Empirical Study of Lazy Multilabel Classification Algorithms

Use of KNN for the Netflix Prize Ted Hong, Dimitris Tsamis Stanford University

Mining High Order Decision Rules

MSA220 - Statistical Learning for Big Data

Sampling PCA, enhancing recovered missing values in large scale matrices. Luis Gabriel De Alba Rivera 80555S

Prognosis of Lung Cancer Using Data Mining Techniques

Collaborative Filtering using a Spreading Activation Approach

Yelp Recommendation System

Review of feature selection techniques in bioinformatics by Yvan Saeys, Iñaki Inza and Pedro Larrañaga.

7. Nearest neighbors. Learning objectives. Centre for Computational Biology, Mines ParisTech

Chapter 12 Feature Selection

Improving the Random Forest Algorithm by Randomly Varying the Size of the Bootstrap Samples for Low Dimensional Data Sets

Variable Selection 6.783, Biomedical Decision Support

Cross-validation and the Bootstrap

Allstate Insurance Claims Severity: A Machine Learning Approach

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018

scikit-learn (Machine Learning in Python)

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables

Internet Path Stability: Exploring the Impact of MPLS. Zakaria Al-Qudah, PhD. Yarmouk University April 2, 2015

A New Maximum-Relevance Criterion for Significant Gene Selection

CSC411 Fall 2014 Machine Learning & Data Mining. Ensemble Methods. Slides by Rich Zemel

Study of Data Mining Algorithm in Social Network Analysis

Package EFS. R topics documented:

Supervised Learning Classification Algorithms Comparison

Mining Significant Graph Patterns by Leap Search

A hybrid of discrete particle swarm optimization and support vector machine for gene selection and molecular classification of cancer

Transcription:

Stability of Feature Selection Algorithms Alexandros Kalousis, Jullien Prados, Phong Nguyen Melanie Hilario Artificial Intelligence Group Department of Computer Science University of Geneva

Stability of Feature Selection Algorithms 1 Motivation Provide a method to study and better understand the behavior of feature selection algorithms. Present users with a quantification of the resilience-robustness of the selected features. A lot of work for classification algorithms but nothing for feature selection algorithms.

Stability of Feature Selection Algorithms 2 Stability of Classification Algorithms Stability has been extensively studied in the context of classification algorithms. The main tool has been the Bias-Variance decomposition of the error. Variance measures sensitivity of a classification algorithm to different training sets. It does not measure how different are the models created from different datasets but only how different their predictions are. Nevertheless: BV needs predictions, feature selection algorithms alone do not provide predictions. We could couple a feature selection with a classification algorithm and perform BV but then we would be measuring their joint BV profile.

Stability of Feature Selection Algorithms 3 Feature Preferences From a vector of features, f = (f 1, f 2,..., f m ), a feature selection algorithm outputs feature preferences which can be: Weightings-scorings of features, w = (w 1, w 2,..., w m ) Rankings of features, w = (r 1, r 2,..., r m ) Subsets of selected features, s = (s 1, s 2,..., s m ), s i {0, 1} This is as close as we get to a classification model but it does not output predictions so if we want to directly quantify the stability of feature selection algorithms we should directly operate in this output.

Stability of Feature Selection Algorithms 4 Stability of feature selection algorithms We define the stability of a feature selection algorithm as the sensitivity of the feature preferences it produces to different training sets drawn from the same generating distribution P (X, C). To measure sensitivity we need similarity measures between feature preferences. We define one for each type of feature preference: S W (w, w ) based on Pearson s correlation coefficient. S R (r, r ) based on Spearman s correlation coefficient. S S (s, s ) based on Tanimoto s distance between sets. The sensitivity of feature selection algorithms that output: Weightings-scorings, can be measured using all three. Rankings, can be measured using S R and S S Subsets of features, can be measured using only S S

Stability of Feature Selection Algorithms 5 Estimating stability Draw k different training sets from P (X, C). Construct the corresponding k feature preferences. Compute the k(k 1) 2 pairwise feature preference similarities. The average, S, of feature preference similarities is the estimated value of stability. Since we do not have enough training sets we rely on resampling. We use 10-fold cross validation, but any other resampling method would do.

Stability of Feature Selection Algorithms 6 Coupling stability with error estimation In practice a feature selection algorithm is applied together with a classification algorithm and we get an error estimation of the coupled application. We want to couple this error estimation with a stability estimation in order to select among the most accurate configurations the one that is most stable. We use 10-fold cross-validation for error estimation and stability is estimated on each fold by an inner 10-fold cross-validation as described previously.

Stability of Feature Selection Algorithms 7 Experiments 11 datasets from proteomics, genomics and text mining. We examined five well known feature selection approaches: Three univariate: Information Gain (IG), CHI-Square (CHI), Symmetrical Uncertainty (SYM), Two multivarate: ReliefF (feature weights are determined based on their contribution on the euclidean distance) and SVM-RFE (linear svm with recursive feature elimination) We also examined the stability of a simple linear support vector machine (SVMONE) to demonstrate that the notion of stability can be also applied to any classification algorithm that produces a set of feature weights. To build the final classification models we used a linear SVM. S W and S R were computed on the complete feature preferences. S S was computed on the set of the ten best features given by each algorithm.

Stability of Feature Selection Algorithms 8 Datasets Proteomics, genomics and text mining datasets. Proteomics and genomics datasets are typical for their high dimensionality small sample size. The text mining datasets have a high dimensionality but also a high number of training instances. dataset class 1 # class 1 class 2 # class 2 # features ovarian normal 91 diseased 162 824 prostate normal 253 diseased 69 2200 stroke normal 101 diseased 107 4928 leukemia ALL 47 AML 25 7131 nervous survival 21 failure 39 7131 colon normal 22 tumor 40 2000 alt relevant 1425 not 2732 2112 disease relevant 631 not 2606 2376 function relevant 818 not 3089 2708 structure relevant 927 not 2621 2368 subcell relevant 1502 not 6475 4031

Stability of Feature Selection Algorithms 9 Can we say which stability measure is more appropriate? S W, S R, are applied on the complete feature preferences; they provide a global view. S S, is applied on the selected set of k features; it provides a focused view. S W, S R, treat all weights-ranks equally. However differences or similarities on the highest weighted-ranked features should be emphasized. S S focuses on what is most important, i.e. the selected features. As a result estimates of stability S W, S R > S S (although it does not make sense to compare their values). In practice Feature subsets are more interesting then Rankings which in turn are more interesting than Weightings-Scorings. S R, S S, can be compared meaningfully among different algorithms, S W cannot due to differences in scales-intervals of weights-scores. So it seems that, at least for the moment, the winner is S S (in the next few slides the feature subset size will be fixed at 10).

Stability of Feature Selection Algorithms 10 Stability behavior of FS algorithms IG CHI SYM ovarian prostate stroke leukemia nervous colon alt disease function structure subcell S.W S.R S.S SVMRFE SVMONE ReliefF S.W S.R S.S S.W S.R S.S S.W S.R S.S S.W S.R S.S S.W S.R S.S

Stability of Feature Selection Algorithms 11 Stability behavior of FS algorithms, averaged per dataset category IG CHI SYM proteomics genomics text S.W S.R S.S SVMRFE SVMONE ReliefF S.W S.R S.S S.W S.R S.S S.W S.R S.S S.W S.R S.S S.W S.R S.S

Stability of Feature Selection Algorithms 12 Stability performance of feature selection algorithms, Comments Stability is not only algorithm dependent but also, and probably mainly, dataset dependent. Strong singal provided that the algorithm is able to capture it will give stable models. Stability behavior of the three univariate FS algorithms is indistinguishable. Poor ranking behavior, S R, of univariate algorithms on proteomics and text is explained by the descritization process, which converts many attributes to single value non-informative attributes. Their good ranking behavior on the proteomics datasets should be seen with caution, heavily affected by the ovarian dataset known to have quality problems. The univariate FS algorithms produce the most stable feature subsets, S S, on the text mining data. SVMRFE does not produce weights this is why we have no performance bars on S W, moreover it does not complete execution on the text mining datasets due to their large size. SVMRFE is more unstable with respect to SVMONE on feature subset selection, S S, can be explained by the recursive nature of SVMRFE. ReliefF seems to produce more stable feature subsets, S S, than SVMONE and SVMRFE.

Stability of Feature Selection Algorithms 13 Visualizing stability performance, S S, and the feature models SVMRFE X-axis=features, Y-axis= Two consecutive horizontal lines contain the ten inner cv- of a single outer cv-fold. A point indicates that the corresponding feature was selected. The more complete vertical lines (i.e. same feature selected among different ) the more stable the algorithm. Clear picture of stability behavior and also of which features are important. 500 1000 1500 2000

Stability of Feature Selection Algorithms 14 Visualizing stability performance, Alt, text dataset, high stability CHI RELIEF SVMONE 500 1000 1500 2000 IG SYM 500 1000 1500 2000 SVMRFE 500 1000 1500 2000 500 1000 1500 2000 500 1000 1500 2000 500 1000 1500 2000

Stability of Feature Selection Algorithms 15 Visualizing stability performance, Central Nervous, genomics dataset, low stability CHI RELIEF SVMONE 1000 2000 3000 4000 5000 6000 7000 IG SYM 1000 2000 3000 4000 5000 6000 7000 SVMRFE 1000 2000 3000 4000 5000 6000 7000 1000 2000 3000 4000 5000 6000 7000 1000 2000 3000 4000 5000 6000 7000 1000 2000 3000 4000 5000 6000 7000

Stability of Feature Selection Algorithms 16 Visualizing stability performance, Prostate, proteomics dataset, average stability CHI RELIEF SVMONE 500 1000 1500 2000 IG SYM 500 1000 1500 2000 SVMRFE 500 1000 1500 2000 500 1000 1500 2000 500 1000 1500 2000 500 1000 1500 2000

Stability of Feature Selection Algorithms 17 A closer look on feature subset stability, S S Remember that S S is given by S S (s, s ) = 1 s + s 2 s s s + s s s. Note that as the number of selected features increases s S and s S then s s S and S S 1 the feature subset stability increases trivially. As a consequence the stability of a random feature selector will increase as the number of selected features approaches S We will create the complete S S stability profile for s = s = 10 step=5 to= S for all feature selection algorithms, including a random one as the baseline.

Stability of Feature Selection Algorithms 18 Stability profiles with S S, text mining datasets S S 0.0 0.2 0.4 0.6 0.8 1.0 CHI IG RANDOM RELIEF SVMONE SVMRFE SYM S S S S 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 structure 0 500 1000 1500 2000 Number of Features subcell 0 1000 2000 3000 4000 Number of Features S S 0.0 0.2 0.4 0.6 0.8 1.0 0 500 1000 1500 2000 0 500 1000 1500 2000 0 500 1000 1500 2000 2500 alt Number of Features disease Number of Features function Number of Features

Stability of Feature Selection Algorithms 19 Stability profiles with S S, genomics and proteomics datasets S S 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 S S 0.0 0.2 0.4 0.6 0.8 1.0 Stroke 0 1000 2000 3000 4000 5000 Number of Features Prostate 0 500 1000 1500 2000 Number of Features Ovarian 0 200 400 600 800 Number of Features Nervous 0 1000 2000 3000 4000 5000 6000 7000 Number of Features Colon 0 500 1000 1500 2000 Number of Features Leukemia 0 1000 2000 3000 4000 5000 6000 7000 Number of Features

Stability of Feature Selection Algorithms 20 Stability profiles with S S, comments Univariate algorithms have indistinguishable profiles. Univariate algorithms exhibit a dramatic drop in stability indicating random addition of features, explained by the discretization process. SVMRFE and SVMONE have very similar profiles and diverge on low feature set cardinalities, reasonable if one considers the similarities of the two methods. ReliefF has one of the most stable behaviors for large range of feature subset cardinalities. All algorithms reach a knot point after which stability grows slowly only because of the increase in s the number of selected features. The knot probably indicates inclusion of the most robust-stable features. After that features are included randomly. Could provide the basis for determining the optimal s.

Stability of Feature Selection Algorithms 21 Stability and Classification performance Stability alone is not enough In real world applications selection of algorithms is guided by classification performance We propose the following selection strategy: Couple the feature selection algorithms with a classification algorithm Estimate stability and classification performance Among the combination that achieve top classification performance and which performances is not significantly different choose the one with the bigger stability.

Stability of Feature Selection Algorithms 22 Stability and Classification performance (Example) Stroke N IG Relief SVM SVMRFE 10 1.5-32.22-0.1847 1.5-30.29-0.3410 1.0-37.02-0.2721 2.0-26.45-0.1678 20 1.0-31.73-0.2612 1.0-28.85-0.3670 1.0-35.10-0.3101 3.0-21.64-0.1679 30 1.5-27.89-0.2944 1.5-27.41-0.3830 1.5-28.37-0.3390 1.5-23.56-0.1802 40 1.5-29.81-0.3261 1.5-25.97-0.3887 1.5-25.00-0.3583 1.5-25.49-0.1886 50 1.5-27.89-0.3576 1.5-28.37-0.4013 1.5-26.45-0.3801 1.5-25.49-0.1997 Ovarian N IG Relief SVM SVMRFE 10 1.0-10.28-0.4948 1.0-10.28-0.7296 1.0-07.11-0.5965 3.0-01.19-0.4680 20 1.0-05.53-0.6111 1.0-05.93-0.6933 1.5-03.95-0.5897 2.5-01.19-0.4749 30 0.0-04.74-0.6567 2.0-01.58-0.6966 2.0-01.19-0.5631 2.0-00.40-0.4498 40 0.5-03.16-0.7011 1.5-01.58-0.7080 2.0-00.40-0.5682 2.0-00.40-0.4401 50 1.5-02.77-0.7496 1.5-01.58-0.7368 1.5-00.40-0.5825 1.5-00.40-0.4473 Prostate N IG Relief SVM SVMRFE 10 1.0-18.64-0.4073 1.0-18.95-0.5842 1.0-18.02-0.5308 3.0-13.05-0.4417 20 1.0-17.71-0.4299 1.0-17.09-0.6044 1.0-16.46-0.5131 3.0-11.50-0.4006 30 1.0-16.46-0.4639 1.0-15.84-0.6170 1.0-14.91-0.5193 3.0-10.87-0.3786 40 1.0-16.15-0.5044 1.0-14.91-0.6214 1.0-13.36-0.5280 3.0-09.01-0.3848 50 1.0-14.60-0.5374 1.0-13.36-0.6304 1.0-13.05-0.5343 3.0-09.32-0.3890 Classification algorithm linear SVM. Cells contain: Class. Perf. Ranking Error Stability (S S ). Class. Perf. Ranking is the sum of significant wins plus 0.5 for each tie. In red all cases with top and not significantly different classification performance.

Stability of Feature Selection Algorithms 23 Stability and Classification performance (Example) Stroke N IG Relief SVM SVMRFE 10 1.5-32.22-0.1847 1.5-30.29-0.3410 1.0-37.02-0.2721 2.0-26.45-0.1678 20 1.0-31.73-0.2612 1.0-28.85-0.3670 1.0-35.10-0.3101 3.0-21.64-0.1679 30 1.5-27.89-0.2944 1.5-27.41-0.3830 1.5-28.37-0.3390 1.5-23.56-0.1802 40 1.5-29.81-0.3261 1.5-25.97-0.3887 1.5-25.00-0.3583 1.5-25.49-0.1886 50 1.5-27.89-0.3576 1.5-28.37-0.4013 1.5-26.45-0.3801 1.5-25.49-0.1997 Ovarian N IG Relief SVM SVMRFE 10 1.0-10.28-0.4948 1.0-10.28-0.7296 1.0-07.11-0.5965 3.0-01.19-0.4680 20 1.0-05.53-0.6111 1.0-05.93-0.6933 1.5-03.95-0.5897 2.5-01.19-0.4749 30 0.0-04.74-0.6567 2.0-01.58-0.6966 2.0-01.19-0.5631 2.0-00.40-0.4498 40 0.5-03.16-0.7011 1.5-01.58-0.7080 2.0-00.40-0.5682 2.0-00.40-0.4401 50 1.5-02.77-0.7496 1.5-01.58-0.7368 1.5-00.40-0.5825 1.5-00.40-0.4473 Prostate N IG Relief SVM SVMRFE 10 1.0-18.64-0.4073 1.0-18.95-0.5842 1.0-18.02-0.5308 3.0-13.05-0.4417 20 1.0-17.71-0.4299 1.0-17.09-0.6044 1.0-16.46-0.5131 3.0-11.50-0.4006 30 1.0-16.46-0.4639 1.0-15.84-0.6170 1.0-14.91-0.5193 3.0-10.87-0.3786 40 1.0-16.15-0.5044 1.0-14.91-0.6214 1.0-13.36-0.5280 3.0-09.01-0.3848 50 1.0-14.60-0.5374 1.0-13.36-0.6304 1.0-13.05-0.5343 3.0-09.32-0.3890 Classification algorithm linear SVM. Cells contain: Class. Perf. Ranking Error Stability (S S ). Class. Perf. Ranking is the sum of significant wins plus 0.5 for each tie. In red all cases with top and not significantly different classification performance. So among the top classification performers choose the most stable

Stability of Feature Selection Algorithms 24 Stability, Classification Performance and Redundancy Stroke N IG Relief SVM SVMRFE 10 1.5-32.22-0.1847 1.5-30.29-0.3410 1.0-37.02-0.2721 2.0-26.45-0.1678 20 1.0-31.73-0.2612 1.0-28.85-0.3670 1.0-35.10-0.3101 3.0-21.64-0.1679 30 1.5-27.89-0.2944 1.5-27.41-0.3830 1.5-28.37-0.3390 1.5-23.56-0.1802 40 1.5-29.81-0.3261 1.5-25.97-0.3887 1.5-25.00-0.3583 1.5-25.49-0.1886 50 1.5-27.89-0.3576 1.5-28.37-0.4013 1.5-26.45-0.3801 1.5-25.49-0.1997 Ovarian N IG Relief SVM SVMRFE 10 1.0-10.28-0.4948 1.0-10.28-0.7296 1.0-07.11-0.5965 3.0-01.19-0.4680 20 1.0-05.53-0.6111 1.0-05.93-0.6933 1.5-03.95-0.5897 2.5-01.19-0.4749 30 0.0-04.74-0.6567 2.0-01.58-0.6966 2.0-01.19-0.5631 2.0-00.40-0.4498 40 0.5-03.16-0.7011 1.5-01.58-0.7080 2.0-00.40-0.5682 2.0-00.40-0.4401 50 1.5-02.77-0.7496 1.5-01.58-0.7368 1.5-00.40-0.5825 1.5-00.40-0.4473 Prostate N IG Relief SVM SVMRFE 10 1.0-18.64-0.4073 1.0-18.95-0.5842 1.0-18.02-0.5308 3.0-13.05-0.4417 20 1.0-17.71-0.4299 1.0-17.09-0.6044 1.0-16.46-0.5131 3.0-11.50-0.4006 30 1.0-16.46-0.4639 1.0-15.84-0.6170 1.0-14.91-0.5193 3.0-10.87-0.3786 40 1.0-16.15-0.5044 1.0-14.91-0.6214 1.0-13.36-0.5280 3.0-09.01-0.3848 50 1.0-14.60-0.5374 1.0-13.36-0.6304 1.0-13.05-0.5343 3.0-09.32-0.3890 low or lower stability is not related to low classification performance. High classification performance does not imply high stability. High classification performance + low stability feature redundancy

Stability of Feature Selection Algorithms 25 Summary Introduce the concept of stability of feature selection algorithms. Examine the stability profiles of different feature selection algorithms. Provide an eloquent visualization of stability performance and feature resilience. Couple classification performance based algorithm selection with stability based algorithm selection.

Stability of Feature Selection Algorithms 26 Future Work back then... Examine how we can use stability to choose the optimal number of features. Examine how we can use stability to detect feature redundancies. Combine feature sets to increase stability (the analogue of ensemble methods in classification algorithms).

Stability of Feature Selection Algorithms 27 Related Work Different stability measures have been proposed Efforts to increase feature selection stability

Stability of Feature Selection Algorithms 28 Sources of Feature Selection Instability Small datasets which result in statistics that are sensitive to data perturbations Low signal in the features Redundant feature sets that can lead to different models of equal predictive power The algorithms themselves: algorithms that depend on initial conditions, multiple local optima more aggressive feature selection algorithms result in more unstable models, e.g. l 1 norm regularization is more unstable compared to l 2. With l 1 out of a set of relevant but redundant features only one will be selected the algorithm does not care which one that will be, in fact in the case of redundant and relevant features there is not a unique solution; with l 2 all of them will be but with lower weights.

Stability of Feature Selection Algorithms 29 Increasing Feature Selection Stability Different methods have been proposed to increase stability Managing feature redundancy in a principled manner: Feature clustering methods, (Yu et al., 2008), which cluster features to groups of similar-redundant features and then apply feature selection in the clustering results. Elastic net, (Zou & Hastie, 2005), which combines l 1 and l 2 regularization that has as result to select all together a group of redundant features or none of them in what is known as the grouping effect. Strict convexity has the grouping effect. Feature model agregation, e.g. ensemble methods, (Sayes et al., 2008; Loscalzo et al., 2009), which agregate feature models obtained from different subsamples of the training set to obtain a single stable feature model. Increasing stability especially if this is done outside the FS algorithm using, e.g. ensemble methods, does not mean we can expect accuracy gains.

Stability of Feature Selection Algorithms 30 Our current work Define more appropriate stability measures Explore the effect of sample size in stability Place the concept of feature stability in a broader concept of model similarity

Stability of Feature Selection Algorithms 31 References Kalousis, A., Prados, J., & Hilario, M. (2007). Stability of feature selection algorithms: a study on high dimensional spaces. Knowledge and Information Systems, 12, 95 116. Loscalzo, S., Yu, L., & Ding, C. (2009). Consensus group based stable feature selection. Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD-2009. Sayes, Y., Abeel, T., & Peer, Y. (2008). Robust feature selection using ensemble feature selection techniques. Proceedings of the European conference on Machine Learning and Knowledge Discovery in Databases. Yu, L., Ding, C., & Loscalzo, S. (2008). Stable feature selection via dense feature groups. Proceedings of the KDD 2008. Zou, H., & Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society B, 67.