ISSN ICIRET-2014

Similar documents
Published by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 1

A Survey on Clustered Feature Selection Algorithms for High Dimensional Data

International Journal of Scientific & Engineering Research, Volume 5, Issue 7, July ISSN

A FAST CLUSTERING-BASED FEATURE SUBSET SELECTION ALGORITHM

BENCHMARKING ATTRIBUTE SELECTION TECHNIQUES FOR MICROARRAY DATA

Feature Subset Selection Algorithm for Elevated Dimensional Data By using Fast Cluster

Filter methods for feature selection. A comparative study

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Efficiently Handling Feature Redundancy in High-Dimensional Data

Feature Subset Selection Algorithms for Irrelevant Removal Using Minimum Spanning Tree Construction

CBFAST- Efficient Clustering Based Extended Fast Feature Subset Selection Algorithm for High Dimensional Data

Redundancy Based Feature Selection for Microarray Data

Graph Clustering and Feature Selection for High Dimensional Data

CLUSTERING-BASED FEATURE SUBSET SELECTION ALGORITHM USING FAST

Correlation Based Feature Selection with Irrelevant Feature Removal

REMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD

Chapter 12 Feature Selection

Searching for Interacting Features

Supervised Variable Clustering for Classification of NIR Spectra

AN ENSEMBLE OF FILTERS AND WRAPPERS FOR MICROARRAY DATA CLASSIFICATION

MANY factors affect the success of data mining algorithms

Clustering-Based Feature Selection Framework for Microarray Data

A Fast Clustering-Based Feature Subset Selection Algorithm for High-Dimensional Data

Forward Feature Selection Using Residual Mutual Information

Analysis of Feature Selection Techniques: A Data Mining Approach

An Empirical Study on feature selection for Data Classification

Weighted Heuristic Ensemble of Filters

Removal of Redundant and Irrelevant attributes for high Dimensional data using Clustering Approach

Feature Ranking in Intrusion Detection Dataset using Combination of Filtering Methods

Prognosis of Lung Cancer Using Data Mining Techniques

University of Florida CISE department Gator Engineering. Data Preprocessing. Dr. Sanjay Ranka

A study of classification algorithms using Rapidminer

Data Preprocessing. Data Preprocessing

Contents. Preface to the Second Edition

Feature Selection for Supervised Classification: A Kolmogorov- Smirnov Class Correlation-Based Filter

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Hybrid Feature Selection for Modeling Intrusion Detection Systems

Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

Feature clustering and mutual information for the selection of variables in spectral data

Cluster based boosting for high dimensional data

An Effective Feature Selection Approach Using the Hybrid Filter Wrapper

Best First and Greedy Search Based CFS and Naïve Bayes Algorithms for Hepatitis Diagnosis

A Detailed Analysis on NSL-KDD Dataset Using Various Machine Learning Techniques for Intrusion Detection

An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm

Hybrid Correlation and Causal Feature Selection for Ensemble Classifiers

Scalable Feature Selection for Multi-class Problems

Information theory methods for feature selection

Survey on Rough Set Feature Selection Using Evolutionary Algorithm

Noise-based Feature Perturbation as a Selection Method for Microarray Data

Stability of Feature Selection Algorithms

9. Conclusions. 9.1 Definition KDD

Effect of Principle Component Analysis and Support Vector Machine in Software Fault Prediction

Fuzzy Entropy based feature selection for classification of hyperspectral data

Keywords Soil classification; Neural networks; fuzzy logic; fuzzy association rule mining; Relief-F

An Efficient Clustering for Crime Analysis

Feature Selection Algorithm with Discretization and PSO Search Methods for Continuous Attributes

A Selective Sampling Approach to Active Feature Selection

Index Terms:- Document classification, document clustering, similarity measure, accuracy, classifiers, clustering algorithms.

A Novel Feature Selection Framework for Automatic Web Page Classification

Semi-Supervised Clustering with Partial Background Information

WEIGHTED K NEAREST NEIGHBOR CLASSIFICATION ON FEATURE PROJECTIONS 1

IJREAT International Journal of Research in Engineering & Advanced Technology, Volume 1, Issue 5, Oct-Nov, ISSN:

Feature Selection Using Modified-MCA Based Scoring Metric for Classification

A Survey on Pre-processing and Post-processing Techniques in Data Mining

Feature Selection and Classification for Small Gene Sets

Collaborative Filtering using a Spreading Activation Approach

A Heart Disease Risk Prediction System Based On Novel Technique Stratified Sampling

Feature Selection Technique to Improve Performance Prediction in a Wafer Fabrication Process

Review of feature selection techniques in bioinformatics by Yvan Saeys, Iñaki Inza and Pedro Larrañaga.

Using Association Rules for Better Treatment of Missing Values

A Monotonic Sequence and Subsequence Approach in Missing Data Statistical Analysis

Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data

An Empirical Study of Lazy Multilabel Classification Algorithms

Study on Classifiers using Genetic Algorithm and Class based Rules Generation

International Journal of Data Mining & Knowledge Management Process (IJDKP) Vol.7, No.3, May Dr.Zakea Il-Agure and Mr.Hicham Noureddine Itani

Feature Selection with Decision Tree Criterion

International Journal of Advance Research in Computer Science and Management Studies

Data Cleaning and Prototyping Using K-Means to Enhance Classification Accuracy

Feature Selection in Knowledge Discovery

Using Google s PageRank Algorithm to Identify Important Attributes of Genes

SCENARIO BASED ADAPTIVE PREPROCESSING FOR STREAM DATA USING SVM CLASSIFIER

International Journal of Advanced Research in Computer Science and Software Engineering

A Weighted Majority Voting based on Normalized Mutual Information for Cluster Analysis

Conditional Mutual Information Based Feature Selection for Classification Task

Data Preparation. Nikola Milikić. Jelena Jovanović

Online Streaming Feature Selection

A Comparison of Classification Accuracy Achieved with Wrappers, Filters and PCA

ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A COMPARATIVE ANALYSIS

Machine Learning Techniques for Data Mining

Categorization of Sequential Data using Associative Classifiers

INFORMATION-THEORETIC OUTLIER DETECTION FOR LARGE-SCALE CATEGORICAL DATA

K-Nearest-Neighbours with a Novel Similarity Measure for Intrusion Detection

Slides for Data Mining by I. H. Witten and E. Frank

Discovery of Agricultural Patterns Using Parallel Hybrid Clustering Paradigm

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

FEATURE SELECTION TECHNIQUES

Chapter 1, Introduction

A STUDY OF SOME DATA MINING CLASSIFICATION TECHNIQUES

A Feature Selection Method to Handle Imbalanced Data in Text Classification

Transcription:

Feature Subset Collection in High Dimensional Data: Fast Technique V.M.Suresh 1, P.Vennila 2 1 Assistant Professor, 2 PG Scholars 1 Department Of Information Technology, 2 Department Of PG-Computer science and Engineering 1, 2 E.G.S.Pillay Engineering College 1, 2 Nagapattinam, Tamilnadu, India 1 vmsureshme@gmail.com, 2 vennilapremi@gmail.com Abstract Feature selection involves identifying a subset of the most useful features that produces compatible results as the original entire set of features. A feature selection algorithm may be evaluated from both the efficiency and effectiveness. A fast clustering-based feature selection algorithm, FAST works in two steps. In the first step, features are divided into clusters by using graph-theoretic clustering methods. In the second step, the most representative feature that is strongly related to target classes is selected from each cluster to form a subset of features. To ensure the efficiency of FAST, Adopt the efficient minimum-spanning tree clustering method feature selection algorithms, namely, FCBF, ReliefF, CFS, Consist, and FOCUS-SF, with respect to four types of wellknown classifiers, namely, the probability-based Naive Bayes,the tree-based C4.5, the instance-based IB1, and the rule-based RIPPER.real-world high dimensional image, microarray, and text data, demonstrate that FAST not only produces smaller subsets of features but also improves the performances of the four types of classifiers. Index Terms Feature subset selection, filter method, feature clustering, graph-based clustering 1. INTRODUCTION Feature subset selection is an effective way for reducing dimensionality, removing irrelevant data, increasing learning accuracy, and improving result comprehensibility. They can be divided into four broad categories: the Embedded, Wrapper, Filter, and Hybrid approaches. Embedded method incorporate feature selection as a part of the training process and are usually specific to given learning algorithms, and more efficient than the other three categories. Traditional machine learning algorithms like decision trees or artificial neural networks are examples of embedded approaches. Wrapper method: This methods use the predictive accuracy of a predetermined learning algorithm to determine the goodness of the selected subsets, the accuracy of the learning algorithms is usually high.the generality of the selected features is limited and the computational complexity is large filter method. This methods are independent of learning algorithms, with good generality. Their computational complexity is low, but the accuracy of the learning algorithms is not guaranteed. filter method to reduce searching space. The filter methods are usually a good choice when then number of features is very large. Thus, we will focus on the filter method in this paper, hybrid method This methods area combination of filter and wrapper methods..in cluster analysis, graph-theoretic methods have been well studied and used in many applications. The general graph-theoretic clustering is simple. we apply graph theoretic clustering methods to features. we adopt the minimum spanning tree (MST) based clustering algorithm propose a Fast clustering-based feature Selection algorithm (FAST).The FAST algorithm works in two steps. In the first step, features are divided into clusters by using graphtheoretic clustering methods. In the second step, the most representative feature that is strongly related to target classes is selected from each cluster to form the final subset of features the clustering-based strategy of FAST has a high probability of producing a subset of useful and independent features. The proposed feature subset selection algorithm FAST was tested image, microarray, and text data sets. the proposed algorithm not only reduces the number of features, but also improves the performances of the four well-known different types of classifiers OBJECTIVE OF THE PAPER i) Among high dimensional data how are selecting relevant features, removal redundant features, irrelevant data. ii) Identify the best subsets by using feature selection. iii) Ensure the efficiency of FAST, adopt the efficient minimum spanning tree method. iv) Dimensionality reduction techniques can be categorized mainly into feature extraction and feature selection. E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 154

2. RELATED WORK Feature subset selection can be viewed as the process of identifying and removing as many irrelevant and redundant features as possible (i) Irrelevant features do not contribute to the predictive accuracy (ii) Redundant features do not redound to getting a better predictor for that they provide many feature subset selection algorithms Many algorithms effectively eliminate irrelevant features but fail to handle redundant feature. A well known example is Relief is ineffective at removing redundant features as two predictive but highly correlated features are likely both to be highly weighted Relief-F extends Relief. RELIEF ALGORITHM ReliefF[13]searches for nearest neighbors of instances of different classes and weights features according to how well they differentiate instances of different classes. The other three feature selection algorithms are based on subset evaluation.this method to work with noisy and incomplete data sets and to deal with multi-class problems, but still cannot identify redundant features also affect the speed and accuracy of learning algorithms. CORRELATION BASED FEATURE SELECTION CFS[5] exploits best-first search based on the evaluation of a subset that contains features highly correlated with the tar- get concept, yet uncorrelated with each other. is achieved by the hypothesis that a good feature subset is one that contains features highly correlated with the target, yet uncorrelated with each other. FAST CORRELATION BASED FILTER SOLUTION FCBF[15],[17] is a fast filter method which can identify relevant features as well as redundancy among relevant features without pairwise correlation analysis., we intro- duce a novel concept, predominant correlation, and propose a fast filter method which can identify relevant features as well as redundancy among relevant features without pairwise correlation analysis. The efficiency and effectiveness of our method is demonstrated through extensive comparisons with other methods using realworld data of high dimensionality. THE CONSIST METHOD The Consist method[3] searches for the minimal subset that separates classes as consistently as the full set can under best-first search strategy. FOCUS-SF FOCUS-SF[1] is a variation of FOCUS. FOCUS has the same evaluation strategy as Consist, but it examines all subsets of features. Considering the time efficiency, FOUCS-SF replaces exhaustive search in FOCUS with sequential forward selection. HIERARCHICAL CLUSTERING Hierarchical clustering has been adopted in word selection in the context of text classification [2],[4].Hierarchical clustering also has been used to select features on spectral data. Van Dijk and Van Hullefor[14] proposed a hybrid filter/wrapper feature subset selection algorithm for regression. Krier et al. [11] presented methodology combining hierarchical constrained clustering of spectral variables and selection of clusters by mutual information. 3. FEATURE SUBSET SELECTION ALGORITHM A) FRAMEWORK AND DEFINITIONS Irrelevant features, along with redundant features, severely affect the accuracy of the learning machines [7], [9]. Good feature subsets contain features highly correlated with (predictive of) the class, yet uncorrelated with (not predictive of) each other. [6] Upload datasets Noise Removal FAST algorithm Preprocessing Irrelevant features removal MST construction Tree partition Redundant data removal Feature Subset Performance Evaluation Fig. No.1: Feature subset selection algorithm John et al. [8] presented a definition of relevant features. Suppose to be the full set of features, # be a feature, = { } and. E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 155

Relevant feature: is relevant to the target concept if and only if there exists some, and, such that, for probability ( =, = ) > 0, ( = =, = ) = ( = = ). Markov blanket: Given a feature, let ( ), is said to be a Markov blanket for if and only if ( { },, )= ( { }, ). Redundant feature: Let be a set of features, a feature in is redundant if and only if it has a Markov Blanket. The symmetric uncertainty ( ) [12] is derived from the mutual information by normalizing it to the entropies of feature values or feature values and target classes. The symmetric uncertainty is defined as follows (, )=2 ( ) ( )+ ( ) H(X)= ( )log2 ( ). ( )= ( ) ( ) = ( ) ( ). ( )= ( ) ( )log2 ( ). T-Relevance: The relevance between the feature and the target concept and, and denoted by (, ). F-Correlation: The correlation between any pair of features is referred to as the T-Relevance of and (, = ) is called the F-Correlation of and, and denoted by (, ). F-Redundancy: Let = { 1, 2,...,,..., < } be a cluster of features. if, (, ) (, ) (, ) > (, ) is always corrected for each. R-Feature: A feature = { 1, 2,..., } ( < ) is a representative feature of the cluster ( i.e. is a R- Feature ) if and only if, = argmax (, ). B) ALGORITHM AND ANALYSIS The proposed FAST algorithm logically consists of tree steps: (i) removing irrelevant features, (ii) constructing a MST from relative ones, and (iii) Partitioning the MST and selecting representative features. 4. PROPOSED SYSTEM A novel algorithm which can efficiently and effectively deal with both irrelevant and redundant features. FAST algorithm, it involves (i) the construction of the minimum spanning tree (MST) from a weighted complete graph; (ii) the partitioning of the MST into a forest with each tree representing a cluster and (iii) the selection of representative features from the clusters. i) The best proportion of selected features, ii) The best runtime, iii) The best classification accuracy Fast Algorithm: inputs: D( 1, 2,...,, ) - the given data set - the T-Relevance threshold. output: S - selected feature subset. Part 1 : Irrelevant Feature Removal 1) for i = 1 to m do 2) T-Relevance = SU (, ) 3) if T-Relevance > then 4) S = S { }; Part 2: Minimum Spanning Tree Construction 5) G = NULL; 6) for each pair of features {, } S do 7) F-Correlation = SU (, ) 8) / h F- Correlation h h h ; 9) minspantree = Prim (G); Part 3 : TreePartition and Representative Feature Selection 10) Forest = minspantree 11) for each edge Forestdo 12) if SU(, ) < SU(, ) SU(, )< SU(, ) then 13) Forest = Forest 14) S = 15) for each tree Forestdo 16) = argmax SU(, ) 17) S = S { }; 18) return S Time complexity analysis: The first part of the algorithm has a linear time complexity ( ).The second part of the algorithm firstly constructs a complete graph from relevant features and the complexity is ( 2), and then generates a MST from the graph using Prim algorithm whose time complexity is ( 2). The third part partitions the MST and chooses the representative features with the complexity of ( ).when =. However, is heuristically set to be lg in the implementation of FAST. So the complexity is ( lg2 ), the time complexity of FAST deviates from ( 2). Thus, on high dimensional data, the time complexity of FAST is far more less than ( 2).The proposed algorithm is compared with five different types of representative feature selection algorithms. They are (i) FCBF [16], [18], (ii) ReliefF [14], (iii) CFS [5], (iv) Consist [3], and (v) FOCUS- SF [1],set the relevance threshold to be the value of the E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 156

/log h ranked feature for each data set. For proposed FAST algorithm, we heuristically set to be the value of the lg h ranked feature for each data set. 5. IMPLEMENTATION Feature subset selection algorithm,obtain M N feature subsets Subset and the corresponding runtime Time with each data set. Average Subset and Time, we obtain the number of selected features further the proportion of selected features and the corresponding runtime for each feature selection algorithm on each data set. For each classification algorithm, we obtain M N classification Accuracy for each feature selection algorithm and each data set. Procedure Experimental Process: 1) M = 5, N = 101 2) DATA = { 1, 2,..., 35} 3) Learners = {NB, C4.5, IB1, RIPPER} 4) FeatureSelectors = {FAST, FCBF, ReliefF, CFS, Consist, FOCUS-SF} 5) for each data DATA do 6) for each times [1, M] 7) randomize instance-order for data 8) generate N bins from the randomized data 9) for each fold [1, N] do 10) TestData = bin[fold] 11) TrainingData = data TestData 12) for each selector Feature Selectors do 13) (Subset,Time)=selector(TrainingData) 14) TrainingData = se 15) lect Subset from TrainingData 16) TestData = select Subset from TestData 17) for each learner Learners do 18) classifier = learner(trainingdata ) 19) Accuracy = apply classifier to TestData 18 Results and Analysis: The experimental results in terms of the proportion of selected features, the time to obtain the feature subset, the classification accuracy, and the Win/Draw/Loss record. TABLE 1: Summary of the 10 benchmark data sets Data ID Data name F I T Domain 1 Chess 37 31 96 Text 2 Mfeat-fourier 77 2000 10 Image face 3 Coil2000 86 98 22 Text 4 Elephant 232 1391 2 5 Fgs-nowe 320 265 2 Image face 6 Colon 2001 62 2 Microarray bio 7 Arrhythmia 280 452 16 Microarray bio 8 Fbis.wc 2001 2463 17 Text 9 Ar10p 2401 130 10 Image face 10 Pie10p 2421 210 10 Image face TABLE 2:proportion of selected features of the six feature selection algorithms Data set FAST FCBF CFS Relief Consist FOCUS- SF Chess 16.22 21.62 10.81 62.16 81.08 18.92 Mfeat-fourier 19.48 49.35 24.68 98.70 15.58 15.58 Coil2000 3.49 8.14 11.63 50.00 37.21 1.16Text Elephant 0.86 8.88 5.60 6.03 0.86 0.86 Fgs-nowe 0.31 2.19 5.63 26.66 4.69 4.69 Colon 0.30 0.75 1.35 39.13 0.30 0.30 Arrhythmia 2.50 4.64 9.29 30.00 8.93 8.93 Fbis.wc 0.80 1.45 2.30 0.95 1.75 1.75 Ar10p 0.21 1.04 2.12 62.89 0.29 0.29 Pie10p 1.07 1.98 2.52 91.00 0.25 0.25 Average(image) 3.59 10.04 6.68 79.85 47.56 3.48 Average(microarray) 0.71 2.34 2.50 52.92 0.91 0.91 Average(text) 2.05 3.25 2.64 10.87 11.46 2.53 Average 1.82 4.27 3.42 42.54 5.44 2.06 Win/draw/loss - 33/0/2 31/0/2 29/1/5 20/2/13 19/2/13 E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 157

TABLE 3:Runtime of the six feature selection algorithms Data set FAST FCBF CFS Relief Consist FOCUS- SF Chess 105 60 352 12660 1999 653 Mfeat-fourier 1472 716 938 13918 3227 660 Coil2000 866 875 1483 304162 53850 1281 Elephant 783 312 905 20991 2439 1098 Fgs-nowe 977 97 736 1072 1360 1032 Colon 166 148 12249 744 1624 960 Arrhythmia 110 115 821 3684 3492 2940 Fbis.wc 14761 16207 66058 79527 579376 479651 Ar10p 706 458 57319 3874 3568 2083 Pie10p 678 1223 77579 7636 4149 2910 Average(image) 1520 4090 905678 3456 2547 2865 Average(microarray) 7543 3567 678945 347256 5907 6834 Average(text) 3792 4617 5902 4167 489032 5734 Average 4532 367832 56743 43167 4321 6743 Win/draw/loss - 33/0/2 31/0/2 29/1/5 20/2/13 34/0/13 TABLE 4:Accuracy of the navie bayes of the six feature selection algorithms Data set FAST FCBF CFS Relief Consist FOCUS- Full set SF Chess 92.92 92.12 90.13 88.56 89.50 94.34 87.68 Mfeat-fourier 19.48 49.35 24.68 98.70 15.58 15.58 76.07 Coil2000 3.49 8.14 11.63 50.00 37.21 1.16 78.04 Elephant 0.86 8.88 5.60 6.03 0.86 0.86 82.34 Fgs-nowe 0.31 2.19 5.63 26.66 4.69 4.69 63.06 Colon 0.30 0.75 1.35 39.13 0.30 0.30 65.61 Arrhythmia 2.50 4.64 9.29 30.00 8.93 8.93 56.33 Fbis.wc 0.80 1.45 2.30 0.95 1.75 1.75 61.89 Ar10p 0.21 1.04 2.12 62.89 0.29 0.29 72.62 Pie10p 1.07 1.98 2.52 91.00 0.25 0.25 90.67 Average(image) 3.59 10.04 6.68 79.85 47.56 3.48 45.46 Average(microarray) 0.71 2.34 2.50 52.92 0.91 0.91 45.23 Average(text) 2.05 3.25 2.64 10.87 11.46 2.53 67.45 Average 1.82 4.27 3.42 42.54 5.44 2.06 78.44 Win/draw/loss - 33/0/2 31/0/2 29/1/5 20/2/13 19/2/13 23/2/13 TABLE 5 : Rank of the six feature selection algorithms under different types of data Image Data Microarray Data FAST FCBF CFS ReliefF Consist FOCUS SF FAST FCBF CFS ReliefF Consist FOCUS SF NB 3 1 2 4 5 5 1 3 2 6 4 4 C4 2 3 1 4 5 5 1 3 2 6 4 4 IB1 4 2 1 3 5 5 1 3 2 4 5 5 RIPPER 1 2 1 6 3 3 1 4 5 6 2 2 Sum 10 8 5 17 18 18 4 13 11 22 15 15 Rank 3 2 1 4 5 5 1 3 2 6 4 4 E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 158

Image Data Microarray Data FAST FCBF CFS ReliefF Consist FOCUS SF FAS FCBF CFS ReliefF Consist FOCUS SF T NB 1 3 2 6 5 4 1 3 2 4 6 5 C4 1 2 6 4 5 1 1 2 4 6 5 3 IB1 3 4 1 6 2 4 2 3 1 6 4 5 RIPPER 5 3 1 6 2 4 1 4 5 6 2 3 Sum 12 12 10 22 14 13 5 12 12 22 17 16 Rank 2 2 1 6 5 4 1 2 2 6 5 4 Fig.No.2: Accuracy differences between FAST and the comparing algorithms 6. CONCLUSION AND FUTURE WORK A novel clustering-based feature subset selection algorithm for high dimensional data. The algorithm involves (i) removing irrelevant features, (ii) constructing a minimum spanning tree from relative ones, and (iii) partitioning the MST and selecting representative features. In the proposed algorithm, a cluster consists of features. Each cluster is treated as a single feature and thus dimensionality is drastically reduced. the performance of the proposed algorithm with those of the five well-known feature selection algorithms FCBF, ReliefF, CFS, Consist, and FOCUS-SF on the image, microarray, and text data.proposed algorithm obtained the best proportion of selected features, the best runtime, and the best classification accuracy for Naive Bayes, C4.5, and RIPPER, and the second best classification accuracy for IB1. The Win/Draw/Loss records confirmed the conclusions.fast obtains the rank of 1 for microarray data, the rank of 2 for text data, and the rank of 3 for image data. For the future work,plan to explore different types of correlation measures, and study some formal properties of feature space. REFERENCE [1] Almuallim H and Dietterich T.G.,Learning Boolean concepts the presence of many Asia conference on knowledge discovery and data -mining pp 109,2000. [2] Basker L.D.andMcCallum A.K.Distributional clustering of words for text classification.,in Proceedings of the 21 st Annual international ACMSIGIR Conference on Research and pacific irrelevant features,aritificial Intelligence,6991-20,pp 279-305,1999 [3] Dash M,Liu H and Motoda H..Consistency based feature selection,in proceedings of the Fourth Development in information Retrieval,pp 96-103,1998 [4] Dhillon I S.,Mallela S and Kumar R.,A Divisive information theoretic clustering algorithm for text classification,j.mach.learn Res.,3,pp 1265-1287,2003 [5] Hall M.A.,Correlation Based Feature Subset Selection for Machine Learning Proceedings of the Fifth IEEE international Confernce on Data Mining pp 581-554 [6] Hall M.A. and Smith L.A., Feature Selection for Machine Learning: Com- paring a Correlation- Based Filter Approach to the Wrapper, In Proceedings of the Twelfth international Florida Artificial intelligence Research Society Conference, pp 235-239, 1999. [7] Hall M.A., Correlation-Based Feature Selection for Discrete and Numeric Class Machine Learning, In Proceedings of 17th International Conference on Machine Learning, pp 359-366, 2000. [8] John G.H., Kohavi R. and Pfleger K.,Irrelevant Features and the Subset Selection Problem, In the Proceedings of the Eleventh International Conference on Machine Learning, pp 121-129, 1994. [9] Kohavi R. and John G.H., Wrappers for feature subset selection, Artif. Intell., 97(1-2), pp 273-324, 1997 [10] Koller D. and Sahami M., Toward optimal feature selection, In Proceedings of International E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 159

Conference on MachineLearning, pp 284-292, 1996. [11] Krier C., Francois D., Rossi F. and Verleysen M., Feature clustering and mutual information for the selection of variables in spectral data, In Proc European Symposium on Artificial Neural Networks Advances in Computational Intelligence and Learning, pp 157-162, 2007. [12] Press W.H., Flannery B.P., Teukolsky S.A. an Vetterling W.T., Numerical recipes in C. Cambridge University Press, Cambridge, 1988. [13] Robnik-Sikonja M. and KononenkoI.Theoretical and empirical analysis of Relief and ReliefF, Machine Learning,53, pp 23-69, 2003. [14] Van Dijk G. and Van Hulle M.M., Speeding Up the Wrapper Feature Subset Selection in Regression by Mutual Information Relevance and Redundancy Analysis, International Conference on Artificial Neural Networks, 2006 [15] Yu L. and Liu H., Feature selection for High dimensional data: a fast correlation-based filter solution, in Proceedings of 20th International Conference on Machine Leaning, 20(2), pp856-863, 2003. [16] Yu L. and Liu H., Redundancy based feature selection for micro array data, In Proceedings of the tenth ACM SIGKDD international conference on Knowledge discovery and data mining, pp 737-742, 2004. [17] Yu L. and Liu H., Efficient feature selection via analysis of relevance and redundancy, Journal of Machine Learning Research, 10(5), pp 12051224,2004. E.G.S.PILLAY ENGINEERING COLLEGE NAGAPATTINAM Page 160