A Comparative Study of Selected Classification Algorithms of Data Mining

Similar documents
Performance Evaluation of Rule Based Classification Algorithms

Ripple Down Rule learner (RIDOR) Classifier for IRIS Dataset

Dr. Prof. El-Bahlul Emhemed Fgee Supervisor, Computer Department, Libyan Academy, Libya

EVALUATING THE EFFICIENCY OF RULE TECHNIQUES FOR FILE CLASSIFICATION

International Journal of Computer Engineering and Applications, Volume XI, Issue XII, Dec. 17, ISSN

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

AGRICULTURAL SOIL LIME STATUS ANALYSIS USING DATA MINING CLASSIFICATION TECHNIQUES

ANALYSIS COMPUTER SCIENCE Discovery Science, Volume 9, Number 20, April 3, Comparative Study of Classification Algorithms Using Data Mining

Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques

Comparative Study on Classification Meta Algorithms

Study on Classifiers using Genetic Algorithm and Class based Rules Generation

Performance Analysis of Data Mining Classification Techniques

COMPARISON OF DIFFERENT CLASSIFICATION TECHNIQUES

International Journal of Computer Science Trends and Technology (IJCST) Volume 5 Issue 4, Jul Aug 2017

ISSN: (Online) Volume 3, Issue 9, September 2015 International Journal of Advance Research in Computer Science and Management Studies

AMOL MUKUND LONDHE, DR.CHELPA LINGAM

Hands on Datamining & Machine Learning with Weka

IMPLEMENTATION OF CLASSIFICATION ALGORITHMS USING WEKA NAÏVE BAYES CLASSIFIER

Classifying Twitter Data in Multiple Classes Based On Sentiment Class Labels

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6)

Correlation Based Feature Selection with Irrelevant Feature Removal

Data Mining Techniques Methods Algorithms and Tools

International Journal of Advanced Research in Computer Science and Software Engineering

Iteration Reduction K Means Clustering Algorithm

Data Mining: An experimental approach with WEKA on UCI Dataset

International Journal of Software and Web Sciences (IJSWS)

Comparative Study of Instance Based Learning and Back Propagation for Classification Problems

Data Mining with Oracle 10g using Clustering and Classification Algorithms Nhamo Mdzingwa September 25, 2005

Weka ( )

A STUDY OF SOME DATA MINING CLASSIFICATION TECHNIQUES

3 Virtual attribute subsetting

Data Mining. 3.3 Rule-Based Classification. Fall Instructor: Dr. Masoud Yaghini. Rule-Based Classification

A study of classification algorithms using Rapidminer

A Supervised Method for Multi-keyword Web Crawling on Web Forums

Estimating Missing Attribute Values Using Dynamically-Ordered Attribute Trees

Data Mining Practical Machine Learning Tools and Techniques. Slides for Chapter 6 of Data Mining by I. H. Witten and E. Frank

REMOVAL OF REDUNDANT AND IRRELEVANT DATA FROM TRAINING DATASETS USING SPEEDY FEATURE SELECTION METHOD

Data Mining Part 5. Prediction

CLUSTERING BIG DATA USING NORMALIZATION BASED k-means ALGORITHM

A New Technique to Optimize User s Browsing Session using Data Mining

An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm

Best First and Greedy Search Based CFS and Naïve Bayes Algorithms for Hepatitis Diagnosis

Comparative Study of J48, Naive Bayes and One-R Classification Technique for Credit Card Fraud Detection using WEKA

A Survey on k-means Clustering Algorithm Using Different Ranking Methods in Data Mining

The Explorer. chapter Getting started

Effect of Principle Component Analysis and Support Vector Machine in Software Fault Prediction

DATA ANALYSIS I. Types of Attributes Sparse, Incomplete, Inaccurate Data

A REVIEW PAPER ON DATA MINING TECHNIQUES, ALGORITHMS AND TOOLS

MetaData for Database Mining

A Critical Study of Selected Classification Algorithms for Liver Disease Diagnosis

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Index Terms Data Mining, Classification, Rapid Miner. Fig.1. RapidMiner User Interface

Classification using Weka (Brain, Computation, and Neural Learning)

Educational Data Mining: Performance Evaluation of Decision Tree and Clustering Techniques using WEKA Platform

Enhancing the Efficiency of Radix Sort by Using Clustering Mechanism

Enhancing K-means Clustering Algorithm with Improved Initial Center

A Comparison of Decision Tree Algorithms For UCI Repository Classification

Keyword Extraction by KNN considering Similarity among Features

A Cloud Based Intrusion Detection System Using BPN Classifier

NORMALIZATION INDEXING BASED ENHANCED GROUPING K-MEAN ALGORITHM

An Improved Document Clustering Approach Using Weighted K-Means Algorithm

Keywords Clustering, Goals of clustering, clustering techniques, clustering algorithms.

Face Recognition Using K-Means and RBFN

S2 Text. Instructions to replicate classification results.

Improving the Random Forest Algorithm by Randomly Varying the Size of the Bootstrap Samples for Low Dimensional Data Sets

Credit Card Fraud Detection Using Historical Transaction Data

The Data Mining Application Based on WEKA: Geographical Original of Music

Uncertain Data Classification Using Decision Tree Classification Tool With Probability Density Function Modeling Technique

A Novel Approach for Removal of Redundant Test Cases using Hash Set Algorithm along with Data Mining Techniques

The digital copy of this thesis is protected by the Copyright Act 1994 (New Zealand).

Infrequent Weighted Itemset Mining Using SVM Classifier in Transaction Dataset

Comparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini*

String Vector based KNN for Text Categorization

A Comparative Study of Locality Preserving Projection and Principle Component Analysis on Classification Performance Using Logistic Regression

Performance Evaluation of Various Classification Algorithms

Rank Measures for Ordering

An Empirical Study on feature selection for Data Classification

INTRODUCTION TO DATA MINING. Daniel Rodríguez, University of Alcalá

Best Combination of Machine Learning Algorithms for Course Recommendation System in E-learning

International Journal of Data Mining & Knowledge Management Process (IJDKP) Vol.7, No.3, May Dr.Zakea Il-Agure and Mr.Hicham Noureddine Itani

Analysis of a Population of Diabetic Patients Databases in Weka Tool P.Yasodha, M. Kannan

Efficient Pairwise Classification

Count based K-Means Clustering Algorithm

International Journal of Computer Engineering and Applications, ICCSTAR-2016, Special Issue, May.16

PROJECT 1 DATA ANALYSIS (KR-VS-KP)

Credit card Fraud Detection using Predictive Modeling: a Review

CHAPTER 3 MACHINE LEARNING MODEL FOR PREDICTION OF PERFORMANCE

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

INTRODUCTION TO MACHINE LEARNING. Measuring model performance or error

Different Classification Technique for Data mining in Insurance Industry using weka

Using Data Mining to Determine User-Specific Movie Ratings

Comparison of FP tree and Apriori Algorithm

Data Mining of Web Access Logs Using Classification Techniques

Optimizing Number of Hidden Nodes for Artificial Neural Network using Competitive Learning Approach

.. Cal Poly CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. for each element of the dataset we are given its class label.

Machine Learning Techniques for Data Mining

A Study of Random Forest Algorithm with implemetation using Weka

What is Learning? CS 343: Artificial Intelligence Machine Learning. Raymond J. Mooney. Problem Solving / Planning / Control.

Chapter 5: Summary and Conclusion CHAPTER 5 SUMMARY AND CONCLUSION. Chapter 1: Introduction

Transcription:

Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 6, June 2015, pg.220 229 RESEARCH ARTICLE ISSN 2320 088X A Comparative Study of Selected Classification Algorithms of Data Mining Ashish Kumar Dogra 1, Tanuj Wala 2 ¹Computer science & Career Point University, India ²Computer science & Career Point University, India 1 Ashish.dogra01@gmail.com; 2 Tanu.thakur52@gmail.com Abstract Data Mining is the process of extracting hidden knowledge from large volumes of raw data. Data mining has been defined as the nontrivial extraction of previously unknown, implicit and potentially useful information from data. Classification techniques are being used in different industry to easily identify the type and group to which a particular tuple belongs. This work explores the five classification algorithms (Zero, PART, OneR, Prism, J48) are compared on the bases of two techniques (cross validation and percentage split). Experimental comparison was performed by considering the car evaluation dataset and analyzing them using data mining open source WEKA tool. Keywords Data Mining, Classification Algorithms, Car evaluation data set INTRODUCTION Data mining is the motion of all methods and techniques, which allow analyzing very large data sets to extract and discover previously unknown structures and relations out of such huge heaps of details. This information is filtered, prepared and classified so that it will be a valuable aid for decisions and strategies [2]. With the enormous amount of data stored in files, databases, and other repositories, it is increasingly important, if not necessary, to develop powerful means for analysis and perhaps interpretation of such data and for the extraction of interesting that could help in decision-making [3]. The title Data mining is the process of extracting patterns from data. As more data are gathered, with the amount of data doubling every three years [1]. Data mining is the process of discovering knowledge from large amounts of data stored either in databases or warehouses [2]. Data mining is becoming an increasingly important tool to transform these data into information. Data mining can also be referred as knowledge mining or knowledge discovery from data. Many techniques are used in data mining to extract patterns from large amount of database [1]. Classification is a data mining (machine learning) technique used to predict group membership for data instances. Classification is a supervised procedure that learns to classify new instances based on the knowledge learnt from a previously classified training set of instances [5]. It takes a set of data already divided into predefined groups and searches for patterns in the data that differentiate those groups supervised learning, pattern recognition and prediction. Classification consists of predicting a certain outcome based on a given input. It is one of the Data 2015, IJCSMC All Rights Reserved 220

Mining techniques that is mainly used to analyze a given dataset and takes each instance of it and assigns this instance to a particular class with the aim of achieving least classification error. It is used to extract models that correctly define important data classes within the given dataset. It is a two-step process. In first step the model is created by applying classification algorithm on training data set. Then in second step, the extracted model is tested against a predefined test dataset to measure the model trained performance and accuracy. So, classification is the process to assign class label for this dataset whose class label is unknown [7]. Find a model that describes the data classes or concepts. This derived model is based on analysis of set of training data. The derived model can be presented in the following forms Classification (IF-THEN) Rules, Decision Trees, Mathematical Formulae and Neural Networks. Classification Algorithm Classification is a classic data mining technique based on machine learning. Basically classification is used to classify each item in a set of data into one of predefined set of classes or groups. Classification method makes use of mathematical techniques such as decision trees, linear programming, neural network and statistics [9]. Classification algorithm finds relationships between the values of the predictors and the values of the target. Components of Data Mining Algorithms i) Model Representation (Knowledge Representation) is the language for describing discoverable patterns / knowledge ii) Model Evaluation is estimating the predictive accuracy of the derived patterns iii) Search methods we have two type of broadly classified search method one is parameter search when the structure of a model is fixed, search for the parameters which optimise the model evaluation criteria and another is model search when the structure of the model is unknown iv) Learning Bias consist of ffeature selection and Pruning algorithm [11]. Data classification is a two steps process in which first step is the training phase where the classifier algorithm builds classifier with the training set of tuples and the second phase is classification phase where the model is used for classification and its performance is analyzed with the testing set of tuples. The algorithm can differ with respect to accuracy, time to completion, and transparency. In practice, it sometimes makes sense to develop several models for each algorithm, select the best model for each algorithm, and then choose the best of those for deployment. Algorithms Evaluated Five Classification algorithms considered for our study are ZeroR Part Prism OneR J48 ZeroR:- ZeroR is the simplest classification method which relies on the target and ignores all predictors. ZeroR classifier simply predicts the majority category (class). Although there is no predictability power in ZeroR, it is useful for determining a baseline performance as a benchmark for other classification methods [10]. It is the simplest method which relies on the frequency of target. ZeroR is only useful for determining a baseline performance for other classification methods. PART:- PART is a partial decision tree algorithm, which is the developed version of C4.5 and RIPPER algorithms.part is a separate-and-conquer rule learner proposed by Eibe and Witten [54]. The algorithm producing sets of rules called decision lists which are ordered set of rules. A new data is compared to each rule in the list in turn, and the item is assigned the category of the first matching rule (a default is applied if no rule successfully matches). PART builds a partial C4.5 decision tree in its each iteration and makes the best leaf into a rule. The algorithm is a combination of C4.5 and RIPPER rule learning [2]. Prism:- This simple and straightforward covering algorithm work by first picking a class from the dataset for which to create a new rule having the class as its conclusion, and selectively adding tests to the condition of the rule, striving for maximum number of instances covered and 100% accuracy. The accuracy of a test is measured by the ratio of the number of positive instances p to the total number of instances covered by the rule. The positive instances covered by the new rule then are removed from the dataset for further rule generation. Then, negative instances should remain in the data set to await a later iteration of the process. This process continues until no more instances remain to covered. 2015, IJCSMC All Rights Reserved 221

OneR:- OneR or One Rule is a simple algorithm proposed by Holt. The OneR builds one rule for each attribute in the training data and then selects the rule with the smallest error rate as its one rule. The algorithm is based on ranking all the attributes based on the error rate.to create a rule for an attribute, the most frequent class for each attribute value must be determined [54]. The most frequent class is simply the class that appears most often for that attribute value. A rule is simply a set of attribute values bound to their majority class. OneR selects the rule with the lowest error rate. In the event that two or more rules have the same error rate, the rule is chosen at random [8]. The OneR algorithm creates a single rule for each attribute of training data and then picks up the rule with the least error rate [4]. J48:- J48 classifier is a simple C4.5 decision tree for classification. It creates a binary tree. The decision tree approach is most useful in classification problem. With this technique, a tree is constructed to model the classification process. Once the tree is built, it is applied to each tuple in the database and results in classification for that tuple[1][3]. Tool Used for Performance Evaluation Platform SOFTWARE USED Operating system Window 8 JAVA version 7 Tool-WEKA 3.7 Introduction to WEKA Waikato Environment for Knowledge Analysis (WEKA) tool is used for the performance evaluation. WEKA is a collection of machine learning algorithms for data mining tasks. WEKA is created by researchers at University of Waikato in New Zealand. It is written in Java and runs on almost any platform. The algorithms in WEKA can either be applied directly to a dataset or called from own Java code. It is also well-suited for developing new machine learning schemes. WEKA is open source software issued under the GNU General public License. It is a collection of open source ML algorithms pre-processing, classifiers, clustering, association rule and visualization [6]. Data Set The dataset is car evaluation data set has been used for the experiment. In this data set their are 1728 number of instances and 7 number of attributes. Result and Discussion Cross Validation Testing Method Cross-Validation is a statistical method of evaluating and comparing learning algorithms by dividing data into two segments: one used to learn or train a model and the other used to validate the model. In cross validation, the training and validation sets must cross-over in successive rounds such that each data point has a chance of being validated against. The basic form of cross-validation is k-fold cross-validation. Other forms of cross- 2015, IJCSMC All Rights Reserved 222

validation are special cases of k-fold cross-validation or involve repeated rounds of k-fold cross-validation. Ten Fold cross validation method is used in our experiment. In ten folds cross-validation; classifier constructed 10 identical instances of the data set and then split the data in each of these instances in 10 % for training and 90% for testing. Number of Instances Classified Three data are used for evaluation of five rule based algorithms. Firstly all these five algorithms are compared on the basis of number of classified instances. Number of instances consists of number of correctly classified instances and number of incorrectly classified instances by algorithms using ten fold cross validation technique in WEKA. Total numbers of instances in car evaluation data set are 1728 instances. Ten fold cross validation technique is used to obtain the number of classified and in classified number of instances in car evaluation data sets. Table 1: Number of Classified Instances for Algorithm Correctly Classified Instances Incorrectly Classified Instances ZeroR 1210 518 PART 1655 73 OneR 1210 518 Prism 1550 54 J48 1596 132 Table 1 shows the number of correctly classified and incorrectly classified instances for five classification algorithms. PART algorithm with the highest number of 1655 classified instances and OneR algorithm with greatest number of incorrectly classified instances i.e 518. OneR 518 incorrect instances are very high as compare to number of incorrectly classified instances of other four studied algorithms. Figure 1: Number of Classified Instances for Data Set 2015, IJCSMC All Rights Reserved 223

From Table 1 and Figure 1 it is evident that from car evaluation data set, PART have highest number of correctly classified instances followed by Prism and j48. ZeroR and OneR have same performance for this dataset. Accuracy Measurement The following table shows the accuracy measure of classification techniques. They are the True Positive rate, F Measure, Precision, Recall, Receiver Operating Characteristics (ROC) Area and Kappa Statistics. Total numbers of instances in car evaluation data set are 1728 instances. Ten fold cross validation is used. Six parameters value is obtained when car evaluation is evaluated using a WEKA for five algorithms. Table 2: Accuracy Parameters for Algorithm Kappa Statistics TP Rate Precision Recall F-Measure ROC area ZeroR 0.7.49.7.577.497 PART.9091.958.959.958.958.99 oner 0 0.7.49 0.7.577.5 Prism 0.915.966.964.966.965.949 J48.8343.924.924.924.924.976 From table 2 it is evident that PART algorithm had best performance for all six studied parameters for accuracy evaluation. OneR with worst performance for all six evaluated parameters Figure 2: Various Accuracy Parameters for Data Set From table 2 and figure 2 it is evident that for car evaluation data set, from over all evaluation Prism and PART algorithm performance good in terms of accuracy. ZeroR shows average performance and OneR and J48 shows poor accuracy performance for all parameters. 2015, IJCSMC All Rights Reserved 224

Error Rate Mean Absolute Error (M.A.E), Root Mean Square Error (R.M.S.E), Relative Absolute Error (R.A.E) and Root Relative Squared Error (R.R.S.R). Four basic error rate parameters for the evaluation of five classification algorithms for car evaluation data set. Table 3: Error Evaluation Parameters for Algorithm MAE RMSE RAE RRSE ZeroR 0.229 0.3381 100 100 PART 0.0241 0.1276 10.5343 37.7421 OneR 0.1499 0.3871 65.4464 114.5017 Prism 0.0168 0.1297 8.4112 42.394 J48 0.0421 0.1718 18.3833 50.8176 Table 3 shows the Prism algorithm had the least value for two parameters i.e MAE and RAE. PART algorithm having least value for other two parameters i.e RMSE and RRSE Figure 3: Various Error Evaluation Parameters for Data Set From Table 3 and Figure.3 it is evident that PART and J48 have minimum error rate and good performance. Prism algorithm have intermediate error rate thus having average performance. Both OneR and ZeroR algorithm have very high error rate and poor performance as compare to other algorithms. Percentage Split Method In percentage split, the database is randomly split in to two disjoint datasets. The first set, which the data mining system tries to extract knowledge from called training set. The extracted knowledge tested against the 2015, IJCSMC All Rights Reserved 225

second set which is called test set, it is common to randomly split a data set under the mining task in to 2 parts [42F]. 66% percentage split is chosen. Objects of the original database are used as a training set and the rest of objects as a test set. Once the tests is carried out using the selected datasets, then using the available classification and 66 % percentage split test mode, results are collected and an overall comparison is conducted. Number of Instances Classified Three data are used for evaluation of five algorithms. Firstly all these five algorithms are compared on the basis of number of classified instances. Number of instances consists of number of correctly classified instances and number of incorrectly classified instances by algorithms using 66% percentage split technique in WEKA. Total number of classified instances for car evaluation dataset is 1728. Number of instances used for training is 1140 i.e 66% of instances used for training and left 33% of total instances i.e 588 instances are used for testing. Table 4: Number of Classified Instances for Algorithm Correctly Classified Instances Incorrectly Classified Instances ZeroR 535 53 PART 559 29 OneR 409 179 Prism 516 18 J48 409 179 Table 4 shows the number of correctly classified and incorrectly classified instances for five algorithms. PART and Ridor algorithm had the highest number of classified instances. OneR algorithm had highest number of incorrectly classified instances. Figure 4: Number of Classified Instances for Data Set 2015, IJCSMC All Rights Reserved 226

From the Table 4 and Figure 4 it is evident that from car evaluation data set PART have highest number of correctly classified instances followed by ZeroR algorithm. All the two algorithms perform well in classifying the instances. Prism algorithm show average performance. Whereas OneR nd J48 algorithm have lowest performance in terms of correctly classification of instances. Accuracy Measure The following table shows the accuracy measure of classification techniques. True Positive rate, F Measure, Precision, Receiver Operating Characteristics (ROC) Area and Kappa Statistics. Total numbers of instances in car evaluation data set are 1728 instances. Algorithm Kappa Statistics Table 5: Accuracy Parameters for TP Rate Precision Recall F-Measure ROC area ZeroR 0 0 0.484 0.696 0.571 0.5 PART 0.8939 0.951 0.953 0.951 0.951 0.986 OneR 0 0.696 0.484 0.696 0.571 0.5 Prism 0.9109 0.966 0.964 0.966 0.964 0.944 J48 0.8088 0.91 0.915 0.91 0.571 0.965 Table 5 shows the six parameters for evaluating accuracy of five classification algorithms. These parameters are Kappa Statics, TP Rate, Precision, Recall, F-measure and ROC area. Part and Prism algorithm had best performance and ZeroR with worst performance for all six evaluated parameters. Figure 5 Various Accuracy Parameters for Data Set 2015, IJCSMC All Rights Reserved 227

From the Table 5 and Figure 5 it is evident that for car evaluation data set, from over all evaluation Prism and PART algorithm performance good in terms of accuracy. J48 algorithm having performance following two algorithm OneR shows average performance and ZeroR shows poor accuracy performance for all parameters. ERROR RATE Mean Absolute Error (M.A.E), Root Mean Square Error (R.M.S.E), Relative Absolute Error (R.A.E) and Root Relative Squared Error (R.R.S.R). Four basic error rate parameters for the evaluation of five classification algorithms foe car evaluation data set. Table 6: Error Evaluation Parameters for Algorithm MAE RMSE RAE RRSE ZeroR 0.2295 0.339 100 100 PART 0.0294 0.1391 12.8137 41.0226 OneR 0.1522 0.3901 66.3341 115.0861 Prism 0.0169 0.1298 8.778 43.7677 J48 0.0509 0.1883 22.1706 55.5396 Table 6 shows the four basic error rate parameters for the evaluation of five classification algorithms. PART algorithm had the least value for all four parameters. OneR algorithm having highest value for two parameters i.e RMSE and RRSE. DTNB algorithm having highest value for other two parameters i.e MAE and RAE. Figure6: Various Error Evaluation Parameters for Data Set 2015, IJCSMC All Rights Reserved 228

From Table 6 and Figure 6 it is evident that PART algorithm has the best performance when compares to other algorithm. PART has minimum error followed by J48 algorithm. ZeroR and OneR algorithm have very high error rate and show poor performance. Summary From all the above figures it can be summon that PART is best algorithm for classification of car evaluation. Over all J48 table and Prism algorithm have average performance and OneR with very low performance for all three data set. Thus PART and Prism perform well for all the three performance measurement parameters i.e number of classified instances, accuracy and error rate. Conclusion & Future Scope This thesis work presented an introduction to data mining and data mining process. A brief introduction to classification methods and algorithms is provided. Five Classification algorithms ZeroR, OneR, PART, Prism, and J48 are introduced and experimentally evaluated with different data set using WEKA tool. The classification algorithms are experimentally compared based on number of classified instances, accuracy and error rate. We used two testing options for our experiment ten fold cross validation and 66% percentage split. From our study we found that PART is best algorithm for classification of car evaluation data set. Prism and J48 had average performance. OneR algorithm with over all low performance and thus least recommended algorithm. After considering and comparing all the tables and graphs in our study PART algorithm is recommended among all five classification algorithms. As per the future scope is considered during thesis work I analysed that a good amount of research work is devoted to various classification algorithms. And this research work can also be extended by testing the selected algorithms for other machine learning tasks such as clustering, association etc. REFERENCES [1] Abdulkarim, Mirza Nazura. Classification and Retrieval of Research Papers: A Semantic Hierarchical Approach.Diss. Christ University, 2010 [2] Ali, Shawkat, and Kate A. Smith. "On learning algorithm selection for classification." Applied Soft Computing, 2006 [3] Apté, Chidanand, and Sholom Weiss. "Data mining with decision trees and decision rules." Future generation computer systems,1997. [4] Buddhinath, Gaya, and Damien Derry."A simple enhancement to One Rule Classification." Department of Computer Science & Software Engineering. University of Melbourne, Australia, 2006. [5] Devasena, C. Lakshmi, et al. "Effectiveness Evaluation of Rule Based Classifiers for the Classification of Iris Data Set." Bonfring International Journal of Man Machine Interface 1.Special Issue Inaugural Special Issue, 2011. [6] Lilla Gula Information Data Management for the Future of Communication,Luleå University of Technology Department of Computer Science, Electrical and Space Engineering. [7] Ragab, Abdul Hamid M., et al. "A Comparative Analysis of Classification Algorithms for Students College Enrollment Approval Using Data Mining", Proceedings of the 2014 Workshop on Interaction Design in Educational Environments. ACM, 2014. [8] Vijayaran S, Sudha. An Effective Classification Rule Technique for Heart Disease Prediction,.International Journal of Engineering Associates (IJEA), ISSN: 2320-0804, Vol.1, Issue 4, 2013. [9] http://www.zentut.com/data-mining, data-mining-techniques,retrived on:26 April 2015. [10] http://www.saedsayad.com/zeror.htm, ZeroR,Retrived on: 14 May 2015 [11] http://www.doc.ic.ac.uk/~yg/course/ida2002/ida-2002-2.ppt, Data Pre-processing,Retrived on: 30 May 2015. 2015, IJCSMC All Rights Reserved 229