A Comparative Study of Locality Preserving Projection and Principle Component Analysis on Classification Performance Using Logistic Regression

Similar documents
CS145: INTRODUCTION TO DATA MINING

Weka ( )

Evaluating Classifiers

Lecture 25: Review I

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

CS6375: Machine Learning Gautam Kunapuli. Mid-Term Review

Evaluating Classifiers

CS4491/CS 7265 BIG DATA ANALYTICS

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

Tutorial on Machine Learning Tools

Evaluation Metrics. (Classifiers) CS229 Section Anand Avati

Evaluation of different biological data and computational classification methods for use in protein interaction prediction.

Evaluating Machine-Learning Methods. Goals for the lecture

DATA MINING AND MACHINE LEARNING. Lecture 6: Data preprocessing and model selection Lecturer: Simone Scardapane

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Classification Part 4

CS249: ADVANCED DATA MINING

Tutorials Case studies

EE795: Computer Vision and Intelligent Systems

Metrics for Performance Evaluation How to evaluate the performance of a model? Methods for Performance Evaluation How to obtain reliable estimates?

List of Exercises: Data Mining 1 December 12th, 2015

Data Mining Classification: Alternative Techniques. Imbalanced Class Problem

Evaluating Classifiers

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

Evaluating Machine Learning Methods: Part 1

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

Use of Synthetic Data in Testing Administrative Records Systems

INTRODUCTION TO DATA MINING. Daniel Rodríguez, University of Alcalá

Ripple Down Rule learner (RIDOR) Classifier for IRIS Dataset

ISyE 6416 Basic Statistical Methods - Spring 2016 Bonus Project: Big Data Analytics Final Report. Team Member Names: Xi Yang, Yi Wen, Xue Zhang

Time Series Clustering Ensemble Algorithm Based on Locality Preserving Projection

INTRODUCTION TO MACHINE LEARNING. Measuring model performance or error

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

SYS 6021 Linear Statistical Models

An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm

Data Mining and Knowledge Discovery: Practice Notes

CSE Data Mining Concepts and Techniques STATISTICAL METHODS (REGRESSION) Professor- Anita Wasilewska. Team 13

CCRMA MIR Workshop 2014 Evaluating Information Retrieval Systems. Leigh M. Smith Humtap Inc.

Evaluation. Evaluate what? For really large amounts of data... A: Use a validation set.

Fraud Detection using Machine Learning

Machine Learning nearest neighbors classification. Luigi Cerulo Department of Science and Technology University of Sannio

MIPE: Model Informing Probability of Eradication of non-indigenous aquatic species. User Manual. Version 2.4

Combination of PCA with SMOTE Resampling to Boost the Prediction Rate in Lung Cancer Dataset

Classification of Subject Motion for Improved Reconstruction of Dynamic Magnetic Resonance Imaging

Improving Positron Emission Tomography Imaging with Machine Learning David Fan-Chung Hsu CS 229 Fall

CLASSIFICATION JELENA JOVANOVIĆ. Web:

Machine Learning in Python. Rohith Mohan GradQuant Spring 2018

Louis Fourrier Fabien Gaie Thomas Rolf

MACHINE LEARNING TOOLBOX. Logistic regression on Sonar

Data Mining and Knowledge Discovery Practice notes 2

Using Real-valued Meta Classifiers to Integrate and Contextualize Binding Site Predictions

Classification. Instructor: Wei Ding

8/3/2017. Contour Assessment for Quality Assurance and Data Mining. Objective. Outline. Tom Purdie, PhD, MCCPM

Data Mining and Knowledge Discovery: Practice Notes

DATA MINING OVERFITTING AND EVALUATION

Event: PASS SQL Saturday - DC 2018 Presenter: Jon Tupitza, CTO Architect

CS145: INTRODUCTION TO DATA MINING

Data Imbalance Problem solving for SMOTE Based Oversampling: Study on Fault Detection Prediction Model in Semiconductor Manufacturing Process

Classification Algorithms in Data Mining

2. On classification and related tasks

ML4Bio Lecture #1: Introduc3on. February 24 th, 2016 Quaid Morris

Study on Classifiers using Genetic Algorithm and Class based Rules Generation

Constrained Classification of Large Imbalanced Data

MEASURING CLASSIFIER PERFORMANCE

Research on Applications of Data Mining in Electronic Commerce. Xiuping YANG 1, a

Fisher Score Dimensionality Reduction for Svm Classification Arunasakthi. K, KamatchiPriya.L, Askerunisa.A

Data Mining Classification: Bayesian Decision Theory

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

Bagging-Based Logistic Regression With Spark: A Medical Data Mining Method

Cross- Valida+on & ROC curve. Anna Helena Reali Costa PCS 5024

SOFTWARE DEFECT PREDICTION USING IMPROVED SUPPORT VECTOR MACHINE CLASSIFIER

Applying Supervised Learning

Further Thoughts on Precision

Why MultiLayer Perceptron/Neural Network? Objective: Attributes:

Performance Evaluation of Various Classification Algorithms

Business Club. Decision Trees

DATA MINING INTRODUCTION TO CLASSIFICATION USING LINEAR CLASSIFIERS

Artificial Intelligence. Programming Styles

Information Management course

A novel supervised learning algorithm and its use for Spam Detection in Social Bookmarking Systems

Fast or furious? - User analysis of SF Express Inc

10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors

A Comparative Study of Selected Classification Algorithms of Data Mining

Impact of Encryption Techniques on Classification Algorithm for Privacy Preservation of Data

Package hmeasure. February 20, 2015

Research Article International Journals of Advanced Research in Computer Science and Software Engineering ISSN: X (Volume-7, Issue-6)

Logistic Regression: Probabilistic Interpretation

Probabilistic Classifiers DWML, /27

CPSC 340: Machine Learning and Data Mining. Non-Parametric Models Fall 2016

Application of Support Vector Machine In Bioinformatics

Comparison of Optimization Methods for L1-regularized Logistic Regression

Pattern recognition (4)

Image-Based Face Recognition using Global Features

Part II: A broader view

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Machine Problem 8 - Mean Field Inference on Boltzman Machine

Data Mining and Knowledge Discovery Practice notes: Numeric Prediction, Association Rules

Global Probability of Boundary

Function Algorithms: Linear Regression, Logistic Regression

1 Machine Learning System Design

Transcription:

Journal of Data Analysis and Information Processing, 2016, 4, 55-63 Published Online May 2016 in SciRes. http://www.scirp.org/journal/jdaip http://dx.doi.org/10.4236/jdaip.2016.42005 A Comparative Study of Locality Preserving Projection and Principle Component Analysis on Classification Performance Using Logistic Regression Azza Kamal Ahmed Abdelmajed Department of Computer Sciences, Faculty of Mathematical and Computer Sciences, University of Gezira, Wad Madani, Sudan Received 6 March 2016; accepted 9 May 2016; published 12 May 2016 Copyright 2016 by author and Scientific Research Publishing Inc. This work is licensed under the Creative Commons Attribution International License (CC BY). http://creativecommons.org/licenses/by/4.0/ Abstract There are a variety of classification techniques such as neural network, decision tree, support vector machine and logistic regression. The problem of dimensionality is pertinent to many learning algorithms, and it denotes the drastic raise of computational complexity, however, we need to use dimensionality reduction methods. These methods include principal component analysis (PCA) and locality preserving projection (LPP). In many real-world classification problems, the local structure is more important than the global structure and dimensionality reduction techniques ignore the local structure and preserve the global structure. The objectives is to compare PCA and LPP in terms of accuracy, to develop appropriate representations of complex data by reducing the dimensions of the data and to explain the importance of using LPP with logistic regression. The results of this paper find that the proposed LPP approach provides a better representation and high accuracy than the PCA approach. Keywords Logistic Regression (LR), Principal Component Analysis (PCA), Locality Preserving Projection (LPP) 1. Introduction Data mining is the extraction and retrieval of useful data and also involves the retrieval and analysis of data that are stored in a data ware house. Some of the major techniques of data mining are classification, association and clustering. Data mining is upcoming research area to solve various problems and classification is one of main How to cite this paper: Abdelmajed, A.K.A. (2016) A Comparative Study of Locality Preserving Projection and Principle Component Analysis on Classification Performance Using Logistic Regression. Journal of Data Analysis and Information Processing, 4, 55-63. http://dx.doi.org/10.4236/jdaip.2016.42005

problems in the field of data mining [1]. Classification predicts categorical class labels and it classifies the data based on the training set and the values in classifying the attributes and uses it in classifying the new data. Data classification is a two-step process consisting of model construction and model usage. Model construction is used for describing predetermined classes. Model usage is used for classifying future or unknown objects [2]. The development of data mining applications such as classification has shown the need for machine learning algorithms to be applied specially in large scale data [3]. Machine learning focuses on the development of computer programs that can teach themselves to grow and change when exposed to new data; there are two types of machine learning: supervised learning and unsupervised learning: the first one is use training data to infer model and apply model to test data and for the other one, there is no training data, and model inference and application both rely on test data exclusively. Modern machine learning techniques are progressively being used by biologists to obtain proper results from the databases. There are a variety of classification techniques such as neural network, decision tree, support vector machine and logistic regression. Logistic regression (LR) is a well-known statistical classification method that has been used widely in a variety of applications including document classification, bioinformatics and analyzing a data set in which there are one or more independent variables that determine an outcome [4]. The advantage of using LR is that deals with dependent variables that are categorical and extension to the multiclass case. Logistic regression not just classes the data but also predicts probabilities and then estimates the parameters of the model using maximum likelihood estimator. Before using the data set in the classification, it needs some preprocesses such as data cleaning, data transformation and data reduction, the last one is very important because it usually represents the dataset in an (n*m) dimensional space, and these (n*m) dimensional spaces are too large, which however need to reduce the size of the dataset before applying a learning algorithm. A common way to attempt to resolve this problem is to use dimensionality reduction techniques [5]. Because principle component analysis (PCA) preserves the global structure of the dataset and ignores the local structure of the dataset. Therefore this paper proposes linear dimensionality reduction algorithm, called locality preserving projections (LPP). LPP is linear projective maps that arise by solving variational problem that optimally preserves the neighborhood structure of the dataset. LPP should be seen as an alternative to principal component analysis (PCA) [6]. This study aims to compare PCA and LPP in terms of accuracy, to develop appropriate representations of complex data by reducing the dimensions of the data and to explain the importance of using LPP with logistic regression. A variety of performance metrics has been utilized: accuracy, sensitivity, specificity, precision, the area under receiver operating characteristic curve (AUC) and the receiver operating characteristic (ROC) analysis. A detailed concept of using statistical analysis in comparing these methods is given. 2. Materials and Methods 2.1. Experimental Setup In order to evaluate a prediction method it is necessary to have different data sets for training and testing, however five datasets will be used and apply the algorithms principle component analysis (PCA) and locality preserving projections (LPP) to reduce the dimensions using dimensionality reduction toolbox (drtoolbox) in matlab software. After the input space is reduced to a lower dimension by applying one of the two methods PCA and LPP, cross-validation method will be applied to this new reduced features space using 10 fold to evaluation the model and then apply logistic regression to classifier the reduced data. All the performance measures: accuracy, sensitivity, specificity, f-score, precision and roc curve will be computed. The ROC analysis is plotted after each cross validation for the two methods using spss software to compute the area under the curve. 2.1.1. Dimensionality Reduction Toolbox (Drtoolbox) The template is used to format your paper and style the text. All margins, column widths, line spaces, and text fonts are prescribed; please do not alter them. You may note peculiarities. For example, the head margin in this template measures proportionately more than is customary. This measurement and others are deliberate, using specifications that anticipate your paper as one part of the entire journals, and not as an independent document. Please do not revise any of the current designations. 2.1.2. Cross Validation Cross validation is a model evaluation method that is better than residuals. The problem with residual evalua- 56

tions is that they do not give an indication of how well the learner will do when it is asked to make new predictions for data it has not already seen. One way to overcome this problem is to not use the entire data set when training a learner. Some of the data is removed before training begins. Then when training is done, the data that was removed can be used to test the performance of the learned model on new data. This is the basic idea for a whole class of model evaluation methods called cross validation [7]. 2.1.3. Performance Measure The measures that are used in this paper depend on matrix called the confusion matrix are as follows in Table 1: Where: TP: true positives (predicted positive, actual positive), TN: true negatives (predicted negative, actual negative), FP: false positives (predicted positive, actual negative), FN: false negatives (predicted negative, actual positive) [8]. Accuracy: Accuracy is the proportion of true results (both true positives and true negatives) in the population [9]. Sensitivity or Recall: Proportion of actual positives which are predicted positive [9] Specificity: Proportion of actual negative which are predicted negative [9] Positive predictive value (PPV): Proportion of predicted positives which are actual positive [9] F Score: Harmonic Mean of Precision and recall [10] TP+TN Accuracy= TP+TN+FP+FN. (2.1) TP Sensitivity =. (2.2) TP + FN TN Specificity =. (2.3) TN + FP TP Precision =. (2.4) TP + FP ( ) 2 Precision recall F Score= Precision + recall. (2.5) ROC analysis: Receiver Operating Characteristics (ROC) graphs are a useful and clear possibility for organizing classifiers and visualizing their quality (performance) [10]. A Roc curve is a plot of TPR vs FPR for different thresholds θ. Receiver operating characteristic analysis is being used with greater frequency as an evaluation powerful methodology in machine learning and pattern recognition. The ROC is a well-known performance metric for evaluating and comparing algorithms. Area under curve (AUC): Table 1. Confusion matrix. Predicted positive Predicted negative Total Actual positive TP FN AP Actual negative FP TN AN Total PP PN N 57

AUC The area under the ROC is between 0 and 1 and increasingly being recognized as a better measure for evaluating algorithm performance than accuracy. A bigger AUC value implies a better ranking performance for a classifier [10]. Figure 1 explains the Experimental Setup of this paper which uses any algorithm alone and then compares them. 3. Results and Discussion 3.1. Performance by Measures Table 2 shows the classification results of the LR on the whole features, while Table 3 shows the classification results of logistic regression with principle component analysis algorithm. Table 4 shows the classification results of logistic regression with locality preserving projection algorithm. 3.2. The ROC Analysis The estimated probability is used to construct the ROC analysis after each cross validation for the two feature selection methods. Figure 2 represents the PCA and LPP Roc curve for climate model simulation crashes dataset while Figure 3 shows the PCA and LPP Roc curve for heart dataset. Figure 4 shows the PCA and LPP Roc curve for Spam Base dataset. Figure 5 shows the PCA and LPP Roc curve for Phishing Websites dataset. Figure 6 shows the PCA and LPP Roc curve for Musk (version 1) dataset. The figures in Table 5 show the features selection methods; including PCA and LPP with LR; it shows that most points of AUCs of these methods for each dataset. However, the AUCs of LPP is higher for all data sets than PCA. 3.3. Discussion Two experiments on five databases have been systematically performed. These experiments reveal a number of interesting points: 1) In all datasets Locality preserving projection approach performed better than principle component analysis. 2) The datasets is downloaded from UCI repository website and I selected this datasets because most of paper apply LPP on face recognition dataset and there in no study using normal datasets with LR and LPP, however it was necessary to compare LPP with another algorithm like PCA to show the different between them and to improve that LPP is the best than PCA. 3) Table 2 shows the results of the performance measures for logistic regression with all variables, while Table 3 shows the results of the performance measures for logistic regression with principle component analysis. Table 4 shows the results of the performance measures for logistic regression with locality preserving projection. From those tables notice the locality preserving projection (LPP) method it has given a better result in all data sets although there are different in the number of Instances, number of attributes and type of attributes if compare to the principle component analysis (PCA) method and In all performance measures (accuracy, sensitivity, Specificity, precision, f-score and roc curve) LPP performs better than PCA. 4) The ROC curves of PCA, LPP with all data set are shown in Figures 2-5 and LPP seems to be the best one. Table 5 represents the AUCs of each data set and the value of the area under the curve in LPP bigger than the value in the PCA and that indicates to LPP is better than PCA. 5) Comparing to PCA method which it preserve the global structure, the LPP method preserving local structure which is more important than the global structure for many reason: it is important to maintain the intrinsic information of high-dimensional data when they are transformed to a low dimensional space for analysis, a single characterization, either global or local, may be insufficient to represent the underlying structures of real world data and the local geometric structure of data can be seen as a data dependent regularization of the transformation matrix, which helps to avoid over fitting, especially when training samples are scarce. 4. Conclusion This paper proposes dimensionality reduction algorithm called LPP and then compares it with another method of dimensionality reduction approach called Principle component analysis for LR classification. The comparison 58

Dataset Features Matlab (dimensionality reduction toolbox) Apply LPP Apply PCA New features reduced dimensions New features reduced dimension ApplyLogistic regression ApplyLogistic regression Calculate the performance measures Calculate the performance measures Compare of LPP & PCA Figure 1. Experimental setup. Table 2. The results of the performance measures for logistic regression with all variables. Datasets Performance measures Accuracy Specify Sensitivity Precision F-score Climate model simulation crashes (540 18) 0.9259 0.7500 0.9400 0.9792 0.9592 Heart (270 13) 0.7778 0.7273 0.8125 0.8125 0.8125 Spambase (4601 57) 0.9067 0.9333 0.8571 0.8734 0.8652 Phishing websites (2456 30) 0.9283 0.9219 0.9350 0.9200 0.9274 Musk (version 1) (476 186) 0.7358 0.7500 0.7143 0.6522 0.6818 Table 3. The results of the performance measures for logistic regression with principle component analysis. Datasets Climate model simulation crashes (540 18) Performance measures Accuracy Specify Sensitivity Precision F-score 0.9444 0.5000 0.9800 0.9608 0.9703 Heart (270 13) 0.8148 0.8333 0.8200 0.8571 0.8276 Spambase (4601 57) 0.9197 0.9296 0.9040 0.8889 0.8964 Phishing websites (2456 30) 0.9323 0.9329 0.9314 0.9048 0.9179 Musk (version 1) (476 186) 0.8113 0.8438 0.7619 0.7619 0.7619 59

Table 4. The results of the performance measures for logistic regression with locality preserving projection algorithm. Datasets Climate model simulation crashes (540 18) Performance measures Accuracy Specify Sensitivity Precision F-score 0.9815 0.8000 0.9900 0.9800 0.9899 Heart (270 13) 0.8889 0.8462 0.9286 0.8667 0.8966 Spambase (4601 57) 0.9284 0.9368 0.9167 0.9119 0.9143 Phishing websites (2456 30) 0.9602 0.9632 0.9565 0.9565 0.9565 Musk (version 1) (476 186) 0.8491 0.8571 0.8333 0.7500 0.7895 Figure 2. Roc curve for climate model simulation crashes dataset. includes several performance measures, which resulted in a valid and reliable conclusion. The performance of these approaches is evaluated in terms of accuracy, sensitivity, specificity, F-score, precision, AUC and ROC analysis. The comparison is done through experiments conducted on various types/sizes of datasets. The comparison shows that LPP gives relatively good result in feature reduction and computational complexity when the training data size is relatively larger in comparison to the number of features. In LR, the features are required to be uncorrelated but not needed to be independent, when PCA and LPP are applied to the datasets with the number of features quite bigger than the data size, the dimension needs to be reduced to a very low dimension, this results in loss of more information. It can be stated that LR has proven to be a powerful classifier for high dimensional data sets and it also gives good efficiency when using the features selection methods. From previous study, LPP performed better in face of recognition and in this paper LPP also performed better in normal datasets by preserving the local structure rather than the global structure. 60

Figure 3. Roc curve for heart dataset. Figure 4. Roc curve for spambase dataset. 61

Figure 5. Roc curve for phishing websites dataset. Figure 6. Roc curve for musk (version 1) dataset. 62

Table 5. Area under the curve. 5. Recommendations Datasets PCA Methods Climate model simulation crashes 0.912 0.948 Heart 0.812 0.904 Spambase 0.847 0.966 Phishing websites 0.873 0.987 Musk (version 1) 0.792 0.899 This paper can be extended to other data mining techniques like clustering, association, it can also be extended for other classification algorithm such as neural network, decision tree and support vector machine and much more datasets should be taken. Moreover, this paper recommends by using more than mathematical model to obtain the best results. References [1] Rohit Arora, S. (2012) Comparative Analysis of Classification Algorithms on Different Datasets Using WEKA. International Journal of Computer Applications, 54, No. 13. [2] Deepajothi, S. and Selvarajan, S. (2012) A Comparative Study of Classification Techniques on Adult Data Set. International Journal of Engineering Research & Technology (IJERT), 1, No. 8. [3] bin Othman, M.F. and Yau, T.M.S. (2007) Comparison of Different Classification Techniques Using WEKA for Breast Cancer. [4] Musa, A.B. (2013) A Comparison of l1-regularizion, PCA, KPCA and ICA for Dimensionality Reduction in Logistic Regression. International Journal of Machine Learning and Cybernetics. [5] He, X.F., Yan, S.C., Hu, Y.X., Niyogi, P. and Zhang, H.-J. (2005) Face Recognition Using Laplacianfaces. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27, No. 3. [6] He, X.F. and Niyogi, P. (2004) Locality Preserving Projections (LPP). [7] Murphy, K.P. (2007) Performance Evaluation of Binary Classifiers. [8] Chaurasia, S., Chakrabarti, P. and Chourasia, N. (2014) Prediction of Breast Cancer Biopsy Outcomes An Approach Using Machine Leaning Perspectives. International Journal of Computer Applications, 100, No. 9. [9] François, D. (2009) Binary Classification Performances Measure Cheat Sheet. [10] Slaby, A. (2007) ROC Analysis with Matlab. 29th International Conference on Information Technology Interfaces, Cavtat, 25-28 June 2007, 191-196. http://dx.doi.org/10.1109/iti.2007.4283768 LPP 63