CHAPTER 4 STOCK PRICE PREDICTION USING MODIFIED K-NEAREST NEIGHBOR (MKNN) ALGORITHM
|
|
- Godwin Sparks
- 6 years ago
- Views:
Transcription
1 CHAPTER 4 STOCK PRICE PREDICTION USING MODIFIED K-NEAREST NEIGHBOR (MKNN) ALGORITHM 4.1 Introduction Nowadays money investment in stock market gains major attention because of its dynamic nature. So the significant issue in market finance is discovering well organized approaches to outline and envision the stock market information to provide individuals or organizations helpful data about the behavior of the market for making decision about investment.the huge amount of important information produced by the stock market has attracted researchers to investigate this issue utilizing distinctive approaches. Since stock markets produce huge datasets it data mining techniques is found to be more efficient.data mining is utilized for excavate data from databases and discover the meaningful patterns from the database. The usefulness of this data makes data mining imperative and necessary.the essentials of data mining in finance are originating from the need to adopt specific well organized criteria to predict exactness, facilitate multi-resolution calculation. 4.2 k- Nearest Neighbor (k-nn) In pattern identification, the KNN is a technique for categorizing items according nearest training samples. KNN is a sort of illustration based learning, or lazy learning where the task is just approximated locally and all calculation is delayed until classification Assumptions in KNN KNN assumes that the information is present in a feature space. Accurately, the data points are in a metric space. Mostly these data are either multidimensional or scalar vectors. Since the points are in feature space, they have a concept of distance. This requirement is not need to be Euclidean distance yet it is used commonly. Every training sample comprises of a vectors set and separate class label corresponding with each vector. These classes may be either positive or negative classes. But KNN have the efficiency to accomplish different tasks with random number of classes.
2 Additionally a single number k is given. This number makes a decision of what numbers of neighbors (where neighbors are defined based on the distance metric) impact the classification. This is typically an odd number if the quantity of classes is 2. In the event that k=1, then the is just called the nearest neighbor Basics of KNN The KNN is the principal and most straightforward classification technique when the information about the distribution of the data is insufficient. This convention basically holds the whole training set during learning and allocates to every query a class characterized by the majority label of its k-nearest neighbors in the training set. The Nearest Neighbor (NN) principle is the least complex type of KNN when K = 1. In this every training samples ought to be grouped to its samples surrounded by it. Subsequently, if the classification of any of the sample data is obscure, then it could be anticipated by considering the classification of its nearest neighbor tests. Given an obscure sample and a training set consisting of samples, all the distances between the obscure sample and the entire sample in the training set can be calculated by utilizing the accompanying mathematical statement (4.1) where, x 1, x 2, x 3,x p are anticipators of the first sample and u 1, u 2,u 3, u p are anticipators of the second sample. If distance is of smallest value, then the samples in the training set is close to the obscure sample. Hence, the obscure sample may be categorized based on this nearest neighbor classification.
3 Known Samples Unknown Samples (a) (b) Fig 4.1 KNN decision rule Fig 4.1 illustrates the KNN decision rule for K= 1 and K= 3 for a set of samples divided into 2 classes.in Fig 4.1(a), an obscure sample (unknown sample) is categorized by using only one known sample; In Fig 4.1(b) more than one known sample is used. In the last case, the parameter K is set to 3, hence the closest three samples is considered for classifying the obscure one. Two of them belong to the same class, whereas only one belongs to the other class. In both cases, the unknown sample is classified as belonging to the class on the left. Fig 4.2 shows the pseudo code for the KNN
4 Input: Finite set A, Finite Set B, k, function c:b->{1,2,.n} Output: r:a->{1,2,..n} Begin For each x in A do Let L<- {} For each b in B add (a(x,b), c(b)) to L Sort the elements in L with the first components Compute the class labels from the first k elements from L Let r(x) be the class containing highest number of occurrences End Return r End Fig 4.2 Pseudo code for KNN The classifier performance is principally controlled by the decision of K and in addition the distance metric applied [20-25]. This evaluation is influenced by the sensitivity of the choosing the neighborhood size K, since local region radius is calculated by the K th nearest neighbor distance to the query and diverse value of K yields various conditional class probabilities Distance Metric KNN makes estimation according to the result of the K neighbors closest to that point. Accordingly, to make estimation with KNN, we have to characterize a metric for measuring the separation between the query point and cases from the samples. A familiar opinion to estimate this distance is known as Euclidean. Different measures include Euclidean square, City-square, and Chebychev. Table 4.1 presents the distance metrics and their formula.
5 Table 4.1 Distance metrics employed in KNN Distance Metric Formula (x- query point, p data point from unknown sample) Euclidean Distance Euclidean Squared City-block Chebychev K-Nearest Neighbor Predictions After choosing the value of K, anticipations are made based on the KNN samples. For regression, KNN prediction is the result of average of the K nearest neighbors: (4.2) Where x i is the i th case of the sample and y is the query point anticipation (result).in classification problem, based on the voting scheme KNN anticipation is performed in which the winner is used to name the query. Generally the K neighbors have equivalent impact on prediction regardless of their relative distance from the query point. An optional methodology is to use randomly large K values with more vitality given to cases nearest to the query point. This is accomplished by using 'distance weighting' Distance Weighting Since KNN forecasts are based on the belief that items close in distance are conceivably similar, it is good to differentiate between the K nearest neighbors during prediction, i.e., let the closest points among the K nearest neighbors have more say in influencing the result of the query point. This can be attained by presenting a set of weights W, one for every nearest neighbor, characterized by the relative closeness of each one neighbor regarding the query point. Thus
6 (4.3) Where is the distance between the query point x and the i th case p i of the sample. It is clear that the weights defined in this manner above will satisfy: (4.4) Thus, for regression problems, we have: (4.5) For classification problems, the highest value of the above equation is taken for every one of class variables. It is obvious from the above equation that when K>1, one can basically characterize the standard deviation for predictions in regression tasks using, (4.6) Some of the KNN merits are depicted as follows: Easy to use; resilient to noisy training samples, particularly if the inverse square of weighted distance is used as the "distance" measure; and Effective if the training data is vast. In spite of these advantages, it has a few demerits such as: a) computationally expensive as it needs to find distance of each one query example to all training sample data; b) The huge memory to execute in extent with size of training set; c) Low precision rate in multidimensional datasets; d) Need to find the parameter value K, the quantity of nearest neighbors; e) Distance based learning is not clear which sort of distance to use; and f) decide which labels are ideal to produce the best results. Therefore, to overcome the low precision rate of KNN, Modified KNN (MKNN) has been proposed in this research work. The MKNN preprocesses the training set before using it and finds the legitimacy of any training data.the final classification is then made by applying weighted KNN which used validity as the multiplicative factor. 4.3 Modified K-Nearest Neighbor (MKNN) In this research Modified K-Nearest Neighbor Algorithm is used for prediction of stock index movement.the fundamental idea of the presented technique is allocating the class label of the
7 queried instance into K validated data training points and the validity of all data tests in the training set is calculated. At that point, a weighted KNN is performed on any trained samples. Fig 4.3 demonstrates the pseudo code of the MKNN. Pseudo-code of the MKNN Algorithm End Output_label:= MKNN ( train_set, test_sample) Begin For i := 1 to train_size Validity(i) := Compute Validity of i-th sample; End for; Output_label:=Weighted_KNN(Validity,test_sample); Return Output_label; Fig 4.3 Pseudo code of the MKNN Data and Sources of Data This exploration inspects the monthly change of closing values of NSE-NIFTY and BSE stock data according to the following predictors: Open price, High price, Low price and Close price. NSE-NIFTY and BSE stock index values are acquired from the NSE and BSE sites separately for the period from Jan'2013 to Dec The data is split into two sub-tests of 80:20 where the in-test sample or preparing data compasses from Jan' 2010 to Dec' 2012 and the data for the remaining period from Jan 2013 to Dec 2013are used for out-of sample or test data Preprocessing the data When the data was gathered at first, all the values of the attributes chosen were continuous numeric values. Data conversion was applied by generalizing the data to a higher-level concept so as all the values got to be discrete. The rule that was made to convert the numeric values of each one attribute to discrete values relied on upon the earlier day closing price of the stock. If in case that the values of the properties open, high, low, and close were more prominent than the estimation of attribute past for the same trading day, the numeric values of the attribute were supplanted by the value positive. In the event that the values of the attributes said above were
8 short of what the value of the attributes used previously, the numeric values of the attributes were supplanted by negative. If the values of those attributes were equal to the value of the attribute previous,then values were replaced by the same equal value Building the Model After the data has been arranged and converted, the upcoming step was to build the forecast model using the MKNN. The MKNN was chosen since the development of MKNN classifiers does not require any domain information, along these lines it is fitting for exploratory learning discovery. Also, it can deal with high dimensional data. In the MKNN, each sample in training set must be validated at the first step. The validity of each one point is found as per its neighbors. The validation procedure is performed for all train samples. To accept a sample point in the training set, the H nearest neighbors of the point is considered. Among the H nearest neighbors of a training test x, validity(x) enumerate the quantity of points with the same name to the label of x. The formula which is proposed to calculate the validity of every point in train set is (4.7) where H is the number of considered neighbors and lbl(x)returns the true class label of the sample x. also, Ni(x) stands for the i th nearest neighbor of the point x. The function S takes into account the similarity between the point x and the i th nearest neighbor. (4.8) Prediction Model The prediction model considers Opening value, High value, Low value and Closing value of the market index as independent variables and the next day s closing value as the dependent variable. The MKNN identifies k nearest neighbors in the training data set in terms of the Euclidean distance with respect to the day for which prediction is to be done. Once k- nearest neighbors are identified, the prediction for that day is computed as the average of the next day s closing prices of those neighbors. The MKNN employs weighted KNN on the test
9 data set for predicting the next day s closing value. The output of the predictive model is compared with the actual values of the test dataset for validation. Applying weighted KNN Each of the K samples is given a weighted vote that is usually equal to some decreasing function of its distance from the unknown sample. For example, the vote might set be equal to 1/(de+1), where de is Euclidian distance. These weighted votes are then summed for each class, and the class with the largest total vote is chosen. This distance weighted KNN technique is very similar to the window technique for estimating density functions. For example, using a weighted of 1/ (de+1) is equivalent to the window technique with a window function of 1/ (de+1) if K is chosen equal to the total number of training samples. In the MKNN method, first the weight of each neighbor is computed using (4.9) Then, the validity of that training sample is multiplied on its raw weight which is based on the Euclidian distance. In the MKNN method, the weight of each neighbor sample is derived according to (4.10) Here v (i) and Val (i) stand for the weight and the validity of the i th nearest sample in the train set Classifier Model The classifier model considers opening value, high value, low value, closing value and returns of the market index as independent variables and the next day s class as the dependent variable. Returns for a day is calculated as (4.11)
10 Where v t is the closing value of the index on the current day and v t-1 is the closing value of the index of previous day. If the next days return is positive, the next day s class is classified as bull otherwise bear.the yield of the classifier is compared with the real classes of the test data set to improve the effectiveness of the approach. 4.4 Empirical Results The examined data sample comprises of daily returns from January 2010 to December 2013 of three stock market indices, BSE oil and gas, CNX-100 and CNX-NIFTY. Data samples are collected from the historical values of NSE- NIFTY and BSE (Bombay Stock Exchange) data. The total data set is split into two one for training the network and remaining for testing the performance of the network. In this experiment, the stock index data from January 2010 to December 2012 is used to train the network and the data from January 2013 to December 2013 is used to test the performance of the proposed approach Performance Measures The following performance measures are used to gauge the performance of the trained forecasting model for the test data: The Mean Squared Error (MSE), Root Mean Squared Error (RMSE), R-Squared (R 2 ), Adjusted R-squared (RA 2 ), Hannan-Quinn Information Criterion (HQ). Table 4.2 illustrates various performance measures that are used to evaluate the effectiveness of the proposed approach.
11 Table 4.2: Performance Criteria and the related formula Performance Criteria Mean Squared Error Formula Root Mean Squared Error (RMSE) R-Squared(R 2 ) Adjusted R-Squared(R A 2 ) = real value, = estimated value, = mean value Hannan-Quinn Information Criterion (HQ) SSR = Results Prediction Model Fig 4.4 presents the results for the returns (close price) for the year 2013 of the BSE Oil and Gas index obtained using Modified KNN (MKNN) and table 4.3 shows the error rate of the proposed approach using various performance measures.
12 Jan Feb Mar April May June July Aug Sep Oct Nov Dec Close Price Actual Predicted Fig 4.4 BSE Predicted Close Price Value Table 4.3 Error Rate of BSE Test Criteria Error Rate (%) Mean Squared Error 3.87 Root Mean Squared Error (RMSE) 5.98 R-Squared(R 2 ) 0.35 Adjusted R-Squared(R 2 A ) 1.67 Hannan-Quinn Information Criterion (HQ) -5.03
13 Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Close Price Fig 4.5 presents the results for the returns (close price) for the year 2013 of the NSE CNX 100 index and table 4.4 shows the error rate of the MKNN approach using various performance measures Actual Predicted Fig 4.5 Predicted Close Price of CNX-100 Stock Index Table 4.4 Error Rate of CNX-100 Stock Index Test Criteria Error Rate (%) Mean Squared Error 3.67 Root Mean Squared Error (RMSE) 4.98 R-Squared(R 2 ) 0.38 Adjusted R-Squared(R 2 A ) 1.98 Hannan-Quinn Information Criterion (HQ) -5.03
14 Jan Feb Mar April May Jun Jul Aug Sep Oct Nov Dec Close Price Fig 4.6 presents the results for the returns (close price) for the year 2013 of the NSE CNX NIFTY index and table 4.5 shows the error rate of the proposed approach using various performance measures Actual Predicted Fig 4.6 Predicted Close Price of CNX-NIFTY Stock Index Table 4.5 Error Rate of CNX-NIFTY Stock Index Test Criteria Error Rate (%) Mean Squared Error 3.89 Root Mean Squared Error (RMSE) 4.43 R-Squared(R 2 ) 0.45 Adjusted R-Squared(R 2 A ) 1.78 Hannan-Quinn Information Criterion (HQ) -5.03
15 Classification Model The results obtained from the two classifying models for BSE oil and gas, CNX-100 and CNX- NIFTY are given below. Table 4.6 Comparison of Classifier Models on the Test Dataset for BSE Oil and Gas k-nn MKNN Instances Accuracy Instances Accuracy Correctly classified Incorrectly classified % % % % Table 4.6 shows that the MKNN rightly classifies the next day s index movement of BSE Oil and Gas Index for 294 instances out of the total of 331 instances with an accuracy rate of 88.8% and misclassifies 37instances with an error rate of 12.2%.But the KNN correctly classifies the next day s index movement only for 258 instances out of the total of 331 instances with an accuracy rate of 77.9% and misclassifies 73 instances with an error rate of 22.1% respectively. Table 4.7 Comparison of Classifier Models on the Test Dataset for CNX-100 k-nn MKNN Instances Accuracy Instances Accuracy Correctly classified Incorrectly classified % % % % Table 4.7 shows that the MKNN rightly classifies the next day s index movement of CNX-100 Index for 294 instances out of the total of 331 instances with an accuracy rate of 88.01% and misclassifies 41 instances with an error rate of 12.79% respectively.but the KNN correctly classifies the next day s index movement only for 254 instances out of the total of 331
16 instances with an accuracy rate of 76.89% and misclassifies 77 instances with an error rate of 22.87% respectively. Table 4.8Comparison of Classifier Models on the Test Dataset for CNX-NIFTY Correctly classified Incorrectly classified k-nn MKNN Instances Accuracy Instances Accuracy % % % % Table 4.8 shows that the MKNN rightly classifies the next day s index movement of CNX-NIFTY Index for 295 instances out of the total of 331 instances with an accuracy rate of 88.57% and misclassifies 36instances with an error rate of 11.98% respectively. But the KNN correctly classifies the next day s index movement only for 256 instances out of the total of 331 instances with an accuracy rate of 77.01% and misclassifies 75 instances with an error rate of 22.45% respectively. Table 4.9Confusion Matrices for BSE Oil and Gas K-NN MKNN Algorithm Actual Class Predicted Class Predicted Class Bull Bear Bull Bear Bull Bear It is seen from the Table 4.9 that MKNN rightly classifies 150 bull class instances out of the total of 164 bull class instances and rightly classifies158 bear class instances out of the total of 167 bear class instances. But the KNN has lower performance compared to the MKNN model.
17 Table 4.10Confusion Matrices for CNX-100 K-NN MKNN Algorithm Actual Class Predicted Class Predicted Class Bull Bear Bull Bear Bull Bear Table 4.10 shows that MKNN rightly classifies 149 bull class instances out of the total of 164 bull class instances while KNN classifies only about 100 instances and MKNN correctly classifies158 bear class instances out of the total of 167 bear class instances which is about 145 instances in bear class. Table 4.11 Confusion Matrices for CNX-NIFTY K-NN MKNN Algorithm Actual Class Predicted Class Predicted Class Bull Bear Bull Bear Bull Bear Table 4.11 shows that MKNN rightly classifies 151 bull class instances out of the total of 164 bull class instances while KNN classifies only about 101 instances and MKNN correctly classifies156 bear class instances out of the total of 167 bear class instances which is about 146 instances in bear class.
Data Preprocessing. Supervised Learning
Supervised Learning Regression Given the value of an input X, the output Y belongs to the set of real values R. The goal is to predict output accurately for a new input. The predictions or outputs y are
More informationData Mining and Machine Learning: Techniques and Algorithms
Instance based classification Data Mining and Machine Learning: Techniques and Algorithms Eneldo Loza Mencía eneldo@ke.tu-darmstadt.de Knowledge Engineering Group, TU Darmstadt International Week 2019,
More informationData mining. Classification k-nn Classifier. Piotr Paszek. (Piotr Paszek) Data mining k-nn 1 / 20
Data mining Piotr Paszek Classification k-nn Classifier (Piotr Paszek) Data mining k-nn 1 / 20 Plan of the lecture 1 Lazy Learner 2 k-nearest Neighbor Classifier 1 Distance (metric) 2 How to Determine
More informationInternational Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X
Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,
More informationCHAPTER 3 A FAST K-MODES CLUSTERING ALGORITHM TO WAREHOUSE VERY LARGE HETEROGENEOUS MEDICAL DATABASES
70 CHAPTER 3 A FAST K-MODES CLUSTERING ALGORITHM TO WAREHOUSE VERY LARGE HETEROGENEOUS MEDICAL DATABASES 3.1 INTRODUCTION In medical science, effective tools are essential to categorize and systematically
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationVoronoi Region. K-means method for Signal Compression: Vector Quantization. Compression Formula 11/20/2013
Voronoi Region K-means method for Signal Compression: Vector Quantization Blocks of signals: A sequence of audio. A block of image pixels. Formally: vector example: (0.2, 0.3, 0.5, 0.1) A vector quantizer
More informationEnhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques
24 Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques Ruxandra PETRE
More information10/5/2017 MIST.6060 Business Intelligence and Data Mining 1. Nearest Neighbors. In a p-dimensional space, the Euclidean distance between two records,
10/5/2017 MIST.6060 Business Intelligence and Data Mining 1 Distance Measures Nearest Neighbors In a p-dimensional space, the Euclidean distance between two records, a = a, a,..., a ) and b = b, b,...,
More informationPerformance Analysis of Data Mining Classification Techniques
Performance Analysis of Data Mining Classification Techniques Tejas Mehta 1, Dr. Dhaval Kathiriya 2 Ph.D. Student, School of Computer Science, Dr. Babasaheb Ambedkar Open University, Gujarat, India 1 Principal
More information6. Dicretization methods 6.1 The purpose of discretization
6. Dicretization methods 6.1 The purpose of discretization Often data are given in the form of continuous values. If their number is huge, model building for such data can be difficult. Moreover, many
More informationCHAPTER 4 DETECTION OF DISEASES IN PLANT LEAF USING IMAGE SEGMENTATION
CHAPTER 4 DETECTION OF DISEASES IN PLANT LEAF USING IMAGE SEGMENTATION 4.1. Introduction Indian economy is highly dependent of agricultural productivity. Therefore, in field of agriculture, detection of
More informationEvaluating Classifiers
Evaluating Classifiers Charles Elkan elkan@cs.ucsd.edu January 18, 2011 In a real-world application of supervised learning, we have a training set of examples with labels, and a test set of examples with
More informationEstablishing Virtual Private Network Bandwidth Requirement at the University of Wisconsin Foundation
Establishing Virtual Private Network Bandwidth Requirement at the University of Wisconsin Foundation by Joe Madden In conjunction with ECE 39 Introduction to Artificial Neural Networks and Fuzzy Systems
More informationClustering and Visualisation of Data
Clustering and Visualisation of Data Hiroshi Shimodaira January-March 28 Cluster analysis aims to partition a data set into meaningful or useful groups, based on distances between data points. In some
More informationData Mining Classification: Alternative Techniques. Lecture Notes for Chapter 4. Instance-Based Learning. Introduction to Data Mining, 2 nd Edition
Data Mining Classification: Alternative Techniques Lecture Notes for Chapter 4 Instance-Based Learning Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Instance Based Classifiers
More informationK-Nearest-Neighbours with a Novel Similarity Measure for Intrusion Detection
K-Nearest-Neighbours with a Novel Similarity Measure for Intrusion Detection Zhenghui Ma School of Computer Science The University of Birmingham Edgbaston, B15 2TT Birmingham, UK Ata Kaban School of Computer
More informationUnsupervised Learning
Unsupervised Learning Unsupervised learning Until now, we have assumed our training samples are labeled by their category membership. Methods that use labeled samples are said to be supervised. However,
More informationCharacter Recognition
Character Recognition 5.1 INTRODUCTION Recognition is one of the important steps in image processing. There are different methods such as Histogram method, Hough transformation, Neural computing approaches
More informationCS178: Machine Learning and Data Mining. Complexity & Nearest Neighbor Methods
+ CS78: Machine Learning and Data Mining Complexity & Nearest Neighbor Methods Prof. Erik Sudderth Some materials courtesy Alex Ihler & Sameer Singh Machine Learning Complexity and Overfitting Nearest
More informationA System to Automatically Index Genealogical Microfilm Titleboards Introduction Preprocessing Method Identification
A System to Automatically Index Genealogical Microfilm Titleboards Samuel James Pinson, Mark Pinson and William Barrett Department of Computer Science Brigham Young University Introduction Millions of
More information1) Give decision trees to represent the following Boolean functions:
1) Give decision trees to represent the following Boolean functions: 1) A B 2) A [B C] 3) A XOR B 4) [A B] [C Dl Answer: 1) A B 2) A [B C] 1 3) A XOR B = (A B) ( A B) 4) [A B] [C D] 2 2) Consider the following
More informationI211: Information infrastructure II
Data Mining: Classifier Evaluation I211: Information infrastructure II 3-nearest neighbor labeled data find class labels for the 4 data points 1 0 0 6 0 0 0 5 17 1.7 1 1 4 1 7.1 1 1 1 0.4 1 2 1 3.0 0 0.1
More informationData Mining and Data Warehousing Classification-Lazy Learners
Motivation Data Mining and Data Warehousing Classification-Lazy Learners Lazy Learners are the most intuitive type of learners and are used in many practical scenarios. The reason of their popularity is
More informationCHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS
CHAPTER 8 COMPOUND CHARACTER RECOGNITION USING VARIOUS MODELS 8.1 Introduction The recognition systems developed so far were for simple characters comprising of consonants and vowels. But there is one
More informationK Nearest Neighbor Wrap Up K- Means Clustering. Slides adapted from Prof. Carpuat
K Nearest Neighbor Wrap Up K- Means Clustering Slides adapted from Prof. Carpuat K Nearest Neighbor classification Classification is based on Test instance with Training Data K: number of neighbors that
More informationUnsupervised Learning
Outline Unsupervised Learning Basic concepts K-means algorithm Representation of clusters Hierarchical clustering Distance functions Which clustering algorithm to use? NN Supervised learning vs. unsupervised
More informationMODULE 7 Nearest Neighbour Classifier and its variants LESSON 11. Nearest Neighbour Classifier. Keywords: K Neighbours, Weighted, Nearest Neighbour
MODULE 7 Nearest Neighbour Classifier and its variants LESSON 11 Nearest Neighbour Classifier Keywords: K Neighbours, Weighted, Nearest Neighbour 1 Nearest neighbour classifiers This is amongst the simplest
More informationCharacterization and Modeling of Deleted Questions on Stack Overflow
Characterization and Modeling of Deleted Questions on Stack Overflow Denzil Correa, Ashish Sureka http://correa.in/ February 16, 2014 Denzil Correa, Ashish Sureka (http://correa.in/) ACM WWW-2014 February
More informationBusiness Club. Decision Trees
Business Club Decision Trees Business Club Analytics Team December 2017 Index 1. Motivation- A Case Study 2. The Trees a. What is a decision tree b. Representation 3. Regression v/s Classification 4. Building
More informationCHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION
CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant
More informationMachine Learning for Signal Processing Clustering. Bhiksha Raj Class Oct 2016
Machine Learning for Signal Processing Clustering Bhiksha Raj Class 11. 13 Oct 2016 1 Statistical Modelling and Latent Structure Much of statistical modelling attempts to identify latent structure in the
More informationNonparametric Methods Recap
Nonparametric Methods Recap Aarti Singh Machine Learning 10-701/15-781 Oct 4, 2010 Nonparametric Methods Kernel Density estimate (also Histogram) Weighted frequency Classification - K-NN Classifier Majority
More informationDATA MINING LECTURE 10B. Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines
DATA MINING LECTURE 10B Classification k-nearest neighbor classifier Naïve Bayes Logistic Regression Support Vector Machines NEAREST NEIGHBOR CLASSIFICATION 10 10 Illustrating Classification Task Tid Attrib1
More informationNORMALIZATION INDEXING BASED ENHANCED GROUPING K-MEAN ALGORITHM
NORMALIZATION INDEXING BASED ENHANCED GROUPING K-MEAN ALGORITHM Saroj 1, Ms. Kavita2 1 Student of Masters of Technology, 2 Assistant Professor Department of Computer Science and Engineering JCDM college
More informationMachine Learning / Jan 27, 2010
Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationNaïve Bayes for text classification
Road Map Basic concepts Decision tree induction Evaluation of classifiers Rule induction Classification using association rules Naïve Bayesian classification Naïve Bayes for text classification Support
More informationIteration Reduction K Means Clustering Algorithm
Iteration Reduction K Means Clustering Algorithm Kedar Sawant 1 and Snehal Bhogan 2 1 Department of Computer Engineering, Agnel Institute of Technology and Design, Assagao, Goa 403507, India 2 Department
More informationTopic 1 Classification Alternatives
Topic 1 Classification Alternatives [Jiawei Han, Micheline Kamber, Jian Pei. 2011. Data Mining Concepts and Techniques. 3 rd Ed. Morgan Kaufmann. ISBN: 9380931913.] 1 Contents 2. Classification Using Frequent
More informationEvaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München
Evaluation Measures Sebastian Pölsterl Computer Aided Medical Procedures Technische Universität München April 28, 2015 Outline 1 Classification 1. Confusion Matrix 2. Receiver operating characteristics
More informationK- Nearest Neighbors(KNN) And Predictive Accuracy
Contact: mailto: Ammar@cu.edu.eg Drammarcu@gmail.com K- Nearest Neighbors(KNN) And Predictive Accuracy Dr. Ammar Mohammed Associate Professor of Computer Science ISSR, Cairo University PhD of CS ( Uni.
More informationToward Part-based Document Image Decoding
2012 10th IAPR International Workshop on Document Analysis Systems Toward Part-based Document Image Decoding Wang Song, Seiichi Uchida Kyushu University, Fukuoka, Japan wangsong@human.ait.kyushu-u.ac.jp,
More informationCHAPTER VII INDEXED K TWIN NEIGHBOUR CLUSTERING ALGORITHM 7.1 INTRODUCTION
CHAPTER VII INDEXED K TWIN NEIGHBOUR CLUSTERING ALGORITHM 7.1 INTRODUCTION Cluster analysis or clustering is the task of grouping a set of objects in such a way that objects in the same group (called cluster)
More informationSand Pit Utilization
Sand Pit Utilization A construction company obtains sand, fine gravel, and coarse gravel from three different sand pits. The pits have different average compositions for the three types of raw materials
More informationThe Comparative Study of Machine Learning Algorithms in Text Data Classification*
The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification
More informationNearest Neighbor Classification
Nearest Neighbor Classification Charles Elkan elkan@cs.ucsd.edu October 9, 2007 The nearest-neighbor method is perhaps the simplest of all algorithms for predicting the class of a test example. The training
More informationAnalyzing Outlier Detection Techniques with Hybrid Method
Analyzing Outlier Detection Techniques with Hybrid Method Shruti Aggarwal Assistant Professor Department of Computer Science and Engineering Sri Guru Granth Sahib World University. (SGGSWU) Fatehgarh Sahib,
More informationA study of classification algorithms using Rapidminer
Volume 119 No. 12 2018, 15977-15988 ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu A study of classification algorithms using Rapidminer Dr.J.Arunadevi 1, S.Ramya 2, M.Ramesh Raja
More informationRETRACTED ARTICLE. Web-Based Data Mining in System Design and Implementation. Open Access. Jianhu Gong 1* and Jianzhi Gong 2
Send Orders for Reprints to reprints@benthamscience.ae The Open Automation and Control Systems Journal, 2014, 6, 1907-1911 1907 Web-Based Data Mining in System Design and Implementation Open Access Jianhu
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationMachine Learning Techniques for Data Mining
Machine Learning Techniques for Data Mining Eibe Frank University of Waikato New Zealand 10/25/2000 1 PART VII Moving on: Engineering the input and output 10/25/2000 2 Applying a learner is not all Already
More informationTHE preceding chapters were all devoted to the analysis of images and signals which
Chapter 5 Segmentation of Color, Texture, and Orientation Images THE preceding chapters were all devoted to the analysis of images and signals which take values in IR. It is often necessary, however, to
More informationDATE OF BIRTH SORTING (DBSORT)
DATE OF BIRTH SORTING (DBSORT) Release 3.1 December 1997 - ii - DBSORT Table of Contents 1 Changes Since Last Release... 1 2 Purpose... 3 3 Limitations... 5 3.1 Command Line Parameters... 5 4 Input...
More informationDistribution-free Predictive Approaches
Distribution-free Predictive Approaches The methods discussed in the previous sections are essentially model-based. Model-free approaches such as tree-based classification also exist and are popular for
More informationCHAPTER 3. Preprocessing and Feature Extraction. Techniques
CHAPTER 3 Preprocessing and Feature Extraction Techniques CHAPTER 3 Preprocessing and Feature Extraction Techniques 3.1 Need for Preprocessing and Feature Extraction schemes for Pattern Recognition and
More informationJarek Szlichta
Jarek Szlichta http://data.science.uoit.ca/ Approximate terminology, though there is some overlap: Data(base) operations Executing specific operations or queries over data Data mining Looking for patterns
More informationFigure (5) Kohonen Self-Organized Map
2- KOHONEN SELF-ORGANIZING MAPS (SOM) - The self-organizing neural networks assume a topological structure among the cluster units. - There are m cluster units, arranged in a one- or two-dimensional array;
More informationData Mining. Part 2. Data Understanding and Preparation. 2.4 Data Transformation. Spring Instructor: Dr. Masoud Yaghini. Data Transformation
Data Mining Part 2. Data Understanding and Preparation 2.4 Spring 2010 Instructor: Dr. Masoud Yaghini Outline Introduction Normalization Attribute Construction Aggregation Attribute Subset Selection Discretization
More informationThe k-means Algorithm and Genetic Algorithm
The k-means Algorithm and Genetic Algorithm k-means algorithm Genetic algorithm Rough set approach Fuzzy set approaches Chapter 8 2 The K-Means Algorithm The K-Means algorithm is a simple yet effective
More informationIdentifying Layout Classes for Mathematical Symbols Using Layout Context
Rochester Institute of Technology RIT Scholar Works Articles 2009 Identifying Layout Classes for Mathematical Symbols Using Layout Context Ling Ouyang Rochester Institute of Technology Richard Zanibbi
More informationLecture-17: Clustering with K-Means (Contd: DT + Random Forest)
Lecture-17: Clustering with K-Means (Contd: DT + Random Forest) Medha Vidyotma April 24, 2018 1 Contd. Random Forest For Example, if there are 50 scholars who take the measurement of the length of the
More informationWithdrawn Equity Offerings: Event Study and Cross-Sectional Regression Analysis Using Eventus Software
Withdrawn Equity Offerings: Event Study and Cross-Sectional Regression Analysis Using Eventus Software Copyright 1998-2001 Cowan Research, L.C. This note demonstrates the use of Eventus software to conduct
More informationPerformance Evaluation of Various Classification Algorithms
Performance Evaluation of Various Classification Algorithms Shafali Deora Amritsar College of Engineering & Technology, Punjab Technical University -----------------------------------------------------------***----------------------------------------------------------
More informationTCA metric #3. TCA and fair execution. The metrics that the FX industry must use.
LMAX Exchange: TCA white paper V1. - May 17 TCA metric #3 TCA and fair execution. The metrics that the FX industry must use. An analysis and comparison of common FX execution quality metrics between last
More informationData mining overview. Data Mining. Data mining overview. Data mining overview. Data mining overview. Data mining overview 3/24/2014
Data Mining Data mining processes What technological infrastructure is required? Data mining is a system of searching through large amounts of data for patterns. It is a relatively new concept which is
More informationNotes and Announcements
Notes and Announcements Midterm exam: Oct 20, Wednesday, In Class Late Homeworks Turn in hardcopies to Michelle. DO NOT ask Michelle for extensions. Note down the date and time of submission. If submitting
More informationUniversity of Florida CISE department Gator Engineering. Data Preprocessing. Dr. Sanjay Ranka
Data Preprocessing Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville ranka@cise.ufl.edu Data Preprocessing What preprocessing step can or should
More informationData Preprocessing. Slides by: Shree Jaswal
Data Preprocessing Slides by: Shree Jaswal Topics to be covered Why Preprocessing? Data Cleaning; Data Integration; Data Reduction: Attribute subset selection, Histograms, Clustering and Sampling; Data
More informationMIT 801. Machine Learning I. [Presented by Anna Bosman] 16 February 2018
MIT 801 [Presented by Anna Bosman] 16 February 2018 Machine Learning What is machine learning? Artificial Intelligence? Yes as we know it. What is intelligence? The ability to acquire and apply knowledge
More informationInternational Journal of Data Mining & Knowledge Management Process (IJDKP) Vol.7, No.3, May Dr.Zakea Il-Agure and Mr.Hicham Noureddine Itani
LINK MINING PROCESS Dr.Zakea Il-Agure and Mr.Hicham Noureddine Itani Higher Colleges of Technology, United Arab Emirates ABSTRACT Many data mining and knowledge discovery methodologies and process models
More informationMining Web Data. Lijun Zhang
Mining Web Data Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Web Crawling and Resource Discovery Search Engine Indexing and Query Processing Ranking Algorithms Recommender Systems
More informationSYS 6021 Linear Statistical Models
SYS 6021 Linear Statistical Models Project 2 Spam Filters Jinghe Zhang Summary The spambase data and time indexed counts of spams and hams are studied to develop accurate spam filters. Static models are
More informationUnsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing
Unsupervised Data Mining: Clustering Izabela Moise, Evangelos Pournaras, Dirk Helbing Izabela Moise, Evangelos Pournaras, Dirk Helbing 1 1. Supervised Data Mining Classification Regression Outlier detection
More informationFor personal use only. Update Event & nearmap Solar
Update Event & nearmap Solar Update Event and nearmap Solar Paul Peterson Senior VP Product & Engineering 2 Current Clear Change Current What s on the ground now Clear Unrivalled clarity Change Monitor
More informationNearest Neighbor Classifiers
Nearest Neighbor Classifiers TNM033 Data Mining Techniques Linköping University 2009-12-04 When I see a bird that walks like a duck and swims like a duck and quacks like a duck, I call that bird a duck.
More informationFootball result prediction using simple classification algorithms, a comparison between k-nearest Neighbor and Linear Regression
EXAMENSARBETE INOM TEKNIK, GRUNDNIVÅ, 15 HP STOCKHOLM, SVERIGE 2016 Football result prediction using simple classification algorithms, a comparison between k-nearest Neighbor and Linear Regression PIERRE
More informationClustering & Classification (chapter 15)
Clustering & Classification (chapter 5) Kai Goebel Bill Cheetham RPI/GE Global Research goebel@cs.rpi.edu cheetham@cs.rpi.edu Outline k-means Fuzzy c-means Mountain Clustering knn Fuzzy knn Hierarchical
More informationData Preprocessing. Data Preprocessing
Data Preprocessing Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville ranka@cise.ufl.edu Data Preprocessing What preprocessing step can or should
More informationChapter 4: Algorithms CS 795
Chapter 4: Algorithms CS 795 Inferring Rudimentary Rules 1R Single rule one level decision tree Pick each attribute and form a single level tree without overfitting and with minimal branches Pick that
More informationModel Assessment and Selection. Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer
Model Assessment and Selection Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Model Training data Testing data Model Testing error rate Training error
More informationClassification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University
Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate
More informationRoute Map (Start September 2012) Year 9
Route Map (Start September 2012) Year 9 3 th 7 th Sept 10 th -14 th Sept 17 th 21 st Sept 24 th 28 th Sept 1 st -5 th Oct 8 th -12 th Oct 15 th 19 th Oct 22 nd -26 th Oct 29 th Oct-2 nd Nov 5 th -9 th
More informationStructural and Syntactic Pattern Recognition
Structural and Syntactic Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent
More informationComparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini*
Comparative analysis of classifier algorithm in data mining Aikjot Kaur Narula#, Dr.Raman Maini* #Student, Department of Computer Engineering, Punjabi university Patiala, India, aikjotnarula@gmail.com
More informationGraphical Analysis of Data using Microsoft Excel [2016 Version]
Graphical Analysis of Data using Microsoft Excel [2016 Version] Introduction In several upcoming labs, a primary goal will be to determine the mathematical relationship between two variable physical parameters.
More informationChapter 2: Classification & Prediction
Chapter 2: Classification & Prediction 2.1 Basic Concepts of Classification and Prediction 2.2 Decision Tree Induction 2.3 Bayes Classification Methods 2.4 Rule Based Classification 2.4.1 The principle
More informationCSC 411: Lecture 05: Nearest Neighbors
CSC 411: Lecture 05: Nearest Neighbors Raquel Urtasun & Rich Zemel University of Toronto Sep 28, 2015 Urtasun & Zemel (UofT) CSC 411: 05-Nearest Neighbors Sep 28, 2015 1 / 13 Today Non-parametric models
More informationAdaptive Metric Nearest Neighbor Classification
Adaptive Metric Nearest Neighbor Classification Carlotta Domeniconi Jing Peng Dimitrios Gunopulos Computer Science Department Computer Science Department Computer Science Department University of California
More informationNonparametric Classification Methods
Nonparametric Classification Methods We now examine some modern, computationally intensive methods for regression and classification. Recall that the LDA approach constructs a line (or plane or hyperplane)
More informationCS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp
CS 229 Final Project - Using machine learning to enhance a collaborative filtering recommendation system for Yelp Chris Guthrie Abstract In this paper I present my investigation of machine learning as
More informationClassification Algorithms in Data Mining
August 9th, 2016 Suhas Mallesh Yash Thakkar Ashok Choudhary CIS660 Data Mining and Big Data Processing -Dr. Sunnie S. Chung Classification Algorithms in Data Mining Deciding on the classification algorithms
More informationEnhancement in Next Web Page Recommendation with the help of Multi- Attribute Weight Prophecy
2017 IJSRST Volume 3 Issue 1 Print ISSN: 2395-6011 Online ISSN: 2395-602X Themed Section: Science and Technology Enhancement in Next Web Page Recommendation with the help of Multi- Attribute Weight Prophecy
More informationA NOVEL APPROACH FOR TEST SUITE PRIORITIZATION
Journal of Computer Science 10 (1): 138-142, 2014 ISSN: 1549-3636 2014 doi:10.3844/jcssp.2014.138.142 Published Online 10 (1) 2014 (http://www.thescipub.com/jcs.toc) A NOVEL APPROACH FOR TEST SUITE PRIORITIZATION
More informationData Preprocessing. Why Data Preprocessing? MIT-652 Data Mining Applications. Chapter 3: Data Preprocessing. Multi-Dimensional Measure of Data Quality
Why Data Preprocessing? Data in the real world is dirty incomplete: lacking attribute values, lacking certain attributes of interest, or containing only aggregate data e.g., occupation = noisy: containing
More information2. On classification and related tasks
2. On classification and related tasks In this part of the course we take a concise bird s-eye view of different central tasks and concepts involved in machine learning and classification particularly.
More informationCHAPTER 5 ANT-FUZZY META HEURISTIC GENETIC SENSOR NETWORK SYSTEM FOR MULTI - SINK AGGREGATED DATA TRANSMISSION
CHAPTER 5 ANT-FUZZY META HEURISTIC GENETIC SENSOR NETWORK SYSTEM FOR MULTI - SINK AGGREGATED DATA TRANSMISSION 5.1 INTRODUCTION Generally, deployment of Wireless Sensor Network (WSN) is based on a many
More informationNearest Neighbor Predictors
Nearest Neighbor Predictors September 2, 2018 Perhaps the simplest machine learning prediction method, from a conceptual point of view, and perhaps also the most unusual, is the nearest-neighbor method,
More informationINF 4300 Classification III Anne Solberg The agenda today:
INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15
More informationData Mining: Classifier Evaluation. CSCI-B490 Seminar in Computer Science (Data Mining)
Data Mining: Classifier Evaluation CSCI-B490 Seminar in Computer Science (Data Mining) Predictor Evaluation 1. Question: how good is our algorithm? how will we estimate its performance? 2. Question: what
More information