DIFFERENCES BETWEEN THE METHOD OF MINI-MODELS AND THE K-NEAREST NEIGHBORS ON EXAMPLE OF MODELING OF UNEMPLOYMENT RATE IN POLAND

Similar documents
Nonlinear ellipsoidal mini-models application for the function approximation task

Performance Analysis of Data Mining Classification Techniques

Weighting and selection of features.

A *69>H>N6 #DJGC6A DG C<>C::G>C<,8>:C8:H /DA 'D 2:6G, ()-"&"3 -"(' ( +-" " " % '.+ % ' -0(+$,

Improving Results and Performance of Collaborative Filtering-based Recommender Systems using Cuckoo Optimization Algorithm

Relative Constraints as Features

RDM interval arithmetic for solving economic problem with uncertain variables 2

Processing Missing Values with Self-Organized Maps

CSE 626: Data mining. Instructor: Sargur N. Srihari. Phone: , ext. 113

Distribution-free Predictive Approaches

Data Imbalance Problem solving for SMOTE Based Oversampling: Study on Fault Detection Prediction Model in Semiconductor Manufacturing Process

ESTIMATING THE COST OF ENERGY USAGE IN SPORT CENTRES: A COMPARATIVE MODELLING APPROACH

Transactions on Information and Communications Technologies vol 19, 1997 WIT Press, ISSN

Liquefaction Analysis in 3D based on Neural Network Algorithm

A study of classification algorithms using Rapidminer

A Lazy Approach for Machine Learning Algorithms

Use of KNN for the Netflix Prize Ted Hong, Dimitris Tsamis Stanford University

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

SERVICE RECOMMENDATION ON WIKI-WS PLATFORM

EVALUATION OF CLASSIFICATION ABILITY OF THE PARAMETERS CHARACTERIZING STEREOMETRIC PROPERTIES OF TECHNICAL SURFACES 1.

Artificial Neural Network-Based Prediction of Human Posture

An Empirical Study of Lazy Multilabel Classification Algorithms

Relation Organization of SOM Initial Map by Improved Node Exchange

A Triangular Fuzzy Model for Assessing Problem Solving Skills

INCREASING CLASSIFICATION QUALITY BY USING FUZZY LOGIC

Using Natural Clusters Information to Build Fuzzy Indexing Structure

Proximity Prestige using Incremental Iteration in Page Rank Algorithm

Naïve Bayes for text classification

Enhancing Forecasting Performance of Naïve-Bayes Classifiers with Discretization Techniques

Global Journal of Engineering Science and Research Management

Extraction of representative learning set from measured geospatial data

An Empirical Study of Hoeffding Racing for Model Selection in k-nearest Neighbor Classification

Feature Selection with Decision Tree Criterion

Available online at ScienceDirect. Procedia Computer Science 35 (2014 )

Telecommunication and Informatics University of North Carolina, Technical University of Gdansk Charlotte, NC 28223, USA

THE IDENTIFICATION OF THE BOUNDARY GEOMETRY WITH CORNER POINTS IN INVERSE TWO-DIMENSIONAL POTENTIAL PROBLEMS

An Improved KNN Classification Algorithm based on Sampling

MetaData for Database Mining

PASS EVALUATING IN SIMULATED SOCCER DOMAIN USING ANT-MINER ALGORITHM

On the Role of Weibull-type Distributions in NHPP-based Software Reliability Modeling

Wind energy production forecasting

Binary Histogram in Image Classification for Retrieval Purposes

Use of Multi-category Proximal SVM for Data Set Reduction

PROBLEM FORMULATION AND RESEARCH METHODOLOGY

The Procedure Proposal of Manufacturing Systems Management by Using of Gained Knowledge from Production Data

Efficient Pairwise Classification

Introduction to Support Vector Machines

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

FITTING PIECEWISE LINEAR FUNCTIONS USING PARTICLE SWARM OPTIMIZATION

Constrained Optimization of the Stress Function for Multidimensional Scaling

Voronoi Region. K-means method for Signal Compression: Vector Quantization. Compression Formula 11/20/2013

Support vector machines

HYBRIDIZED MODEL FOR EFFICIENT MATCHING AND DATA PREDICTION IN INFORMATION RETRIEVAL

CHAPTER 4 STOCK PRICE PREDICTION USING MODIFIED K-NEAREST NEIGHBOR (MKNN) ALGORITHM

Nonparametric Regression

Outlier Ensembles. Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY Keynote, Outlier Detection and Description Workshop, 2013

A Learning Algorithm for Piecewise Linear Regression

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

A Comparative Study of LOWESS and RBF Approximations for Visualization

Using Genetic Algorithms to Improve Pattern Classification Performance

Constrained Clustering with Interactive Similarity Learning

An FPTAS for minimizing the product of two non-negative linear cost functions

Perceptron as a graph

Data Mining: Models and Methods

Some doubts about the objectivity of logical determination of the uniqueness of the elementary process in the Function Point Analysis

Review on Methods of Selecting Number of Hidden Nodes in Artificial Neural Network

A Comparative Study of Selected Classification Algorithms of Data Mining

Nonparametric Estimation of Distribution Function using Bezier Curve

Recommender Systems New Approaches with Netflix Dataset

A Fuzzy C-means Clustering Algorithm Based on Pseudo-nearest-neighbor Intervals for Incomplete Data

International Journal of Advanced Research in Computer Science and Software Engineering

ARMA MODEL SELECTION USING PARTICLE SWARM OPTIMIZATION AND AIC CRITERIA. Mark S. Voss a b. and Xin Feng.

Fast Associative Memory

Instance-based Learning

Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data

Using Decision Boundary to Analyze Classifiers

16.1 Runge-Kutta Method

Mathematics of Data. INFO-4604, Applied Machine Learning University of Colorado Boulder. September 5, 2017 Prof. Michael Paul

IMPROVING THE PARTICLE SWARM OPTIMIZATION ALGORITHM USING THE SIMPLEX METHOD AT LATE STAGE

Dynamic Data in terms of Data Mining Streams

On Estimation of Optimal Process Parameters of the Abrasive Waterjet Machining

Introduction to Machine Learning. Xiaojin Zhu

Predicting User Ratings Using Status Models on Amazon.com

The Effects of Dimensionality Curse in High Dimensional knn Search

Resolving the Conflict Between Competitive and Cooperative Behavior in Michigan-Type Fuzzy Classifier Systems

Using Text Learning to help Web browsing

TiP: Analyzing Periodic Time Series Patterns

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms

ARTIFICIAL NEURAL NETWORKS FOR THE COMPUTATION OF TRAFFIC QUEUES

Vocabulary: Data Distributions

Revision of a Floating-Point Genetic Algorithm GENOCOP V for Nonlinear Programming Problems

Recommender Systems (RSs)

Chap.12 Kernel methods [Book, Chap.7]

SVM Classification in Multiclass Letter Recognition System

Contributions to Economics

International Journal of Digital Application & Contemporary research Website: (Volume 1, Issue 7, February 2013)

Introduction to Supervised Learning

SECURE MULTI-KEYWORD TOP KEY RANKED SEARCH SCHEME OVER ENCRYPTED CLOUD DATA

Bootstrap Confidence Intervals for Regression Error Characteristic Curves Evaluating the Prediction Error of Software Cost Estimation Models

Individualized Error Estimation for Classification and Regression Models

Transcription:

DIFFERENCES BETWEEN THE METHOD OF MINI-MODELS AND THE K-NEAREST NEIGHBORS ON EXAMPLE OF MODELING OF UNEMPLOYMENT RATE IN POLAND Andrzej Piegat a), b), Barbara Wąsikowska b), Marcin Korzeń a) a) Faculty of Computer Science and Information Systems, Westpomeranian University of Technology, 7-2 Szczecin, Zolnierska 49, Poland b) Faculty of Economics and Management, University of Szczecin, 7- Szczecin, Mickiewicza 64, Poland Abstract. The paper presents in a possibly reader-friendly way, in the 2Dspace, the method of mini-models, which suits very good for modeling economic dependencies, where frequently a part of explanatory variables influencing the resulting variable is not known (lack of data). Experiments realized by authors confirmed superiority of mini-models over such modeling methods as polynomials, neural network, and the method of K-nearest neighbors (KNN). Because the method of mini-models is frequently mistaken for the KNN-method the authors explain in the paper the significant difference between the both competitive methods. The indicated difference is also the main reason of superiority of mini-models over the KNN-method. Accuracy of both methods has been compared experimentally on example of modeling unemployment rate in Poland and also on examples of other economic dependencies. Keywords. Quantitative methods in data analyses and modeling, modeling of business processes.. INTRODUCTION The concept and formulas for mini-models have been developed by Andrzej Piegat [6] and experiments were carried out by Barbara Wąsikowska and Marcin Korzeń. Method of mini-models serves for calculation of numerical answers (predictions) for numerical questions concerning the explained variable y for a given vector of explanatory variables x, x 2,..., x n. Thus, the task of a mini-model is the same as of a usual global model of the dependence y=f(x, x 2,..., x n ). The difference consists in the fact that the global model encircles the full space X X 2... X n of the explanatory variables and mini-model encircles only a certain local part of the full space, however this part in which the question is located.

Owing to this, the mini-model can have a simple mathematical form and its learning process can be considerably shorter than of the global model, e.g. of the global neural-network model. A mini-model can have both linear and non-linear form. In this paper the most simple linear-segment mini-models in the 2-D space will be presented. In [6] were presented triangle-linear mini-models. In 3D-space also tetragonal, pentagonal, and more complicated mini-models can be used. Figure shows example of a triangle, linear, constrained mini-model, which has the task to calculate an answer for the following question: What does y * amount to if {x *, x 2 * }={8, 8}? Y* Population Figure. Example of a mini-model learned with samples to calculate the answer for the question What does y amounts to if {x *, x 2 * }={8, 6}?. The answer calculated by the tuned mini-model is y * = 2. Source: own preparation. The aim of this paper is not to explain the construction and the algorithm for training of mini-models with samples because of the paper-volume limitation, but to explain the difference between mini-models and the KNN-method of modeling. The aim is also to explain advantages of mini-models. Mini-models learn and calculate the numerical answer very quickly in comparison with global models. The reason of this fact is that they use only limited number of learning samples that lie in the neighborhood of the question point and also because mathematical form of mini-models is very simple. If new samples are achieved they can be quickly introduced to the data bank of the problem and used to easily actualize the previous 2

mini-model and the answer of it. Global models encircle large spaces and are frequently strongly non-linear. Therefore their learning is much longer and sometimes not effective because of the local-minima phenomenon occurring in the learning process. Almost all experiments realized by the authors showed that minimodels are superior to such modeling methods as polynomials, neural network of the GRRN-type, and the KNN-method. Especially the comparison with the KNNmethod is very important because the both methods applied for regression task calculate the answer for the question: What does y * amounts to if {x *, x 2 * }={...,...}? on the basis of samples from the neighborhood of the question point {x *, x 2 * } where x * and x 2 * are numerical values of the explanatory variables x and x 2. The KNN-method has been described in many books, e.g. in [,3,5,8]. It determines the answer for a question on the basis of values of y i, i=,2,...,k of the k-nearest neighbors (samples) of the question {x *, x 2 * }. However, the neighbor samples are multiplied by weights of contribution. A common weighting scheme is to give each neighbor a weight of l/d, where d is the distance of the question point {x *, x 2 * } to the neighbor in the space of the explanatory variables. The KNNmethod is evaluated by many scientists as very effective and some of them are of the opinion that other methods are not necessary [5] ( Do we need whatever more than KNN? ). Experimental comparisons of the KNN-method and of the minimodel method have shown better accuracy of mini-models. But this superiority can also be explained essentially, what will be done in next chapters. 2. EXPERIMENTS OF MODELING THE DEPENDENCE y=f(x) IN 2D-SPACE (UNEMPLOYMENT RATE y IN POLAND AS A FUNCTION OF MONEY SUPPLY x) The question, which factors influence the unemployment rate, is important and interesting for the government of any country. Investigations on the unemployment rate in Poland in years 99-999 were made in University of Szczecin and published in [7]. The investigations shown that the unemployment rate y depends on 7 factors x, which are ranked below from the most to the less significant one: x money supply, x 2 population, x 3 dollar/zloty rate of exchange, x 4 import value, x 5 rediscount rate, x 6 household expenditures, x 7 number of high school graduates. In [6] the authors presented results of modeling the unemployment rate y =f(x, x 2 ) in 3-D space. In this paper modeling of the dependence y=f(x ) will be shown, were x was the most significant and the strongest factor influencing the unemployment rate in Poland in years 99-999. The strength of this factor is testified by the fact that it delivers modeling errors of the order of few percents. 3

Below there are shown modeling errors that were achieved with various modeling methods: a) modeling with polynomials y = a + a x + a 2 x 2 + a 3 x 3 + a 4 x 4 + a 5 x 5 of order 5, the mean absolute error (MAE) =.725 b) modeling with the GRNN-neural network (General Regression Neural Networks), MAE=.47 c) modeling with the K-nearest neighbors (KNN) method, MAE=.93. d) modeling with linear-segment mini-models, MAE=.2. Also in other experiments of modeling, which will not be presented here, minimodels achieved larger accuracy than other modeling methods. In the next chapter there will be presented 3 steps of the learning process of the mini-model in 2Dspace to enable better understanding of this process. 3. ILLUSTRATIVE EXAMPLE OF THE MINI-MODEL LEARNING IN THE 2D- SPACE The task of the mini-model in the presented example was to calculate the answer for the question: What does the unemployment rate y * amounts to if the money supply x * =,28? To calculate the possibly credible answer the mini-model has at first to tune itself on the basis of samples of the dependence y = f(x), which encircle the question point QP = {8,?} in the X-space, Figure 2..9.7.5.3 QP=( 8,?). x* = 8..3.5.7.9 Figure 2. Example of a mini-model learned with samples to calculate the answer for the question What does y * amounts to if x * = 8?. Source: own preparation. In the learning process with samples the mini-model have to find its optimal 4

length, gradient (inclination) and position in the X Y-space. These are 3 factors that have to be determined. In the case of the KNN-method only one factor (position) is determined. Only when all the 3 factors have been determined the mini-model calculates the answer y * for the question point x *. In Figures 3a, 3b, 3c three steps chosen from many steps of the learning process in 2D-space were shown. Figure 3a presents the start position of the linear-segment mini-model. The start position, inclination and length of the segment was chosen by the computer program by chance. The only condition which has to be satisfied is that the question point (x * ), which in this case is {8} has always be encircled by the mini-model range. a).9.7.5.3 QP=(8,?). b)..3.5.7.9.2 QP=(8,?) -..3.5.7.9 5

c).9.7 The answer of mini-model: x*=8, y*=4 The mini-model: y=.576-3.288*x.5.3 The precise answer: x=8, y=2...3.5.7.9 Figure 3a. The start position of the mini-model which has to calculate the answer for the question What does y amounts to if x * = 8?. Figure 3b. One of the intermediate positions of the mini-models assumed by it in the learning process. Figure 3c. The end, optimal position of the mini-models, y=.58 3.29x, QP = [8, 4]. Source: own preparation. After finishing the learning process the mini-model calculated the answer y * =,64. The precise and correct answer was in this case known (y =,62) because the question referred to one of the samples, which was taken out of the learning process (the sample gave values of unemployment rate and money supply from July 995). The method of mini-models also was tested with the known crossvalidation method leave one out [2] for each of the 96 samples representing particular months of the period 99-999. The mean absolute error (MAE) was equal to.2 at the variables x and y normalized to the interval [, ]. In the next chapter the difference between the method of mini-models and of KNN will be discussed. 4. DIFFERENCES BETWEEN MINI-MODELS AND THE K-NEAREST NEIGHBOR METHOD The method of mini-models calculates the numerical value of the explained variable y * on the basis of samples, which have the x-value positioned near to the 6

x-value of the question. For this reason the Reader may mistake this method as a method being identical with the KNN-one, that is commonly known in the scientific world, good tested and verified. Therefore using mini-models could appear nonsensical. Such opinion would be highly erroneous because, as it will be shown, the mini-model method is a much more refined one and in most cases delivers more precise calculation results than the KNN method. To enable the Reader better understanding mini-models specially the 2-dimensional dependence y=f(x) has been chosen. In Figure 4a the example question What does y amounts to if x * = 2? and the answer y * calculated by the KNN-method is presented. a).9.7 The KNN method.5.3. The answer of KNN method: x*=2, y*=66 The neighborhood The question point x*=2, y*=?..3.5.7.9 b).2 The method of mini-models The mini-model: y = -.38 +.97 * x The question point x*=2, y*=? -..3.5.7.9 7

c).2 The method of mini-models The answer of mini-model: x*=2, y*=54 The question point x*=2, y*=? -..3.5.7.9 Figure 4. Illustration of differences between the KNN (figure 4a) and the mini-model method (figure 4b, 4c) in calculating the predicted y-value in the case of existence of information gaps in learning samples. Source: own preparation. The question point x * = 2 presented in Figure 4 is positioned in an information gap, in the range with local lack of learning samples. However, the point is positioned nearer the right side than the left side of the gap. In this situation the KNN-method calculates the answer y * for the question point QP=[8,?] mainly on the basis of the nearest samples, which are situated on the right side of the question. The calculated answer (prediction) is y * = 66. As can be seen the KNN-method does not take under consideration (or takes very weakly) the learning samples that lie on the left side of the question point, on the left side of the information gap. However, the mini-model method does it, Figure 4b. Therefore its result, taking only human intuition into account, is more credible. This credibility and accuracy of mini-models can not only intuitively but also objectively be verified with the cross-validation method leave one out [2] when for the query point one of samples is chosen and query is positioned among x * samples. It this case accurate answer y * should be equal to the position of y sample. For testing both methods a sample representing December 994 was chosen. Both methods were given the task to calculate answer for question point equal to the x-value of the sample taken out. The results achieved are shown 8

below. The tested sample: x (money supply) = 5, y (unemployment rate) =. The question: What does y * amounts to if x * = 5? The KNN-method: the answer y * =.4, the mean absolute error MAE = 296 The mini-model method: the answer y * =.774, MAE =.375. The above example shows that the KNN-method is not effective regression method in information gaps contrary to mini-models. If the question point lies nearer to one border of an information gap the KNN-method chooses for calculation of the answer y samples of the nearest border. Thus, this method loses information about the tendency of changes dy/dx of the modeled dependence. And the mini-model method identifies this tendency and takes it into account at calculating the answer y *. Additionally, the length of the mini-model is the third information identified by the mini-model (information about the validity range of the mini-model) apart from its position and inclination dy/dx. Thus, the information delivered by the mini-model is much richer than that delivered by the KNN-method. Additionally, the superiority of mini-models is supported by the fact that in the cross-validation test for all 96 samples (not only for one sample as in the example above) the mini-model method has shown in most experiments higher prediction precision than the KNN-method. In the experiment described in this paper the mean absolute error MAE equals for the mini-model.2 whereas the KNN method had the error equal to.93 for the unemployment rate y normalized to the interval [,]. On example of samples shown in Figure one could think that existence of information gaps in samples is a rare event. Unfortunately, the truth is the other way round. In the case of dependence y=f(x) investigated in this paper, 96 samples seems to be a considerable number, which tightly covers the interval [, ] of variable x (normalized money supply). However, when the unemployment rate is modeled in 3D space, y=f(x, x 2 ), where x money supply, x 2 population number, the same 96 samples have to cover the area [, ] [, ] of the input space X X 2. This number of samples covers input space only fragmentary and a larger part of this space is without samples and creates great information gap. Then no questions concerning the unemployment rate are allowed in this space part. Participation of the space without any samples in the full input space drastically increases with the number of explanatory variables x taken into account in modeling, e.g. when the unemployment rate is modeled in 4D or larger space. As investigations described in [4] have shown modeling multidimensional dependencies almost always is connected with considerable information gaps. For this reason the mini-model method increases its superiority over the KNN-method owing to its capability of identifying tendencies dy/dx in samples and of the minimodel-lengths. 9

5. CONCLUSION The mini-model method is an effective method of modeling that gives good prediction results in regression tasks. The authors have not tested classification applications of this method until now. Mini-models have some similarity with the popular KNN-method and can easily be mistaken with it. The paper presented main differences between both methods and also explained why mini-models are more precise than models delivered by the KNN-method. The authors have investigated modeling capabilities of mini-models in 2D- and 3D-space and at present they are preparing computer programs and experiments with this method for higher spaces. REFERENCES [] Cherkassky V., Mulier F. (27) Learning from data, Wiley-Interscience, IEEE Press. [2] Geissler S. (993) Predictive inference. Chapman and Hall, New York. [3] Hand D., Mannila H., Smyth P. (2) Principles of data mining. Massachusetts Institute of Technology, USA. [4] Klęsk P. (28) Construction of a neurofuzzy network capable of extrapolating (and interpolating) with respect to the convex hull of a set of input samples in Rn. IEEE Transactions on Fuzzy Systems, vol 6, issue 5, pp.6-79. [5] Kordos M., Blachnik M., Strzempa D. (2) Do we need whatever more than k-nn? Proceedings of -th International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, June 3-7, 2, Springer, Germany, pp.44-42. [6] Piegat A., Wąsikowska B., Korzeń M. (2) Zastosowania samouczącego się, 3-punktowego minimodelu do modelowania stopy bezrobocia w Polsce. Artykuł ukaże się w 2 roku w zeszycie naukowym Studia Informatica Uniwersytetu Szczecińskiego. [7] Rejer I. (23) Metody modelowania wielkowymiarowego systemu z użyciem metod sztucznej inteligencji na przykładzie bezrobocia w Polsce. Wydawnictwo Naukowe Uniwersytetu Szczecińskiego, Szczecin. [8] Witten I.A., Frank E. (25) Data mining. Morgan Kaufmann Publishers, San Francisco.