DATA DEPTH AND ITS APPLICATIONS IN CLASSIFICATION

Size: px
Start display at page:

Download "DATA DEPTH AND ITS APPLICATIONS IN CLASSIFICATION"

Transcription

1 DATA DEPTH AND ITS APPLICATIONS IN CLASSIFICATION Ondrej Vencalek Department of Mathematical Analysis and Applications of Mathematics Palacky University Olomouc, CZECH REPUBLIC Abstract Concept of data depth provides one possible approach to the analysis of multivariate data. Among other it can be also used for classification purposes. The present paper summarizes main ideas how the depth can be used in classification. Two step scheme of the typical depth-based classifiers is discussed. Different choices of the depth function and classification procedure used in this scheme lead to different classifiers with different properties. Different distributional assumptions might lead to the preference of either global or local depth functions and classification procedures. 1 Introduction Depth function is basically any function which provides ordering (or quasi-ordering) of points in multidimensional space. Existence of ordering enables generalization of many nonparametric techniques proposed for univariate variables and thus the data depth creates one possible basis of nonparametric multivariate data analysis. Data depth has been also applied in classification. The aim of classification is to create a rule for allocation of new observations into one of two (or more) groups. Several such rules, known as classifiers, based on data depth were proposed since The aim of the present paper is to summarize main ideas how the depth can be used in classification and to present new trends in this area. 2 Concept of data depth There are several depth functions commonly used in applications halfspace depth, projection depth, spatial depth, Mahalanobis depth, zonoid depth, simplicial depth and some others. Their survey can be found in [8]. We recall here the first three of the depth functions listed above since they are used most frequently in context of classification: The halfspace depth of a point x in R d with respect to a probability measure P is defined as the minimum probability mass carried by any closed halfspace containing x, that is D(x, P ) = inf H { } P(H) : H a closed halfspace in R d : x H. 74

2 The projection depth of a point x in R d with respect to a probability measure P is defined as D(x, P ) = 1 u T x µ Pu, where O(x, P ) = sup, 1 + O(x, P ) u =1 σ Pu where µ Pu is some location and σ Pu is some scale measure of distribution of random vector u T X, usually the median and MAD. The spatial depth (also called L 1 -depth) of a point x in R d with respect to a probability measure P with variance matrix Σ is defined as Σ 1/2 (x X) D(x, P ) = 1 E Σ 1/2. (x X) All three depth functions listed above have desirable properties like affine invariance, maximality at a point of symmetry (if the distribution is symmetric in some sense, e.g. angularly), monotonicity on rays from the point with the maximal depth so called deepest point, which can be considered as multivariate analogy to median. See [10] for the discussion of desirable properties of the depth functions. The very important difference among the considered depth functions consists in different behaviour of their empirical versions. While the empirical halfspace depth is equal to zero for any point which lies outside of the convex hull of the data, the empirical projection depth (as well as empirical spatial depth) are nonzero everywhere. The depth of a point is a characteristic of the point specifying its centrality or outlyingness with respect to the considered distribution. Since the whole distribution is considered, the depth is said to be a global characteristic of the point. However, in recent years there have been attempts to localize depth, see [7] or [3]. Later we will discuss importance of these attempts for classification purposes. 3 Maximal depth classifier Let us first shortly recall the two-classes classification problem. We consider two unknown absolutely continuous probability distributions P 1 and P 2 on R d. Independent random samples from these distributions are available. Together they constitute so called training set. Empirical distributions based on the training set are denoted P 1 and P 2. The class to which the observation x is assigned is denoted by d(x). The maximal depth classifier the very first depth-based classifier is simple: d(x) = arg max i=1,2 D(x; P i ) (1) The points close to the centre (multivariate median) of some distribution have high depth with respect to this distribution and it seems to be natural to classify them to 75

3 Figure 1: Scheme of typical depth-based classifier. this distribution. The idea of the maximal depth classifier is thus in accordance to common sense. However Ghosh and Chaudhuri [2] proved that the maximal depth attains minimal possible probability of misclassification (known as Bayes risk) only in very special cases they showed optimality when the considered distributions are elliptically symmetric with the density decreasing from the centre, differing only in location and having equal prior probabilities. The optimality is lost even if only one of these assumptions is not fulfilled. 4 Advanced depth-based classifiers The paper by Ghosh and Chaudhuri [2] started the search for depth-based classifiers which would be applicable in broader class of distributional settings. Typical depthbased classifier can be described as a two-steps procedure: 1. The first step consists in computation of depths of the new observation x with respect to both parts of the training set. Each point is characterized by a pair of depths, these pairs lies in so called DD-space (depth-versus-depth space). Typically the DD-space is subset of [0, 1] [0, 1] R 2 and thus the first step can be usually considered as reduction of dimensionality from R d to the compact subset of R 2. This step is connected with the question Which depth function should be used? 2. The second step consists in application of some classification procedure in the DDspace. This step is connected with the question Which classification procedure should be applied in the DD-space? The scheme of typical depth-based classification procedure is shown in Figure 1. The difference among the classifiers proposed in literature consists mainly in different answers to the two questions connected with the two steps different depth functions can be applied in the first step and different classification procedures can be applied in the second step. 4.1 Which depth function should be used The classifiers which use any global depth function perform well only if the considered distributions have some global properties like symmetry or unimodality. In more gen- 76

4 eral settings, for example in the case of multimodality or nonconvexity of levelsets of density, some local depth should be used - see e.g. [1] or [4]. 4.2 Which classification procedure should be applied in the DD-space Procedures that are applied in the DD-space can be also either global or local in nature. As the global we denote the procedures that take into account all points of the training set when constructing the classifier while the local procedures take into account only points of the training set close to the point which is classified. Let us mention at least some of the both global and local procedures: The DD-classifier proposed by Li et al. [6] belongs to the global depth-based classifiers. The idea of the classifier is to separate the groups in DD-space by line (or more generally by a polynom) to minimize number of errors when classifying points from the training set. The DD-alpha procedure proposed by Lange et al. [5] is another global depthbased classifier. Instead of the pair [D(x, P 1 ), D(x, P 2 )], it works with a vector z := [D(x, P 1 ), D(x, P 2 ), D(x, P 1 ) D(x, P 2 ), D(x, P 1 ) 2, D(x, P 2 ) 2 ]. Lange et al. proposed a heuristic for finding proper parameters which specify separating hyperplane given by the equation ad(x, P 1 ) + bd(x, P 2 ) + cd(x, P 1 )D(x, P 2 ) + dd(x, P 1 ) 2 + ed(x, P 2 ) 2 = 0. The procedure was successfully tested on many real datasets leading usually to low misclassification rates. The procedure was implemented in the R-package ddalpha. The classifier which uses kernel density estimation proposed by Ghosh and Chaudhuri [2] is a local depth-based classifier. In the case of elliptically symmetric distribution the classifier which minimizes probability of misclassification can be expressed as d(x) = arg max i=1,2 π iθ i (D(x, P i )), where θ i, i = 1, 2 are some unknown real functions that can be estimated by kernel density estimation. This procedure can be used even if the distributions are nonelliptical. The k depth-nearest neighbour classifier used by Vencalek [9] is another local depth-based procedure. The idea is quite simple to use the well known k- nearest neighbour procedure in the DD-space. The question is which metric should be used to measure distances between distinct points. 5 Conclusion The data depth provides basis for nonparametric inference on multidimensional data. It can be also applied in classification. Although one can expect broad applicability of 77

5 the nonparametric depth-based classifiers their optimality can be guaranteed usually only under some restrictive assumptions. Global depth functions and global classification techniques applied on the DD-space lead to good results only if the considered distributions have some global properties like unimodality. In more general settings localization is needed one can use local depth functions or local classifiers used in the DD-space. Acknowledgement The research was supported by the grant of Czech Science Foundation GA S. References [1] Dutta, S., Chaudhuri, P., Ghosh, A. K. (2012). Classification using Localized Spatial Depth with Multiple Localization. Communicated for publication. [2] Ghosh, A. K., Chaudhuri, P. (2005). On maximum depth and related classifiers. Scandinavian Journal of Statistics. Vol. 32, pp [3] Hlubinka, D., Kotik, L., Vencalek, O. (2010). Weighted data depth. Kybernetika, Vol. 46, No. 1, pp [4] Hlubinka, D., Venclek, O. (2013). Depth-Based Classification for Distributions with Nonconvex Support. Journal of Probability and Statistics [5] Lange, T., Mosler, K., Mozharovskyi, P. (2014). Fast nonparametric classification based on data depth. Statistical Papers. Vol. 55, No. 1, pp [6] Li, J., Cuesta-Albertos, J. A., Liu, R. (2012). DD-classifier: nonparametric classification procedure based on DD-plot. Journal of the American Statistical Association. Vol. 107, No. 498, pp [7] Paindaveine, D., Van Bever, G. (2013). From depth to local depth: a focus on centrality. Journal of the American Statistical Association. Vol. 108, No. 503, pp [8] Vencalek, O. (2011). Weighted data depth and depth based classification. PhD thesis. URL: venco2am/datadepth.html [9] Vencalek, O. (2014). New depth-based modification of the k-nearest neighbour method. SOP Transactions on Statistics and Analysis, [10] Zuo, Y., Serfling, R. (2000). General notion of statistical depth function. Annals of Statistics. Vol. 28, pp

NONPARAMETRIC CLASSIFICATION OF DIRECTIONAL DATA THROUGH DEPTH FUNCTIONS

NONPARAMETRIC CLASSIFICATION OF DIRECTIONAL DATA THROUGH DEPTH FUNCTIONS NONPARAMETRIC CLASSIFICATION OF DIRECTIONAL DATA THROUGH DEPTH FUNCTIONS Houyem Demni 1 Amor Messaoud 1 Giovanni C.Porzio 2 1 Tunis Business School University of Tunis, Tunisia 2 University of Cassino,

More information

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution

More information

IcETRAN Proceedings of papers. International Conference on Electrical, Electronic and Computing Engineering

IcETRAN Proceedings of papers. International Conference on Electrical, Electronic and Computing Engineering IcETRAN 2015 Proceedings of papers International Conference on Electrical, Electronic and Computing Engineering Silver Lake (Srebrno jezero), Serbia, June 8 11, 2015 ISBN 978-86-80509-71-6 Data Centrality

More information

The alpha-procedure - a nonparametric invariant method for automatic classification of multi-dimensional objects

The alpha-procedure - a nonparametric invariant method for automatic classification of multi-dimensional objects The alpha-procedure - a nonparametric invariant method for automatic classification of multi-dimensional objects Tatjana Lange Pavlo Mozharovskyi Hochschule Merseburg, 06217 Merseburg, Germany Universität

More information

Contributions to depth-based classification and computation of the Tukey depth

Contributions to depth-based classification and computation of the Tukey depth Contributions to depth-based classification and computation of the Tukey depth Inauguraldissertation zur Erlangung des Doktorgrades der Wirtschafts- und Sozialwissenschaftlichen Fakultät der Universität

More information

Shape fitting and non convex data analysis

Shape fitting and non convex data analysis Shape fitting and non convex data analysis Petra Surynková, Zbyněk Šír Faculty of Mathematics and Physics, Charles University in Prague Sokolovská 83, 186 7 Praha 8, Czech Republic email: petra.surynkova@mff.cuni.cz,

More information

ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART

ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART ECONOMIC DESIGN OF STATISTICAL PROCESS CONTROL USING PRINCIPAL COMPONENTS ANALYSIS AND THE SIMPLICIAL DEPTH RANK CONTROL CHART Vadhana Jayathavaj Rangsit University, Thailand vadhana.j@rsu.ac.th Adisak

More information

A new depth-based approach for detecting outlying curves

A new depth-based approach for detecting outlying curves A new depth-based approach for detecting outlying curves Mia Hubert, Department of Mathematics & LStat, KU Leuven, mia.hubert@wis.kuleuven.be Gerda Claeskens, ORSTAT & LStat, KU Leuven, gerda.claeskens@econ.kuleuven.be

More information

On Liu s Simplicial Depth and Randles Interdirections

On Liu s Simplicial Depth and Randles Interdirections On Liu s Simplicial Depth and Randles Interdirections Robert Serfling a,, Yunfei Wang a a Department of Mathematical Sciences, University of Texas at Dallas, Richardson, Texas 75080, USA Abstract At about

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

Chart 3. Synthetic indicator for trade and trade output (year to year) DATA DEPTH CONCEPT IN MULTIVARIATE TIME SERIES

Chart 3. Synthetic indicator for trade and trade output (year to year) DATA DEPTH CONCEPT IN MULTIVARIATE TIME SERIES Chart 3. Synthetic indicator for trade and trade output (year to year) Conclusion: The synthetic indicators created for Ukrainian economy sectors on the basis of BTS indicators much better reflect the

More information

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures: Homework Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression 3.0-3.2 Pod-cast lecture on-line Next lectures: I posted a rough plan. It is flexible though so please come with suggestions Bayes

More information

Depth Functions in Multivariate & Other Data Settings: Concepts, Perspectives, Challenges

Depth Functions in Multivariate & Other Data Settings: Concepts, Perspectives, Challenges Depth Functions in Multivariate & Other Data Settings: Concepts, Perspectives, Challenges Robert Serfling 1 Serfling & Thompson Statistical Consulting and Tutoring PIMS Lecture University of Manitoba,

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

Density estimation. In density estimation problems, we are given a random from an unknown density. Our objective is to estimate

Density estimation. In density estimation problems, we are given a random from an unknown density. Our objective is to estimate Density estimation In density estimation problems, we are given a random sample from an unknown density Our objective is to estimate? Applications Classification If we estimate the density for each class,

More information

INF 4300 Classification III Anne Solberg The agenda today:

INF 4300 Classification III Anne Solberg The agenda today: INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 6

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 6 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 6 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 19, 2012 Andre Tkacenko

More information

Computational Statistics and Mathematics for Cyber Security

Computational Statistics and Mathematics for Cyber Security and Mathematics for Cyber Security David J. Marchette Sept, 0 Acknowledgment: This work funded in part by the NSWC In-House Laboratory Independent Research (ILIR) program. NSWCDD-PN--00 Topics NSWCDD-PN--00

More information

Interpolation is a basic tool used extensively in tasks such as zooming, shrinking, rotating, and geometric corrections.

Interpolation is a basic tool used extensively in tasks such as zooming, shrinking, rotating, and geometric corrections. Image Interpolation 48 Interpolation is a basic tool used extensively in tasks such as zooming, shrinking, rotating, and geometric corrections. Fundamentally, interpolation is the process of using known

More information

MSA220 - Statistical Learning for Big Data

MSA220 - Statistical Learning for Big Data MSA220 - Statistical Learning for Big Data Lecture 13 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Clustering Explorative analysis - finding groups

More information

Dimension Reduction of Image Manifolds

Dimension Reduction of Image Manifolds Dimension Reduction of Image Manifolds Arian Maleki Department of Electrical Engineering Stanford University Stanford, CA, 9435, USA E-mail: arianm@stanford.edu I. INTRODUCTION Dimension reduction of datasets

More information

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster

More information

A Fast Clustering Algorithm with Application to Cosmology. Woncheol Jang

A Fast Clustering Algorithm with Application to Cosmology. Woncheol Jang A Fast Clustering Algorithm with Application to Cosmology Woncheol Jang May 5, 2004 Abstract We present a fast clustering algorithm for density contour clusters (Hartigan, 1975) that is a modified version

More information

Density estimation. In density estimation problems, we are given a random from an unknown density. Our objective is to estimate

Density estimation. In density estimation problems, we are given a random from an unknown density. Our objective is to estimate Density estimation In density estimation problems, we are given a random sample from an unknown density Our objective is to estimate? Applications Classification If we estimate the density for each class,

More information

Evgeny Maksakov Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages:

Evgeny Maksakov Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages: Advantages and disadvantages: Today Problems with visualizing high dimensional data Problem Overview Direct Visualization Approaches High dimensionality Visual cluttering Clarity of representation Visualization is time consuming Dimensional

More information

Convexity: an introduction

Convexity: an introduction Convexity: an introduction Geir Dahl CMA, Dept. of Mathematics and Dept. of Informatics University of Oslo 1 / 74 1. Introduction 1. Introduction what is convexity where does it arise main concepts and

More information

Two approaches for solving tasks of pattern recognition and reconstruction of functional dependencies

Two approaches for solving tasks of pattern recognition and reconstruction of functional dependencies Two approaches for solving tasks of pattern recognition and reconstruction of functional dependencies Tatjana Lange 1, Pavlo Mozharovskyi 2, Gabor Barath 3 1,3 University of Applied Sciences, Merseburg,

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

Expectation-Maximization. Nuno Vasconcelos ECE Department, UCSD

Expectation-Maximization. Nuno Vasconcelos ECE Department, UCSD Expectation-Maximization Nuno Vasconcelos ECE Department, UCSD Plan for today last time we started talking about mixture models we introduced the main ideas behind EM to motivate EM, we looked at classification-maximization

More information

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2015

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2015 Instance-based Learning CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2015 Outline Non-parametric approach Unsupervised: Non-parametric density estimation Parzen Windows K-Nearest

More information

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering Introduction to Pattern Recognition Part II Selim Aksoy Bilkent University Department of Computer Engineering saksoy@cs.bilkent.edu.tr RETINA Pattern Recognition Tutorial, Summer 2005 Overview Statistical

More information

Conditional Volatility Estimation by. Conditional Quantile Autoregression

Conditional Volatility Estimation by. Conditional Quantile Autoregression International Journal of Mathematical Analysis Vol. 8, 2014, no. 41, 2033-2046 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ijma.2014.47210 Conditional Volatility Estimation by Conditional Quantile

More information

arxiv: v10 [stat.co] 31 Mar 2018

arxiv: v10 [stat.co] 31 Mar 2018 DepthProc: An R Package for Robust Exploration of Multidimensional Economic Phenomena Daniel Kosiorowski Cracow University of Economics Zygmunt Zawadzki Cracow University of Economics Abstract arxiv:1408.4542v10

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Restricted-Orientation Convexity in Higher-Dimensional Spaces

Restricted-Orientation Convexity in Higher-Dimensional Spaces Restricted-Orientation Convexity in Higher-Dimensional Spaces ABSTRACT Eugene Fink Derick Wood University of Waterloo, Waterloo, Ont, Canada N2L3G1 {efink, dwood}@violetwaterlooedu A restricted-oriented

More information

Xin Dang 1. University of Mississippi. and. Robert Serfling 2. University of Texas at Dallas. September 2009

Xin Dang 1. University of Mississippi. and. Robert Serfling 2. University of Texas at Dallas. September 2009 A numerical study of multiple imputation methods using nonparametric multivariate outlier identifiers and depth-based performance criteria with clinical laboratory data Xin Dang 1 University of Mississippi

More information

Mixture Models and the EM Algorithm

Mixture Models and the EM Algorithm Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is

More information

Machine Learning : Clustering, Self-Organizing Maps

Machine Learning : Clustering, Self-Organizing Maps Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps Clustering The task: partition a set of objects into meaningful subsets (clusters). The

More information

9881 Project Deliverable #1

9881 Project Deliverable #1 Memorial University of Newfoundland Pattern Recognition Lecture 3 May 9, 2006 http://www.engr.mun.ca/~charlesr Office Hours: Tuesdays & Thursdays 8:30-9:30 PM EN-3026 9881 Project Deliverable #1 Report

More information

Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection

Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection Flexible-Hybrid Sequential Floating Search in Statistical Feature Selection Petr Somol 1,2, Jana Novovičová 1,2, and Pavel Pudil 2,1 1 Dept. of Pattern Recognition, Institute of Information Theory and

More information

General properties of staircase and convex dual feasible functions

General properties of staircase and convex dual feasible functions General properties of staircase and convex dual feasible functions JÜRGEN RIETZ, CLÁUDIO ALVES, J. M. VALÉRIO de CARVALHO Centro de Investigação Algoritmi da Universidade do Minho, Escola de Engenharia

More information

arxiv: v3 [stat.me] 7 Jul 2016

arxiv: v3 [stat.me] 7 Jul 2016 Multivariate and functional classification using depth and distance Mia Hubert, Peter J. Rousseeuw, Pieter Segaert arxiv:1504.01128v3 [stat.me] 7 Jul 2016 July 7, 2016 Abstract We construct classifiers

More information

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences UNIVERSITY OF OSLO Faculty of Mathematics and Natural Sciences Exam: INF 4300 / INF 9305 Digital image analysis Date: Thursday December 21, 2017 Exam hours: 09.00-13.00 (4 hours) Number of pages: 8 pages

More information

OUTLIER MINING IN HIGH DIMENSIONAL DATASETS

OUTLIER MINING IN HIGH DIMENSIONAL DATASETS OUTLIER MINING IN HIGH DIMENSIONAL DATASETS DATA MINING DISCUSSION GROUP OUTLINE MOTIVATION OUTLIERS IN MULTIVARIATE DATA OUTLIERS IN HIGH DIMENSIONAL DATA Distribution-based Distance-based NN-based Density-based

More information

MS1b Statistical Data Mining Part 3: Supervised Learning Nonparametric Methods

MS1b Statistical Data Mining Part 3: Supervised Learning Nonparametric Methods MS1b Statistical Data Mining Part 3: Supervised Learning Nonparametric Methods Yee Whye Teh Department of Statistics Oxford http://www.stats.ox.ac.uk/~teh/datamining.html Outline Supervised Learning: Nonparametric

More information

Some questions of consensus building using co-association

Some questions of consensus building using co-association Some questions of consensus building using co-association VITALIY TAYANOV Polish-Japanese High School of Computer Technics Aleja Legionow, 4190, Bytom POLAND vtayanov@yahoo.com Abstract: In this paper

More information

A noninformative Bayesian approach to small area estimation

A noninformative Bayesian approach to small area estimation A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported

More information

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions Thomas Giraud Simon Chabot October 12, 2013 Contents 1 Discriminant analysis 3 1.1 Main idea................................

More information

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves

More information

Summer School in Statistics for Astronomers & Physicists June 15-17, Cluster Analysis

Summer School in Statistics for Astronomers & Physicists June 15-17, Cluster Analysis Summer School in Statistics for Astronomers & Physicists June 15-17, 2005 Session on Computational Algorithms for Astrostatistics Cluster Analysis Max Buot Department of Statistics Carnegie-Mellon University

More information

Discovery of Agricultural Patterns Using Parallel Hybrid Clustering Paradigm

Discovery of Agricultural Patterns Using Parallel Hybrid Clustering Paradigm IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 PP 10-15 www.iosrjen.org Discovery of Agricultural Patterns Using Parallel Hybrid Clustering Paradigm P.Arun, M.Phil, Dr.A.Senthilkumar

More information

A Scale Invariant Clustering using Minimum Volume Ellipsoids

A Scale Invariant Clustering using Minimum Volume Ellipsoids A Scale Invariant Clustering using Minimum Volume Ellipsoids Mahesh Kumar MSIS Dept., Rutgers Business School Rutgers University, Newark, NJ 07102 maheshk@rutgers.edu James B. Orlin Operations Research

More information

Machine Learning. Nonparametric methods for Classification. Eric Xing , Fall Lecture 2, September 12, 2016

Machine Learning. Nonparametric methods for Classification. Eric Xing , Fall Lecture 2, September 12, 2016 Machine Learning 10-701, Fall 2016 Nonparametric methods for Classification Eric Xing Lecture 2, September 12, 2016 Reading: 1 Classification Representing data: Hypothesis (classifier) 2 Clustering 3 Supervised

More information

MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS

MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS MODELLING DOCUMENT CATEGORIES BY EVOLUTIONARY LEARNING OF TEXT CENTROIDS J.I. Serrano M.D. Del Castillo Instituto de Automática Industrial CSIC. Ctra. Campo Real km.0 200. La Poveda. Arganda del Rey. 28500

More information

Identifying Layout Classes for Mathematical Symbols Using Layout Context

Identifying Layout Classes for Mathematical Symbols Using Layout Context Rochester Institute of Technology RIT Scholar Works Articles 2009 Identifying Layout Classes for Mathematical Symbols Using Layout Context Ling Ouyang Rochester Institute of Technology Richard Zanibbi

More information

A PROOF OF THE LOWER BOUND CONJECTURE FOR CONVEX POLYTOPES

A PROOF OF THE LOWER BOUND CONJECTURE FOR CONVEX POLYTOPES PACIFIC JOURNAL OF MATHEMATICS Vol. 46, No. 2, 1973 A PROOF OF THE LOWER BOUND CONJECTURE FOR CONVEX POLYTOPES DAVID BARNETTE A d polytope is defined to be a cz-dimensional set that is the convex hull

More information

Instance-based Learning

Instance-based Learning Instance-based Learning Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University February 19 th, 2007 2005-2007 Carlos Guestrin 1 Why not just use Linear Regression? 2005-2007 Carlos Guestrin

More information

Lecture 3: Convex sets

Lecture 3: Convex sets Lecture 3: Convex sets Rajat Mittal IIT Kanpur We denote the set of real numbers as R. Most of the time we will be working with space R n and its elements will be called vectors. Remember that a subspace

More information

Robust Shape Retrieval Using Maximum Likelihood Theory

Robust Shape Retrieval Using Maximum Likelihood Theory Robust Shape Retrieval Using Maximum Likelihood Theory Naif Alajlan 1, Paul Fieguth 2, and Mohamed Kamel 1 1 PAMI Lab, E & CE Dept., UW, Waterloo, ON, N2L 3G1, Canada. naif, mkamel@pami.uwaterloo.ca 2

More information

We use non-bold capital letters for all random variables in these notes, whether they are scalar-, vector-, matrix-, or whatever-valued.

We use non-bold capital letters for all random variables in these notes, whether they are scalar-, vector-, matrix-, or whatever-valued. The Bayes Classifier We have been starting to look at the supervised classification problem: we are given data (x i, y i ) for i = 1,..., n, where x i R d, and y i {1,..., K}. In this section, we suppose

More information

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering.

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering. Chapter 4 Fuzzy Segmentation 4. Introduction. The segmentation of objects whose color-composition is not common represents a difficult task, due to the illumination and the appropriate threshold selection

More information

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop Machine Learning Algorithms (IFT6266 A7) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

More information

A Feature Selection Method to Handle Imbalanced Data in Text Classification

A Feature Selection Method to Handle Imbalanced Data in Text Classification A Feature Selection Method to Handle Imbalanced Data in Text Classification Fengxiang Chang 1*, Jun Guo 1, Weiran Xu 1, Kejun Yao 2 1 School of Information and Communication Engineering Beijing University

More information

Supervised Learning: K-Nearest Neighbors and Decision Trees

Supervised Learning: K-Nearest Neighbors and Decision Trees Supervised Learning: K-Nearest Neighbors and Decision Trees Piyush Rai CS5350/6350: Machine Learning August 25, 2011 (CS5350/6350) K-NN and DT August 25, 2011 1 / 20 Supervised Learning Given training

More information

Chapter 4: Non-Parametric Techniques

Chapter 4: Non-Parametric Techniques Chapter 4: Non-Parametric Techniques Introduction Density Estimation Parzen Windows Kn-Nearest Neighbor Density Estimation K-Nearest Neighbor (KNN) Decision Rule Supervised Learning How to fit a density

More information

Characterization of Super Strongly Perfect Graphs in Chordal and Strongly Chordal Graphs

Characterization of Super Strongly Perfect Graphs in Chordal and Strongly Chordal Graphs ISSN 0975-3303 Mapana J Sci, 11, 4(2012), 121-131 https://doi.org/10.12725/mjs.23.10 Characterization of Super Strongly Perfect Graphs in Chordal and Strongly Chordal Graphs R Mary Jeya Jothi * and A Amutha

More information

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

FACE RECOGNITION USING SUPPORT VECTOR MACHINES FACE RECOGNITION USING SUPPORT VECTOR MACHINES Ashwin Swaminathan ashwins@umd.edu ENEE633: Statistical and Neural Pattern Recognition Instructor : Prof. Rama Chellappa Project 2, Part (b) 1. INTRODUCTION

More information

Quadratic and cubic b-splines by generalizing higher-order voronoi diagrams

Quadratic and cubic b-splines by generalizing higher-order voronoi diagrams Quadratic and cubic b-splines by generalizing higher-order voronoi diagrams Yuanxin Liu and Jack Snoeyink Joshua Levine April 18, 2007 Computer Science and Engineering, The Ohio State University 1 / 24

More information

Robot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning

Robot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning Robot Learning 1 General Pipeline 1. Data acquisition (e.g., from 3D sensors) 2. Feature extraction and representation construction 3. Robot learning: e.g., classification (recognition) or clustering (knowledge

More information

Supervised Learning Classification Algorithms Comparison

Supervised Learning Classification Algorithms Comparison Supervised Learning Classification Algorithms Comparison Aditya Singh Rathore B.Tech, J.K. Lakshmipat University -------------------------------------------------------------***---------------------------------------------------------

More information

R u t c o r Research R e p o r t. Scale-invariant Clustering with Minimum Volume Ellipsoids. Mahesh Kumar a. James B. Orlin b

R u t c o r Research R e p o r t. Scale-invariant Clustering with Minimum Volume Ellipsoids. Mahesh Kumar a. James B. Orlin b R u t c o r Research R e p o r t Scale-invariant Clustering with Minimum Volume Ellipsoids Mahesh Kumar a James B. Orlin b RRR 28-2005, September, 2005 RUTCOR Rutgers Center for Operations Research Rutgers

More information

Lecture 25 Nonlinear Programming. November 9, 2009

Lecture 25 Nonlinear Programming. November 9, 2009 Nonlinear Programming November 9, 2009 Outline Nonlinear Programming Another example of NLP problem What makes these problems complex Scalar Function Unconstrained Problem Local and global optima: definition,

More information

Programs for MDE Modeling and Conditional Distribution Calculation

Programs for MDE Modeling and Conditional Distribution Calculation Programs for MDE Modeling and Conditional Distribution Calculation Sahyun Hong and Clayton V. Deutsch Improved numerical reservoir models are constructed when all available diverse data sources are accounted

More information

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

The Comparative Study of Machine Learning Algorithms in Text Data Classification* The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification

More information

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression John Fox Department of Sociology McMaster University 1280 Main Street West Hamilton, Ontario Canada L8S 4M4 jfox@mcmaster.ca February 2004 Abstract Nonparametric regression analysis

More information

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a Week 9 Based in part on slides from textbook, slides of Susan Holmes Part I December 2, 2012 Hierarchical Clustering 1 / 1 Produces a set of nested clusters organized as a Hierarchical hierarchical clustering

More information

Classification with Diffuse or Incomplete Information

Classification with Diffuse or Incomplete Information Classification with Diffuse or Incomplete Information AMAURY CABALLERO, KANG YEN Florida International University Abstract. In many different fields like finance, business, pattern recognition, communication

More information

Elemental Set Methods. David Banks Duke University

Elemental Set Methods. David Banks Duke University Elemental Set Methods David Banks Duke University 1 1. Introduction Data mining deals with complex, high-dimensional data. This means that datasets often combine different kinds of structure. For example:

More information

2. Convex sets. x 1. x 2. affine set: contains the line through any two distinct points in the set

2. Convex sets. x 1. x 2. affine set: contains the line through any two distinct points in the set 2. Convex sets Convex Optimization Boyd & Vandenberghe affine and convex sets some important examples operations that preserve convexity generalized inequalities separating and supporting hyperplanes dual

More information

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Cluster Analysis Mu-Chun Su Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Introduction Cluster analysis is the formal study of algorithms and methods

More information

Application of Support Vector Machine Algorithm in Spam Filtering

Application of Support Vector Machine Algorithm in  Spam Filtering Application of Support Vector Machine Algorithm in E-Mail Spam Filtering Julia Bluszcz, Daria Fitisova, Alexander Hamann, Alexey Trifonov, Advisor: Patrick Jähnichen Abstract The problem of spam classification

More information

Analysis of high dimensional data via Topology. Louis Xiang. Oak Ridge National Laboratory. Oak Ridge, Tennessee

Analysis of high dimensional data via Topology. Louis Xiang. Oak Ridge National Laboratory. Oak Ridge, Tennessee Analysis of high dimensional data via Topology Louis Xiang Oak Ridge National Laboratory Oak Ridge, Tennessee Contents Abstract iii 1 Overview 1 2 Data Set 1 3 Simplicial Complex 5 4 Computation of homology

More information

Convex Optimization. Convex Sets. ENSAE: Optimisation 1/24

Convex Optimization. Convex Sets. ENSAE: Optimisation 1/24 Convex Optimization Convex Sets ENSAE: Optimisation 1/24 Today affine and convex sets some important examples operations that preserve convexity generalized inequalities separating and supporting hyperplanes

More information

Lifting Transform, Voronoi, Delaunay, Convex Hulls

Lifting Transform, Voronoi, Delaunay, Convex Hulls Lifting Transform, Voronoi, Delaunay, Convex Hulls Subhash Suri Department of Computer Science University of California Santa Barbara, CA 93106 1 Lifting Transform (A combination of Pless notes and my

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

Application of Characteristic Function Method in Target Detection

Application of Characteristic Function Method in Target Detection Application of Characteristic Function Method in Target Detection Mohammad H Marhaban and Josef Kittler Centre for Vision, Speech and Signal Processing University of Surrey Surrey, GU2 7XH, UK eep5mm@ee.surrey.ac.uk

More information

Nearest neighbor classification DSE 220

Nearest neighbor classification DSE 220 Nearest neighbor classification DSE 220 Decision Trees Target variable Label Dependent variable Output space Person ID Age Gender Income Balance Mortgag e payment 123213 32 F 25000 32000 Y 17824 49 M 12000-3000

More information

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER A.Shabbir 1, 2 and G.Verdoolaege 1, 3 1 Department of Applied Physics, Ghent University, B-9000 Ghent, Belgium 2 Max Planck Institute

More information

Simplicial Global Optimization

Simplicial Global Optimization Simplicial Global Optimization Julius Žilinskas Vilnius University, Lithuania September, 7 http://web.vu.lt/mii/j.zilinskas Global optimization Find f = min x A f (x) and x A, f (x ) = f, where A R n.

More information

Learning a Manifold as an Atlas Supplementary Material

Learning a Manifold as an Atlas Supplementary Material Learning a Manifold as an Atlas Supplementary Material Nikolaos Pitelis Chris Russell School of EECS, Queen Mary, University of London [nikolaos.pitelis,chrisr,lourdes]@eecs.qmul.ac.uk Lourdes Agapito

More information

Exact adaptive parallel algorithms for data depth problems. Vera Rosta Department of Mathematics and Statistics McGill University, Montreal

Exact adaptive parallel algorithms for data depth problems. Vera Rosta Department of Mathematics and Statistics McGill University, Montreal Exact adaptive parallel algorithms for data depth problems Vera Rosta Department of Mathematics and Statistics McGill University, Montreal joint work with Komei Fukuda School of Computer Science McGill

More information

A Novel Criterion Function in Feature Evaluation. Application to the Classification of Corks.

A Novel Criterion Function in Feature Evaluation. Application to the Classification of Corks. A Novel Criterion Function in Feature Evaluation. Application to the Classification of Corks. X. Lladó, J. Martí, J. Freixenet, Ll. Pacheco Computer Vision and Robotics Group Institute of Informatics and

More information

Machine Learning and Pervasive Computing

Machine Learning and Pervasive Computing Stephan Sigg Georg-August-University Goettingen, Computer Networks 17.12.2014 Overview and Structure 22.10.2014 Organisation 22.10.3014 Introduction (Def.: Machine learning, Supervised/Unsupervised, Examples)

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,

More information

arxiv: v3 [math.mg] 4 Feb 2016

arxiv: v3 [math.mg] 4 Feb 2016 A Classification of Non-Compact Coxeter Polytopes with n + Facets and One Non-Simple Vertex Mike Roberts Liverpool University michael.roberts@liverpool.ac.uk arxiv:.8v [math.mg] Feb Abstract In this paper

More information

MATH3016: OPTIMIZATION

MATH3016: OPTIMIZATION MATH3016: OPTIMIZATION Lecturer: Dr Huifu Xu School of Mathematics University of Southampton Highfield SO17 1BJ Southampton Email: h.xu@soton.ac.uk 1 Introduction What is optimization? Optimization is

More information

On data depth and distribution-free discriminant analysis using separating surfaces

On data depth and distribution-free discriminant analysis using separating surfaces Bernoulli 11(1), 2005, 1 27 On data depth and distribution-free discriminant analysis using separating surfaces ANIL K. GHOSH* and PROBAL CHAUDHURI** Theoretical Statistics and Mathematics Unit, Indian

More information

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X

International Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,

More information