Experimental Analysis of GTM

Size: px
Start display at page:

Download "Experimental Analysis of GTM"

Transcription

1 Experimental Analysis of GTM Elias Pampalk In the past years many different data mining techniques have been developed. The goal of the seminar Kosice-Vienna is to compare some of them to determine which of them is preferable for which type of datasets. Therefor we analyze three different datasets with different properties regarding dimensionality and amount of data. I will summarize the generative topographic mapping algorithm and discuss the results of the experiments and show the similarity to related architectures. 1. INTRODUCTION Not linear methods for statistical data analysis have become more and more popular thanks to the rapid development of computers. The fields in which they are applied to are as various as the methods them self. Generative topographic mapping (GTM) has been developed by [Bishop et al. 1997] as a principal alternative to the self-organizing map (SOM) algorithm [Kohonen 1982] in which a set of unlabelled data vectors is summarized in terms of a set to reference vectors having a spatial organization corresponding to a (generally) two-dimensional sheet. While the SOM algorithm has achieved many successes in practical applications, it also suffers from the significant deficiencies, many of which highlighted in [Kohonen 1995]. They include: the absence of a cost function, the lack of theoretical basis for choosing learning rate parameter schedules and neighborhood parameters to ensure topographic ordering, the absence of any general proofs of convergence, and the fact that the model does not define a probability density. These Problems can all be traced to the heuristic origins of the SOM algorithm. The GTM algorithm overcomes most of the limitations of the SOM while introducing no significant disadvantages. The datasets used to analyze the strengths and weaknesses of GTM range from low dimensional with few vectors to very high dimensional with many vectors. In Section 2 I give an overview of related work. I summarize the method in Section 3. In Section 4 I describe the experiments I made and discuss the similarities to the SOM. Conclusions are presented in Section RELATED WORK A lot of research has been done in the field of statistical data analysis. GTM belongs to the family of unsupervised methods. Another representative is for example k-means clustering. An important aspect of unsupervised training is data visualization. The high dimensional data space is mapped on to a mostly two dimensional space. The main two criteria are preserving the topology and clustering the data. The most common tool is the SOM but there are also others as for example Sammons mapping. The GTM is a rather new robabilistic re-formulation of the SOM. The developers of the GTM algorithm have compared it with a batch algorithm of the SOM and obtained the result that the computational cost of the

2 2 Elias Pampalk GTM algorithm is about a third higher. 3. THE METHOD 3.1 Principals GTM consists of a constrained mixture of Gaussians in which the model parameters are determined by maximum likelihood using the EM algorithm. It is defined by specifying a set of points {x i } in latent space, together with a set of basis functions {Φ j (x)}. The adaptive parameters W and β define a constrained mixture of Gaussians with centers W Φ(x i ) and a common covariance matrix given by β 1 I. After initializing W and β, training involves alternating between the E-step in which the posterior probabilities are evaluated, and the M-step in which W and β are re-estimated. 3.2 Details The continuous function y = f(x, W ) defines a mapping from the latent space into the data space. Where x is a latent variable (vector) and W are the parameters of the mapping (matrix). And y is a vector in the higher dimensional data space. The transformation f(x, W ) maps the latent-variable space into an non-euclidean manifold embedded within the data space. Defining a probability distribution p(x) on the latent-variable space induces a corresponding distribution p(y W ) in the data space. Since in reality the data t will only approximately live on a lower-dimensional manifold, it is appropriate to include a noise model. A radially-symmetric Gaussian distribution centered on f(x, W ) is chosen p(t x, W, β) = ( β 2π ) D 2 exp{ β 2 f(x, W ) t 2 }. Where D is the dimension of the data space and β 1 is the variance of the Gaussian distribution. The distribution in the data space, for a given value of W, is then obtained by integration over the x-distribution p(t W, β) = p(t x, W, β)p(x)dx. For a given data set D = (t 1,.., t N ) of N data points, the parameters W and β can be determined using maximum likelihood. The log likelihood function is given by L(W, β) = ln N p(t n W, β). n=1 The (prior) distribution p(x) is chosen so that the integral over x can be solved p(x) = 1 K Where K is the number of latent points. K δ(x x i ). i=1

3 Experimental Analysis of GTM 3 Each point x i is mapped to a corresponding point f(x i, W ) in the data space, which forms the center of a Gaussian density function p(t W, β) = 1 K K p(t x i, W, β) i=1 and the log likelihood function becomes L(W, β) = N ln{ 1 K n=1 K p(t n x i, W, β)}. i=1 This corresponds to a constrained Gaussian mixture model, since the centers of the Gaussians, given by f(x i, W ), cannot move independently but are related through the function f(x, W ). If the mapping function f(x, W ) is smooth and continuous, the projected points f(x i, W ) will necessarily have a topographic ordering in the data space. If now a particular function parametrized form for f(x, W ) is chosen which is a differntiable function of W (for example a feed-forward network with sigmoidal hidden units) then any standard non-linear optimization method, such as conjugate gradients can be used. However, since the model consists of a mixture distribution the EM algorithm is used. f(x, W ) is chosen to be given by a generalized linear regression model of the form f(x, W ) = W Φ(x) where the elements of Φ(x) consist of M fixed basis functions Φ j (x), and W is a DxM matrix. 3.3 Matlab Toolbox The GTM toolbox for Matlab [Svensen 1999] provides a set of functions to generate GTMs and use them for visualization of data. There are standard functions for the two main steps: setup and training. Setup refers to a process of generating an initial GTM model, made up by a set of components (Matlab matrices). Training refers to adapting the initial model to a data set, in order to improve the fit to that data. The standard initialization (gtm stp2) consists of the following steps: First a latent variable sample is generated. Then the centers of the basis functions are generated and the activations in the basis functions are computed, given the latent variable sample. At last an initial weight matrix mapping from the output of the basis functions to the data space, and an initial value for the inverse variance of the Gaussian mixture is computed useing the two first principal components. The standard training function (gtm trn) basically consists of two steps: In the first E-step a matrix is calculated, containing the responsibilities assumed by each Gaussian mixture component for each of the data points. These responsibilities are used in the second M-step for calculating new parameters of the Gaussian mixture. For each training cycle both steps are calculated.

4 4 Elias Pampalk 4. EXPERIMENTS 4.1 Introduction Diagrams. GTM defines a distribution for each data point in the latent space. There are a few ways to visualize this. The first would be to look at the distribution of each single data point. Normally this is not desired. The second way would be to plot the means of the data points. However the distribution could be multi-modal in which case the mean can give a very misleading summary of the distribution. So another way would be too plot the means and the corresponding modes. The problem with this approach however is that with many data points it becomes difficult to recognize anything. Therefor for each experiment I plot a diagram of the means to represent the data points and to easily recognize clusters. And I plot the means with their corresponding modes to indicate if the distributions are multi-modal. Further I plot the sum of the distribution of all points in the data set. This is easily done by adding the single distributions and normalizing the result. The forth plot I always make is the log-likelihood. At [ elan/elias/kosice/] a complete description of the experiments including the used Matlab scripts can be found Parameters. Following parameters must be specified using the GTM toolbox. A more detailed description can be found in [Svensen 1999]. Number of Latent Points Using a L-dimensional latent space for visualization it is recommended to set the number of latent points to O(10 L ) in the support of each basis function. The latent points lie within a regular square grid. Settings as 10x5 are not possible. With too few points per basis function the smoothness of the mapping is lost. The number of latent points is limited computationally, but a very high number would be ideal. Number of Basis Functions A limited number of basis functions will necessarily restrict the possible forms that the mapping can take. The amount of basis functions must be a square of a whole number. Settings as 2x3 are not possible. Sigma Denotes a scalar giving the relative width of the basis functions. The absolute width is calculated as sigma times the distance between two neighboring basis function centers. When basis functions overlap their response will be correlated, which causes the smoothness of the mapping. More or narrower basis functions will allow a more flexible mapping, while fewer or broader functions will prescribe a smoother mapping. Sigma = 1.0 is a good starting point. Cycles The GTM implemented in Matlab uses a batch algorithm. With the testruns I made the log likelihood had converged after 5 to 30 cycles. Lambda The weight regulation factor governs the degree of weight decay applied during training. It controls the scaling, by restricting the magnitude of the weights. It is recommended to set it to All experiments I made had this setting. 4.2 Animals Description. A toy example, very useful to test the various parameters: small, easy to handle, and intuitively interpretable: 16 animals, described by 13 attributes.

5 Experimental Analysis of GTM Raw Data Experiments. The following four diagrams illustrate the effects of the GTM parameters. Fig. 1. Legend for the figures 2, 3 and 4. Fig. 2. Parameters: 3x3 latent points, 2x2 basis functions, 1.0 sigma, 10 cycles. A low number of latent points generates a low resolution of the distribution. Since this dataset only contains few basic clusters a low number of latent points is sufficient. In figure 2 (means) you can see that the clusters are separated. Horse, zebra and cow are mapped to the upper left. Dove, hen duck and goose are mapped to the upper right. Tiger and lion are mapped to the middle left, cat is mapped to the center. Owl, hawk and eagle are mapped to the middle and lower right. Fox, dog and wolf can be found on the lower left. Notice that convergence is reached after about 5 cycles. Figure 3 shows the effects of a higher number of latent points: the form of the distribution has a higher resolution. Compared to figure 2 a clearer picture of the data structure is revealed. The low number of basis functions and the high sigma cause the smoothness of the mapping. Each hill in the distribution represents a cluster. The peek at (-1,1) is caused by two identical vectors. Because the standard setup procedure uses a principal component initialization the clusters are almost at the same location as in figure 2.

6 6 Elias Pampalk Fig. 3. Parameters: 20x20 latent points, 2x2 basis functions, 3.0 sigma, 10 cycles. Fig. 4. Parameters: 20x20 latent points, 2x2 basis functions, 0.2 sigma, 10 cycles. Decreasing sigma reduces the smoothness of the mapping as can be seen in figure 4. The hills in the distribution are much higher and rougher as in figure 3. Notice that the distributions of the data points are now multi-modal (modes diagram). Increasing the number of basis functions dramatically increases the flexibility of the mapping. Figure 5 shows the result. The distribution is spiked. Because there are more basis functions than data points the log-likelihood diagram seems strange. Even with these parameter settings topology and clusters are basically correct.

7 Experimental Analysis of GTM 7 Fig. 5. Parameters: 20x20 latent points, 5x5 basis functions, 0.2 sigma, 10 cycles Evaluation. I encountered no problems with this dataset. The results made sense and the standard functions worked fine. Time was no problem. The clusters generated look very nice. Finding good parameters was not a problem. 4.3 MIS Description. This data describes characteristics of software modules (size, complexity measures,...), medium-sized data set with low dimensionality: 420 vectors, described by 13 attributes. Fig. 6. Legend for the figures 7, 8 and 9. The numbers represent how often the source code has been modified. A plotted circle represents a source code which has been modified up to 50 times and so on Raw Data Experiments. It is difficult to find parameters with which a result other than a mapping on to one point is generated. Figure 7 shows one of my best tries. Notice the log-likelihood diagram. I suspect a numerical error. On the right side of the means diagram there are mainly data points which represent source codes which have been modified allot. On the left side there is a big cluster of the rest.

8 8 Elias Pampalk Fig. 7. Parameters: 20x20 latent points, 10x10 basis functions, 2.5 sigma, 10 cycles. Fig. 8. Parameters: 30x30 latent points, 2x2 basis functions, 1.0 sigma, 20 cycles Normalized by Attribute Experiments. In Figure 8 again mainly data points which represent a high modification are on the right side of the means diagram and there is one big cluster of the rest. Notice that the distribution has a form similar to waves. Possibly this is a side effect of the GTM algorithm. This strange form was generated with almost any parameter setting.

9 Experimental Analysis of GTM 9 Fig. 9. Parameters: 20x20 latent points, 2x2 basis functions, 4.0 sigma, 20 cycles Vector Length Normalized Experiments. In Figure 9 there seems to be no left right separation between data points with a low and a high modification count. What can be seen is that there are small clusters of similar data points. One example is the top right, where many squares occupy the same space. This can also be seen in the distribution: there is a high peek at (-1, -1) Evaluation. I used the standard procedures. Time was not a real problem. But I was not able to produce good looking results with the raw data. The convergence problems I had might be caused by numerical errors. With the vector length normalized a very strange form was generated. The form reminded me of waves in water after dropping a stone. Normalizing the attributes caused more or less random results, at least I could not detect any structure. 4.4 TIME Magazine Description. Newspaper articles of the TIME Magazine from the 1960 s, medium-sized data set with very high dimensionality: 420 vectors (articles) described by 5923 attributes Vector Length Normalized Experiments. Matlab does not support functions that would make it possible to easily analyze maps with many (420) vectors. To read a document plotted at a certain point on the map I have to find out by hand which vector it is. One way to do it is to mark one vector after the other with a special symbol (for example + instead of o ) which makes it possible to distinguish the vector from the others. It is impossible to analyze a map that way within acceptable time limits. The approach I chose was a comparison. I used the results of Michael Dittenbach. He has trained a flat SOM and labeled the nodes. I took some nodes that looked good. I marked these clusters with symbols so I could recognize them in the diagrams.

10 10 Elias Pampalk viet(1) SOM node: (1, 8); Content: South Vietnam, religious crisis. viet(2) SOM node: (1, 7); Content: South Vietnam. viet(3) SOM node: (1, 6); Content: South Vietnam, military. moscow,... SOM node: (11, 7); Content: Russia, communism, Khrushchev. khrushch(1) SOM node: (11, 8); Content: Khrushchev, Cold War. khrushch(2) SOM node: (10, 7); Content: Russia, economy. khrushch(3) SOM node: (10, 8); Content: Russia, culture. zanzibar, kenya SOM node: (3, 4); Content: Kenya, Zanzibar. nato, nuclear SOM node: (2, 11); Content: NATO, nuclear weapons. Fig. 10. Legend for the figures 11 and 12. Figure 11 shows a mapping for the TIME magazine documents. The clusters are basically identical with those found by the SOM. Especially the South Vietnam cluster is separated very nicely. Overlap of the symbols + and x causes a * as seen on the left top in the means diagram. Because of the low number of basis functions the clusters in the center are cramped. The connection between the triangles pointing down and the pentagrams came out very nicely. Notice the convergence after the second cycle. This can be observed by all other parameter settings as well. Figure 12 shows a mapping with more basis functions. The clusters are the same again, except that the Russian documents have moved closer together. It is hard to recognize in this plot but the triangles pointing down [krushch(1)] are not all in the same region. All but one are on the top left with the other documents on Russia. The remainder is on the top right with the documents on nato and nuclear weapons. The reason therefor is that the document is about Russian nuclear weapons. The South Vietnam cluster is now located at the lower right. While with the MIS and animals experiments principal component analysis was used to initialize the variables with the TIME dataset I hade to use a random initialization Evaluation. The quality of the mappings seems to be good. The results are very similar to those of the flat SOM. I have observed 9 clusters that have been found by the SOM. GTM found the same clusters and topology with almost any parameter settings. For practical use with text analysis it would be necessary to develop a better user interface. It is difficult to find the corresponding documents to the plotted diagrams.

11 Experimental Analysis of GTM 11 Fig. 11. Parameters: 10x10 latent points, 3x3 basis functions, 2.0 sigma, 5 cycles. Fig. 12. Parameters: 10x10 latent points, 4x4 basis functions, 4.0 sigma, 5 cycles. I encountered several problems while working with this dataset. First of all it was not possible to use the standard procedures because of the high dimensionality. The principal component analysis which is used by default to initialize the data took too long. And in my case also too much memory (over 500MB). The GTM toolbox also provides the possibility to use a random initialization which I used. The next problem where numerical errors. Some of the functions in the GTM toolbox by default use algorithms which are fast but not very precise. This caused

12 12 Elias Pampalk divisions by zero when the means were calculated. The next problem was the efficiency of some calculations. I was able to solve the problem by changing some matrix multiplications into loops. Clearly this toolbox was not developed for high dimensional data. The remaining problem still is the time it takes to initialize, train and visualize a map. With average parameters it takes me about 20 minutes for one run. Almost halve of the time is spent calculating the distribution of the entire dataset (if plotting modes and means is enough a run only takes halve of the time). 5. CONCLUSION I have presented the results of experiments done with different datasets. I have explained the problems of the GTM toolbox for Matlab and I have shown the similarity between GTM and SOM. For datasets with only few clusters the GTM is a good alternative to the SOM. For datasets with many unclear clusters, as in text data mining, it is necessary to develop a better interface to work with the results. REFERENCES Bishop, C. et al GTM: The Generative Topographic Mapping. Kohonen, T Self-Organizing Maps. Kohonen, T Self-organized formation of topologically correct feature maps. Svensen, M The GTM Toolbox - User s Guide.

t 1 y(x;w) x 2 t 2 t 3 x 1

t 1 y(x;w) x 2 t 2 t 3 x 1 Neural Computing Research Group Dept of Computer Science & Applied Mathematics Aston University Birmingham B4 7ET United Kingdom Tel: +44 (0)121 333 4631 Fax: +44 (0)121 333 4586 http://www.ncrg.aston.ac.uk/

More information

Generative and discriminative classification techniques

Generative and discriminative classification techniques Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14

More information

Clustering with Reinforcement Learning

Clustering with Reinforcement Learning Clustering with Reinforcement Learning Wesam Barbakh and Colin Fyfe, The University of Paisley, Scotland. email:wesam.barbakh,colin.fyfe@paisley.ac.uk Abstract We show how a previously derived method of

More information

Graph projection techniques for Self-Organizing Maps

Graph projection techniques for Self-Organizing Maps Graph projection techniques for Self-Organizing Maps Georg Pölzlbauer 1, Andreas Rauber 1, Michael Dittenbach 2 1- Vienna University of Technology - Department of Software Technology Favoritenstr. 9 11

More information

Semi-Supervised Construction of General Visualization Hierarchies

Semi-Supervised Construction of General Visualization Hierarchies Semi-Supervised Construction of General Visualization Hierarchies Peter Tiňo Yi Sun Ian Nabney Aston University, Aston Triangle, Birmingham, B4 7ET United Kingdom Abstract We have recently developed a

More information

COMPUTATIONAL INTELLIGENCE

COMPUTATIONAL INTELLIGENCE COMPUTATIONAL INTELLIGENCE Radial Basis Function Networks Adrian Horzyk Preface Radial Basis Function Networks (RBFN) are a kind of artificial neural networks that use radial basis functions (RBF) as activation

More information

Function approximation using RBF network. 10 basis functions and 25 data points.

Function approximation using RBF network. 10 basis functions and 25 data points. 1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data

More information

Clustering Lecture 5: Mixture Model

Clustering Lecture 5: Mixture Model Clustering Lecture 5: Mixture Model Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced topics

More information

GTM: The Generative Topographic Mapping

GTM: The Generative Topographic Mapping GTM: The Generative Topographic Mapping Christopher M. Bishop, Markus Svensén Microsoft Research 7 J J Thomson Avenue Cambridge, CB3 0FB, U.K. {cmbishop,markussv}@microsoft.com http://research.microsoft.com/{

More information

Tight Clusters and Smooth Manifolds with the Harmonic Topographic Map.

Tight Clusters and Smooth Manifolds with the Harmonic Topographic Map. Proceedings of the th WSEAS Int. Conf. on SIMULATION, MODELING AND OPTIMIZATION, Corfu, Greece, August -9, (pp8-) Tight Clusters and Smooth Manifolds with the Harmonic Topographic Map. MARIAN PEÑA AND

More information

3 Nonlinear Regression

3 Nonlinear Regression CSC 4 / CSC D / CSC C 3 Sometimes linear models are not sufficient to capture the real-world phenomena, and thus nonlinear models are necessary. In regression, all such models will have the same basic

More information

3 Nonlinear Regression

3 Nonlinear Regression 3 Linear models are often insufficient to capture the real-world phenomena. That is, the relation between the inputs and the outputs we want to be able to predict are not linear. As a consequence, nonlinear

More information

Unsupervised Learning

Unsupervised Learning Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Seismic facies analysis using generative topographic mapping

Seismic facies analysis using generative topographic mapping Satinder Chopra + * and Kurt J. Marfurt + Arcis Seismic Solutions, Calgary; The University of Oklahoma, Norman Summary Seismic facies analysis is commonly carried out by classifying seismic waveforms based

More information

CS Introduction to Data Mining Instructor: Abdullah Mueen

CS Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts

More information

Free Projection SOM: A New Method For SOM-Based Cluster Visualization

Free Projection SOM: A New Method For SOM-Based Cluster Visualization Free Projection SOM: A New Method For SOM-Based Cluster Visualization 1 ABDEL-BADEEH M. SALEM 1, EMAD MONIER, KHALED NAGATY Computer Science Department Ain Shams University Faculty of Computer & Information

More information

GTM: The Generative Topographic Mapping

GTM: The Generative Topographic Mapping Communicated by Helge Ritter GTM: The Generative Topographic Mapping Christopher M. Bishop Markus Svensén Christopher K. I. Williams Neural Computing Research Group, Department of Computer Science and

More information

Ordered Vector Quantization for the Integrated Analysis of Geochemical and Geoscientific Data Sets

Ordered Vector Quantization for the Integrated Analysis of Geochemical and Geoscientific Data Sets Ordered Vector Quantization for the Integrated Analysis of Geochemical and Geoscientific Data Sets Stephen Fraser 1 & Bruce Dickson 2 We are drowning in information and starving for knowledge. Rutherford

More information

Expectation Maximization (EM) and Gaussian Mixture Models

Expectation Maximization (EM) and Gaussian Mixture Models Expectation Maximization (EM) and Gaussian Mixture Models Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 2 3 4 5 6 7 8 Unsupervised Learning Motivation

More information

Self-Organizing Maps (SOM)

Self-Organizing Maps (SOM) Overview (SOM) Basics Sequential Training (On-Line Learning) Batch SOM Visualizing the SOM - SOM Grid - Music Description Map (MDM) - Bar Plots and Chernoff's Faces - U-Matrix and Distance Matrix - Smoothed

More information

Advanced visualization techniques for Self-Organizing Maps with graph-based methods

Advanced visualization techniques for Self-Organizing Maps with graph-based methods Advanced visualization techniques for Self-Organizing Maps with graph-based methods Georg Pölzlbauer 1, Andreas Rauber 1, and Michael Dittenbach 2 1 Department of Software Technology Vienna University

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

A SOM-view of oilfield data: A novel vector field visualization for Self-Organizing Maps and its applications in the petroleum industry

A SOM-view of oilfield data: A novel vector field visualization for Self-Organizing Maps and its applications in the petroleum industry A SOM-view of oilfield data: A novel vector field visualization for Self-Organizing Maps and its applications in the petroleum industry Georg Pölzlbauer, Andreas Rauber (Department of Software Technology

More information

Gaussian Process Latent Variable Models for Visualisation of High Dimensional Data

Gaussian Process Latent Variable Models for Visualisation of High Dimensional Data Gaussian Process Latent Variable Models for Visualisation of High Dimensional Data Neil D. Lawrence Department of Computer Science University of Sheffield Regent Court, 211 Portobello Street, Sheffield,

More information

Content-based image and video analysis. Machine learning

Content-based image and video analysis. Machine learning Content-based image and video analysis Machine learning for multimedia retrieval 04.05.2009 What is machine learning? Some problems are very hard to solve by writing a computer program by hand Almost all

More information

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm DATA MINING LECTURE 7 Hierarchical Clustering, DBSCAN The EM Algorithm CLUSTERING What is a Clustering? In general a grouping of objects such that the objects in a group (cluster) are similar (or related)

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,

More information

Generative and discriminative classification techniques

Generative and discriminative classification techniques Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15

More information

Machine Learning Methods in Visualisation for Big Data 2018

Machine Learning Methods in Visualisation for Big Data 2018 Machine Learning Methods in Visualisation for Big Data 2018 Daniel Archambault1 Ian Nabney2 Jaakko Peltonen3 1 Swansea University 2 University of Bristol 3 University of Tampere, Aalto University Evaluating

More information

K-Means and Gaussian Mixture Models

K-Means and Gaussian Mixture Models K-Means and Gaussian Mixture Models David Rosenberg New York University June 15, 2015 David Rosenberg (New York University) DS-GA 1003 June 15, 2015 1 / 43 K-Means Clustering Example: Old Faithful Geyser

More information

A Population Based Convergence Criterion for Self-Organizing Maps

A Population Based Convergence Criterion for Self-Organizing Maps A Population Based Convergence Criterion for Self-Organizing Maps Lutz Hamel and Benjamin Ott Department of Computer Science and Statistics, University of Rhode Island, Kingston, RI 02881, USA. Email:

More information

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013 Machine Learning Topic 5: Linear Discriminants Bryan Pardo, EECS 349 Machine Learning, 2013 Thanks to Mark Cartwright for his extensive contributions to these slides Thanks to Alpaydin, Bishop, and Duda/Hart/Stork

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

What is machine learning?

What is machine learning? Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Applying Supervised Learning

Applying Supervised Learning Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains

More information

Machine Learning : Clustering, Self-Organizing Maps

Machine Learning : Clustering, Self-Organizing Maps Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps Clustering The task: partition a set of objects into meaningful subsets (clusters). The

More information

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation

More information

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017 CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2017 Assignment 3: 2 late days to hand in tonight. Admin Assignment 4: Due Friday of next week. Last Time: MAP Estimation MAP

More information

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Your Name: Your student id: Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Problem 1 [5+?]: Hypothesis Classes Problem 2 [8]: Losses and Risks Problem 3 [11]: Model Generation

More information

Unsupervised Learning: Clustering

Unsupervised Learning: Clustering Unsupervised Learning: Clustering Vibhav Gogate The University of Texas at Dallas Slides adapted from Carlos Guestrin, Dan Klein & Luke Zettlemoyer Machine Learning Supervised Learning Unsupervised Learning

More information

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering SYDE 372 - Winter 2011 Introduction to Pattern Recognition Clustering Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 5 All the approaches we have learned

More information

Lecture on Modeling Tools for Clustering & Regression

Lecture on Modeling Tools for Clustering & Regression Lecture on Modeling Tools for Clustering & Regression CS 590.21 Analysis and Modeling of Brain Networks Department of Computer Science University of Crete Data Clustering Overview Organizing data into

More information

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions Thomas Giraud Simon Chabot October 12, 2013 Contents 1 Discriminant analysis 3 1.1 Main idea................................

More information

Self-organizing mixture models

Self-organizing mixture models Self-organizing mixture models Jakob Verbeek, Nikos Vlassis, Ben Krose To cite this version: Jakob Verbeek, Nikos Vlassis, Ben Krose. Self-organizing mixture models. Neurocomputing / EEG Neurocomputing,

More information

Overview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010

Overview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010 INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,

More information

A Principled Approach to Interactive Hierarchical Non-Linear Visualization of High-Dimensional Data

A Principled Approach to Interactive Hierarchical Non-Linear Visualization of High-Dimensional Data A Principled Approach to Interactive Hierarchical Non-Linear Visualization of High-Dimensional Data Peter Tiňo, Ian Nabney, Yi Sun Neural Computing Research Group Aston University, Birmingham, B4 7ET,

More information

COMPRESSED DETECTION VIA MANIFOLD LEARNING. Hyun Jeong Cho, Kuang-Hung Liu, Jae Young Park. { zzon, khliu, jaeypark

COMPRESSED DETECTION VIA MANIFOLD LEARNING. Hyun Jeong Cho, Kuang-Hung Liu, Jae Young Park.   { zzon, khliu, jaeypark COMPRESSED DETECTION VIA MANIFOLD LEARNING Hyun Jeong Cho, Kuang-Hung Liu, Jae Young Park Email : { zzon, khliu, jaeypark } @umich.edu 1. INTRODUCTION In many imaging applications such as Computed Tomography

More information

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A. 205-206 Pietro Guccione, PhD DEI - DIPARTIMENTO DI INGEGNERIA ELETTRICA E DELL INFORMAZIONE POLITECNICO DI BARI

More information

A Stochastic Optimization Approach for Unsupervised Kernel Regression

A Stochastic Optimization Approach for Unsupervised Kernel Regression A Stochastic Optimization Approach for Unsupervised Kernel Regression Oliver Kramer Institute of Structural Mechanics Bauhaus-University Weimar oliver.kramer@uni-weimar.de Fabian Gieseke Institute of Structural

More information

Points Lines Connected points X-Y Scatter. X-Y Matrix Star Plot Histogram Box Plot. Bar Group Bar Stacked H-Bar Grouped H-Bar Stacked

Points Lines Connected points X-Y Scatter. X-Y Matrix Star Plot Histogram Box Plot. Bar Group Bar Stacked H-Bar Grouped H-Bar Stacked Plotting Menu: QCExpert Plotting Module graphs offers various tools for visualization of uni- and multivariate data. Settings and options in different types of graphs allow for modifications and customizations

More information

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of

More information

Artificial Neural Networks Unsupervised learning: SOM

Artificial Neural Networks Unsupervised learning: SOM Artificial Neural Networks Unsupervised learning: SOM 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001

More information

Homework #4 Programming Assignment Due: 11:59 pm, November 4, 2018

Homework #4 Programming Assignment Due: 11:59 pm, November 4, 2018 CSCI 567, Fall 18 Haipeng Luo Homework #4 Programming Assignment Due: 11:59 pm, ovember 4, 2018 General instructions Your repository will have now a directory P4/. Please do not change the name of this

More information

Fall 09, Homework 5

Fall 09, Homework 5 5-38 Fall 09, Homework 5 Due: Wednesday, November 8th, beginning of the class You can work in a group of up to two people. This group does not need to be the same group as for the other homeworks. You

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2008 CS 551, Spring 2008 c 2008, Selim Aksoy (Bilkent University)

More information

SGN (4 cr) Chapter 11

SGN (4 cr) Chapter 11 SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter

More information

A vector field visualization technique for Self-Organizing Maps

A vector field visualization technique for Self-Organizing Maps A vector field visualization technique for Self-Organizing Maps Georg Pölzlbauer 1, Andreas Rauber 1, and Michael Dittenbach 2 1 Department of Software Technology Vienna University of Technology Favoritenstr.

More information

ABSTRACT. Keywords: visual training, unsupervised learning, lumber inspection, projection 1. INTRODUCTION

ABSTRACT. Keywords: visual training, unsupervised learning, lumber inspection, projection 1. INTRODUCTION Comparison of Dimensionality Reduction Methods for Wood Surface Inspection Matti Niskanen and Olli Silvén Machine Vision Group, Infotech Oulu, University of Oulu, Finland ABSTRACT Dimensionality reduction

More information

A Topography-Preserving Latent Variable Model with Learning Metrics

A Topography-Preserving Latent Variable Model with Learning Metrics A Topography-Preserving Latent Variable Model with Learning Metrics Samuel Kaski and Janne Sinkkonen Helsinki University of Technology Neural Networks Research Centre P.O. Box 5400, FIN-02015 HUT, Finland

More information

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures:

Homework. Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression Pod-cast lecture on-line. Next lectures: Homework Gaussian, Bishop 2.3 Non-parametric, Bishop 2.5 Linear regression 3.0-3.2 Pod-cast lecture on-line Next lectures: I posted a rough plan. It is flexible though so please come with suggestions Bayes

More information

Unsupervised Learning : Clustering

Unsupervised Learning : Clustering Unsupervised Learning : Clustering Things to be Addressed Traditional Learning Models. Cluster Analysis K-means Clustering Algorithm Drawbacks of traditional clustering algorithms. Clustering as a complex

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves

More information

A Taxonomy of Semi-Supervised Learning Algorithms

A Taxonomy of Semi-Supervised Learning Algorithms A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph

More information

Introduction to Machine Learning

Introduction to Machine Learning Department of Computer Science, University of Helsinki Autumn 2009, second term Session 8, November 27 th, 2009 1 2 3 Multiplicative Updates for L1-Regularized Linear and Logistic Last time I gave you

More information

Using Machine Learning to Optimize Storage Systems

Using Machine Learning to Optimize Storage Systems Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)

More information

Bastian Wormuth. Version About this Manual

Bastian Wormuth. Version About this Manual Elba User Manual Table of Contents Bastian Wormuth Version 0.1 1 About this Manual...1 2 Overview...2 3 Starting Elba...3 4 Establishing the database connection... 3 5 Elba's Main Window... 5 6 Creating

More information

MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION. Steve Tjoa June 25, 2014

MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION. Steve Tjoa June 25, 2014 MACHINE LEARNING: CLUSTERING, AND CLASSIFICATION Steve Tjoa kiemyang@gmail.com June 25, 2014 Review from Day 2 Supervised vs. Unsupervised Unsupervised - clustering Supervised binary classifiers (2 classes)

More information

CSE 6242 A / CX 4242 DVA. March 6, Dimension Reduction. Guest Lecturer: Jaegul Choo

CSE 6242 A / CX 4242 DVA. March 6, Dimension Reduction. Guest Lecturer: Jaegul Choo CSE 6242 A / CX 4242 DVA March 6, 2014 Dimension Reduction Guest Lecturer: Jaegul Choo Data is Too Big To Analyze! Limited memory size! Data may not be fitted to the memory of your machine! Slow computation!

More information

University of Florida CISE department Gator Engineering. Clustering Part 5

University of Florida CISE department Gator Engineering. Clustering Part 5 Clustering Part 5 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville SNN Approach to Clustering Ordinary distance measures have problems Euclidean

More information

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Exploratory data analysis tasks Examine the data, in search of structures

More information

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components Review Lecture 14 ! PRINCIPAL COMPONENT ANALYSIS Eigenvectors of the covariance matrix are the principal components 1. =cov X Top K principal components are the eigenvectors with K largest eigenvalues

More information

3. Data Structures for Image Analysis L AK S H M O U. E D U

3. Data Structures for Image Analysis L AK S H M O U. E D U 3. Data Structures for Image Analysis L AK S H M AN @ O U. E D U Different formulations Can be advantageous to treat a spatial grid as a: Levelset Matrix Markov chain Topographic map Relational structure

More information

Figure (5) Kohonen Self-Organized Map

Figure (5) Kohonen Self-Organized Map 2- KOHONEN SELF-ORGANIZING MAPS (SOM) - The self-organizing neural networks assume a topological structure among the cluster units. - There are m cluster units, arranged in a one- or two-dimensional array;

More information

Clustering. Lecture 6, 1/24/03 ECS289A

Clustering. Lecture 6, 1/24/03 ECS289A Clustering Lecture 6, 1/24/03 What is Clustering? Given n objects, assign them to groups (clusters) based on their similarity Unsupervised Machine Learning Class Discovery Difficult, and maybe ill-posed

More information

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Assignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018

Assignment 2. Unsupervised & Probabilistic Learning. Maneesh Sahani Due: Monday Nov 5, 2018 Assignment 2 Unsupervised & Probabilistic Learning Maneesh Sahani Due: Monday Nov 5, 2018 Note: Assignments are due at 11:00 AM (the start of lecture) on the date above. he usual College late assignments

More information

Data visualisation and exploration with prior knowledge

Data visualisation and exploration with prior knowledge Data visualisation and exploration with prior knowledge Martin Schroeder, Dan Cornford, Ian T. Nabney Aston University, NCRG, Aston Triangle, Birmingham, B4 7ET, UK shroderm@aston.ac.uk Abstract. Visualising

More information

I How does the formulation (5) serve the purpose of the composite parameterization

I How does the formulation (5) serve the purpose of the composite parameterization Supplemental Material to Identifying Alzheimer s Disease-Related Brain Regions from Multi-Modality Neuroimaging Data using Sparse Composite Linear Discrimination Analysis I How does the formulation (5)

More information

08 An Introduction to Dense Continuous Robotic Mapping

08 An Introduction to Dense Continuous Robotic Mapping NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy

More information

CSC 2515 Introduction to Machine Learning Assignment 2

CSC 2515 Introduction to Machine Learning Assignment 2 CSC 2515 Introduction to Machine Learning Assignment 2 Zhongtian Qiu(1002274530) Problem 1 See attached scan files for question 1. 2. Neural Network 2.1 Examine the statistics and plots of training error

More information

Breaking it Down: The World as Legos Benjamin Savage, Eric Chu

Breaking it Down: The World as Legos Benjamin Savage, Eric Chu Breaking it Down: The World as Legos Benjamin Savage, Eric Chu To devise a general formalization for identifying objects via image processing, we suggest a two-pronged approach of identifying principal

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised

More information

Dynamic Thresholding for Image Analysis

Dynamic Thresholding for Image Analysis Dynamic Thresholding for Image Analysis Statistical Consulting Report for Edward Chan Clean Energy Research Center University of British Columbia by Libo Lu Department of Statistics University of British

More information

Image Processing. Image Features

Image Processing. Image Features Image Processing Image Features Preliminaries 2 What are Image Features? Anything. What they are used for? Some statements about image fragments (patches) recognition Search for similar patches matching

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Robust cartogram visualization of outliers in manifold learning

Robust cartogram visualization of outliers in manifold learning Robust cartogram visualization of outliers in manifold learning Alessandra Tosi 1 and Alfredo Vellido 1 1- Llenguatges i Sistemes Informàtics, Universitat Politècnica de Catalunya, Edifici Omega, Campus

More information

Abstractacceptedforpresentationatthe2018SEGConventionatAnaheim,California.Presentationtobemadeinsesion

Abstractacceptedforpresentationatthe2018SEGConventionatAnaheim,California.Presentationtobemadeinsesion Abstractacceptedforpresentationatthe2018SEGConventionatAnaheim,California.Presentationtobemadeinsesion MLDA3:FaciesClasificationandReservoirProperties2,onOctober17,2018from 11:25am to11:50am inroom 204B

More information

CHAPTER 4: CLUSTER ANALYSIS

CHAPTER 4: CLUSTER ANALYSIS CHAPTER 4: CLUSTER ANALYSIS WHAT IS CLUSTER ANALYSIS? A cluster is a collection of data-objects similar to one another within the same group & dissimilar to the objects in other groups. Cluster analysis

More information

ECS 234: Data Analysis: Clustering ECS 234

ECS 234: Data Analysis: Clustering ECS 234 : Data Analysis: Clustering What is Clustering? Given n objects, assign them to groups (clusters) based on their similarity Unsupervised Machine Learning Class Discovery Difficult, and maybe ill-posed

More information

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample CS 1675 Introduction to Machine Learning Lecture 18 Clustering Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square Clustering Groups together similar instances in the data sample Basic clustering problem:

More information

Clustering web search results

Clustering web search results Clustering K-means Machine Learning CSE546 Emily Fox University of Washington November 4, 2013 1 Clustering images Set of Images [Goldberger et al.] 2 1 Clustering web search results 3 Some Data 4 2 K-means

More information

MTTTS17 Dimensionality Reduction and Visualization. Spring 2018 Jaakko Peltonen. Lecture 11: Neighbor Embedding Methods continued

MTTTS17 Dimensionality Reduction and Visualization. Spring 2018 Jaakko Peltonen. Lecture 11: Neighbor Embedding Methods continued MTTTS17 Dimensionality Reduction and Visualization Spring 2018 Jaakko Peltonen Lecture 11: Neighbor Embedding Methods continued This Lecture Neighbor embedding by generative modeling Some supervised neighbor

More information