A New Implementation of the co-vat Algorithm for Visual Assessment of Clusters in Rectangular Relational Data

Size: px
Start display at page:

Download "A New Implementation of the co-vat Algorithm for Visual Assessment of Clusters in Rectangular Relational Data"

Transcription

1 A New Implementation of the co-vat Algorithm for Visual Assessment of Clusters in Rectangular Relational Data Timothy C. Havens 1, James C. Bezdek 1, and James M. Keller 1 Department of Electrical and Computer Engineering University of Missouri Columbia, MO 65211, USA ( havenst@gmail.com, jcbezdek@gmail.com, kellerj@missouri.edu) Abstract. This paper presents a new implementation of the co-vat algorithm. We assume we have an m n matrix D, where the elements of D are pair-wise dissimilarities between m row objects O r and n column objects O c. The union of these disjoint sets are (N = m + n) objects O. Clustering tendency assessment is the process by which a data set is analyzed to determine the number(s) of clusters present. In 2007, the co-visual Assessment of Tendency (co-vat) algorithm was proposed for rectangular data such as these. co-vat is a visual approach that addresses four clustering tendency questions: i) How many clusters are in the row objects O r? ii) How many clusters are in the column objects O c? iii) How many clusters are in the union of the row and column objects O r O c? And, iv) How many (co)-clusters are there that contain at least one of each type? co-vat first imputes pair-wise dissimilarity values among the row objects, the square relational matrix D r, and the column objects, the square relational matrix D c, and then builds a larger square dissimilarity matrix D r c. The clustering questions can then be addressed by using the VAT algorithm on D r, D c, and D r c; D is reordered by shuffling the reordering indices of D r c. Subsequently, the co-vat image of D may show tendency for co-clusters (problem iv). We first discuss a different way to construct this image, and then we also extend a path-based distance transform, which is used in the ivat algorithm, to co-vat. The new algorithm, co-ivat, shows dramatic improvement in the ability of co-vat to show cluster tendency in rectangular dissimilarity data. 1 Introduction Consider a set of objects O = {o 1,...,o N }. These objects can represent virtually anything vintage bass guitars, pure-bred cats, cancer genes expressed in a microarray experiment, cake recipes, or web-pages. The object set O is unlabeled data; that is, each object has no associated class label. However, it is assumed that there are subsets of similar objects in O. These subsets are called clusters. Numerical object data is represented as X = {x 1,...,x N } IR p, where each dimension of the vector x i is a feature value of the associated object o i. These features can be a veritable cornucopia of numerical descriptions, i.e., RGB values, gene

2 expression, year of manufacture, number of stripes, etc. Another way to represent the objects in O is with numerical relational data, which consist of N 2 values that represent the (dis)similarity between pairs of objects. These data are commonly represented by a relational matrix R =[r ij = relation(o i,o j ) 1 i, j N]. The relational matrix often takes the form of a dissimilarity matrix D. Dissimilarity can be interpreted as a distance between objects. For instance, numerical data X can always be converted to D by d ij = x i x j (any vector norm on IR p ). There are, however, similarity and dissimilarity relational data sets that do not begin as numerical object data; for these, there is no choice but to use a relational algorithm. Hence, relational data represent the most general form of input data. An even more general form of relational data is rectangular. These data are represented by an m n dissimilarity matrix D, where the entries are the pair-wise dissimilarity values between m row objects O r and n column objects O c. An example comes from web-document analysis, where the row objects are m web-pages, the columns are n words, and the (dis)similarity entries are occurrence measures of words in web-pages [1]. In each case, the row and column objects are non-intersecting sets, such that the pair-wise relation among row (or column) objects is unknown. Conventional relational clustering algorithms are ill-equipped to deal with rectangular data. Additionally, the definition of a cluster as a group of similar objects takes on a new meaning. There can be groups of similar objects that are composed of only row objects, of only column objects, or of mixed objects (often called co-clusters). In this paper we consider these four types of clusters in rectangular data. Clustering is the process of grouping the objects in O in a sensible manner. This process is often performed to elucidate the similarity and dissimilarity among and between the grouped objects. Clustering has also been called unsupervised learning, typology, and partitioning [2]. Although clustering is typically thought of as only the act of separating objects into the proper groups, cluster analysis actually consists of three concise questions: i) cluster tendency how many clusters are there?; ii) partitioning which objects belong to which cluster and to what degree?; and iii) cluster validity are the partitions good?. The VAT [3] and co-vat [4] algorithms address problem i). 1.1 VAT and ivat Algorithms The VAT algorithm displays an image of reordered and scaled dissimilarity data [3]. Each pixel of the grayscale VAT image I(D ) displays the scaled dissimilarity value of two objects. White pixels represent high dissimilarity, while black represents low dissimilarity. Each object is exactly similar with itself, which results in zero-valued (black) diagonal elements of I(D ). The off-diagonal elements of I(D ) are scaled to the range [0, 1]. A dark block along the diagonal of I(D ) is a sub-matrix of similarly small dissimilarity values; hence, the dark block represents a cluster of objects that are relatively similar to each other. Thus, cluster tendency is shown by the number of dark blocks along the diagonal of the VAT image. The VAT algorithm is based on Prim s algorithm [5] for finding the minimum spanning tree (MST) of a weighted connected graph [3]. Algorithm 1 illustrates the steps of the VAT algorithm. The resulting VAT-reordered dissimilarity matrix D can be nor-

3 malized and mapped to a gray-scale image with black representing the minimum dissimilarity and white the maximum. Algorithm 1: VAT Ordering Algorithm [3] Input: D n n dissimilarity matrix Data: K = {1, 2,...,n}; I = J = ; P =(0, 0,...,0). Select (i, j) arg max p K,q K D pq. Set P (1) = i; I = {i}; and J = K {i}. for r =2,...,ndo Select (i, j) arg min p I,q J D pq. Set P (r) =j; Replace I I {j} and J J {j}. Obtain the ordered dissimilarity matrix D using the ordering array P as: D pq = D P (p),p (q), for 1 p, q n. Output: Reordered dissimilarity D Reference [6] proposed an improved VAT (ivat) algorithm that uses a path-based distance measure from [7]. Consider D to represent the weights of the edges of a fullyconnected graph. The path-based distance is defined as D ij =min max D p[h]p[h+1], (1) 1 h< p p P ij where p P ij is an acyclic path in the set of all acyclic paths between vertex i (o i ) and vertex j (o j ), p[h] is the index of the hth vertex along path p, and p is the number of vertexes along the path. Hence, D p[h]p[h+1] is the weight of the hth edge along path p. Essentially the cost of each path p is the maximum weight of its p edges. The distance between i and j is the cost of the minimum-cost path in P ij. The authors of [6] first transform D into D with (1), then they use VAT on the transformed dissimilarity matrix. The ivat images show considerable improvement over VAT images in showing the cluster tendency for tough cases. Note that computing D exhaustively is very computationally expensive. We can show that i) the ivat dissimilarity matrix D can be computed recursively from the VAT-reordered data D (see Algorithm 2) and, ii) the matrix D computed by Algorithm 2 is already in VAT-order (we have an article in preparation that includes proofs of these assertions). We will denote D as D to indicate that it is a VAT-ordered matrix. Essentially, ivat is a distance transform that improves the visual contrast of the dark blocks along the VAT diagonal. 1.2 co-vat Algorithm The co-vat algorithm begins by creating a square matrix D r c, part of which is composed of the rectangular dissimilarity matrix D. D r c is created by first estimating the dissimilarity matrices D r and D c, which are, respectively, square dissimilarity matrices that relate the objects in O r and O c to themselves i.e. [D r ] ij d(o i,o j ) and [D c ] ij d(o m+i,o m+j ). D r c is organized as in Eq. (2).

4 Algorithm 2: Recursive calculation of ivat image Input: D - VAT-reordered dissimilarity matrix Data: D =[0] n n for r =2,...,ndo 1 j = arg min k=1,...,r 1 Drk 2 D rc = Drc,c= j D rc =max { 3 Drj,D jc},c=1,...,r 1,c j D is symmetric, thus D rc = D cr. [ ] Dr D D r c = D T D c d(o 1,o 1 ) d(o 1,o m ) d(o 1,o m+1 ) d(o 1,o m+n ) d(o m,o 1 ) d(o m,o m ) d(o m,o m+1 ) d(o m,o m+n ) d(o 1,o m+1 ) d(o m,o m+n ) d(o m+1,o m+1 ) d(o m+1,o m+n ) d(o 1,o m+n ) d(o m,o m+n ) d(o m+n,o m+1 ) d(o m+n,o m+n ) (2) The elements in D r and D c are estimated from D using any vector norms on IR n and IR m, [D r ] ij = λ r d i d j, 1 i, j m, (3) [D c ] ij = λ c d i d j, 1 i, j n, (4) where d i is the ith row of D, d j is the jth column of D, and λ r and λ c are scale factors such that the mean of the off-diagonal elements of D r and D c match the mean of D. Section 2.1 presents a new method for finding the reordering of the rectangular dissimilarity matrix D. Section 2.2 adapts the ivat distance transform to co-vat, which shows improved performance over the standard co-vat method. Section 3 presents a numerical example of the new implementations of co-vat and Section 4 concludes this paper. 2 New Implementations of co-vat 2.1 Alternate co-vat reordering scheme The original co-vat algorithm, outlined in Algorithm 3, reorders the rectangular matrix D by shuffling the VAT-reordering indexes of D r c. Thus, co-vat is very dependent on the construction of D r c. We have discovered that the original co-vat fails to show

5 Algorithm 3: co-vat Algorithm [4] Input: D - m n rectangular dissimilarity matrix Build estimates of D r and D c using Eqs.(3) and (4), respectively. Build D r c using Eq.(2). Run VAT on D r c, saving permutation array P r c = {P (1),...,P(m + n)} Initialize rc = cc =0; RP = CP =0. 1 for t =1,...,m+ n do 2 if P (t) m then 3 rc = rc +1, rc is row component 4 RP (rc) =P (t), RP are row indexes else 5 cc = cc +1, cc is column component 6 CP(cc) =P (t) m, CP are column indexes Form the co-vat ordered rectangular dissimilarity matrix, D =[D ij] =[D RP (i)cp(j) ], 1 i m; 1 j n Output: Reordered dissimilarity matrices D, D r, D c, and D r c cluster tendency in certain cases; we have an upcoming paper that discusses these cases in detail. Algorithm 4 presents a reordering scheme that is not dependent on the reordering of D r c this matrix does not even need to be constructed. However, we still need the matrix of the union if we intend to assess cluster tendency in O r c. Essentially, the reordering of the row indexes of D are taken from the VAT-reordering of D r and the reordering of the column indexes are taken from the VAT-reordering of D c. Another advantage of this alternate reordering scheme is that the scale factors, λ r and λ c in (3) and (4), can be ignored. Algorithm 4: Alternate co-vat Reordering Scheme Input: D - m n rectangular dissimilarity matrix Build estimates of D r and D c using Eqs.(3) and (4), respectively. 1 Run VAT on D r, saving permutation array, RP = {RP (1),...,RP(m)} 2 Run VAT on D c, saving permutation array, CP = {CP(1),...,CP(n)} 3 Form the co-vat ordered rectangular dissimilarity matrix, D =[Dij] =[D RP (i)cp(j) ], 1 i m; 1 j n Output: Reordered dissimilarity matrices, D, D r, and D c 2.2 Using the ivat distance transform in co-vat We apply the ivat distance transform in (1) to the three square co-vat matrices, D r, D c, and D r c, using the recursive formulation in Algorithm 2. We denote the transformed matrices as D r, D c, and D r c (examples of these matrices are shown in Figs. 2(a,c,d), respectively). Although, by definition, (1) could applied to the rectangular dissimilarity matrix D by considering D to represent a partially-connected graph

6 (edges only exist between row objects and column objects), applying this transform directly is computationally expensive. However, if we consider the elements of D r c that correspond to elements of the rectangular matrix, we can build the reordered rectangular matrix D from D r c. The rectangular co-ivat image is created as follows: 1. Build D r c and run VAT to produce D r c, where the reordering indexes are P r c = {P (1),...,P(m + n)}. 2. Compute D r c from D r c using the recursive ivat distance transform outlined in Algorithm Build the rectangular co-ivat image D from the corresponding elements of D r c. First, create the reordering indexes K and L, where K are the indexes of the elements of P r c m and L are the indexes of the elements of P r c >m(mis the number of row objects). Then create D by D = [ D ij ] ] = [(D r c) K(i),L(j), 1 i m, 1 j n. (5) We have also adapted ivat to the new reordering scheme of co-vat presented in Algorithm 4. This adaptation requires the construction of D r c, as above, and the corresponding elements of D r c are extracted to create the reordered rectangular matrix D. We denote the co-vat matrices built with the ivat distance transform as co-ivat images. Next we present an example that shows the effectiveness of the new implementations of co-vat. 3 Numerical Example Our numerical example is composed of 360 row objects and 360 column objects, as displayed in Fig. 1(a). Note that, although the number of row objects N equals the number column objects M, this data set is rectangular data because O r O c =Ø. The associated dissimilarity data, calculated using Euclidean distance, is shown in Fig. 1(b). The column objects (shown as green squares) are composed of three groups, the two groups of 50 objects located around coordinates (1.5,3) and (4.5,3), and the 260 objects organized along the curved line extending from the upper-left to the lower-right. The row objects (shown as blue circles) are composed of three groups, the two groups of 50 objects located around coordinates (1.5,0.5) and (3.5,5.5), and the 260 objects organized along the curved line extending from the upper-left to the lower-right. Hence, this example has a preferable cluster tendency of 3 clusters of row objects, 3 clusters composed of column objects, 5 clusters in the union of the row and column objects, and 1 co-cluster. Figure 1(c,d) shows that both co-vat and the new co-vat display the 1 co-cluster as a diagonal band in the upper-left of the image, with both giving approximately equally pleasing results. The co-vat images of D r and D c in Figs. 1(e,f), respectively, clearly show the smaller 2 clusters in each of the row objects and column objects as 2 smaller dark blocks in the lower-right of the image. Again the large co-cluster is shown as a dark diagonal band in the upper-left. The image of D r c is shown in Fig. 1(g); it

7 6 Y coordinate Row objects Column objects X coordinate (a) Object Data (b) Dissimilarity Data - D (c) co-vat image - D (d) New co-vat image - D (e) co-vat image - D r (f) co-vat image - D c (g) co-vat image - D r c Fig. 1: co-vat images of 360 row objects and 360 column objects represented by rectangular dissimilarity data D

8 (a) co-ivat image - D (b) New co-ivat image - D (c) co-ivat image - D r (d) co-ivat image - D c (e) co-ivat image - D r c Fig. 2: co-ivat images of 360 row objects and 360 column objects represented by rectangular dissimilarity data D shows the 5 clusters in O r O c as the four dark blocks in the lower-right and the dark diagonal band in the upper-left. While we hesitate to say that co-vat and the new co-vat have failed for this example, we believe that the large diagonal band leads to ambiguity as to the cluster tendency of these data. Further more, the contrast in Figs. 1(c,d) make it difficult to determine the number of co-clusters. Figure 2 shows the corresponding co-ivat images of the rectangular dissimilarity data shown in Fig. 1. The co-ivat images give a very clear view of the cluster tendency for each of the four types of clusters: D shows 1 co-cluster, D r shows 3 row-clusters, D c shows 3 column-clusters, and D r c shows 5 clusters in O r O c. 4 Conclusion This paper presented a new implementation of the co-vat algorithm with two innovations; a new reordering scheme was presented in Algorithm 4 and each co-vat algorithm was adapted to include the distance transform in (1). Although the numerical example shown in Fig. 1 does not clearly show the strength of the alternate reordering scheme presented in Algorithm 4, we are currently preparing a paper that will present several examples, many of which show that the alternate reordering scheme performs well in cases where the original co-vat fails. Moreover, we emphasize that the alter-

9 nate reordering scheme is computationally less expensive as D r c does not need to be constructed or VAT-reordered. Figure 2 shows that the co-ivat images are clearly superior to the original co-vat images in showing the cluster tendency of the four types of clusters in the rectangular data. Additionally, due to our recursive implementation in Algorithm 2, these improved images come at very little additional computational cost. We left many questions unanswered in this paper and are currently penning an article that describes, in detail, each of the following: 1. What are the advantages and disadvantages of the alternate reordering scheme presented in Section 2.1? We have discovered pure relational data (data for which object data X does not exist) on which the original co-vat formulation fails and on which the alternate reordering scheme is successful. Additionally, we have revisited the normalization of D r and D c in (3) and (4) and have devised different ways to normalize these matrices which show improvement in performance. 2. We have developed proofs of our assertions on ivat presented in Section 1.1 and the recursive formulation of (1) outlined in Algorithm The path-based distance transform can be applied to the rectangular dissimilarity data directly. We are working on an algorithm, similar to Algorithm 2, that will compute the distance transform of the rectangular dissimilarity data, without having to exhaustively search through every path in the partially connected tree. 4. The sco-vat [8] algorithm performs the operations of co-vat for very-large (unloadable) data. We are extending the co-vat implementation presented here to the sco-vat algorithm. 5. As always, we wish to demonstrate the performance of our algorithms on real data. However, gold-standard rectangular data-sets are not as ubiquitous as square relational data-sets. Hence, we are identifying data on which we can validate our work.

10 Bibliography [1] Dhillon, I.: Co-clustering documents and words using bipartite spectral graph partitioning. In: Proc. 7th ACM SIGKDD Int. Conf. on Knowledge Discovery Data Mining, San Francisco, CA (2001) [2] Theodoridis, S., Koutroumbas, K.: Pattern Recognition. 3rd edn. Academic Press, San Diego, CA (2006) [3] Bezdek, J., Hathaway, R.: VAT: A tool for visual assessment of (cluster) tendency. In: Proc. IJCNN 2002, Honolulu, HI (2002) [4] Bezdek, J., Hathaway, R., Huband, J.: Visual assessment of clustering tendency for rectangular dissimilarity matrices. IEEE Trans. Fuzzy Systems 15(5) (October 2007) [5] Prim, R.: Shortest connection networks and some generalisations. Bell System Tech. J. 36 (1957) [6] Wang, L., Nguyen, T., Bezdek, J., Leckie, C., Ramamohanarao, K.: ivat and avat: enhanced visual analysis for cluster tendency assessment. in review (2009) [7] Fisher, B., Zoller, T., Buhmann, J.: Path based pairwise data clustering with application to texture segmentation. Energy Minimization Methods in Computer Vision and Pattern Recognition 2134 (2001) [8] Park, L., Bezdek, J., Leckie, C.: Visualization of clusters in very large rectangular dissimilarity data. In Gupta, G.S., Mukhopadhyay, S., eds.: Proc. 4th Int. Conf. Autonomous Robots and Agents. (February 2009)

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM 96 CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM Clustering is the process of combining a set of relevant information in the same group. In this process KM algorithm plays

More information

Enhancing Clustering Results In Hierarchical Approach By Mvs Measures

Enhancing Clustering Results In Hierarchical Approach By Mvs Measures International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 10, Issue 6 (June 2014), PP.25-30 Enhancing Clustering Results In Hierarchical Approach

More information

REDUCING RUNTIME VALUES IN MINIMUM SPANNING TREE BASED CLUSTERING BY VISUAL ACCESS TENDENCY

REDUCING RUNTIME VALUES IN MINIMUM SPANNING TREE BASED CLUSTERING BY VISUAL ACCESS TENDENCY REDUCING RUNTIME VALUES IN MINIMUM SPANNING TREE BASED CLUSTERING BY VISUAL ACCESS TENDENCY Dr. B. Eswara Reddy 1 and K. Rajendra Prasad 2 1 Associate Professor, Head of CSE Department, JNTUA College of

More information

Bipartite Graph Partitioning and Content-based Image Clustering

Bipartite Graph Partitioning and Content-based Image Clustering Bipartite Graph Partitioning and Content-based Image Clustering Guoping Qiu School of Computer Science The University of Nottingham qiu @ cs.nott.ac.uk Abstract This paper presents a method to model the

More information

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs Graphs and Network Flows IE411 Lecture 21 Dr. Ted Ralphs IE411 Lecture 21 1 Combinatorial Optimization and Network Flows In general, most combinatorial optimization and integer programming problems are

More information

Biclustering Bioinformatics Data Sets. A Possibilistic Approach

Biclustering Bioinformatics Data Sets. A Possibilistic Approach Possibilistic algorithm Bioinformatics Data Sets: A Possibilistic Approach Dept Computer and Information Sciences, University of Genova ITALY EMFCSC Erice 20/4/2007 Bioinformatics Data Sets Outline Introduction

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2006-010 Revisiting hypergraph models for sparse matrix decomposition by Cevdet Aykanat, Bora Ucar Mathematics and Computer Science EMORY UNIVERSITY REVISITING HYPERGRAPH MODELS FOR

More information

Reflexive Regular Equivalence for Bipartite Data

Reflexive Regular Equivalence for Bipartite Data Reflexive Regular Equivalence for Bipartite Data Aaron Gerow 1, Mingyang Zhou 2, Stan Matwin 1, and Feng Shi 3 1 Faculty of Computer Science, Dalhousie University, Halifax, NS, Canada 2 Department of Computer

More information

Supervised vs. Unsupervised Learning

Supervised vs. Unsupervised Learning Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now

More information

An Approach to Identify the Number of Clusters

An Approach to Identify the Number of Clusters An Approach to Identify the Number of Clusters Katelyn Gao Heather Hardeman Edward Lim Cristian Potter Carl Meyer Ralph Abbey July 11, 212 Abstract In this technological age, vast amounts of data are generated.

More information

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING Idan Ram, Michael Elad and Israel Cohen Department of Electrical Engineering Department of Computer Science Technion - Israel Institute of Technology

More information

REGULAR GRAPHS OF GIVEN GIRTH. Contents

REGULAR GRAPHS OF GIVEN GIRTH. Contents REGULAR GRAPHS OF GIVEN GIRTH BROOKE ULLERY Contents 1. Introduction This paper gives an introduction to the area of graph theory dealing with properties of regular graphs of given girth. A large portion

More information

6. Concluding Remarks

6. Concluding Remarks [8] K. J. Supowit, The relative neighborhood graph with an application to minimum spanning trees, Tech. Rept., Department of Computer Science, University of Illinois, Urbana-Champaign, August 1980, also

More information

Implementation of Fuzzy C-Means and Possibilistic C-Means Clustering Algorithms, Cluster Tendency Analysis and Cluster Validation

Implementation of Fuzzy C-Means and Possibilistic C-Means Clustering Algorithms, Cluster Tendency Analysis and Cluster Validation Implementation of Fuzzy C-Means and Possibilistic C-Means Clustering Algorithms, Cluster Tendency Analysis and Cluster Validation Md. Abu Bakr Siddiue *, Rezoana Bente Arif #, Mohammad Mahmudur Rahman

More information

Math 485, Graph Theory: Homework #3

Math 485, Graph Theory: Homework #3 Math 485, Graph Theory: Homework #3 Stephen G Simpson Due Monday, October 26, 2009 The assignment consists of Exercises 2129, 2135, 2137, 2218, 238, 2310, 2313, 2314, 2315 in the West textbook, plus the

More information

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering.

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering. Chapter 4 Fuzzy Segmentation 4. Introduction. The segmentation of objects whose color-composition is not common represents a difficult task, due to the illumination and the appropriate threshold selection

More information

arxiv: v5 [cs.dm] 9 May 2016

arxiv: v5 [cs.dm] 9 May 2016 Tree spanners of bounded degree graphs Ioannis Papoutsakis Kastelli Pediados, Heraklion, Crete, reece, 700 06 October 21, 2018 arxiv:1503.06822v5 [cs.dm] 9 May 2016 Abstract A tree t-spanner of a graph

More information

AMS /672: Graph Theory Homework Problems - Week V. Problems to be handed in on Wednesday, March 2: 6, 8, 9, 11, 12.

AMS /672: Graph Theory Homework Problems - Week V. Problems to be handed in on Wednesday, March 2: 6, 8, 9, 11, 12. AMS 550.47/67: Graph Theory Homework Problems - Week V Problems to be handed in on Wednesday, March : 6, 8, 9,,.. Assignment Problem. Suppose we have a set {J, J,..., J r } of r jobs to be filled by a

More information

Research Question Presentation on the Edge Clique Covers of a Complete Multipartite Graph. Nechama Florans. Mentor: Dr. Boram Park

Research Question Presentation on the Edge Clique Covers of a Complete Multipartite Graph. Nechama Florans. Mentor: Dr. Boram Park Research Question Presentation on the Edge Clique Covers of a Complete Multipartite Graph Nechama Florans Mentor: Dr. Boram Park G: V 5 Vertex Clique Covers and Edge Clique Covers: Suppose we have a graph

More information

Efficient Semi-supervised Spectral Co-clustering with Constraints

Efficient Semi-supervised Spectral Co-clustering with Constraints 2010 IEEE International Conference on Data Mining Efficient Semi-supervised Spectral Co-clustering with Constraints Xiaoxiao Shi, Wei Fan, Philip S. Yu Department of Computer Science, University of Illinois

More information

Incorporating Known Pathways into Gene Clustering Algorithms for Genetic Expression Data

Incorporating Known Pathways into Gene Clustering Algorithms for Genetic Expression Data Incorporating Known Pathways into Gene Clustering Algorithms for Genetic Expression Data Ryan Atallah, John Ryan, David Aeschlimann December 14, 2013 Abstract In this project, we study the problem of classifying

More information

Graph Laplacian Kernels for Object Classification from a Single Example

Graph Laplacian Kernels for Object Classification from a Single Example Graph Laplacian Kernels for Object Classification from a Single Example Hong Chang & Dit-Yan Yeung Department of Computer Science, Hong Kong University of Science and Technology {hongch,dyyeung}@cs.ust.hk

More information

A Computer Vision System for Graphical Pattern Recognition and Semantic Object Detection

A Computer Vision System for Graphical Pattern Recognition and Semantic Object Detection A Computer Vision System for Graphical Pattern Recognition and Semantic Object Detection Tudor Barbu Institute of Computer Science, Iaşi, Romania Abstract We have focused on a set of problems related to

More information

Visual Representations for Machine Learning

Visual Representations for Machine Learning Visual Representations for Machine Learning Spectral Clustering and Channel Representations Lecture 1 Spectral Clustering: introduction and confusion Michael Felsberg Klas Nordberg The Spectral Clustering

More information

An Algorithm for Clustering Tendency Assessment

An Algorithm for Clustering Tendency Assessment An Algorithm for Clustering Tendency Assessment Yingkang Hu Georgia Southern University Department of Mathematical Sciences Statesboro, GA 3-93 USA yhu@georgiasouthern.edu Richard J. Hathaway Georgia Southern

More information

Lab # 2 - ACS I Part I - DATA COMPRESSION in IMAGE PROCESSING using SVD

Lab # 2 - ACS I Part I - DATA COMPRESSION in IMAGE PROCESSING using SVD Lab # 2 - ACS I Part I - DATA COMPRESSION in IMAGE PROCESSING using SVD Goals. The goal of the first part of this lab is to demonstrate how the SVD can be used to remove redundancies in data; in this example

More information

Minimum-Spanning-Tree problem. Minimum Spanning Trees (Forests) Minimum-Spanning-Tree problem

Minimum-Spanning-Tree problem. Minimum Spanning Trees (Forests) Minimum-Spanning-Tree problem Minimum Spanning Trees (Forests) Given an undirected graph G=(V,E) with each edge e having a weight w(e) : Find a subgraph T of G of minimum total weight s.t. every pair of vertices connected in G are

More information

Big Data Analytics. Special Topics for Computer Science CSE CSE Feb 11

Big Data Analytics. Special Topics for Computer Science CSE CSE Feb 11 Big Data Analytics Special Topics for Computer Science CSE 4095-001 CSE 5095-005 Feb 11 Fei Wang Associate Professor Department of Computer Science and Engineering fei_wang@uconn.edu Clustering II Spectral

More information

Mining Social Network Graphs

Mining Social Network Graphs Mining Social Network Graphs Analysis of Large Graphs: Community Detection Rafael Ferreira da Silva rafsilva@isi.edu http://rafaelsilva.com Note to other teachers and users of these slides: We would be

More information

Motion Detection. Final project by. Neta Sokolovsky

Motion Detection. Final project by. Neta Sokolovsky Motion Detection Final project by Neta Sokolovsky Introduction The goal of this project is to recognize a motion of objects found in the two given images. This functionality is useful in the video processing

More information

PACKING DIGRAPHS WITH DIRECTED CLOSED TRAILS

PACKING DIGRAPHS WITH DIRECTED CLOSED TRAILS PACKING DIGRAPHS WITH DIRECTED CLOSED TRAILS PAUL BALISTER Abstract It has been shown [Balister, 2001] that if n is odd and m 1,, m t are integers with m i 3 and t i=1 m i = E(K n) then K n can be decomposed

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

CS 787: Assignment 4, Stereo Vision: Block Matching and Dynamic Programming Due: 12:00noon, Fri. Mar. 30, 2007.

CS 787: Assignment 4, Stereo Vision: Block Matching and Dynamic Programming Due: 12:00noon, Fri. Mar. 30, 2007. CS 787: Assignment 4, Stereo Vision: Block Matching and Dynamic Programming Due: 12:00noon, Fri. Mar. 30, 2007. In this assignment you will implement and test some simple stereo algorithms discussed in

More information

A graph is finite if its vertex set and edge set are finite. We call a graph with just one vertex trivial and all other graphs nontrivial.

A graph is finite if its vertex set and edge set are finite. We call a graph with just one vertex trivial and all other graphs nontrivial. 2301-670 Graph theory 1.1 What is a graph? 1 st semester 2550 1 1.1. What is a graph? 1.1.2. Definition. A graph G is a triple (V(G), E(G), ψ G ) consisting of V(G) of vertices, a set E(G), disjoint from

More information

CS 5540 Spring 2013 Assignment 3, v1.0 Due: Apr. 24th 11:59PM

CS 5540 Spring 2013 Assignment 3, v1.0 Due: Apr. 24th 11:59PM 1 Introduction In this programming project, we are going to do a simple image segmentation task. Given a grayscale image with a bright object against a dark background and we are going to do a binary decision

More information

Digital Halftoning Algorithm Based o Space-Filling Curve

Digital Halftoning Algorithm Based o Space-Filling Curve JAIST Reposi https://dspace.j Title Digital Halftoning Algorithm Based o Space-Filling Curve Author(s)ASANO, Tetsuo Citation IEICE TRANSACTIONS on Fundamentals o Electronics, Communications and Comp Sciences,

More information

CSE 547: Machine Learning for Big Data Spring Problem Set 2. Please read the homework submission policies.

CSE 547: Machine Learning for Big Data Spring Problem Set 2. Please read the homework submission policies. CSE 547: Machine Learning for Big Data Spring 2019 Problem Set 2 Please read the homework submission policies. 1 Principal Component Analysis and Reconstruction (25 points) Let s do PCA and reconstruct

More information

by conservation of flow, hence the cancelation. Similarly, we have

by conservation of flow, hence the cancelation. Similarly, we have Chapter 13: Network Flows and Applications Network: directed graph with source S and target T. Non-negative edge weights represent capacities. Assume no edges into S or out of T. (If necessary, we can

More information

CIS 121 Data Structures and Algorithms Minimum Spanning Trees

CIS 121 Data Structures and Algorithms Minimum Spanning Trees CIS 121 Data Structures and Algorithms Minimum Spanning Trees March 19, 2019 Introduction and Background Consider a very natural problem: we are given a set of locations V = {v 1, v 2,..., v n }. We want

More information

SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM. Olga Kouropteva, Oleg Okun and Matti Pietikäinen

SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM. Olga Kouropteva, Oleg Okun and Matti Pietikäinen SELECTION OF THE OPTIMAL PARAMETER VALUE FOR THE LOCALLY LINEAR EMBEDDING ALGORITHM Olga Kouropteva, Oleg Okun and Matti Pietikäinen Machine Vision Group, Infotech Oulu and Department of Electrical and

More information

The Ordered Covering Problem

The Ordered Covering Problem The Ordered Covering Problem Uriel Feige Yael Hitron November 8, 2016 Abstract We introduce the Ordered Covering (OC) problem. The input is a finite set of n elements X, a color function c : X {0, 1} and

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK REVIEW ON CONTENT BASED IMAGE RETRIEVAL BY USING VISUAL SEARCH RANKING MS. PRAGATI

More information

Properly even harmonious labelings of disconnected graphs

Properly even harmonious labelings of disconnected graphs Available online at www.sciencedirect.com ScienceDirect AKCE International Journal of Graphs and Combinatorics 12 (2015) 193 203 www.elsevier.com/locate/akcej Properly even harmonious labelings of disconnected

More information

Rectangular Partitioning

Rectangular Partitioning Rectangular Partitioning Joe Forsmann and Rock Hymas Introduction/Abstract We will look at a problem that I (Rock) had to solve in the course of my work. Given a set of non-overlapping rectangles each

More information

Non-exhaustive, Overlapping k-means

Non-exhaustive, Overlapping k-means Non-exhaustive, Overlapping k-means J. J. Whang, I. S. Dhilon, and D. F. Gleich Teresa Lebair University of Maryland, Baltimore County October 29th, 2015 Teresa Lebair UMBC 1/38 Outline Introduction NEO-K-Means

More information

ON THE STRUCTURE OF SELF-COMPLEMENTARY GRAPHS ROBERT MOLINA DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE ALMA COLLEGE ABSTRACT

ON THE STRUCTURE OF SELF-COMPLEMENTARY GRAPHS ROBERT MOLINA DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE ALMA COLLEGE ABSTRACT ON THE STRUCTURE OF SELF-COMPLEMENTARY GRAPHS ROBERT MOLINA DEPARTMENT OF MATHEMATICS AND COMPUTER SCIENCE ALMA COLLEGE ABSTRACT A graph G is self complementary if it is isomorphic to its complement G.

More information

Similarity-Driven Cluster Merging Method for Unsupervised Fuzzy Clustering

Similarity-Driven Cluster Merging Method for Unsupervised Fuzzy Clustering Similarity-Driven Cluster Merging Method for Unsupervised Fuzzy Clustering Xuejian Xiong, Kian Lee Tan Singapore-MIT Alliance E4-04-10, 4 Engineering Drive 3 Singapore 117576 Abstract In this paper, a

More information

A REVIEW ON IMAGE RETRIEVAL USING HYPERGRAPH

A REVIEW ON IMAGE RETRIEVAL USING HYPERGRAPH A REVIEW ON IMAGE RETRIEVAL USING HYPERGRAPH Sandhya V. Kawale Prof. Dr. S. M. Kamalapur M.E. Student Associate Professor Deparment of Computer Engineering, Deparment of Computer Engineering, K. K. Wagh

More information

Efficient Algorithms for Solving Shortest Paths on Nearly Acyclic Directed Graphs

Efficient Algorithms for Solving Shortest Paths on Nearly Acyclic Directed Graphs Efficient Algorithms for Solving Shortest Paths on Nearly Acyclic Directed Graphs Shane Saunders Tadao Takaoka Department of Computer Science and Software Engineering, University of Canterbury, Christchurch,

More information

ON THE STRONGLY REGULAR GRAPH OF PARAMETERS

ON THE STRONGLY REGULAR GRAPH OF PARAMETERS ON THE STRONGLY REGULAR GRAPH OF PARAMETERS (99, 14, 1, 2) SUZY LOU AND MAX MURIN Abstract. In an attempt to find a strongly regular graph of parameters (99, 14, 1, 2) or to disprove its existence, we

More information

On the Relationships between Zero Forcing Numbers and Certain Graph Coverings

On the Relationships between Zero Forcing Numbers and Certain Graph Coverings On the Relationships between Zero Forcing Numbers and Certain Graph Coverings Fatemeh Alinaghipour Taklimi, Shaun Fallat 1,, Karen Meagher 2 Department of Mathematics and Statistics, University of Regina,

More information

Recognizing Interval Bigraphs by Forbidden Patterns

Recognizing Interval Bigraphs by Forbidden Patterns Recognizing Interval Bigraphs by Forbidden Patterns Arash Rafiey Simon Fraser University, Vancouver, Canada, and Indiana State University, IN, USA arashr@sfu.ca, arash.rafiey@indstate.edu Abstract Let

More information

Elastic Partial Matching of Time Series

Elastic Partial Matching of Time Series Elastic Partial Matching of Time Series L. J. Latecki 1, V. Megalooikonomou 1, Q. Wang 1, R. Lakaemper 1, C. A. Ratanamahatana 2, and E. Keogh 2 1 Computer and Information Sciences Dept. Temple University,

More information

The Clustering Validity with Silhouette and Sum of Squared Errors

The Clustering Validity with Silhouette and Sum of Squared Errors Proceedings of the 3rd International Conference on Industrial Application Engineering 2015 The Clustering Validity with Silhouette and Sum of Squared Errors Tippaya Thinsungnoen a*, Nuntawut Kaoungku b,

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

LATIN SQUARES AND THEIR APPLICATION TO THE FEASIBLE SET FOR ASSIGNMENT PROBLEMS

LATIN SQUARES AND THEIR APPLICATION TO THE FEASIBLE SET FOR ASSIGNMENT PROBLEMS LATIN SQUARES AND THEIR APPLICATION TO THE FEASIBLE SET FOR ASSIGNMENT PROBLEMS TIMOTHY L. VIS Abstract. A significant problem in finite optimization is the assignment problem. In essence, the assignment

More information

A fuzzy k-modes algorithm for clustering categorical data. Citation IEEE Transactions on Fuzzy Systems, 1999, v. 7 n. 4, p.

A fuzzy k-modes algorithm for clustering categorical data. Citation IEEE Transactions on Fuzzy Systems, 1999, v. 7 n. 4, p. Title A fuzzy k-modes algorithm for clustering categorical data Author(s) Huang, Z; Ng, MKP Citation IEEE Transactions on Fuzzy Systems, 1999, v. 7 n. 4, p. 446-452 Issued Date 1999 URL http://hdl.handle.net/10722/42992

More information

Lorentzian Distance Classifier for Multiple Features

Lorentzian Distance Classifier for Multiple Features Yerzhan Kerimbekov 1 and Hasan Şakir Bilge 2 1 Department of Computer Engineering, Ahmet Yesevi University, Ankara, Turkey 2 Department of Electrical-Electronics Engineering, Gazi University, Ankara, Turkey

More information

Graphs, graph algorithms (for image segmentation),... in progress

Graphs, graph algorithms (for image segmentation),... in progress Graphs, graph algorithms (for image segmentation),... in progress Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 66 36 Prague 6, Jugoslávských

More information

Decreasing a key FIB-HEAP-DECREASE-KEY(,, ) 3.. NIL. 2. error new key is greater than current key 6. CASCADING-CUT(, )

Decreasing a key FIB-HEAP-DECREASE-KEY(,, ) 3.. NIL. 2. error new key is greater than current key 6. CASCADING-CUT(, ) Decreasing a key FIB-HEAP-DECREASE-KEY(,, ) 1. if >. 2. error new key is greater than current key 3.. 4.. 5. if NIL and.

More information

Clustering and Visualisation of Data

Clustering and Visualisation of Data Clustering and Visualisation of Data Hiroshi Shimodaira January-March 28 Cluster analysis aims to partition a data set into meaningful or useful groups, based on distances between data points. In some

More information

Graph Theory S 1 I 2 I 1 S 2 I 1 I 2

Graph Theory S 1 I 2 I 1 S 2 I 1 I 2 Graph Theory S I I S S I I S Graphs Definition A graph G is a pair consisting of a vertex set V (G), and an edge set E(G) ( ) V (G). x and y are the endpoints of edge e = {x, y}. They are called adjacent

More information

Pattern Recognition Methods for Object Boundary Detection

Pattern Recognition Methods for Object Boundary Detection Pattern Recognition Methods for Object Boundary Detection Arnaldo J. Abrantesy and Jorge S. Marquesz yelectronic and Communication Engineering Instituto Superior Eng. de Lisboa R. Conselheiro Emídio Navarror

More information

Texture Segmentation by Windowed Projection

Texture Segmentation by Windowed Projection Texture Segmentation by Windowed Projection 1, 2 Fan-Chen Tseng, 2 Ching-Chi Hsu, 2 Chiou-Shann Fuh 1 Department of Electronic Engineering National I-Lan Institute of Technology e-mail : fctseng@ccmail.ilantech.edu.tw

More information

Unsupervised Learning

Unsupervised Learning Outline Unsupervised Learning Basic concepts K-means algorithm Representation of clusters Hierarchical clustering Distance functions Which clustering algorithm to use? NN Supervised learning vs. unsupervised

More information

Clustering. Supervised vs. Unsupervised Learning

Clustering. Supervised vs. Unsupervised Learning Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now

More information

An Introduction to Graph Theory

An Introduction to Graph Theory An Introduction to Graph Theory CIS008-2 Logic and Foundations of Mathematics David Goodwin david.goodwin@perisic.com 12:00, Friday 17 th February 2012 Outline 1 Graphs 2 Paths and cycles 3 Graphs and

More information

2.3 Algorithms Using Map-Reduce

2.3 Algorithms Using Map-Reduce 28 CHAPTER 2. MAP-REDUCE AND THE NEW SOFTWARE STACK one becomes available. The Master must also inform each Reduce task that the location of its input from that Map task has changed. Dealing with a failure

More information

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.

More information

International Journal of Advanced Research in Computer Science and Software Engineering

International Journal of Advanced Research in Computer Science and Software Engineering Volume 2, Issue 10, October 2012 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A New Efficient

More information

A Novel Approach for Minimum Spanning Tree Based Clustering Algorithm

A Novel Approach for Minimum Spanning Tree Based Clustering Algorithm IJCSES International Journal of Computer Sciences and Engineering Systems, Vol. 5, No. 2, April 2011 CSES International 2011 ISSN 0973-4406 A Novel Approach for Minimum Spanning Tree Based Clustering Algorithm

More information

Semi-Supervised Clustering with Partial Background Information

Semi-Supervised Clustering with Partial Background Information Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject

More information

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural

VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD. Ertem Tuncel and Levent Onural VIDEO OBJECT SEGMENTATION BY EXTENDED RECURSIVE-SHORTEST-SPANNING-TREE METHOD Ertem Tuncel and Levent Onural Electrical and Electronics Engineering Department, Bilkent University, TR-06533, Ankara, Turkey

More information

CS 341: Algorithms. Douglas R. Stinson. David R. Cheriton School of Computer Science University of Waterloo. February 26, 2019

CS 341: Algorithms. Douglas R. Stinson. David R. Cheriton School of Computer Science University of Waterloo. February 26, 2019 CS 341: Algorithms Douglas R. Stinson David R. Cheriton School of Computer Science University of Waterloo February 26, 2019 D.R. Stinson (SCS) CS 341 February 26, 2019 1 / 296 1 Course Information 2 Introduction

More information

Handling Missing Values via Decomposition of the Conditioned Set

Handling Missing Values via Decomposition of the Conditioned Set Handling Missing Values via Decomposition of the Conditioned Set Mei-Ling Shyu, Indika Priyantha Kuruppu-Appuhamilage Department of Electrical and Computer Engineering, University of Miami Coral Gables,

More information

Algebraic Graph Theory- Adjacency Matrix and Spectrum

Algebraic Graph Theory- Adjacency Matrix and Spectrum Algebraic Graph Theory- Adjacency Matrix and Spectrum Michael Levet December 24, 2013 Introduction This tutorial will introduce the adjacency matrix, as well as spectral graph theory. For those familiar

More information

Stats 170A: Project in Data Science Exploratory Data Analysis: Clustering Algorithms

Stats 170A: Project in Data Science Exploratory Data Analysis: Clustering Algorithms Stats 170A: Project in Data Science Exploratory Data Analysis: Clustering Algorithms Padhraic Smyth Department of Computer Science Bren School of Information and Computer Sciences University of California,

More information

Math 778S Spectral Graph Theory Handout #2: Basic graph theory

Math 778S Spectral Graph Theory Handout #2: Basic graph theory Math 778S Spectral Graph Theory Handout #: Basic graph theory Graph theory was founded by the great Swiss mathematician Leonhard Euler (1707-178) after he solved the Königsberg Bridge problem: Is it possible

More information

Image Classification through Dynamic Hyper Graph Learning

Image Classification through Dynamic Hyper Graph Learning Image Classification through Dynamic Hyper Graph Learning Ms. Govada Sahitya, Dept of ECE, St. Ann's College of Engineering and Technology,chirala. J. Lakshmi Narayana,(Ph.D), Associate Professor, Dept

More information

6. Lecture notes on matroid intersection

6. Lecture notes on matroid intersection Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans May 2, 2017 6. Lecture notes on matroid intersection One nice feature about matroids is that a simple greedy algorithm

More information

General Instructions. Questions

General Instructions. Questions CS246: Mining Massive Data Sets Winter 2018 Problem Set 2 Due 11:59pm February 8, 2018 Only one late period is allowed for this homework (11:59pm 2/13). General Instructions Submission instructions: These

More information

Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots

Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots Variable-resolution Velocity Roadmap Generation Considering Safety Constraints for Mobile Robots Jingyu Xiang, Yuichi Tazaki, Tatsuya Suzuki and B. Levedahl Abstract This research develops a new roadmap

More information

Progress Report: Collaborative Filtering Using Bregman Co-clustering

Progress Report: Collaborative Filtering Using Bregman Co-clustering Progress Report: Collaborative Filtering Using Bregman Co-clustering Wei Tang, Srivatsan Ramanujam, and Andrew Dreher April 4, 2008 1 Introduction Analytics are becoming increasingly important for business

More information

Controlling the spread of dynamic self-organising maps

Controlling the spread of dynamic self-organising maps Neural Comput & Applic (2004) 13: 168 174 DOI 10.1007/s00521-004-0419-y ORIGINAL ARTICLE L. D. Alahakoon Controlling the spread of dynamic self-organising maps Received: 7 April 2004 / Accepted: 20 April

More information

Structural and Syntactic Pattern Recognition

Structural and Syntactic Pattern Recognition Structural and Syntactic Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent

More information

9/29/13. Outline Data mining tasks. Clustering algorithms. Applications of clustering in biology

9/29/13. Outline Data mining tasks. Clustering algorithms. Applications of clustering in biology 9/9/ I9 Introduction to Bioinformatics, Clustering algorithms Yuzhen Ye (yye@indiana.edu) School of Informatics & Computing, IUB Outline Data mining tasks Predictive tasks vs descriptive tasks Example

More information

Chapter 18 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal.

Chapter 18 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal. Chapter 8 out of 7 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal 8 Matrices Definitions and Basic Operations Matrix algebra is also known

More information

Manuscript Click here to download Manuscript: Interval neutrosophic MST clustering algorithm and its an application to taxonomy.

Manuscript Click here to download Manuscript: Interval neutrosophic MST clustering algorithm and its an application to taxonomy. Manuscript Click here to download Manuscript: Interval neutrosophic MST clustering algorithm and its an application to taxonomy.pdf 0 0 0 0 0 INTERVAL NEUTROSOPHIC MST CLUSTERING ALGORITHM AND ITS AN APPLICATION

More information

Matrix Multiplication

Matrix Multiplication Matrix Multiplication CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Matrix Multiplication Spring 2013 1 / 32 Outline 1 Matrix operations Importance Dense and sparse

More information

Basic relations between pixels (Chapter 2)

Basic relations between pixels (Chapter 2) Basic relations between pixels (Chapter 2) Lecture 3 Basic Relationships Between Pixels Definitions: f(x,y): digital image Pixels: q, p (p,q f) A subset of pixels of f(x,y): S A typology of relations:

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

8. Clustering: Pattern Classification by Distance Functions

8. Clustering: Pattern Classification by Distance Functions CEE 6: Digital Image Processing Topic : Clustering/Unsupervised Classification - W. Philpot, Cornell University, January 0. Clustering: Pattern Classification by Distance Functions The premise in clustering

More information

CS Introduction to Data Mining Instructor: Abdullah Mueen

CS Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts

More information

Image Segmentation. Selim Aksoy. Bilkent University

Image Segmentation. Selim Aksoy. Bilkent University Image Segmentation Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Examples of grouping in vision [http://poseidon.csd.auth.gr/lab_research/latest/imgs/s peakdepvidindex_img2.jpg]

More information

18.S34 (FALL 2007) PROBLEMS ON HIDDEN INDEPENDENCE AND UNIFORMITY

18.S34 (FALL 2007) PROBLEMS ON HIDDEN INDEPENDENCE AND UNIFORMITY 18.S34 (FALL 2007) PROBLEMS ON HIDDEN INDEPENDENCE AND UNIFORMITY All the problems below (with the possible exception of the last one), when looked at the right way, can be solved by elegant arguments

More information

Digital Image Processing. Introduction

Digital Image Processing. Introduction Digital Image Processing Introduction Digital Image Definition An image can be defined as a twodimensional function f(x,y) x,y: Spatial coordinate F: the amplitude of any pair of coordinate x,y, which

More information

CS 664 Slides #11 Image Segmentation. Prof. Dan Huttenlocher Fall 2003

CS 664 Slides #11 Image Segmentation. Prof. Dan Huttenlocher Fall 2003 CS 664 Slides #11 Image Segmentation Prof. Dan Huttenlocher Fall 2003 Image Segmentation Find regions of image that are coherent Dual of edge detection Regions vs. boundaries Related to clustering problems

More information

Figure 2.1: A bipartite graph.

Figure 2.1: A bipartite graph. Matching problems The dance-class problem. A group of boys and girls, with just as many boys as girls, want to dance together; hence, they have to be matched in couples. Each boy prefers to dance with

More information

Clustering of Data with Mixed Attributes based on Unified Similarity Metric

Clustering of Data with Mixed Attributes based on Unified Similarity Metric Clustering of Data with Mixed Attributes based on Unified Similarity Metric M.Soundaryadevi 1, Dr.L.S.Jayashree 2 Dept of CSE, RVS College of Engineering and Technology, Coimbatore, Tamilnadu, India 1

More information

Analysis of Dendrogram Tree for Identifying and Visualizing Trends in Multi-attribute Transactional Data

Analysis of Dendrogram Tree for Identifying and Visualizing Trends in Multi-attribute Transactional Data Analysis of Dendrogram Tree for Identifying and Visualizing Trends in Multi-attribute Transactional Data D.Radha Rani 1, A.Vini Bharati 2, P.Lakshmi Durga Madhuri 3, M.Phaneendra Babu 4, A.Sravani 5 Department

More information