Gene expression & Clustering (Chapter 10)

Similar documents
Clustering Techniques

CLUSTERING IN BIOINFORMATICS

EECS730: Introduction to Bioinformatics

9/29/13. Outline Data mining tasks. Clustering algorithms. Applications of clustering in biology

11/17/2009 Comp 590/Comp Fall

4/4/16 Comp 555 Spring

Lecture 20: Clustering and Evolution

Lecture 20: Clustering and Evolution

Gene Clustering & Classification

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clustering. RNA-seq: What is it good for? Finding Similarly Expressed Genes. Data... And Lots of It!

SVM Classification in -Arrays

Clustering. Lecture 6, 1/24/03 ECS289A

SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

Clustering CS 550: Machine Learning

CSE 5243 INTRO. TO DATA MINING

10701 Machine Learning. Clustering

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Contents. ! Data sets. ! Distance and similarity metrics. ! K-means clustering. ! Hierarchical clustering. ! Evaluation of clustering results

Clustering Jacques van Helden

Exploratory data analysis for microarrays

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search

Evolutionary tree reconstruction (Chapter 10)

Hierarchical Clustering

CSE 5243 INTRO. TO DATA MINING

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010

ECLT 5810 Clustering

ECLT 5810 Clustering

CS Introduction to Data Mining Instructor: Abdullah Mueen

ECS 234: Data Analysis: Clustering ECS 234

10601 Machine Learning. Hierarchical clustering. Reading: Bishop: 9-9.2

Clustering: Centroid-Based Partitioning

Lecture 5 Finding meaningful clusters in data. 5.1 Kleinberg s axiomatic framework for clustering

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Clustering: K-means and Kernel K-means

10 Clustering and Trees

Cluster Analysis. Ying Shen, SSE, Tongji University

University of Florida CISE department Gator Engineering. Clustering Part 4

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

Clustering Part 4 DBSCAN

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1

TELCOM2125: Network Science and Analysis

How do microarrays work

Dimension reduction : PCA and Clustering

Biclustering for Microarray Data: A Short and Comprehensive Tutorial

Comparisons and validation of statistical clustering techniques for microarray gene expression data. Outline. Microarrays.

Triclustering in Gene Expression Data Analysis: A Selected Survey

Distance based tree reconstruction. Hierarchical clustering (UPGMA) Neighbor-Joining (NJ)

Clustering: Overview and K-means algorithm

Types of general clustering methods. Clustering Algorithms for general similarity measures. Similarity between clusters

Clustering Algorithms for general similarity measures

Cluster analysis. Agnieszka Nowak - Brzezinska

Clustering analysis of gene expression data

University of Florida CISE department Gator Engineering. Clustering Part 2

Theorem 2.9: nearest addition algorithm

Clustering: Overview and K-means algorithm

Giri Narasimhan. CAP 5510: Introduction to Bioinformatics. ECS 254; Phone: x3748

Clustering. Unsupervised Learning

Biclustering Bioinformatics Data Sets. A Possibilistic Approach

Based on Raymond J. Mooney s slides

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Slides From Lecture Notes for Chapter 8. Introduction to Data Mining

Hierarchical Clustering

Clustering COMS 4771

Clustering gene expression data

Introduction to Mobile Robotics

Unsupervised Learning : Clustering

Machine Learning using MapReduce

Microarray data analysis

MICROARRAY IMAGE SEGMENTATION USING CLUSTERING METHODS

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Olmo S. Zavala Romero. Clustering Hierarchical Distance Group Dist. K-means. Center of Atmospheric Sciences, UNAM.

Cluster Analysis for Microarray Data

Clustering algorithms and introduction to persistent homology

Multiple Sequence Alignment Sum-of-Pairs and ClustalW. Ulf Leser

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

Cluster Analysis. Angela Montanari and Laura Anderlucci

Data Clustering. Algorithmic Thinking Luay Nakhleh Department of Computer Science Rice University

Clustering, cont. Genome 373 Genomic Informatics Elhanan Borenstein. Some slides adapted from Jacques van Helden

Clustering. Unsupervised Learning

Clustering. Unsupervised Learning

Multivariate Analysis

Feature Extractors. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. The Perceptron Update Rule.

Clustering in R d. Clustering. Widely-used clustering methods. The k-means optimization problem CSE 250B

2. Background. 2.1 Clustering

Clustering. (Part 2)

Applications. Foreground / background segmentation Finding skin-colored regions. Finding the moving objects. Intelligent scissors

Lecture on Modeling Tools for Clustering & Regression

Variable Selection 6.783, Biomedical Decision Support

Machine Learning (BSMC-GA 4439) Wenke Liu

Clustering k-mean clustering

Unsupervised Learning and Clustering

CSE 527. CAST: a clustering method with a graph-theoretic basis. Larry Ruzzo

4. Ad-hoc I: Hierarchical clustering

Clustering Color/Intensity. Group together pixels of similar color/intensity.

6. Dicretization methods 6.1 The purpose of discretization

Finding and Visualizing Graph Clusters Using PageRank Optimization. Fan Chung and Alexander Tsiatas, UCSD WAW 2010

Greedy Approximations

k-means, k-means++ Barna Saha March 8, 2016

Transcription:

Gene expression & Clustering (Chapter 10)

Determining gene function Sequence comparison tells us if a gene is similar to another gene, e.g., in a new species Dynamic programming Approximate pattern matching Genes with similar sequence likely to have similar function Doesn t always work. Homologous genes may not be similar enough at the sequence level, to be detected this way New method to determine gene function: directly measure gene activity (DNA arrays)

An Introduction to Bioinformatics Algorithms www.bioalgorithms.info DNA Arrays--Technical Foundations An array works by exploiting the ability of a given mrna molecule to hybridize to the DNA template. Using an array containing many DNA samples in an experiment, the expression levels of hundreds or thousands genes within a cell by measuring the amount of mrna bound to each site on the array. With the aid of a computer, the amount of mrna bound to the spots on the microarray is precisely measured, generating a profile of gene expression in the cell. May, 11, 2004 http://www.ncbi.nih.gov/about/primer/microarrays.html 3

An experiment on a microarray In this schematic: GREEN represents Control DNA RED represents Sample DNA YELLOW represents a combination of Control and Sample DNA BLACK represents areas where neither the Control nor Sample DNA Each color in an array represents either healthy (control) or diseased (sample) tissue. The location and intensity of a color tell us whether the gene is present in the control and/or sample DNA. May 11,2004 http://www.ncbi.nih.gov/about/primer/microarrays.htm 10 l

An Introduction to Bioinformatics Algorithms www.bioalgorithms.info DNA Microarray Millions of DNA strands build up on each location. May, 11, 2004 Tagged probes become hybridized to the DNA chip s microarray. http://www.affymetrix.com/corporate/media/image_library/image_library_1.affx 5

Gene expression Microarray gives us an n x m expression matrix I Each of n rows corresponds to a gene Each of m columns corresponds to a condition or time point Each column comes from one microarray I(j,k) is the expression level of gene j in condition/experiment k If two genes (rows) have similar expression profiles, then they may be related in function they may be co-regulated

Clustering Find groups of genes that have similar expression profiles to one another Such groups may be functionally related, and/or co-regulated Compute pairwise distance metric d(i,j) for every pair of genes i and j This gives an n x n distance matrix d

Goal of clustering To group together genes into clusters such that Genes within a cluster have highly similar expression profiles (small d(i,j)): homogeneity Genes in different clusters have very different expression profiles (large d(i,j)): separation Good clustering is one that adheres to these goals A really good clustering is decided by biological interpretation of the clusters

Clustering of Microarray Data Clusters

Clustering problems How to measure distance/similarity? How many clusters? Very large data sets: ~10,000 gene, ~100 conditions create computational difficulties

Hierarchical clustering One approach to clustering Does not explicitly partition genes into groups Organizes genes into a tree; genes are at the leaves of the tree Edges have lengths Total path length between two genes (leaves) correlates with the distance between the genes

Hierarchical Clustering

Hierarchical Clustering: Example

Hierarchical Clustering: Example

Hierarchical Clustering: Example

Hierarchical Clustering: Example

Hierarchical Clustering: Example

Hierarchical Clustering Algorithm 1. Hierarchical Clustering (d, n) 2. Form n clusters each with one element 3. Construct a graph T by assigning one vertex to each cluster 4. while there is more than one cluster 5. Find the two closest clusters C 1 and C 2 6. Merge C 1 and C 2 into new cluster C with C 1 + C 2 elements 7. Compute distance from C to all other clusters 8. Add a new vertex C to T and connect to vertices C 1 and C 2 9. Remove rows and columns of d corresponding to C 1 and C 2 10. Add a row and column to d corresponding to the new cluster C 11. return T Different ways to define distances between clusters may lead to different clusterings

Hierarchical Clustering: Recomputing Distances d min (C, C * ) = min d(x,y) for all elements x in C and y in C * Distance between two clusters is the smallest distance between any pair of their elements d avg (C, C * ) = (1 / C * C ) d(x,y) for all elements x in C and y in C * Distance between two clusters is the average distance between all pairs of their elements

K-means clustering Another popular solution to the clustering problem Guess a number k, which is the number of clusters that will be reported Finds explicit clusters (unlike hierarchical clustering)

Objective function Let V = (v 1,v 2, v n ) be the n data points Each data point for is a vector in an m-dimensional space Let X be a set of k points in the same vector space For each v i, find the x X that is closest to it, i.e., the Euclidian distance d(v i,x) is least Sum the square of this Euclidian distance, over all v i Formally, d(v i, X) = min x "X d(v i, x) d(v,x) = n # i d(v i,x) 2 n

Objective function Find X such that d(v,x) is minimized Input: A set, V, consisting of n points and a parameter k Output: A set X consisting of k points (cluster centers) that minimizes the d(v,x) over all possible choices of X An NP-complete problem for k > 1 Easy for k=1!

K-Means Clustering: Lloyd Algorithm 1. Lloyd Algorithm 2. Arbitrarily assign the k cluster centers 3. while the cluster centers keep changing 4. Assign each data point to the cluster C i corresponding to the closest cluster representative (center) (1 i k) 5. After the assignment of all data points, compute new cluster representatives according to the center of gravity of each cluster, that is, the new cluster representative is v \ C for all v in C for every cluster C *This may lead to merely a locally optimal clustering.

Conservative K-Means Algorithm Lloyd algorithm is fast but in each iteration it moves many data points, not necessarily causing better convergence. A more conservative method would be to move one point at a time only if it improves the overall clustering cost The smaller the clustering cost of a partition of data points is, the better that clustering is Different methods (e.g.,d(v,x) we saw earlier) can be used to measure this clustering cost

K-Means Greedy Algorithm 1. ProgressiveGreedyK-Means(k) 2. Select an arbitrary partition P into k clusters 3. while forever 4. bestchange 0 5. for every cluster C 6. for every element i not in C 7. if moving i to cluster C reduces its clustering cost 8. if (cost(p) cost(p i C ) > bestchange 9. bestchange cost(p) cost(p i C ) 10. i * i 11. C * C 12. if bestchange > 0 13. Change partition P by moving i * to C * 14. else 15. return P

Corrupted Cliques problem Another approach to clustering Based on the concept of cliques Uses threshold on pairwise distance to decide which genes are close and which genes are distant Builds a graph with edge between every pair of close genes

Clique Graphs A clique is a graph with every vertex connected to every other vertex A clique graph is a graph where each connected component is a clique

Clique Graphs (cont d)

Corrupted Cliques Problem Input: A graph G Output: The smallest number of additions and removals of edges that will transform G into a clique graph

Corrupted Cliques & Clustering: Idea Distance graph G, vertices are genes and edge between genes i and j if d i,j <θ A clique graph satisfies homogeneity and separation criteria. However, G is typically not a clique graph, presumably due to (a) noise in measurement, (b) absence of a «universally» good threshold θ Therefore, find the fewest changes to G that make it a clique graph. The cliques are the desired clusters

Heuristics for Corrupted Clique Problem Corrupted Cliques problem is NP-Hard, some heuristics exist to approximately solve it: PCC (Parallel classification with cores): a time consuming algorithm CAST (Cluster Affinity Search Technique): a practical and fast algorithm.

PCC Given a set of genes S and its distance graph G. Suppose we knew the correct clustering for a subset S of S. Could we extend this to a clustering of S? For every gene j S \ S, and for every cluster C i, calculate the affinity of j to C i : N( j,c i ) C i where N(j, C i ) is the number of edges (in G) between j and genes in C i. Add every unclustered gene j to its maximum affinity cluster If S is sufficiently large, and clustering of S is correct, then the above approach is likely to be correct.

PCC We assumed we knew the correct clustering for the subset S of S. How to get around this? Generate all possible clusterings of S! This takes time O(k S ). Choose a small subset S. But that leads to a less accurate solution (we assumed also that S was large enough ) Tradeoff between speed and accuracy Authors of this algorithm used S = log log n. This leads to k S = (log n) c for some constant c. For each such clustering, O(n 2 ) time to extend. Therefore O(n 2 (log n) c ) time complexity. Too slow.

CAST CAST (Cluster Affinity Search Technique): a practical and fast algorithm: CAST is based on the notion of genes close to cluster C or distant from cluster C Distance between gene i and cluster C: d(i,c) = average distance between gene i and all genes in C

CAST Algorithm 1. CAST(S, G, θ) 2. P Ø 3. while S Ø 4. V vertex of maximal degree in the distance graph G 5. C {V} 6. while a close gene i not in C or distant gene i in C exists 7. Find the nearest close gene i not in C and add it to C 8. Remove the farthest distant gene i in C 9. Add cluster C to partition P 10. S S \ C 11. Remove vertices of cluster C from the distance graph G 12. return P S set of elements, G distance graph, θ - distance threshold