Model-Based Clustering for Online Crisis Identification in Distributed Computing
|
|
- Garey Blankenship
- 5 years ago
- Views:
Transcription
1 Model-Based Clustering for Crisis Identification in Distributed Computing Dawn Woodard Operations Research and Information Engineering Cornell University with Moises Goldszmidt Microsoft Research 1
2 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
3 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
4 Distributed Computing Commercial distributed computing providers: Offer remotely-hosted computing services E.g. Microsoft s Exchange Hosted Services (EHS) 24/7 processing incl. spam filtering, encryption 4
5 Distributed Computing This processing is performed by farming out to many servers Often, tens of thousands of servers in multiple locations Server 1 Client Provider Server 2 Server 3 5
6 Distributed Computing 6
7 Distributed Computing Can have occasional severe violation of performance goals ( crises ) E.g. due to: servers becoming overloaded in periods of high demand performance problems in lower-level computing centers on which the servers rely (e.g. for performing authentication) If the problem lasts for more than a few minutes, must pay cash penalties to clients, have potential loss of contracts 7
8 Distributed Computing % of servers violating a performance goal, for a 10-day period in EHS: KPI KPI Time Exceeding the dotted line constitutes a crisis. Metr
9 Distributed Computing Need to rapidly recognize the recurrence of a problem If an effective intervention is known for this problem, can apply it Due to large scale and interdependence, manual problem diagnosis is difficult and slow Have a set of status measurements for each server. E.g., for EHS: CPU utilization Memory utilization For each spam filter, the length of the queue and the throughput... 9
10 Distributed Computing Goal: Match a currently occurring (i.e., incompletely observed) crisis to previous crises of mixed known and unknown cause I.e., are any previous crises of the same type as the new crisis? Which ones? This is an online clustering problem with: partial labeling incomplete data for the new crisis We use model-based clustering based on a Dirichlet process mixture (e.g. Escobar & West 1995) The evolution of each process is modeled as a time series 10
11 Cost-Optimal Decision Making Wish to perform optimal (expected-cost-minimizing) decision making during a crisis......while accounting for uncertainty in the crisis type assignments and the parameters of those types This requires fully Bayesian inference 11
12 Fully Bayesian Inference We apply fully Bayesian inference (via MCMC) in the long periods between crises Due to posterior multimodality, we combine a collapsed-space split-merge method with parallel tempering As a new crisis begins, update rapidly using an approximation 12
13 Related Work Ours is the first instance of fully Bayesian online clustering model-based clustering was performed by Zhang, Ghahramani, and Yang (2004) for documents Obtain a single cluster assignment based on the posterior; insufficient for optimal decision making Fully Bayesian clustering: Bensmail, Celeux, Raftery, and Robert (1997); Pritchard, Stephens, and Donnelly (2000); Lau and Green (2007) Many examples of fully Bayesian mixture modeling 13
14 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
15 Data Medians of 3 metrics across servers, for a 10-day period (EHS): Time 15
16 Data Crises are highlighted; color indicates their known type: Time 16
17 Data The medians of the metrics are very informative as to crisis type specifically, whether the median is low, normal, or high We fit our models to the median values of the metrics, discretized into 1: low, 2: normal, and 3: high 17
18 Crisis Time series model for crisis evolution: Y ilj: value of metric j in the lth time period after the start of crisis i Assume that metrics are independent conditional on the crisis type For crisis type k, Y i1j is drawn from a discrete dist n with probability vector γ (jk)...and Y ilj evolves according to a Markov chain with transition matrix T (jk) 18
19 Crisis Complete-data likelihood fn: π D {Z i} I i=1, {γ (jk), T (jk) } j,k = Q " γ (j Z i) t i,j,t 1(Y i1j =t) Q s T (j Z i) st conditioning on the unknown type indicators Z i of each crisis i = 1,..., I. nijst # n ijst : the number of transitions of the jth metric from state s to state t during crisis i. (1) 19
20 Cluster Dirichlet process mixture (DPM) prior: Natural for online clustering Allows number of clusters to increase with the number of crises Crises are exchangeable Parameterized by α: controls the expected number of clusters occurring in a fixed number of crises G 0 : the prior G 0 (d{γ (jk), T (jk) } j ) for the parameters associated with each cluster k 20
21 Cluster The DPM prior for the cluster indicators {Z i} I i=1 and the cluster parameters γ (jk), T (jk) : π({z i } I i=1) = Q I π(z i {Z i } i <i ) i=1 " Q = I α α+i 1 1(Z i=m i 1 +1)+ 1 P α+i 1 i=1 i <i # 1(Z i =Z i ) (2) where m i = max{z i : i i} for i > 0 and m 0 = 0, and π d{γ (jk),t (jk) }j,k {Z i } I i=1 = m I Q k=1 G 0 d{γ (jk),t (jk) }j. (3) 21
22 Cluster Also called the Chinese Restaurant Process : π (Z i = k {Z i } i <i) 8 >< >: α : k is a new type P 1 (Z i = k) : else i <i Each observation i is a new guest who either sits at an occupied table with prob. proportional to the number of guests at that table, or sits at an empty table: 22
23 Cluster Now we can evaluate the posterior density (up to a normalizing constant): ( ) π {Z i } I i=1, {γ(jk), T (jk) } j,k D π ( {Z i } I ) ) ( ) i=1 π ({γ (jk), T (jk) } j,k {Z i } I i=1 π D {Z i } I i=1, {γ(jk), T (jk) } j,k 23
24 Cluster Partially labeled case: We have given the prior for the case where none of the crisis types Z i are known If we know that Z i = Z i for some crises i i, multiply (2) by Q i i 1(Z i = Z i ) 24
25 Cluster G 0 : Independent Dirichlet priors for γ (jk) for each j Independent product Dirichlet priors for T (jk) for each j 25
26 Computation Computation Decision Making Outline 1 Background and Overview 2 3 Computation Computation Decision Making
27 Computation Computation Decision Making Outline 1 Background and Overview 2 3 Computation Computation Decision Making
28 Computation Computation Decision Making Computation The cluster parameters {γ (jk), T (jk) } j,k can be integrated analytically out of the posterior Run a Markov chain with target dist n π({z i} I i=1 D) Jain and Neal (2004) use a Gibbs sampler, with an additional split-merge move on clusters We add parallel tempering (Geyer 1991) 28
29 Computation Computation Decision Making Outline 1 Background and Overview 2 3 Computation Computation Decision Making
30 Computation Computation Decision Making Inference Wish to identify a crisis in real time Have data D from previous crises and data D new so far for the new crisis E.g., wish to estimate π(z new = Z i D, D new) for each previous crisis i = 1,..., I...and π(z new Z i i D, D new) 30
31 Computation Computation Decision Making Exact Inference Method 1: Just apply the Markov chain method to the data from the I + 1 crises Gives posterior sample vectors {Z (l) i } I i=1, Z (l) for l = 1,..., L new Monte Carlo estimates of the desired probabilities: ˆπ(Z new = Z i D, D new) = 1 L ˆπ(Z new Z i i D, D new) = 1 L LP l=1 LP l=1 1(Z (l) new = Z (l) i ) 1(Z new (l) Z (l) i i) But running the Markov chain is too slow for real-time decision making! 31
32 Computation Computation Decision Making Approximate Inference We give a method using the approximation: π(z new = Z i D, D new) = X π(z new = Z i {Z i } I i=1, D, Dnew)π({Z i} I i=1 D, Dnew) {Z i } I i=1 X {Z i } I i=1 π(z new = Z i {Z i } I i=1, D, Dnew)π({Z i} I i=1 D) * Assumes that D new does not tell us much about the past crisis types 32
33 Computation Computation Decision Making Approximate Inference Method 2: Approximate Inference 1 After the end of each crisis, rerun the Markov chain, yielding sample vectors {Z (l) i } I i=1 from the posterior π({z i} I i=1 D). 2 When a new crisis begins, use its data D new to calculate the Monte Carlo estimates: ˆπ(Z new = Z i D, D new) = 1 L ˆπ(Z new Z i i D, D new) = 1 L LX l=1 LX l=1 π(z new = Z (l) i π(z new Z (l) i {Z (l) i } I i =1, D, D new) i {Z (l) i } I i =1, D, D new). 33
34 Computation Computation Decision Making Approximate Inference Part 2 is O(LIJ), very fast 34
35 Computation Computation Decision Making Outline 1 Background and Overview 2 3 Computation Computation Decision Making
36 Computation Computation Decision Making Optimal Decision Making Want expected-cost-minimizing decision making during a crisis The total cost of the new crisis is a function C ˆφ, {Z i } I i=1, Z new of: The intervention φ The true type Znew of the current crisis The vector of past crisis types {Z i } I i=1, which give the context for Z new Finding the expected cost of the crisis for intervention φ requires integrating C over the posterior distribution of `{Z i} I i=1, Z new Can be done exactly using Method 1, or approximately using Method 2 36
37 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
38 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
39 : Simulate I crises from model Compare MBC with distance-based clustering 39
40 Accuracy Criteria: 1 Pairwise Sensitivity: For pairs of crises of the same type, % assigned to the same cluster for MBC, having prob. > 0.5 of being in the same cluster. 2 Pairwise Specificity: For pairs of crises not of the same type, % assigned to different clusters for MBC, having prob. 0.5 of being in the same cluster. 3 Error of No. Crisis Types: The % error of the estimated number of crisis types for MBC, post. mean is used to estimate No. of types. 40
41 No. Crises No. Metrics Method Pairwise Pairwise % Error Sensitivity Specificity No. Types MBC 94.6 (2.08) 99.0 (0.50) 9.3 (1.87) K-Means (4.26) 95.3 (0.57) K-Means (5.39) 77.9 (1.73) MBC 99.0 (1.00) 99.4 (0.41) 3.7 (0.95) K-Means (4.76) 97.0 (0.54) K-Means (4.01) 78.2 (2.13) MBC 91.9 (1.88) 98.8 (0.40) 7.4 (1.58) K-Means (3.19) 95.5 (0.54) K-Means (4.01) 82.9 (1.16) MBC 99.6 (0.23) 99.9 (0.05) 3.5 (1.13) K-Means (3.76) 95.8 (0.57) K-Means (4.76) 83.0 (1.83) MBC 97.6 (0.65) 99.8 (0.08) 6.4 (1.81) K-Means (3.43) 95.9 (0.48) K-Means (3.93) 83.9 (1.15) MBC 99.5 (0.24) 99.9 (0.03) 3.4 (0.67) K-Means (4.07) 97.8 (0.27) K-Means (4.74) 86.7 (1.48) 41
42 MBC does far better than K-means More metrics better accuracy of MBC More crises better accuracy of MBC 42
43 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
44 : Compare Method 1 ( MBC-EX ) to Method 2 ( MBC ) 44
45 Accuracy Criteria: 1 Full-data misclassification rate: % of crises with incorrect predicted type, using all of the data for the new crisis. 2 p-period misclassification rate: % of crises with incorrect predicted type, using the first p time periods of data for the new crisis. 3 Average time to correct identification: Avg. No. of time periods required to obtain the correct identification ( correct predicted type: ˆπ(Z new Z i i D, D new) > 0.5 if Znew Zi ˆπ(Z new = Z i D, D new) > 0.5 for some i I such that Znew = Zi ) i and otherwise 45
46 Accuracy: No. No. Method Full-data 3-period Avg. Time to Crises Metrics Misclassification Misclassification Identification MBC 6.7 (3.0) 10.7 (4.5) 1.31 (0.11) MBC-EX 8 (2.5) 10.7 (4.5) MBC 6.7 (5.2) 9.3 (6.2) 1.13 (0.08) MBC-EX 5.3 (3.9) 8.0 (4.9) MBC 13.6 (2.7) 15.2 (2.7) 1.33 (0.13) MBC-EX 9.6 (2.0) 15.2 (3.4) MBC 2.4 (1.6) 4.0 (1.8) 1.15 (0.06) MBC-EX 3.2 (1.5) 3.2 (1.5) 46
47 Classification accuracy high (> 80%) for both MBC & MBC-EX MBC not significantly worse than MBC-EX 3-period misclassification is not much > than full-data misclassification Very early identification! 47
48 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
49 Application to EHS 27 crises in EHS during Jan-Apr The causes of some of these were diagnosed later: ID Cause No. of known crises A overloaded front-end 2 B overloaded back-end 8 C database configuration error 1 D configuration error 1 E performance issue 1 F middle-tier issue 1 G whole DC turned off and on 1 H workload spike 1 I request routing error 1 49
50 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
51 Application to EHS Apply the Markov chain method to the set of 27 crises without the labels Compare to those labels 51
52 Application to EHS Trace plots of parallel tempering Markov chain samples of Z 22: beta = beta = beta = Geweke diag. p-value: 0.44 Gelman-Rubin scale factor:
53 Application to EHS Post. mode cluster assignment has 58% prob. Sizes of clusters: ID Cause No. of known No. identified No. MBC crises crises by MBC matching known A overloaded front-end B overloaded back-end C database configuration error D configuration error (labeled as A) E performance issue (labeled as B) F middle-tier issue (labeled as I) G whole DC turned off and on (labeled as B) H workload spike I request routing error
54 Application to EHS Post. mode crisis labels mostly match known clusters The largest 5 clusters are correctly labelled Four uncommon crisis types are clustered with more common types Crises having different causes can have the same patterns in their metrics Need to add metrics that distinguish these types effectively 54
55 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
56 Application to EHS Evaluate online accuracy, treating the posterior mode from the offline context as the gold standard. Original ordering: 1 Full-data misclassification: 7.4% 2 3-period misclassification: 14.8% 3 Avg. time to correct iden.: 1.81 Permuting the crises: 1 Full-data misclassification: 5.9% (SE =3.4%) 2 3-period misclassification: 11.8% (SE =3.2%) 3 Avg. time to correct iden.: 1.56 (SE =0.07) 56
57 Outline 1 Background and Overview 2 3 Computation Computation Decision Making
58 Gave a method for fully Bayesian real-time crisis identification in distributed computing Described how to use this to perform rapid expected-cost-minimizing crisis intervention Very accurate on both simulated data and data from a production computing center A copy of this paper and seminar are available at: 58
59 References Escobar, M. D. and West, M. (1995). Bayesian density estimation and inference using mixtures. Journal of the American Statistical Association, 90, Geyer, C. J. (1991). Markov chain Monte Carlo maximum likelihood. in Computing Science and Statistics, Vol. 23: Proc. of the 23rd Symp. on the Interface, ed. E. Keramidas, pp Jain, S. and Neal, R. M. (2004). A split-merge Markov chain Monte Carlo procedure for the Dirichlet process mixture model. Journal of Computational and Graphical Statistics, 13, Lau, J. W. and Green, P. J. (2007). Bayesian model-based clustering procedures. Journal of Computational and Graphical Statistics, 16, Zhang, J., Ghahramani, Z., and Yang, Y. (2004). A probabilistic model for online document clustering with application to novelty detection. in Advances in Neural Information Processing Systems, ed. Y. Weiss. 59
60 Prior Constants Prior hyperparameters chosen by combining information in data with expert opinion Reflect the fact that the server status measurements are chosen to be indicative of crisis type Results far better than a default prior specification, which contradicts data and experts 60
61 Prior Constants α: Prob. that 2 randomly chosen crises are of same type: 1/(α + 1) EHS experts estimate as 0.1, giving α = 9 13 types in 27 crises γ (jk) Dir(a (j) ). To choose a (j) : Prior mean of γ (jk) taken as empirical dist n of Y i1j over i and j Substantial prob. that one of the γ (jk) is close to 1: π (γ (jk) 1 >.85) OR (γ (jk) 2 >.95) OR (γ (jk) 3 >.85) = 0.5 Analogous for T (jk) 61
62 Optimal Decision Making Want expected-cost-minimizing decision making during a crisis The total cost of the new crisis is a function C ˆφ, {Zi } I i=1, Znew of: The intervention φ The true type Znew of the current crisis The vector of past crisis types {Zi }I i=1, which give the context for Z new 62
63 Optimal Decision Making If we knew C, given posterior sample vectors 1... {Z (l) i...the expected cost can be estimated as: } I i=1, Z (l) from the exact Method new E(C) 1 L LX h C φ, ({Z (l) i } I i=1, Z (l) l=1 new) i. Have a similar expression for approximate inferences from Method 2 63
64 Optimal Decision Making Don t know C in practice For interventions φ taken during previous crises can estimate C from realized costs Otherwise can estimate C from expert knowledge 64
65 Optimal Decision Making Since the goal is optimal intervention...and since this requires the entire posterior distribution over `{Zi} I i=1, Z new... we will avoid choosing a best cluster assignment instead focusing on the accuracy of the soft identification, i.e. the posterior distribution over `{Z i} I i=1, Z new 65
66 K-means: Criteria for choosing the number of clusters do not work well in our context So we apply K-means using the true number of clusters ( K-means 1 ) and half the true number of clusters ( K-means 2 ) This is unrealistically optimistic......but K-means still does terribly 66
Clustering Relational Data using the Infinite Relational Model
Clustering Relational Data using the Infinite Relational Model Ana Daglis Supervised by: Matthew Ludkin September 4, 2015 Ana Daglis Clustering Data using the Infinite Relational Model September 4, 2015
More informationMCMC Diagnostics. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) MCMC Diagnostics MATH / 24
MCMC Diagnostics Yingbo Li Clemson University MATH 9810 Yingbo Li (Clemson) MCMC Diagnostics MATH 9810 1 / 24 Convergence to Posterior Distribution Theory proves that if a Gibbs sampler iterates enough,
More informationMCMC Methods for data modeling
MCMC Methods for data modeling Kenneth Scerri Department of Automatic Control and Systems Engineering Introduction 1. Symposium on Data Modelling 2. Outline: a. Definition and uses of MCMC b. MCMC algorithms
More informationFrom Bayesian Analysis of Item Response Theory Models Using SAS. Full book available for purchase here.
From Bayesian Analysis of Item Response Theory Models Using SAS. Full book available for purchase here. Contents About this Book...ix About the Authors... xiii Acknowledgments... xv Chapter 1: Item Response
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationA Short History of Markov Chain Monte Carlo
A Short History of Markov Chain Monte Carlo Christian Robert and George Casella 2010 Introduction Lack of computing machinery, or background on Markov chains, or hesitation to trust in the practicality
More informationCS281 Section 9: Graph Models and Practical MCMC
CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs
More informationMarkov chain Monte Carlo methods
Markov chain Monte Carlo methods (supplementary material) see also the applet http://www.lbreyer.com/classic.html February 9 6 Independent Hastings Metropolis Sampler Outline Independent Hastings Metropolis
More informationSummary: A Tutorial on Learning With Bayesian Networks
Summary: A Tutorial on Learning With Bayesian Networks Markus Kalisch May 5, 2006 We primarily summarize [4]. When we think that it is appropriate, we comment on additional facts and more recent developments.
More informationLinear Modeling with Bayesian Statistics
Linear Modeling with Bayesian Statistics Bayesian Approach I I I I I Estimate probability of a parameter State degree of believe in specific parameter values Evaluate probability of hypothesis given the
More informationA Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data
A Nonparametric Bayesian Approach to Detecting Spatial Activation Patterns in fmri Data Seyoung Kim, Padhraic Smyth, and Hal Stern Bren School of Information and Computer Sciences University of California,
More informationClustering Sequences with Hidden. Markov Models. Padhraic Smyth CA Abstract
Clustering Sequences with Hidden Markov Models Padhraic Smyth Information and Computer Science University of California, Irvine CA 92697-3425 smyth@ics.uci.edu Abstract This paper discusses a probabilistic
More informationDiscussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg
Discussion on Bayesian Model Selection and Parameter Estimation in Extragalactic Astronomy by Martin Weinberg Phil Gregory Physics and Astronomy Univ. of British Columbia Introduction Martin Weinberg reported
More informationAnalysis of Incomplete Multivariate Data
Analysis of Incomplete Multivariate Data J. L. Schafer Department of Statistics The Pennsylvania State University USA CHAPMAN & HALL/CRC A CR.C Press Company Boca Raton London New York Washington, D.C.
More informationEstimating the Information Rate of Noisy Two-Dimensional Constrained Channels
Estimating the Information Rate of Noisy Two-Dimensional Constrained Channels Mehdi Molkaraie and Hans-Andrea Loeliger Dept. of Information Technology and Electrical Engineering ETH Zurich, Switzerland
More informationApproximate Bayesian Computation. Alireza Shafaei - April 2016
Approximate Bayesian Computation Alireza Shafaei - April 2016 The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested in. The Problem Given a dataset, we are interested
More informationComputer vision: models, learning and inference. Chapter 10 Graphical Models
Computer vision: models, learning and inference Chapter 10 Graphical Models Independence Two variables x 1 and x 2 are independent if their joint probability distribution factorizes as Pr(x 1, x 2 )=Pr(x
More informationMonte Carlo for Spatial Models
Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing
More informationMaximum Likelihood Network Topology Identification from Edge-based Unicast Measurements. About the Authors. Motivation and Main Point.
Maximum Likelihood Network Topology Identification from Edge-based Unicast Measurements Mark Coates McGill University Presented by Chunling Hu Rui Castro, Robert Nowak Rice University About the Authors
More informationQuantitative Biology II!
Quantitative Biology II! Lecture 3: Markov Chain Monte Carlo! March 9, 2015! 2! Plan for Today!! Introduction to Sampling!! Introduction to MCMC!! Metropolis Algorithm!! Metropolis-Hastings Algorithm!!
More informationCase Study IV: Bayesian clustering of Alzheimer patients
Case Study IV: Bayesian clustering of Alzheimer patients Mike Wiper and Conchi Ausín Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School 2nd - 6th
More informationNetwork-based auto-probit modeling for protein function prediction
Network-based auto-probit modeling for protein function prediction Supplementary material Xiaoyu Jiang, David Gold, Eric D. Kolaczyk Derivation of Markov Chain Monte Carlo algorithm with the GO annotation
More informationMarkov Chain Monte Carlo (part 1)
Markov Chain Monte Carlo (part 1) Edps 590BAY Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2018 Depending on the book that you select for
More informationIssues in MCMC use for Bayesian model fitting. Practical Considerations for WinBUGS Users
Practical Considerations for WinBUGS Users Kate Cowles, Ph.D. Department of Statistics and Actuarial Science University of Iowa 22S:138 Lecture 12 Oct. 3, 2003 Issues in MCMC use for Bayesian model fitting
More informationThe Comparative Study of Machine Learning Algorithms in Text Data Classification*
The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification
More informationRandom projection for non-gaussian mixture models
Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,
More informationProbabilistic Graphical Models
Overview of Part Two Probabilistic Graphical Models Part Two: Inference and Learning Christopher M. Bishop Exact inference and the junction tree MCMC Variational methods and EM Example General variational
More informationSampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation
Sampling informative/complex a priori probability distributions using Gibbs sampling assisted by sequential simulation Thomas Mejer Hansen, Klaus Mosegaard, and Knud Skou Cordua 1 1 Center for Energy Resources
More informationSemi-Supervised Clustering with Partial Background Information
Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject
More informationUnsupervised Learning: Clustering
Unsupervised Learning: Clustering Vibhav Gogate The University of Texas at Dallas Slides adapted from Carlos Guestrin, Dan Klein & Luke Zettlemoyer Machine Learning Supervised Learning Unsupervised Learning
More informationTutorial using BEAST v2.4.1 Troubleshooting David A. Rasmussen
Tutorial using BEAST v2.4.1 Troubleshooting David A. Rasmussen 1 Background The primary goal of most phylogenetic analyses in BEAST is to infer the posterior distribution of trees and associated model
More informationNested Sampling: Introduction and Implementation
UNIVERSITY OF TEXAS AT SAN ANTONIO Nested Sampling: Introduction and Implementation Liang Jing May 2009 1 1 ABSTRACT Nested Sampling is a new technique to calculate the evidence, Z = P(D M) = p(d θ, M)p(θ
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationA GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM
A GENERAL GIBBS SAMPLING ALGORITHM FOR ANALYZING LINEAR MODELS USING THE SAS SYSTEM Jayawant Mandrekar, Daniel J. Sargent, Paul J. Novotny, Jeff A. Sloan Mayo Clinic, Rochester, MN 55905 ABSTRACT A general
More informationLevel-set MCMC Curve Sampling and Geometric Conditional Simulation
Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve
More informationBayesian Statistics Group 8th March Slice samplers. (A very brief introduction) The basic idea
Bayesian Statistics Group 8th March 2000 Slice samplers (A very brief introduction) The basic idea lacements To sample from a distribution, simply sample uniformly from the region under the density function
More informationClustering: Classic Methods and Modern Views
Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering
More information1 Methods for Posterior Simulation
1 Methods for Posterior Simulation Let p(θ y) be the posterior. simulation. Koop presents four methods for (posterior) 1. Monte Carlo integration: draw from p(θ y). 2. Gibbs sampler: sequentially drawing
More informationClustering CS 550: Machine Learning
Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf
More informationModeling and Reasoning with Bayesian Networks. Adnan Darwiche University of California Los Angeles, CA
Modeling and Reasoning with Bayesian Networks Adnan Darwiche University of California Los Angeles, CA darwiche@cs.ucla.edu June 24, 2008 Contents Preface 1 1 Introduction 1 1.1 Automated Reasoning........................
More informationMachine Learning and Data Mining. Clustering (1): Basics. Kalev Kask
Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of
More informationA Deterministic Global Optimization Method for Variational Inference
A Deterministic Global Optimization Method for Variational Inference Hachem Saddiki Mathematics and Statistics University of Massachusetts, Amherst saddiki@math.umass.edu Andrew C. Trapp Operations and
More informationApplied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University
Applied Bayesian Nonparametrics 5. Spatial Models via Gaussian Processes, not MRFs Tutorial at CVPR 2012 Erik Sudderth Brown University NIPS 2008: E. Sudderth & M. Jordan, Shared Segmentation of Natural
More informationParallel Gibbs Sampling From Colored Fields to Thin Junction Trees
Parallel Gibbs Sampling From Colored Fields to Thin Junction Trees Joseph Gonzalez Yucheng Low Arthur Gretton Carlos Guestrin Draw Samples Sampling as an Inference Procedure Suppose we wanted to know the
More informationAn Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework
IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster
More informationCurve Sampling and Geometric Conditional Simulation
Curve Sampling and Geometric Conditional Simulation Ayres Fan Joint work with John Fisher, William Wells, Jonathan Kane, and Alan Willsky S S G Massachusetts Institute of Technology September 19, 2007
More informationFeature LDA: a Supervised Topic Model for Automatic Detection of Web API Documentations from the Web
Feature LDA: a Supervised Topic Model for Automatic Detection of Web API Documentations from the Web Chenghua Lin, Yulan He, Carlos Pedrinaci, and John Domingue Knowledge Media Institute, The Open University
More informationHierarchical Bayesian Modeling with Ensemble MCMC. Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014
Hierarchical Bayesian Modeling with Ensemble MCMC Eric B. Ford (Penn State) Bayesian Computing for Astronomical Data Analysis June 12, 2014 Simple Markov Chain Monte Carlo Initialise chain with θ 0 (initial
More informationPackage atmcmc. February 19, 2015
Type Package Package atmcmc February 19, 2015 Title Automatically Tuned Markov Chain Monte Carlo Version 1.0 Date 2014-09-16 Author Jinyoung Yang Maintainer Jinyoung Yang
More informationSimulating from the Polya posterior by Glen Meeden, March 06
1 Introduction Simulating from the Polya posterior by Glen Meeden, glen@stat.umn.edu March 06 The Polya posterior is an objective Bayesian approach to finite population sampling. In its simplest form it
More informationShort-Cut MCMC: An Alternative to Adaptation
Short-Cut MCMC: An Alternative to Adaptation Radford M. Neal Dept. of Statistics and Dept. of Computer Science University of Toronto http://www.cs.utoronto.ca/ radford/ Third Workshop on Monte Carlo Methods,
More informationCodon models. In reality we use codon model Amino acid substitution rates meet nucleotide models Codon(nucleotide triplet)
Phylogeny Codon models Last lecture: poor man s way of calculating dn/ds (Ka/Ks) Tabulate synonymous/non- synonymous substitutions Normalize by the possibilities Transform to genetic distance K JC or K
More informationBayesian Modelling with JAGS and R
Bayesian Modelling with JAGS and R Martyn Plummer International Agency for Research on Cancer Rencontres R, 3 July 2012 CRAN Task View Bayesian Inference The CRAN Task View Bayesian Inference is maintained
More informationRJaCGH, a package for analysis of
RJaCGH, a package for analysis of CGH arrays with Reversible Jump MCMC 1. CGH Arrays: Biological problem: Changes in number of DNA copies are associated to cancer activity. Microarray technology: Oscar
More informationA Sample of Monte Carlo Methods in Robotics and Vision. Credits. Outline. Structure from Motion. without Correspondences
A Sample of Monte Carlo Methods in Robotics and Vision Frank Dellaert College of Computing Georgia Institute of Technology Credits Zia Khan Tucker Balch Michael Kaess Rafal Zboinski Ananth Ranganathan
More informationHidden Markov Models in the context of genetic analysis
Hidden Markov Models in the context of genetic analysis Vincent Plagnol UCL Genetics Institute November 22, 2012 Outline 1 Introduction 2 Two basic problems Forward/backward Baum-Welch algorithm Viterbi
More informationBayesian Estimation for Skew Normal Distributions Using Data Augmentation
The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC
More informationEfficient Feature Learning Using Perturb-and-MAP
Efficient Feature Learning Using Perturb-and-MAP Ke Li, Kevin Swersky, Richard Zemel Dept. of Computer Science, University of Toronto {keli,kswersky,zemel}@cs.toronto.edu Abstract Perturb-and-MAP [1] is
More informationCOPULA MODELS FOR BIG DATA USING DATA SHUFFLING
COPULA MODELS FOR BIG DATA USING DATA SHUFFLING Krish Muralidhar, Rathindra Sarathy Department of Marketing & Supply Chain Management, Price College of Business, University of Oklahoma, Norman OK 73019
More informationScalable Bayes Clustering for Outlier Detection Under Informative Sampling
Scalable Bayes Clustering for Outlier Detection Under Informative Sampling Based on JMLR paper of T. D. Savitsky Terrance D. Savitsky Office of Survey Methods Research FCSM - 2018 March 7-9, 2018 1 / 21
More informationNetwork Lasso: Clustering and Optimization in Large Graphs
Network Lasso: Clustering and Optimization in Large Graphs David Hallac, Jure Leskovec, Stephen Boyd Stanford University September 28, 2015 Convex optimization Convex optimization is everywhere Introduction
More informationStephen Scott.
1 / 33 sscott@cse.unl.edu 2 / 33 Start with a set of sequences In each column, residues are homolgous Residues occupy similar positions in 3D structure Residues diverge from a common ancestral residue
More informationDynamic Bayesian network (DBN)
Readings: K&F: 18.1, 18.2, 18.3, 18.4 ynamic Bayesian Networks Beyond 10708 Graphical Models 10708 Carlos Guestrin Carnegie Mellon University ecember 1 st, 2006 1 ynamic Bayesian network (BN) HMM defined
More informationMachine Learning (BSMC-GA 4439) Wenke Liu
Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning
More informationPost-Processing for MCMC
ost-rocessing for MCMC Edwin D. de Jong Marco A. Wiering Mădălina M. Drugan institute of information and computing sciences, utrecht university technical report UU-CS-23-2 www.cs.uu.nl ost-rocessing for
More informationGiRaF: a toolbox for Gibbs Random Fields analysis
GiRaF: a toolbox for Gibbs Random Fields analysis Julien Stoehr *1, Pierre Pudlo 2, and Nial Friel 1 1 University College Dublin 2 Aix-Marseille Université February 24, 2016 Abstract GiRaF package offers
More informationLecture 21 : A Hybrid: Deep Learning and Graphical Models
10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation
More informationProject Report: "Bayesian Spam Filter"
Humboldt-Universität zu Berlin Lehrstuhl für Maschinelles Lernen Sommersemester 2016 Maschinelles Lernen 1 Project Report: "Bayesian E-Mail Spam Filter" The Bayesians Sabine Bertram, Carolina Gumuljo,
More informationHandling Data with Three Types of Missing Values:
Handling Data with Three Types of Missing Values: A Simulation Study Jennifer Boyko Advisor: Ofer Harel Department of Statistics University of Connecticut Storrs, CT May 21, 2013 Jennifer Boyko Handling
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationBayesian Robust Inference of Differential Gene Expression The bridge package
Bayesian Robust Inference of Differential Gene Expression The bridge package Raphael Gottardo October 30, 2017 Contents Department Statistics, University of Washington http://www.rglab.org raph@stat.washington.edu
More informationThe Multi Stage Gibbs Sampling: Data Augmentation Dutch Example
The Multi Stage Gibbs Sampling: Data Augmentation Dutch Example Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 8 1 Example: Data augmentation / Auxiliary variables A commonly-used
More informationProbabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation
Probabilistic Abstraction Lattices: A Computationally Efficient Model for Conditional Probability Estimation Daniel Lowd January 14, 2004 1 Introduction Probabilistic models have shown increasing popularity
More informationBootstrapping Methods
Bootstrapping Methods example of a Monte Carlo method these are one Monte Carlo statistical method some Bayesian statistical methods are Monte Carlo we can also simulate models using Monte Carlo methods
More informationLearning stick-figure models using nonparametric Bayesian priors over trees
Learning stick-figure models using nonparametric Bayesian priors over trees Edward W. Meeds, David A. Ross, Richard S. Zemel, and Sam T. Roweis Department of Computer Science University of Toronto {ewm,
More informationCSCI 599 Class Presenta/on. Zach Levine. Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates
CSCI 599 Class Presenta/on Zach Levine Markov Chain Monte Carlo (MCMC) HMM Parameter Es/mates April 26 th, 2012 Topics Covered in this Presenta2on A (Brief) Review of HMMs HMM Parameter Learning Expecta2on-
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationThe Cross-Entropy Method for Mathematical Programming
The Cross-Entropy Method for Mathematical Programming Dirk P. Kroese Reuven Y. Rubinstein Department of Mathematics, The University of Queensland, Australia Faculty of Industrial Engineering and Management,
More informationModeling Dyadic Data with Binary Latent Factors
Modeling Dyadic Data with Binary Latent Factors Edward Meeds Department of Computer Science University of Toronto ewm@cs.toronto.edu Radford Neal Department of Computer Science University of Toronto radford@cs.toronto.edu
More informationMissing Data and Imputation
Missing Data and Imputation NINA ORWITZ OCTOBER 30 TH, 2017 Outline Types of missing data Simple methods for dealing with missing data Single and multiple imputation R example Missing data is a complex
More informationCHAPTER 1 INTRODUCTION
Introduction CHAPTER 1 INTRODUCTION Mplus is a statistical modeling program that provides researchers with a flexible tool to analyze their data. Mplus offers researchers a wide choice of models, estimators,
More informationHierarchical Mixture Models for Nested Data Structures
Hierarchical Mixture Models for Nested Data Structures Jeroen K. Vermunt 1 and Jay Magidson 2 1 Department of Methodology and Statistics, Tilburg University, PO Box 90153, 5000 LE Tilburg, Netherlands
More informationOverview. Monte Carlo Methods. Statistics & Bayesian Inference Lecture 3. Situation At End Of Last Week
Statistics & Bayesian Inference Lecture 3 Joe Zuntz Overview Overview & Motivation Metropolis Hastings Monte Carlo Methods Importance sampling Direct sampling Gibbs sampling Monte-Carlo Markov Chains Emcee
More informationBAYESIAN OUTPUT ANALYSIS PROGRAM (BOA) VERSION 1.0 USER S MANUAL
BAYESIAN OUTPUT ANALYSIS PROGRAM (BOA) VERSION 1.0 USER S MANUAL Brian J. Smith January 8, 2003 Contents 1 Getting Started 4 1.1 Hardware/Software Requirements.................... 4 1.2 Obtaining BOA..............................
More informationVariational Methods for Graphical Models
Chapter 2 Variational Methods for Graphical Models 2.1 Introduction The problem of probabb1istic inference in graphical models is the problem of computing a conditional probability distribution over the
More informationADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION
ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION CHRISTOPHER A. SIMS Abstract. A new algorithm for sampling from an arbitrary pdf. 1. Introduction Consider the standard problem of
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14
More informationSearch trees, tree, B+tree Marko Berezovský Radek Mařík PAL 2012
Search trees, 2-3-4 tree, B+tree Marko Berezovský Radek Mařík PL 2012 p 2
More informationPackage DPBBM. September 29, 2016
Type Package Title Dirichlet Process Beta-Binomial Mixture Version 0.2.5 Date 2016-09-21 Author Lin Zhang Package DPBBM September 29, 2016 Maintainer Lin Zhang Depends R (>= 3.1.0)
More informationKeeping flexible active contours on track using Metropolis updates
Keeping flexible active contours on track using Metropolis updates Trausti T. Kristjansson University of Waterloo ttkri stj @uwater l oo. ca Brendan J. Frey University of Waterloo f r ey@uwater l oo. ca
More informationReddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011
Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 1. Introduction Reddit is one of the most popular online social news websites with millions
More informationAn Introduction to Markov Chain Monte Carlo
An Introduction to Markov Chain Monte Carlo Markov Chain Monte Carlo (MCMC) refers to a suite of processes for simulating a posterior distribution based on a random (ie. monte carlo) process. In other
More informationIntegrating Dirichlet Reputation into Usage Control
Integrating Dirichlet Reputation into Usage Control Li Yang and Alma Cemerlic University of Tennessee at Chattanooga Cyber Security and Information Intelligence Research Workshop 2009 Motivation of the
More informationA Dynamic Bayesian Network Click Model for Web Search Ranking
A Dynamic Bayesian Network Click Model for Web Search Ranking Olivier Chapelle and Anne Ya Zhang Apr 22, 2009 18th International World Wide Web Conference Introduction Motivation Clicks provide valuable
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-
More informationA Capacity Planning Methodology for Distributed E-Commerce Applications
A Capacity Planning Methodology for Distributed E-Commerce Applications I. Introduction Most of today s e-commerce environments are based on distributed, multi-tiered, component-based architectures. The
More informationStatistical techniques for data analysis in Cosmology
Statistical techniques for data analysis in Cosmology arxiv:0712.3028; arxiv:0911.3105 Numerical recipes (the bible ) Licia Verde ICREA & ICC UB-IEEC http://icc.ub.edu/~liciaverde outline Lecture 1: Introduction
More informationModeling Criminal Careers as Departures From a Unimodal Population Age-Crime Curve: The Case of Marijuana Use
Modeling Criminal Careers as Departures From a Unimodal Population Curve: The Case of Marijuana Use Donatello Telesca, Elena A. Erosheva, Derek A. Kreader, & Ross Matsueda April 15, 2014 extends Telesca
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 17 EM CS/CNS/EE 155 Andreas Krause Announcements Project poster session on Thursday Dec 3, 4-6pm in Annenberg 2 nd floor atrium! Easels, poster boards and cookies
More informationA Bayesian approach to artificial neural network model selection
A Bayesian approach to artificial neural network model selection Kingston, G. B., H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School of Civil and Environmental Engineering,
More information