Epitomic Analysis of Human Motion

Size: px
Start display at page:

Download "Epitomic Analysis of Human Motion"

Transcription

1 Epitomic Analysis of Human Motion Wooyoung Kim James M. Rehg Department of Computer Science Georgia Institute of Technology Atlanta, GA {wooyoung, Abstract Epitomic analysis is a recent statistical approach to form a generative model. It has been recently applied to image, video, and audio data. In this paper, we describe a novel application of epitomic analysis to motion capture data. We show how to construct a motion epitome, and use it for various applications in animation such as motion synthesis, reconstruction and classification of motions. We discuss the use of motion epitome for measuring motion naturalness. 1. Introduction The image epitome was introduced by Jojic, Frey and Kannan in [4] as a novel generative image model. The epitome representation of a source image is much smaller than the source but retains most of the constitutive elements needed to reconstruct it. Applications of image epitomes include denoising, detection, indexing and retrieval, and segmentation. More recently, epitomic models have been developed for video [1] and audio data [5]. This paper introduces a novel epitomic model for motion capture data which we call the motion epitome. Epitomic motion analysis results in a generative time series model which compresses sequences of motion capture data into a compact representation of their most salient constituent elements. In comparison to conventional generative models of human motion such as a Hidden Markov Model or Switching Linear Dynamic System, the motion epitome offers the potential to simultaneously represent correlations at multiple time scales and complex patterns of dependency between states in the model. Used in conjunction with non-parametric motion representations (such as motion graphs [6]), the motion epitome offers the potential to generate more accurate transitions between motion sequences. We describe a procedure for learning motion epitomes from a corpus of motion capture data, and present the use of motion epitomes in the reconstruction of motions,the synthesis of long motion sequences, and the classification of motions. We provide experimental results of synthesis and classification of motion sequences. 2. Motivation The motivation for this work is the existence, for the first time, of large and diverse collections of motion capture data. For example, the CMU corpus (available at contains 1648 motions and includes 23 different types of motions. We believe these collections can enable a large-scale investigation of data driven approaches to the generation and analysis of human movement. A basic question is whether we can identify a canonical finite set of primitive movement elements from which all human movement can be constructed. One approach to this problem is to design a statistical generative model and examine the representations that it constructs when it is trained on a sufficiently large corpus of motion data. In this work we develop the motion epitome as a generative statistical model for motion capture data. 1

2 3. Motion Capture Data Motion capture data is obtained by tracking markers attached to the figure from multiple cameras at Hz. Using commercial software (such as IQ), marker positions in 3D are converted into rotation angles, φ, θ, ψ with respect to an underlying 3D skeleton model as shown in Figure 1(a). We make an analogy between motion images and intensity images, and propose to learn epitomic representations of motion using the same framework as image epitomes. Figure 1(b) gives an example of a motion image constructed from joint angle data. In the motion image, each row consists of a feature F = {T x, T y, T z, R φ, R θ, R ψ, {φ i, θ i, ψ i } J i=1 }, where T xt y R z, T φ R θ R ψ are a root position and a root rotation, and φ i θ i ψ i is a relative rotation for the ith joint angle. Hence, each row gives the pose of the entire figure at a specific instant time t. Figure 1: (a) is the skeleton of 29 joints; (b) is motion image and a sequence of motion. The motion image of left is un-normalized law data. Motion capture data is represented in the Euler angle system in which a 3D rotation is described as φ, θ, ψ rotation angles. This causes a problem with epitomic analysis, since it considers the same angles with periodic degree 360 as different ones. In order to solve this problem, we first convert the Euler angles to quaternion which use an axis ν in 3D and a rotation angle α. After conversion to quaternion, we scale translation data so that all of the data can be in the same range of [-1,1]. Then for the better visualization, we only extract rotation angles from it to make a new motion image. The new motion images are shown in Figure 3, and Figure Motion Epitome Motion epitome is a generative model of motion capture data which is much more compact in time, but retains most of the patterns in the input motion. This is a probabilistic model of a copying and averaging process inspired 2

3 by the paper in [2]. Figure 3 shows motion epitome extracted from a set of sequences of motions. We first describe how to generate a motion from a motion epitome, then explain inference and the learning process Generating a motion with motion epitome We regard a sequence of motion as a 2D array of features F = {T x, T y, T z, R φ, R θ, R ψ, {φ i, θ i, ψ i } J i=1 } at each time t, defined in Section 3, with t {1,..., T } where T is the number of time frames of the motion. Then the epitome e is a set of probability distributions of size T e where T e < T, so that an instance of a feature vector at time t is estimated from any of the probability distributions in e. Here we use a Gaussian distribution parameterized by a mean and covariance for each entry of the epitome. Hence, at the epitome coordinate t e, the probability density of the feature vector stored in the entry t of the motion is p(x t e te ) = N (x t ; µ te, φ te ), where µ te is the mean and φ te is the covariance. (a) (b) Graphical model Generative process Figure 2: Graphical model of epitomic analysis(a) and generative process (b). Figure 2(b) illustrates the generative procedure consistent with the graphical model in Figure 2(a). Given e, the goal is to generate an output time series X. An intermediate step is to generate overlapping patches, z k by sampling from the epitome representation. For x i at time i to be generated, we first consider all possible time coordinates sets that include i. Representative sets k, s, h are shown in Figure 2 (b). With each time coordinate set k, we randomly choose patch τ k from the epitome. The distribution of τ k could be learned or assumed to be 3

4 uniform. After that we generate a predicted patch z k using the distribution, k p(z k τ k, e) = N (z j,k ; µ τk(j), φ τk(j) ) (1) j where µ τk(j), φ τk(j) are the mean and variances stored in the patch τ k at j. Also note that k(j) refers to the jth coordinate in k, and z j,k to the value at k(j) in z k. Then all overlapping patches z k s are combined to make a single prediction for the observed value x i. Here we assume that in order to generate a consistent value x i, the contributions from all overlapping patches {z k i k} are averaged and Gaussian noise with variance ψ i is added; p(x i {z k : i k}) = N (x i ; where N i is the number of coordinate sets that overlap coordinate i Inference with motion epitome 1 z i,k, ψ i ) (2) N i k:i k Epitomic analysis allows arbitrary patches from the model in the generative process. Although this makes it possible to generate more wide range of motions than we are interested in, we constraint that the motions to be generated from patches agree at time i if they are shared. Therefore, we assume that all values that share a time coordinate i are the same value ζ i. Then the inference process is narrowed as the following; Given an epitome e and an observation X, we want to infer the hidden value ζ i at time i which is shared by all patches {z k : i k} predicted from e. This is obtained by maximizing the log likelihood of the data; where log p(x) = log p(x, {z k, τ k }) {z k } {τ k } {z k } {τ k } q({z k, τ k }) log p(x, {z k, τ k }) q({z k, τ k }). (3) p(x, {z k, τ k }) = p({z k, τ k })p(x {z k, τ k }) n = p({z k, τ k }) p(x i {z k, τ k }) = p({z k, τ k }) = k i=1 n p(x i {z k }) i=1 [p(τ k )p(z k τ k )] n p(x i {z k }). (4) Note that the equation (4) can be fully expressed with epitome parameters by substituting equation (1), (2) into (4). Here, q({z k, τ k }) is the auxiliary joint probability distribution over the set of epitome patches {τ k } and the set of patches {z k } for all coordinate patches k. Following the approximate and efficient inference and learning algorithm proposed in [3], we decouple the variables in the q-distribution; i=1 q({z k }, {τ k }) = Π k q(τ k )q(z k ). (5) 4

5 Since we assumed that the value of z k in the overlapping regions agree, we can express q(z k ) as: q(z k ) = Π j δ(z j,k ζ k(j) ). Then the estimate of ζ i is obtained by setting to zero the derivative of equation (3) with respect to ζ i : ζ i = x i ψ i + k,j:k(j)=j τ k q(τ k ) µ τ k (j) 1 ψ i + k,j:k(j)=j φ τk (j) τ k q(τ k ) 1 φ τk (j) Likewise, we can obtain q(τ k ) by setting zero the derivative of (3) with respect to q(τ k ):. (6) q(τ k ) p(τ k ) i N (ζ i,k ; µ τk(i), φ k(i) ). (7) Inference consists of iterating the above two updates, (6) and (7), until they converge. All the motion applications we discuss in this paper use this inference algorithm Learning motion epitome Now we suppose that we are leaning parameters of a motion epitome given an observation X. The learning process then requires estimating the parameters µ j and ψ j as well. These are obtained by setting to zero the derivative of (3) with respect to µ j, ψ j respectively; µ j = k k(i)=i k k(i)=i τ k,τ k (i)=j q(τ k)ζ i τ k,τ k (i)=j q(τ k) (8) φ j = k k(i)=i k τ k,τ k (i)=j q(τ k)(ζ i µ j ) 2 τ k,τ k (i)=j q(τ. (9) k) k(i)=i Here, estimating ζ i and the parameters of the epitome is the M-step, and computing q(τ k ) which is an approximation to the posterior is the E-step. Hence, we learn the epitome by iterating these EM steps. Figure 3: Learning motion epitome from an input motion Figure 3 illustrates the learning process of motion epitome with motion images. We also provide in Figure 4 examples of motion epitomes learned from a set of basketball-dribbling (a), and from a set of jogging motions (b). The left of each figure is the set of training data and right is the motion epitome of jogging. Clearly the size of epitome is much smaller but contains most salient patterns according to their types. Note that the motion image is not a raw image as shown in Figure 1(b). We converted the raw data to quaternion system from Euler angle and scaled the translation data for efficient use of epitome. The motion images in Figure 3 and 4 are obtained by 5

6 Figure 4: Examples of motion epitome learned from training data of (a) basketball dribble and (b) jogging motions extracting only rotation angle after converting and scaling, for better representation. To see the details of motion visualization, visit wooyoung/motionvisual/motionvisual.htm. In this site, we showed that motion types can be distinguished with this motion image. 5. Applications of Motion Epitome Motion epitome can be used for several motion applications including motion synthesis, key-frame interpolation, motion editing, transitions in motion blending and motion classification. Here we present three of those applications, motion fill-in, motion classification, measuring motion naturalness. We use the same graphical model for these tasks as in Figure 2, and the same notations in Section 3. The core algorithm in each of these applications is first to reconstruct the true motion from an input motion sequence, given a previously trained epitome model, then to compute the similarity between the reconstructed motion and an input Motion Reconstruction We first consider the general case of motion reconstruction with motion epitome. The process follows the inference algorithm as in Section 4.2. Given an epitomic model e and an input motion X, we update (6), (7) iteratively until it converges. Then the sequence ζ = (ζ i ) is the resulting reconstructed motion. We can observe that, from equation (2), the generated value x i will be dominated by Gaussian noise ψ i if it is very high. Then the inferred feature value ζ i shared for the time i is determined by the epitome predicted patches as shown in (6). We also see that if 6

7 the noise ψ i is very low, ζ i is dominated by x i. From these observations, we can set the ψ i high if x i is unreliable. If the noise variance ψ i is known, then we only iterate equation (7) and (6) to obtain ζ. In the motion application in section 5.3 and 5.4, we deal with this problem when the noise variance is unknown and not uniform Motion Fill-In Figure 5: Synthesized frames using epitomic analysis and non-parametric methods. Left-bottom is the result of epitomic analysis and right-bottom is the one of non-parametric synthesis. Compared with the original frame shown middle-bottom, the reconstructed frames using epitome show more smooth transition than the other. Motion type Error with Epitomic analysis Error with Non-parametric synthesis jog and left turn 1.34*1e *1e+03 jog and right turn 4.30*1e *1e+03 jog and walking turn 1.01*1e *1e+03 dribbling basketball 8.40*1e *1e+03 sitting and walking 1.40*1e *1e+03 Table 1: Results of filling-in with motion epitome and non-parametric synthesis The purpose of the task is to fill in several missing frames of a given input motion. The feature values of missing frames are modelled to have infinite variance in the measurement noise. Hence, we set ψ i = for each missing frame i. Then we only need to reconstruct ζ i s of missing frames by iterating equation (7) and (6). In this experiment, first we construct several types of motion epitome, such as, dribbling basketball, turning while jogging, walking, standing or walking after sitting. Then, a walking motion with missing frames is reconstructed from the walking epitome, a jogging motion from the jogging epitome, etc. Here we present the results in Table 1 with the comparison of non-parametric synthesis as appeared in [2]. Non-parametric synthesis searches for 7

8 windows of data in the motion library which best match the neighborhood of the missing frames, and does fill-in by copying. One example is shown in Figure 5. The synthesized frames using epitomic analysis are more similar to the original frames than those of non-parametric sampling. As we can see in table 1 the numbers indicating errors, the synthesized motion with epitomic analysis have less errors than with non-parametric synthesis overall. The results assure that we can fill-in frames only from the patches generated from epitome motion whose size is very small, rather than searching huge data library for most similar patches. We plan to explore this experiment to the application of creating smooth motion transition appeared in [6] Motion Classification Figure 6: Charts of jump model and other types of test motion including jump. Figure 7: Charts of boxing model and other types of test motion including boxing. The basic approach to motion classification is to find an appropriate similarity measurements between given an input and a type of motion epitome. Hence, we estimate the measurement by maximizing probability of an input given the epitome, P (X e) = P (X, e)/p (e). Since epitome e is already given, that leads to finding the joint 8

9 Figure 8: Charts of walking model and other types of test motion including walking. Figure 9: Charts of running model and other types of test motion including running. probability of X, e. For this purpose, we take the log function to the joint probability as the following: log P (X, e) = log P (X, e, T, Z) T,Z = log P (X, {z k, τ k }, e) {z k,τ k } {τ k } {z k,τ k } {τ k } q({z k, τ k }) log P (X, {z k, τ k }, e) q({z k, τ k }) (10) We choose the q-distribution as 7 of Section 4.2. By substituting equation (5) into (10) to find the optimal solution, we eventually get q(τ k ), ζ i as same as equation (6), and ψ i = (x i ζ i ) 2 since ψ i is also unknown. We update ζ i, ψ i and q(τ k ) iteratively until it converges, and replace it to equation (10), then the measure for 9

10 Figure 10: Charts of basketball-dribbling model and other types of test motion including basketball-dribbling. motion classification is; n log N (x i ; ζ i, ψ i ) + log p(e). (11) i=1 In our experiment, we first learned 5 types of motion epitomes - dribbling basketball, boxing, walking, run/jogging, jumping. Next, we reconstruct a true motion given a test motion and an epitome. Then the score that can tell each type of the test data is obtained from equation (11). Figure 6 to 10 show the results. Each figure shows the scores of seven different sets of data against each epitomic model-walk, run, jump, boxing, and basketball. The graph on the left of each figure presents scores of each set on a curve. Each point of a curve is a single motion sequence. For the purpose of representing each data set according to their scores, we added a histogram graph on the right side of each figure. To better understand the graphs, take Figure 8 as an example. The given epitomic model type is walking and seven different test sets are scored against it. walk is a set of walking motions that are excluded from the training data when we learn the epitome, while walk2 is a training set. These two sets are much higher scores than other types of set overall, so that we can draw a boundary for motion classification Motion naturalness Ever since motion capture data has been used for synthesizing human motion, natural-looking has been the important measurement for the quality of the output. Several possible measurements are proposed in [7], where the authors present a statistical model of natural motions using mixtures of Gaussians, hidden Markov models and switching linear dynamic systems. In this paper, we present epitomic analysis as another approach to measuring motion naturalness. The procedure mostly follows that of motion classification, except that we train a classifier to distinguish between natural and unnatural movement based on the natural training data. In this experiment, we take two types of motions - walk and run - and provide a way of discriminating unnaturalness with natural epitomic models. First, we construct three types of epitome models only with natural motions - walk, run, and a combination of walk and run motions. Second, given an input motion sequence, we infer a true motion based on the each model, then compute a score using the measurements in equation (11). We present the results in Figure 11, 12 and 13. In Figure 11, the left graph illustrates that most of the natural walking motions have higher score than the unnatural ones with natural walk model. Similar results appeared with run motions against run model at the left graph of Figure 12, but the classifier is less clear than with walk motions. This phenomenon can be explained as the following reason. The unnatural motion capture data of testing sets were obtained from a number 10

11 Figure 11: Walking motions are measured via two different types of epitome model. The chart on the left shows scores of natural and unnatural walk motions against natural walk epitome model. The right chart is the scores against natural run epitome model. of sources, including editing, key-framing, noise-added, bad motion transitions and insufficiently cleaned motion capture data. Most of the unnatural run motions tested, unlike walk motions obtained from mostly from the bad editing process, are collected by bad key-framing. The set of this category is likely to have very short sequences of unnaturalness so that a natural model cannot catch the differences easily. We also present another interesting results on the right graphs of Figure 11and 12. As we can see at the graph on the right of Figure 11, natural run epitomic model does not distinguish unnatural walks from natural ones. This is the same as in the case of unnatural run against natural walk model. The results tell us that each type of epitomic model captures the differences of motion types more sensitively than the subtle changes between naturalness and unnaturalness. We can, however, overcome this limitation if we combine all different types of natural motions and build a natural epitomic model. We present the possibility in Figure 13, where unnatural walk and runs are scored against the combination of natural walk and run model. We plan to extend this experiment with more robust framework in the future. 6. Conclusion We have shown that the motion epitome is an innovative representative model of motion capture data, and performed several experiments of motion applications with epitome. The results of filling in and classification are quite successful. Furthermore, we provided the potentials of epitomic analysis toward measuring motion naturalness. It has been proved that motion epitome does not just tell between natural and natural motions, but also gives a way of fixing the unnaturalness of given input by reconstructing. More results are available at wooyoung/motionepitome/motionepitome.htm. There are two contributions in our work. For one, in the machine learning point of view, we have explored the use of epitomic analysis to the new problem domain, motion capture data. And as in the graphics point of view, epitomic analysis has the potential of producing richer and more realistic continuous animation than other conventional generative models, such as, HMM, SLDS. This is due to the fact that epitomic analysis is a patchbased algorithm. Motion epitome, however, has to deal with high dimensional feature space unlike image and video epitomes. For efficiency purpose, we need to find a way of reducing the dimension, such as, combining PCA with eptiome. 11

12 Figure 12: Running motions are measured via two different types of epitome model. The chart on the left shows scores of natural and unnatural run motions against natural run epitome model. The right chart is the scores against natural walk epitome model. Another way is to use a subset of joints instead of the whole body, so that we can perform it in the lower dimensional feature space. With more efficient motion epitome, we would like to compare the results of motion naturalness as in the paper [7]. References [1] V. Cheung, B.J. Frey, N.Jojic. Video Epitomes, In Proc. IEEE Intern. Conf. on Computer Vision and Pattern Recognition. Los Alamitos, CA:IEEE Computer Society Press, 2005 [2] A. Efros and T.K. Leung. Texture Synthesis by Non-parametric Sampling, ICCV [3] B.J. Frey, N.Jojic. Advances in algorithms for inference and learning in complex probability models, IEEE Trans. PAMI, [4] N.Jojic, B.J. Frey, and A. Kannan.(2003) Epitomic Analysis of appearance and shape, In Proc. IEEE Intern. Conf. on Computer Vision, 2003 [5] Ashish Kapoor and Sumit Basu. The Audio Epitome: A New Representation for Modeling and Classifying Auditory Phenomena,International Conference on Acoustics, Speech and Signal Processing, ICASSP [6] L. Kovar, M. Gleicher, and F. Pighin.Motion Graphs, In ACM Transactions on Graphics 21(3),Proceedings of SIG- GRAPH [7] Liu Ren,Alton Patrick,Alexei A. Efros, Jessica K. Hodgins,James M. Rehg. A Data-Driven Approach to Quantifying Natural Human Motion, SIGGRAPH

13 Figure 13: Walking and running motions are measured via the combination of walk and run epitome model. The chart on the left shows scores of walk motions. The right chart is the scores of run motions. 13

Detecting unnatural motion with Epitomic analysis : Unsupervised learning

Detecting unnatural motion with Epitomic analysis : Unsupervised learning Detecting unnatural motion with Epitomic analysis : Unsupervised learning Final project Pattern Recognition, Spring 2007 Wooyoung Kim wooyoung@cc.gatech.edu 1. Overview Motion synthesis and creation are

More information

Patch-Based Image Classification Using Image Epitomes

Patch-Based Image Classification Using Image Epitomes Patch-Based Image Classification Using Image Epitomes David Andrzejewski CS 766 - Final Project December 19, 2005 Abstract Automatic image classification has many practical applications, including photo

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

Motion Interpretation and Synthesis by ICA

Motion Interpretation and Synthesis by ICA Motion Interpretation and Synthesis by ICA Renqiang Min Department of Computer Science, University of Toronto, 1 King s College Road, Toronto, ON M5S3G4, Canada Abstract. It is known that high-dimensional

More information

Motion Texture. Harriet Pashley Advisor: Yanxi Liu Ph.D. Student: James Hays. 1. Introduction

Motion Texture. Harriet Pashley Advisor: Yanxi Liu Ph.D. Student: James Hays. 1. Introduction Motion Texture Harriet Pashley Advisor: Yanxi Liu Ph.D. Student: James Hays 1. Introduction Motion capture data is often used in movies and video games because it is able to realistically depict human

More information

3D Human Motion Analysis and Manifolds

3D Human Motion Analysis and Manifolds D E P A R T M E N T O F C O M P U T E R S C I E N C E U N I V E R S I T Y O F C O P E N H A G E N 3D Human Motion Analysis and Manifolds Kim Steenstrup Pedersen DIKU Image group and E-Science center Motivation

More information

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation

Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation Human Motion Synthesis by Motion Manifold Learning and Motion Primitive Segmentation Chan-Su Lee and Ahmed Elgammal Rutgers University, Piscataway, NJ, USA {chansu, elgammal}@cs.rutgers.edu Abstract. We

More information

Mixture Models and EM

Mixture Models and EM Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering

More information

Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00. Topic (Research Paper):

Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00. Topic (Research Paper): Thiruvarangan Ramaraj CS525 Graphics & Scientific Visualization Spring 2007, Presentation I, February 28 th 2007, 14:10 15:00 Topic (Research Paper): Jinxian Chai and Jessica K. Hodgins, Performance Animation

More information

Particle Filtering. CS6240 Multimedia Analysis. Leow Wee Kheng. Department of Computer Science School of Computing National University of Singapore

Particle Filtering. CS6240 Multimedia Analysis. Leow Wee Kheng. Department of Computer Science School of Computing National University of Singapore Particle Filtering CS6240 Multimedia Analysis Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore (CS6240) Particle Filtering 1 / 28 Introduction Introduction

More information

Generating Different Realistic Humanoid Motion

Generating Different Realistic Humanoid Motion Generating Different Realistic Humanoid Motion Zhenbo Li,2,3, Yu Deng,2,3, and Hua Li,2,3 Key Lab. of Computer System and Architecture, Institute of Computing Technology, Chinese Academy of Sciences, Beijing

More information

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves

More information

Motion Synthesis and Editing. Yisheng Chen

Motion Synthesis and Editing. Yisheng Chen Motion Synthesis and Editing Yisheng Chen Overview Data driven motion synthesis automatically generate motion from a motion capture database, offline or interactive User inputs Large, high-dimensional

More information

Lecture 21 : A Hybrid: Deep Learning and Graphical Models

Lecture 21 : A Hybrid: Deep Learning and Graphical Models 10-708: Probabilistic Graphical Models, Spring 2018 Lecture 21 : A Hybrid: Deep Learning and Graphical Models Lecturer: Kayhan Batmanghelich Scribes: Paul Liang, Anirudha Rayasam 1 Introduction and Motivation

More information

Term Project Final Report for CPSC526 Statistical Models of Poses Using Inverse Kinematics

Term Project Final Report for CPSC526 Statistical Models of Poses Using Inverse Kinematics Term Project Final Report for CPSC526 Statistical Models of Poses Using Inverse Kinematics Department of Computer Science The University of British Columbia duanx@cs.ubc.ca, lili1987@cs.ubc.ca Abstract

More information

Monocular Human Motion Capture with a Mixture of Regressors. Ankur Agarwal and Bill Triggs GRAVIR-INRIA-CNRS, Grenoble, France

Monocular Human Motion Capture with a Mixture of Regressors. Ankur Agarwal and Bill Triggs GRAVIR-INRIA-CNRS, Grenoble, France Monocular Human Motion Capture with a Mixture of Regressors Ankur Agarwal and Bill Triggs GRAVIR-INRIA-CNRS, Grenoble, France IEEE Workshop on Vision for Human-Computer Interaction, 21 June 2005 Visual

More information

Expectation-Maximization. Nuno Vasconcelos ECE Department, UCSD

Expectation-Maximization. Nuno Vasconcelos ECE Department, UCSD Expectation-Maximization Nuno Vasconcelos ECE Department, UCSD Plan for today last time we started talking about mixture models we introduced the main ideas behind EM to motivate EM, we looked at classification-maximization

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Conditional Random Fields for Object Recognition

Conditional Random Fields for Object Recognition Conditional Random Fields for Object Recognition Ariadna Quattoni Michael Collins Trevor Darrell MIT Computer Science and Artificial Intelligence Laboratory Cambridge, MA 02139 {ariadna, mcollins, trevor}@csail.mit.edu

More information

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models

A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models A Visualization Tool to Improve the Performance of a Classifier Based on Hidden Markov Models Gleidson Pegoretti da Silva, Masaki Nakagawa Department of Computer and Information Sciences Tokyo University

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

Mixture Models and the EM Algorithm

Mixture Models and the EM Algorithm Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is

More information

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov

ECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern

More information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information

Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Probabilistic Facial Feature Extraction Using Joint Distribution of Location and Texture Information Mustafa Berkay Yilmaz, Hakan Erdogan, Mustafa Unel Sabanci University, Faculty of Engineering and Natural

More information

Robust Model-Free Tracking of Non-Rigid Shape. Abstract

Robust Model-Free Tracking of Non-Rigid Shape. Abstract Robust Model-Free Tracking of Non-Rigid Shape Lorenzo Torresani Stanford University ltorresa@cs.stanford.edu Christoph Bregler New York University chris.bregler@nyu.edu New York University CS TR2003-840

More information

Markov Random Fields for Recognizing Textures modeled by Feature Vectors

Markov Random Fields for Recognizing Textures modeled by Feature Vectors Markov Random Fields for Recognizing Textures modeled by Feature Vectors Juliette Blanchet, Florence Forbes, and Cordelia Schmid Inria Rhône-Alpes, 655 avenue de l Europe, Montbonnot, 38334 Saint Ismier

More information

A Two-stage Scheme for Dynamic Hand Gesture Recognition

A Two-stage Scheme for Dynamic Hand Gesture Recognition A Two-stage Scheme for Dynamic Hand Gesture Recognition James P. Mammen, Subhasis Chaudhuri and Tushar Agrawal (james,sc,tush)@ee.iitb.ac.in Department of Electrical Engg. Indian Institute of Technology,

More information

Identifying Humans by Their Walk and Generating New Motions Using Hidden Markov Models

Identifying Humans by Their Walk and Generating New Motions Using Hidden Markov Models Identifying Humans by Their Walk and Generating New Motions Using Hidden Markov Models CPSC 532A Topics in AI: Graphical Models and CPSC 526 Computer Animation December 15, 2004 Andrew Adam andyadam@cs.ubc.ca

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Adding Hand Motion to the Motion Capture Based Character Animation

Adding Hand Motion to the Motion Capture Based Character Animation Adding Hand Motion to the Motion Capture Based Character Animation Ge Jin and James Hahn Computer Science Department, George Washington University, Washington DC 20052 {jinge, hahn}@gwu.edu Abstract. Most

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

Visual Motion Analysis and Tracking Part II

Visual Motion Analysis and Tracking Part II Visual Motion Analysis and Tracking Part II David J Fleet and Allan D Jepson CIAR NCAP Summer School July 12-16, 16, 2005 Outline Optical Flow and Tracking: Optical flow estimation (robust, iterative refinement,

More information

Graph-based High Level Motion Segmentation using Normalized Cuts

Graph-based High Level Motion Segmentation using Normalized Cuts Graph-based High Level Motion Segmentation using Normalized Cuts Sungju Yun, Anjin Park and Keechul Jung Abstract Motion capture devices have been utilized in producing several contents, such as movies

More information

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction

Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction Invariant Recognition of Hand-Drawn Pictograms Using HMMs with a Rotating Feature Extraction Stefan Müller, Gerhard Rigoll, Andreas Kosmala and Denis Mazurenok Department of Computer Science, Faculty of

More information

ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION

ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION ADAPTIVE METROPOLIS-HASTINGS SAMPLING, OR MONTE CARLO KERNEL ESTIMATION CHRISTOPHER A. SIMS Abstract. A new algorithm for sampling from an arbitrary pdf. 1. Introduction Consider the standard problem of

More information

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning Unsupervised Learning Clustering and the EM Algorithm Susanna Ricco Supervised Learning Given data in the form < x, y >, y is the target to learn. Good news: Easy to tell if our algorithm is giving the

More information

Shweta Gandhi, Dr.D.M.Yadav JSPM S Bhivarabai sawant Institute of technology & research Electronics and telecom.dept, Wagholi, Pune

Shweta Gandhi, Dr.D.M.Yadav JSPM S Bhivarabai sawant Institute of technology & research Electronics and telecom.dept, Wagholi, Pune Face sketch photo synthesis Shweta Gandhi, Dr.D.M.Yadav JSPM S Bhivarabai sawant Institute of technology & research Electronics and telecom.dept, Wagholi, Pune Abstract Face sketch to photo synthesis has

More information

ISSN: (Online) Volume 2, Issue 5, May 2014 International Journal of Advance Research in Computer Science and Management Studies

ISSN: (Online) Volume 2, Issue 5, May 2014 International Journal of Advance Research in Computer Science and Management Studies ISSN: 2321-7782 (Online) Volume 2, Issue 5, May 2014 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at:

More information

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme

Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Unsupervised Human Members Tracking Based on an Silhouette Detection and Analysis Scheme Costas Panagiotakis and Anastasios Doulamis Abstract In this paper, an unsupervised, automatic video human members(human

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

Visual Recognition: Image Formation

Visual Recognition: Image Formation Visual Recognition: Image Formation Raquel Urtasun TTI Chicago Jan 5, 2012 Raquel Urtasun (TTI-C) Visual Recognition Jan 5, 2012 1 / 61 Today s lecture... Fundamentals of image formation You should know

More information

Dynamic Human Shape Description and Characterization

Dynamic Human Shape Description and Characterization Dynamic Human Shape Description and Characterization Z. Cheng*, S. Mosher, Jeanne Smith H. Cheng, and K. Robinette Infoscitex Corporation, Dayton, Ohio, USA 711 th Human Performance Wing, Air Force Research

More information

Learnt Inverse Kinematics for Animation Synthesis

Learnt Inverse Kinematics for Animation Synthesis VVG (5) (Editors) Inverse Kinematics for Animation Synthesis Anonymous Abstract Existing work on animation synthesis can be roughly split into two approaches, those that combine segments of motion capture

More information

Final Exam Schedule. Final exam has been scheduled. 12:30 pm 3:00 pm, May 7. Location: INNOVA It will cover all the topics discussed in class

Final Exam Schedule. Final exam has been scheduled. 12:30 pm 3:00 pm, May 7. Location: INNOVA It will cover all the topics discussed in class Final Exam Schedule Final exam has been scheduled 12:30 pm 3:00 pm, May 7 Location: INNOVA 1400 It will cover all the topics discussed in class One page double-sided cheat sheet is allowed A calculator

More information

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework

An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework IEEE SIGNAL PROCESSING LETTERS, VOL. XX, NO. XX, XXX 23 An Efficient Model Selection for Gaussian Mixture Model in a Bayesian Framework Ji Won Yoon arxiv:37.99v [cs.lg] 3 Jul 23 Abstract In order to cluster

More information

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos Machine Learning for Computer Vision 1 22 October, 2012 MVA ENS Cachan Lecture 5: Introduction to generative models Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Center for Visual Computing Ecole Centrale Paris

More information

Tracking Algorithms. Lecture16: Visual Tracking I. Probabilistic Tracking. Joint Probability and Graphical Model. Deterministic methods

Tracking Algorithms. Lecture16: Visual Tracking I. Probabilistic Tracking. Joint Probability and Graphical Model. Deterministic methods Tracking Algorithms CSED441:Introduction to Computer Vision (2017F) Lecture16: Visual Tracking I Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Deterministic methods Given input video and current state,

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover

More information

Video Texture. A.A. Efros

Video Texture. A.A. Efros Video Texture A.A. Efros 15-463: Computational Photography Alexei Efros, CMU, Fall 2005 Weather Forecasting for Dummies Let s predict weather: Given today s weather only, we want to know tomorrow s Suppose

More information

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed

More information

Multiple Model Estimation : The EM Algorithm & Applications

Multiple Model Estimation : The EM Algorithm & Applications Multiple Model Estimation : The EM Algorithm & Applications Princeton University COS 429 Lecture Dec. 4, 2008 Harpreet S. Sawhney hsawhney@sarnoff.com Plan IBR / Rendering applications of motion / pose

More information

Segmentation and Tracking of Partial Planar Templates

Segmentation and Tracking of Partial Planar Templates Segmentation and Tracking of Partial Planar Templates Abdelsalam Masoud William Hoff Colorado School of Mines Colorado School of Mines Golden, CO 800 Golden, CO 800 amasoud@mines.edu whoff@mines.edu Abstract

More information

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim,

More information

Inferring 3D People from 2D Images

Inferring 3D People from 2D Images Inferring 3D People from 2D Images Department of Computer Science Brown University http://www.cs.brown.edu/~black Collaborators Hedvig Sidenbladh, Swedish Defense Research Inst. Leon Sigal, Brown University

More information

Normalized Texture Motifs and Their Application to Statistical Object Modeling

Normalized Texture Motifs and Their Application to Statistical Object Modeling Normalized Texture Motifs and Their Application to Statistical Obect Modeling S. D. Newsam B. S. Manunath Center for Applied Scientific Computing Electrical and Computer Engineering Lawrence Livermore

More information

Quaternion Rotations AUI Course Denbigh Starkey

Quaternion Rotations AUI Course Denbigh Starkey Major points of these notes: Quaternion Rotations AUI Course Denbigh Starkey. What I will and won t be doing. Definition of a quaternion and notation 3 3. Using quaternions to rotate any point around an

More information

Graphical Models & HMMs

Graphical Models & HMMs Graphical Models & HMMs Henrik I. Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I. Christensen (RIM@GT) Graphical Models

More information

This chapter explains two techniques which are frequently used throughout

This chapter explains two techniques which are frequently used throughout Chapter 2 Basic Techniques This chapter explains two techniques which are frequently used throughout this thesis. First, we will introduce the concept of particle filters. A particle filter is a recursive

More information

Nonparametric Bayesian Texture Learning and Synthesis

Nonparametric Bayesian Texture Learning and Synthesis Appears in Advances in Neural Information Processing Systems (NIPS) 2009. Nonparametric Bayesian Texture Learning and Synthesis Long (Leo) Zhu 1 Yuanhao Chen 2 William Freeman 1 Antonio Torralba 1 1 CSAIL,

More information

Thesis Proposal : Switching Linear Dynamic Systems with Higher-order Temporal Structure. Sang Min Oh

Thesis Proposal : Switching Linear Dynamic Systems with Higher-order Temporal Structure. Sang Min Oh Thesis Proposal : Switching Linear Dynamic Systems with Higher-order Temporal Structure Sang Min Oh sangmin@cc.gatech.edu 28th May 2008 Contents 1 Introduction 1 1.1 Automated Temporal Sequence Analysis.................................

More information

Image Coding with Active Appearance Models

Image Coding with Active Appearance Models Image Coding with Active Appearance Models Simon Baker, Iain Matthews, and Jeff Schneider CMU-RI-TR-03-13 The Robotics Institute Carnegie Mellon University Abstract Image coding is the task of representing

More information

Expectation Maximization (EM) and Gaussian Mixture Models

Expectation Maximization (EM) and Gaussian Mixture Models Expectation Maximization (EM) and Gaussian Mixture Models Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 2 3 4 5 6 7 8 Unsupervised Learning Motivation

More information

Robust Event Boundary Detection in Sensor Networks A Mixture Model Based Approach

Robust Event Boundary Detection in Sensor Networks A Mixture Model Based Approach Robust Event Boundary Detection in Sensor Networks A Mixture Model Based Approach Min Ding Department of Computer Science The George Washington University Washington DC 20052, USA Email: minding@gwu.edu

More information

CHAPTER 9. Classification Scheme Using Modified Photometric. Stereo and 2D Spectra Comparison

CHAPTER 9. Classification Scheme Using Modified Photometric. Stereo and 2D Spectra Comparison CHAPTER 9 Classification Scheme Using Modified Photometric Stereo and 2D Spectra Comparison 9.1. Introduction In Chapter 8, even we combine more feature spaces and more feature generators, we note that

More information

A Neural Classifier for Anomaly Detection in Magnetic Motion Capture

A Neural Classifier for Anomaly Detection in Magnetic Motion Capture A Neural Classifier for Anomaly Detection in Magnetic Motion Capture Iain Miller 1 and Stephen McGlinchey 2 1 University of Paisley, Paisley. PA1 2BE, UK iain.miller@paisley.ac.uk, 2 stephen.mcglinchey@paisley.ac.uk

More information

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images

Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images Applying Synthetic Images to Learning Grasping Orientation from Single Monocular Images 1 Introduction - Steve Chuang and Eric Shan - Determining object orientation in images is a well-established topic

More information

Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing

Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Image Segmentation Using Iterated Graph Cuts Based on Multi-scale Smoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200,

More information

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints

Last week. Multi-Frame Structure from Motion: Multi-View Stereo. Unknown camera viewpoints Last week Multi-Frame Structure from Motion: Multi-View Stereo Unknown camera viewpoints Last week PCA Today Recognition Today Recognition Recognition problems What is it? Object detection Who is it? Recognizing

More information

Generative and discriminative classification techniques

Generative and discriminative classification techniques Generative and discriminative classification techniques Machine Learning and Category Representation 2014-2015 Jakob Verbeek, November 28, 2014 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.14.15

More information

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Your Name: Your student id: Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Problem 1 [5+?]: Hypothesis Classes Problem 2 [8]: Losses and Risks Problem 3 [11]: Model Generation

More information

ICA mixture models for image processing

ICA mixture models for image processing I999 6th Joint Sy~nposiurn orz Neural Computation Proceedings ICA mixture models for image processing Te-Won Lee Michael S. Lewicki The Salk Institute, CNL Carnegie Mellon University, CS & CNBC 10010 N.

More information

Machine Learning for Video-Based Rendering

Machine Learning for Video-Based Rendering Machine Learning for Video-Based Rendering Arno Schadl arno@schoedl. org Irfan Essa irjan@cc.gatech.edu Georgia Institute of Technology GVU Center / College of Computing Atlanta, GA 30332-0280, USA. Abstract

More information

Region-based Segmentation

Region-based Segmentation Region-based Segmentation Image Segmentation Group similar components (such as, pixels in an image, image frames in a video) to obtain a compact representation. Applications: Finding tumors, veins, etc.

More information

SCAPE: Shape Completion and Animation of People

SCAPE: Shape Completion and Animation of People SCAPE: Shape Completion and Animation of People By Dragomir Anguelov, Praveen Srinivasan, Daphne Koller, Sebastian Thrun, Jim Rodgers, James Davis From SIGGRAPH 2005 Presentation for CS468 by Emilio Antúnez

More information

Topological Mapping. Discrete Bayes Filter

Topological Mapping. Discrete Bayes Filter Topological Mapping Discrete Bayes Filter Vision Based Localization Given a image(s) acquired by moving camera determine the robot s location and pose? Towards localization without odometry What can be

More information

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING Jianzhou Feng Li Song Xiaog Huo Xiaokang Yang Wenjun Zhang Shanghai Digital Media Processing Transmission Key Lab, Shanghai Jiaotong University

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing

Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Image Segmentation Using Iterated Graph Cuts BasedonMulti-scaleSmoothing Tomoyuki Nagahashi 1, Hironobu Fujiyoshi 1, and Takeo Kanade 2 1 Dept. of Computer Science, Chubu University. Matsumoto 1200, Kasugai,

More information

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin Clustering K-means Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, 2014 Carlos Guestrin 2005-2014 1 Clustering images Set of Images [Goldberger et al.] Carlos Guestrin 2005-2014

More information

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin

Clustering K-means. Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, Carlos Guestrin Clustering K-means Machine Learning CSEP546 Carlos Guestrin University of Washington February 18, 2014 Carlos Guestrin 2005-2014 1 Clustering images Set of Images [Goldberger et al.] Carlos Guestrin 2005-2014

More information

Supervised vs. Unsupervised Learning

Supervised vs. Unsupervised Learning Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now

More information

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis

Image analysis. Computer Vision and Classification Image Segmentation. 7 Image analysis 7 Computer Vision and Classification 413 / 458 Computer Vision and Classification The k-nearest-neighbor method The k-nearest-neighbor (knn) procedure has been used in data analysis and machine learning

More information

SD 372 Pattern Recognition

SD 372 Pattern Recognition SD 372 Pattern Recognition Lab 2: Model Estimation and Discriminant Functions 1 Purpose This lab examines the areas of statistical model estimation and classifier aggregation. Model estimation will be

More information

An Introduction to Pattern Recognition

An Introduction to Pattern Recognition An Introduction to Pattern Recognition Speaker : Wei lun Chao Advisor : Prof. Jian-jiun Ding DISP Lab Graduate Institute of Communication Engineering 1 Abstract Not a new research field Wide range included

More information

INFOMCANIM Computer Animation Motion Synthesis. Christyowidiasmoro (Chris)

INFOMCANIM Computer Animation Motion Synthesis. Christyowidiasmoro (Chris) INFOMCANIM Computer Animation Motion Synthesis Christyowidiasmoro (Chris) Why Motion Synthesis? We don t have the specific motion / animation We don t have the skeleton and motion for specific characters

More information

Texture Synthesis and Manipulation Project Proposal. Douglas Lanman EN 256: Computer Vision 19 October 2006

Texture Synthesis and Manipulation Project Proposal. Douglas Lanman EN 256: Computer Vision 19 October 2006 Texture Synthesis and Manipulation Project Proposal Douglas Lanman EN 256: Computer Vision 19 October 2006 1 Outline Introduction to Texture Synthesis Previous Work Project Goals and Timeline Douglas Lanman

More information

Missing variable problems

Missing variable problems Missing variable problems In many vision problems, if some variables were known the maximum likelihood inference problem would be easy fitting; if we knew which line each token came from, it would be easy

More information

Joint design of data analysis algorithms and user interface for video applications

Joint design of data analysis algorithms and user interface for video applications Joint design of data analysis algorithms and user interface for video applications Nebojsa Jojic Microsoft Research Sumit Basu Microsoft Research Nemanja Petrovic University of Illinois Brendan Frey University

More information

Reconstruction of Images Distorted by Water Waves

Reconstruction of Images Distorted by Water Waves Reconstruction of Images Distorted by Water Waves Arturo Donate and Eraldo Ribeiro Computer Vision Group Outline of the talk Introduction Analysis Background Method Experiments Conclusions Future Work

More information

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation

Problem Set 4. Assigned: March 23, 2006 Due: April 17, (6.882) Belief Propagation for Segmentation 6.098/6.882 Computational Photography 1 Problem Set 4 Assigned: March 23, 2006 Due: April 17, 2006 Problem 1 (6.882) Belief Propagation for Segmentation In this problem you will set-up a Markov Random

More information

Object and Action Detection from a Single Example

Object and Action Detection from a Single Example Object and Action Detection from a Single Example Peyman Milanfar* EE Department University of California, Santa Cruz *Joint work with Hae Jong Seo AFOSR Program Review, June 4-5, 29 Take a look at this:

More information

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries

Improving Latent Fingerprint Matching Performance by Orientation Field Estimation using Localized Dictionaries Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 11, November 2014,

More information

Audio-Video Sensor Fusion with Probabilistic Graphical Models

Audio-Video Sensor Fusion with Probabilistic Graphical Models Audio-Video Sensor Fusion with Probabilistic Graphical Models Matthew J. Beal 1,2, Hagai Attias 1, and Nebojsa Jojic 1 1 Microsoft Research, 1 Microsoft Way, Redmond, WA 98052, USA {hagaia,jojic}@microsoft.com

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Inference and Representation

Inference and Representation Inference and Representation Rachel Hodos New York University Lecture 5, October 6, 2015 Rachel Hodos Lecture 5: Inference and Representation Today: Learning with hidden variables Outline: Unsupervised

More information

Comparing Dropout Nets to Sum-Product Networks for Predicting Molecular Activity

Comparing Dropout Nets to Sum-Product Networks for Predicting Molecular Activity 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System Proc. of IEEE Conference on Computer Vision and Pattern Recognition, vol.2, II-131 II-137, Dec. 2001. Production of Video Images by Computer Controlled Cameras and Its Application to TV Conference System

More information

Clustering. Supervised vs. Unsupervised Learning

Clustering. Supervised vs. Unsupervised Learning Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now

More information

From Gaze to Focus of Attention

From Gaze to Focus of Attention From Gaze to Focus of Attention Rainer Stiefelhagen, Michael Finke, Jie Yang, Alex Waibel stiefel@ira.uka.de, finkem@cs.cmu.edu, yang+@cs.cmu.edu, ahw@cs.cmu.edu Interactive Systems Laboratories University

More information

Level-set MCMC Curve Sampling and Geometric Conditional Simulation

Level-set MCMC Curve Sampling and Geometric Conditional Simulation Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve

More information