Predictive Discrete Latent Factor Models for Large Scale Dyadic Data Deepak Agarwal, Srujana Merugu Yahoo! Research
|
|
- Cordelia Reeves
- 5 years ago
- Views:
Transcription
1 Predictive Discrete Latent Factor Models for Large Scale Dyadic Data Deepak Agarwal, Srujana Merugu Yahoo! Research March 5, 2009
2 Outline 1 Motivation 2 Background 3 Algorithm 4 Analysis
3 Recommender System Problem. i 2 i 3...i m u 2 u 3... u n Properties Rows represent users Columns represent items Matrix is sparse, most elements are unknown How to estimate unknown elements
4 Problems Sparsity Signal to Noise Ratio is very low Mean centric approaches do not work very well Overfitting PDLF Algorithm Captures global and local structure by simultaneously using a supervised and unsupervised approach.
5 Capturing Global Structure Supervised Approach Example Multinomial logistic regression Select β features for each class Fails to capture local interaction between dyads PDLF The PDLF algorithm captures global structure through the weights associated with the β features.
6 Capturing Local Structure Unsupervised approach Hard Assignment Co-Clustering Cluster n rows into k row clusters Cluster m rows into l column clusters Minimize the Mutual Information Loss
7 Exponential Family Distribution Model The algorithm trains a model from an exponential family for each dyad f(x;θ) = exp(θ t(x) ψ(θ))p 0 (x)) Terms ψ, cumulant generating function θ, natural parameter t(x), sufficent statistic p 0 (x), normalizing term
8 General Linear Models GLM Maps the β features to the model parameter in the exponential family. The response variable Y is linear combination of input random variables {X 1,X 2,...X n } The link function g(θ) = β t x The response function f(β t x) = θ Relation to PDLF The GLM s response function maps β t x + δ to the model parameter of the exponential family distribution.
9 Hard Assignment PDLF β : Features capture global structure : Interaction Effects capture local structure ψ: cumulant generating function of the exponential family ρ: row cluster assignments for each user γ: column cluster assignments for each item/movie x ij : feature values of a given dyad y ij : response variable of the dyad (rating) k: number of row clusters l: number of column clusters
10 Hard Assignment PDLF Generalized M-Step Step 1. Update Interaction Effects Step 2. Update Regression Coefficients Generalized E-Step Step 3. Update Row Cluster Assignments δ I,J argmax y ij δ ψ(β t x ij + δ) δ i I,j J β argmax y ij β t x ij ψ(β t x ij + δ) β ij ρ(i) argmax I Step 4. Update Column Cluster Assignments Repeat until convergence γ(j) argmax J (y ij δ Iγ(j) ψ(β t x ij + δ Iγ(j) )) j (y ij δ ρ(i)j ψ(β t x ij + δ ρ(i)j )) i
11 Algorithm Analysis Algorithm Analysis Algorithm returns β,,ρ,γ Prediction for a dyad: f ψ (:;B t x ij + δ ρ(i)γ(j) ) Algorithm runs in O(Nkl)
12 Questions? References Predictive Discrete Latent Factor Models for Large Scale Dyadic Data,Deepak Agarawal, Srujana Merugu, KDD 2007 Clustering with Bregman Divergences, Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon, Joydeep Ghosh, JMLR 2005 Information-Theoretic co-clustering, I. Dhillon, S. Mallela, D. Hodha KDD
Progress Report: Collaborative Filtering Using Bregman Co-clustering
Progress Report: Collaborative Filtering Using Bregman Co-clustering Wei Tang, Srivatsan Ramanujam, and Andrew Dreher April 4, 2008 1 Introduction Analytics are becoming increasingly important for business
More informationA Memetic Heuristic for the Co-clustering Problem
A Memetic Heuristic for the Co-clustering Problem Mohammad Khoshneshin 1, Mahtab Ghazizadeh 2, W. Nick Street 1, and Jeffrey W. Ohlmann 1 1 The University of Iowa, Iowa City IA 52242, USA {mohammad-khoshneshin,nick-street,jeffrey-ohlmann}@uiowa.edu
More informationMultiplicative Mixture Models for Overlapping Clustering
Multiplicative Mixture Models for Overlapping Clustering Qiang Fu Dept of Computer Science & Engineering University of Minnesota, Twin Cities qifu@cs.umn.edu Arindam Banerjee Dept of Computer Science &
More informationScalable Data Analysis
Scalable Data Analysis David M. Blei April 26, 2012 1 Why? Olden days: Small data sets Careful experimental design Challenge: Squeeze as much as we can out of the data Modern data analysis: Very large
More informationRelaxed Oracles for Semi-Supervised Clustering
Relaxed Oracles for Semi-Supervised Clustering Taewan Kim The University of Texas at Austin twankim@utexas.edu Joydeep Ghosh The University of Texas at Austin jghosh@utexas.edu Abstract Pairwise same-cluster
More informationData Clustering. Danushka Bollegala
Data Clustering Danushka Bollegala Outline Why cluster data? Clustering as unsupervised learning Clustering algorithms k-means, k-medoids agglomerative clustering Brown s clustering Spectral clustering
More informationLiangjie Hong*, Dawei Yin*, Jian Guo, Brian D. Davison*
Tracking Trends: Incorporating Term Volume into Temporal Topic Models Liangjie Hong*, Dawei Yin*, Jian Guo, Brian D. Davison* Dept. of Computer Science and Engineering, Lehigh University, Bethlehem, PA,
More informationCo-Clustering with Generative Models Danial Lashkari and Polina Golland
Computer Science and Artificial Intelligence Laboratory Technical Report MIT-CSAIL-TR-2009-054 November 3, 2009 Co-Clustering with Generative Models Danial Lashkari and Polina Golland massachusetts institute
More informationLecture on Modeling Tools for Clustering & Regression
Lecture on Modeling Tools for Clustering & Regression CS 590.21 Analysis and Modeling of Brain Networks Department of Computer Science University of Crete Data Clustering Overview Organizing data into
More informationGeoff McLachlan and Angus Ng. University of Queensland. Schlumberger Chaired Professor Univ. of Texas at Austin. + Chris Bishop
EM Algorithm Geoff McLachlan and Angus Ng Department of Mathematics & Institute for Molecular Bioscience University of Queensland Adapted by Joydeep Ghosh Schlumberger Chaired Professor Univ. of Texas
More informationECE521: Week 11, Lecture March 2017: HMM learning/inference. With thanks to Russ Salakhutdinov
ECE521: Week 11, Lecture 20 27 March 2017: HMM learning/inference With thanks to Russ Salakhutdinov Examples of other perspectives Murphy 17.4 End of Russell & Norvig 15.2 (Artificial Intelligence: A Modern
More informationConstrained Co-clustering for Textual Documents
Constrained Co-clustering for Textual Documents Yangqiu Song Shimei Pan Shixia Liu Furu Wei Michelle X. Zhou Weihong Qian {yqsong,liusx,weifuru,qianwh}@cn.ibm.com; shimei@us.ibm.com; mzhou@us.ibm.com IBM
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationGAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.
GAMs semi-parametric GLMs Simon Wood Mathematical Sciences, University of Bath, U.K. Generalized linear models, GLM 1. A GLM models a univariate response, y i as g{e(y i )} = X i β where y i Exponential
More informationMissing Data Estimation in Microarrays Using Multi-Organism Approach
Missing Data Estimation in Microarrays Using Multi-Organism Approach Marcel Nassar and Hady Zeineddine Progress Report: Data Mining Course Project, Spring 2008 Prof. Inderjit S. Dhillon April 02, 2008
More informationOutline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff
Compressed Sensing David L Donoho Presented by: Nitesh Shroff University of Maryland Outline 1 Introduction Compressed Sensing 2 Problem Formulation Sparse Signal Problem Statement 3 Proposed Solution
More informationA tutorial for blockcluster R package Version 4
A tutorial for blockcluster R package Version 4 Parmeet Singh Bhatia, Serge Iovleff ❸ January 24, 2018 Contents 1 Introduction 2 2 Package details 2 2.1 cocluster function....................................
More informationMultiresponse Sparse Regression with Application to Multidimensional Scaling
Multiresponse Sparse Regression with Application to Multidimensional Scaling Timo Similä and Jarkko Tikka Helsinki University of Technology, Laboratory of Computer and Information Science P.O. Box 54,
More informationECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 22, 2016 Course Information Website: http://www.stat.ucdavis.edu/~chohsieh/teaching/ ECS289G_Fall2016/main.html My office: Mathematical Sciences
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:
More informationAdvances in Fuzzy Clustering and Its Applications. J. Valente de Oliveira and W. Pedrycz (Editors)
Advances in Fuzzy Clustering and Its Applications J. Valente de Oliveira and W. Pedrycz (Editors) Contents Preface 3 1 Soft Cluster Ensembles 1 1.1 Introduction................................... 1 1.1.1
More informationStatistical Methods for Network Analysis: Exponential Random Graph Models
Day 2: Network Modeling Statistical Methods for Network Analysis: Exponential Random Graph Models NMID workshop September 17 21, 2012 Prof. Martina Morris Prof. Steven Goodreau Supported by the US National
More informationCurve fitting using linear models
Curve fitting using linear models Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark September 28, 2012 1 / 12 Outline for today linear models and basis functions polynomial regression
More informationContents. Preface to the Second Edition
Preface to the Second Edition v 1 Introduction 1 1.1 What Is Data Mining?....................... 4 1.2 Motivating Challenges....................... 5 1.3 The Origins of Data Mining....................
More informationParallel and Distributed Sparse Optimization Algorithms
Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2017 Assignment 3: 2 late days to hand in tonight. Admin Assignment 4: Due Friday of next week. Last Time: MAP Estimation MAP
More informationFeature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani
Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate
More informationCS839: Probabilistic Graphical Models. Lecture 10: Learning with Partially Observed Data. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 10: Learning with Partially Observed Data Theo Rekatsinas 1 Partially Observed GMs Speech recognition 2 Partially Observed GMs Evolution 3 Partially Observed
More informationCPSC 340: Machine Learning and Data Mining. Recommender Systems Fall 2017
CPSC 340: Machine Learning and Data Mining Recommender Systems Fall 2017 Assignment 4: Admin Due tonight, 1 late day for Monday, 2 late days for Wednesday. Assignment 5: Posted, due Monday of last week
More informationCo-ClusterD: A Distributed Framework for Data Co-Clustering with Sequential Updates
Co-ClusterD: A Distributed Framework for Data Co-Clustering with Sequential Updates Sen Su, Xiang Cheng State Key Laboratory of Networking and Switching Technology Beijing University of Posts and Telecommunications
More informationCo-clustering by Block Value Decomposition
Co-clustering by Block Value Decomposition Bo Long Computer Science Dept. SUNY Binghamton Binghamton, NY 13902 blong1@binghamton.edu Zhongfei (Mark) Zhang Computer Science Dept. SUNY Binghamton Binghamton,
More informationRegularization and model selection
CS229 Lecture notes Andrew Ng Part VI Regularization and model selection Suppose we are trying select among several different models for a learning problem. For instance, we might be using a polynomial
More informationLDA for Big Data - Outline
LDA FOR BIG DATA 1 LDA for Big Data - Outline Quick review of LDA model clustering words-in-context Parallel LDA ~= IPM Fast sampling tricks for LDA Sparsified sampler Alias table Fenwick trees LDA for
More informationA Scalable Framework for Discovering Coherent Co-clusters in Noisy Data
A Scalable Framework for Discovering Coherent Co-clusters in Noisy Data Meghana Deodhar, Gunjan Gupta, Joydeep Ghosh deodhar,ggupta,ghosh@ece.utexas.edu Department of ECE, University of Texas at Austin,
More informationMultiple imputation using chained equations: Issues and guidance for practice
Multiple imputation using chained equations: Issues and guidance for practice Ian R. White, Patrick Royston and Angela M. Wood http://onlinelibrary.wiley.com/doi/10.1002/sim.4067/full By Gabrielle Simoneau
More informationHierarchical Co-Clustering Based on Entropy Splitting
Hierarchical Co-Clustering Based on Entropy Splitting Wei Cheng 1, Xiang Zhang 2, Feng Pan 3, and Wei Wang 4 1 Department of Computer Science, University of North Carolina at Chapel Hill, 2 Department
More informationLecture 10: PGM Structure Estimation
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Heuristics 1 Heuristics Manually specify a fixed graph Use a simple rule 2 from labels only from both labels and features
More informationSimultaneous Co-clustering and Learning with Dimensionality Reduction
1 Simultaneous Co-clustering and earning with Dimensionality Reduction Harish Ganapathy CS 229 Project Report E-mail: harishg@utexas.edu Abstract In this project, we are interested in solving predictions
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationC. Poultney S. Cho pra (NYU Courant Institute) Y. LeCun
Efficient Learning of Sparse Overcomplete Representations with an Energy-Based Model Marc'Aurelio Ranzato C. Poultney S. Cho pra (NYU Courant Institute) Y. LeCun CIAR Summer School Toronto 2006 Why Extracting
More informationDistributed Hierarchical Co-clustering and Collaborative Filtering Algorithm
Distributed Hierarchical Co-clustering and Collaborative Filtering Algorithm Ankur Narang, Abhinav Srivastava IBM India Research Laboratory New Delhi, India (annarang,abhin122)@in.ibm.com Naga Praveen
More informationMachine Learning Feature Creation and Selection
Machine Learning Feature Creation and Selection Jeff Howbert Introduction to Machine Learning Winter 2012 1 Feature creation Well-conceived new features can sometimes capture the important information
More informationREGRESSION ANALYSIS : LINEAR BY MAUAJAMA FIRDAUS & TULIKA SAHA
REGRESSION ANALYSIS : LINEAR BY MAUAJAMA FIRDAUS & TULIKA SAHA MACHINE LEARNING It is the science of getting computer to learn without being explicitly programmed. Machine learning is an area of artificial
More informationML4Bio Lecture #1: Introduc3on. February 24 th, 2016 Quaid Morris
ML4Bio Lecture #1: Introduc3on February 24 th, 216 Quaid Morris Course goals Prac3cal introduc3on to ML Having a basic grounding in the terminology and important concepts in ML; to permit self- study,
More informationClustering with Reinforcement Learning
Clustering with Reinforcement Learning Wesam Barbakh and Colin Fyfe, The University of Paisley, Scotland. email:wesam.barbakh,colin.fyfe@paisley.ac.uk Abstract We show how a previously derived method of
More informationConvex Clustering with Exemplar-Based Models
Convex Clustering with Exemplar-Based Models Danial Lashkari Polina Golland Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, MA 2139 {danial, polina}@csail.mit.edu
More informationExplore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan
Explore Co-clustering on Job Applications Qingyun Wan SUNet ID:qywan 1 Introduction In the job marketplace, the supply side represents the job postings posted by job posters and the demand side presents
More informationBenchmarking Spark ML using BigBench. Sweta Singh TPCTC 2016
Benchmarking Spark ML using BigBench Sweta Singh singhswe@us.ibm.com TPCTC 2016 Motivation Study the performance of Machine Learning use cases on large data warehouses in context of assessing Alternate
More informationStatistical Matching using Fractional Imputation
Statistical Matching using Fractional Imputation Jae-Kwang Kim 1 Iowa State University 1 Joint work with Emily Berg and Taesung Park 1 Introduction 2 Classical Approaches 3 Proposed method 4 Application:
More informationMPMA: Mixture Probabilistic Matrix Approximation for Collaborative Filtering
Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) MPMA: Mixture Probabilistic Matrix Approximation for Collaborative Filtering Chao Chen, 1, Dongsheng
More informationExemplar-based Robust Coherent Biclustering
Exemplar-based Robust Coherent Biclustering Kewei Tu Xixiu Ouyang Dingyi Han Yong Yu Vasant Honavar Abstract The biclustering, co-clustering, or subspace clustering problem involves simultaneously grouping
More informationInformation-Theoretic Co-clustering
Information-Theoretic Co-clustering Authors: I. S. Dhillon, S. Mallela, and D. S. Modha. MALNIS Presentation Qiufen Qi, Zheyuan Yu 20 May 2004 Outline 1. Introduction 2. Information Theory Concepts 3.
More informationTransfer Learning Algorithms for Image Classification
Transfer Learning Algorithms for Image Classification Ariadna Quattoni MIT, CSAIL Advisors: Michael Collins Trevor Darrell 1 Motivation Goal: We want to be able to build classifiers for thousands of visual
More informationExpectation Maximization (EM) and Gaussian Mixture Models
Expectation Maximization (EM) and Gaussian Mixture Models Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 2 3 4 5 6 7 8 Unsupervised Learning Motivation
More informationPenalizied Logistic Regression for Classification
Penalizied Logistic Regression for Classification Gennady G. Pekhimenko Department of Computer Science University of Toronto Toronto, ON M5S3L1 pgen@cs.toronto.edu Abstract Investigation for using different
More informationHybrid Hierarchical Clustering: Forming a Tree From Multiple Views
Hybrid Hierarchical Clustering: Forming a Tree From Multiple Views Anjum Gupta University of California, San Diego. 9500 Gilman Drive, La Jolla, CA 92093-0114 Sanjoy Dasgupta University of California,
More informationClustering: Classic Methods and Modern Views
Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering
More informationIntro to Random Graphs and Exponential Random Graph Models
Intro to Random Graphs and Exponential Random Graph Models Danielle Larcomb University of Denver Danielle Larcomb Random Graphs 1/26 Necessity of Random Graphs The study of complex networks plays an increasingly
More informationCS 179 Lecture 16. Logistic Regression & Parallel SGD
CS 179 Lecture 16 Logistic Regression & Parallel SGD 1 Outline logistic regression (stochastic) gradient descent parallelizing SGD for neural nets (with emphasis on Google s distributed neural net implementation)
More informationSymposium on Optimal Transport and Information Geometry
Symposium on Optimal Transport and Information Geometry Computational information geometry for audio signal processing (Based on a tutorial presented at DAFx 2012) Arnaud Dessein Institut de Recherche
More informationDS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University
DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University January 24 2019 Logistics HW 1 is due on Friday 01/25 Project proposal: due Feb 21 1 page description
More informationEfficient Feature Learning Using Perturb-and-MAP
Efficient Feature Learning Using Perturb-and-MAP Ke Li, Kevin Swersky, Richard Zemel Dept. of Computer Science, University of Toronto {keli,kswersky,zemel}@cs.toronto.edu Abstract Perturb-and-MAP [1] is
More informationDOCUMENT CLUSTERING USING HIERARCHICAL METHODS. 1. Dr.R.V.Krishnaiah 2. Katta Sharath Kumar. 3. P.Praveen Kumar. achieved.
DOCUMENT CLUSTERING USING HIERARCHICAL METHODS 1. Dr.R.V.Krishnaiah 2. Katta Sharath Kumar 3. P.Praveen Kumar ABSTRACT: Cluster is a term used regularly in our life is nothing but a group. In the view
More informationLearning to Track Motion
Learning to Track Motion Maitreyi Nanjanath Amit Bose CSci 8980 Course Project April 25, 2006 Background Vision sensor can provide a great deal of information in a short sequence of images Useful for determining
More informationTopic mash II: assortativity, resilience, link prediction CS224W
Topic mash II: assortativity, resilience, link prediction CS224W Outline Node vs. edge percolation Resilience of randomly vs. preferentially grown networks Resilience in real-world networks network resilience
More informationClassification. 1 o Semestre 2007/2008
Classification Departamento de Engenharia Informática Instituto Superior Técnico 1 o Semestre 2007/2008 Slides baseados nos slides oficiais do livro Mining the Web c Soumen Chakrabarti. Outline 1 2 3 Single-Class
More informationSemi-Supervised Clustering with Partial Background Information
Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject
More informationA tutorial for blockcluster R package Version 1.01
A tutorial for blockcluster R package Version 1.01 Parmeet Singh Bhatia INRIA-Lille, parmeet.bhatia@inria.fr Contents 1 Introduction 1 2 Package details 2 2.1 cocluster function....................................
More informationLearning Two-View Stereo Matching
Learning Two-View Stereo Matching Jianxiong Xiao Jingni Chen Dit-Yan Yeung Long Quan Department of Computer Science and Engineering The Hong Kong University of Science and Technology The 10th European
More informationMachine Learning (BSMC-GA 4439) Wenke Liu
Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning
More informationSupervised Learning for Image Segmentation
Supervised Learning for Image Segmentation Raphael Meier 06.10.2016 Raphael Meier MIA 2016 06.10.2016 1 / 52 References A. Ng, Machine Learning lecture, Stanford University. A. Criminisi, J. Shotton, E.
More informationA General Model for Relational Clustering
Bo Long Zhongfei (Mark) Zhang Computer Science Dept., SUNY Binghamton, Binghamton, NY 392 USA Xiaoyun Wu Yahoo! Inc, 7 First Avenue, Sunnyvale, CA 9489 USA Philip S. Yu IBM Watson Research Center, 9 skyline
More informationA Neuro Probabilistic Language Model Bengio et. al. 2003
A Neuro Probabilistic Language Model Bengio et. al. 2003 Class Discussion Notes Scribe: Olivia Winn February 1, 2016 Opening thoughts (or why this paper is interesting): Word embeddings currently have
More informationOnline Social Networks and Media
Online Social Networks and Media Absorbing Random Walks Link Prediction Why does the Power Method work? If a matrix R is real and symmetric, it has real eigenvalues and eigenvectors: λ, w, λ 2, w 2,, (λ
More informationLudwig Fahrmeir Gerhard Tute. Statistical odelling Based on Generalized Linear Model. íecond Edition. . Springer
Ludwig Fahrmeir Gerhard Tute Statistical odelling Based on Generalized Linear Model íecond Edition. Springer Preface to the Second Edition Preface to the First Edition List of Examples List of Figures
More informationMining Web Data. Lijun Zhang
Mining Web Data Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Web Crawling and Resource Discovery Search Engine Indexing and Query Processing Ranking Algorithms Recommender Systems
More informationA Topography-Preserving Latent Variable Model with Learning Metrics
A Topography-Preserving Latent Variable Model with Learning Metrics Samuel Kaski and Janne Sinkkonen Helsinki University of Technology Neural Networks Research Centre P.O. Box 5400, FIN-02015 HUT, Finland
More information1 Training/Validation/Testing
CPSC 340 Final (Fall 2015) Name: Student Number: Please enter your information above, turn off cellphones, space yourselves out throughout the room, and wait until the official start of the exam to begin.
More informationCluster Validation. Ke Chen. Reading: [25.1.2, KPM], [Wang et al., 2009], [Yang & Chen, 2011] COMP24111 Machine Learning
Cluster Validation Ke Chen Reading: [5.., KPM], [Wang et al., 9], [Yang & Chen, ] COMP4 Machine Learning Outline Motivation and Background Internal index Motivation and general ideas Variance-based internal
More informationEfficient Mining of Heterogeneous Star-Structured Data
Int J Software Informatics, Vol.2, No.2, December 2008, pp. 141 161 E-mail: ijsi@iscas.ac.cn International Journal of Software and Informatics, ISSN 1673-7288 http://www.ijsi.org 2008 by Institute of Software,
More informationMachine Learning: Think Big and Parallel
Day 1 Inderjit S. Dhillon Dept of Computer Science UT Austin CS395T: Topics in Multicore Programming Oct 1, 2013 Outline Scikit-learn: Machine Learning in Python Supervised Learning day1 Regression: Least
More informationExtracting and Composing Robust Features with Denoising Autoencoders
Presenter: Alexander Truong March 16, 2017 Extracting and Composing Robust Features with Denoising Autoencoders Pascal Vincent, Hugo Larochelle, Yoshua Bengio, Pierre-Antoine Manzagol 1 Outline Introduction
More informationAn Objective Evaluation Criterion for Clustering
An Objective Evaluation Criterion for Clustering ABSTRACT Arindam Banerjee Dept of ECE University of Texas at Austin Austin, TX, USA abanerje@ece.utexas.edu We propose and test an objective criterion for
More informationConvex Clustering with Exemplar-Based Models
Convex Clustering with Exemplar-Based Models Danial Lashkari Polina Golland Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge, MA 2139 {danial, polina}@csail.mit.edu
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More informationBen Baumer Instructor
MULTIPLE AND LOGISTIC REGRESSION What is logistic regression? Ben Baumer Instructor A categorical response variable ggplot(data = hearttr, aes(x = age, y = survived)) + geom_jitter(width = 0, height =
More informationCluster Analysis (b) Lijun Zhang
Cluster Analysis (b) Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Grid-Based and Density-Based Algorithms Graph-Based Algorithms Non-negative Matrix Factorization Cluster Validation Summary
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University March 4, 2015 Today: Graphical models Bayes Nets: EM Mixture of Gaussian clustering Learning Bayes Net structure
More informationLasso. November 14, 2017
Lasso November 14, 2017 Contents 1 Case Study: Least Absolute Shrinkage and Selection Operator (LASSO) 1 1.1 The Lasso Estimator.................................... 1 1.2 Computation of the Lasso Solution............................
More informationConstrained Convolutional Neural Networks for Weakly Supervised Segmentation. Deepak Pathak, Philipp Krähenbühl and Trevor Darrell
Constrained Convolutional Neural Networks for Weakly Supervised Segmentation Deepak Pathak, Philipp Krähenbühl and Trevor Darrell 1 Multi-class Image Segmentation Assign a class label to each pixel in
More informationModel-based Overlapping Clustering
Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining Chicago, IL, pp. 532--537, August 2005. Model-based Overlapping Clustering Arindam Banerjee Chase
More informationSpectral Clustering for Multi-type Relational Data
Bo Long Zhongfei (Mark) Zhang Computer Science Dept., SUNY Binghamton, Binghamton, NY 392 USA iaoyun Wu Yahoo! Inc, 7 First Avenue, Sunnyvale, CA 9489 USA Philip S. Yu IBM Watson Research Center, 9 skyline
More informationICRA 2016 Tutorial on SLAM. Graph-Based SLAM and Sparsity. Cyrill Stachniss
ICRA 2016 Tutorial on SLAM Graph-Based SLAM and Sparsity Cyrill Stachniss 1 Graph-Based SLAM?? 2 Graph-Based SLAM?? SLAM = simultaneous localization and mapping 3 Graph-Based SLAM?? SLAM = simultaneous
More informationSemi-supervised Learning
Semi-supervised Learning Piyush Rai CS5350/6350: Machine Learning November 8, 2011 Semi-supervised Learning Supervised Learning models require labeled data Learning a reliable model usually requires plenty
More informationEfficient Semi-supervised Spectral Co-clustering with Constraints
2010 IEEE International Conference on Data Mining Efficient Semi-supervised Spectral Co-clustering with Constraints Xiaoxiao Shi, Wei Fan, Philip S. Yu Department of Computer Science, University of Illinois
More informationSCUBA DIVER: SUBSPACE CLUSTERING OF WEB SEARCH RESULTS
SCUBA DIVER: SUBSPACE CLUSTERING OF WEB SEARCH RESULTS Fatih Gelgi, Srinivas Vadrevu, Hasan Davulcu Department of Computer Science and Engineering, Arizona State University, Tempe, AZ fagelgi@asu.edu,
More informationAn MM Algorithm for Multicategory Vertex Discriminant Analysis
An MM Algorithm for Multicategory Vertex Discriminant Analysis Tong Tong Wu Department of Epidemiology and Biostatistics University of Maryland, College Park May 22, 2008 Joint work with Professor Kenneth
More informationLearning from High Dimensional fmri Data using Random Projections
Learning from High Dimensional fmri Data using Random Projections Author: Madhu Advani December 16, 011 Introduction The term the Curse of Dimensionality refers to the difficulty of organizing and applying
More informationCS281 Section 9: Graph Models and Practical MCMC
CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs
More informationGeodesics in heat: A new approach to computing distance
Geodesics in heat: A new approach to computing distance based on heat flow Diana Papyan Faculty of Informatics - Technische Universität München Abstract In this report we are going to introduce new method
More information