Where we are. Exploratory Graph Analysis (40 min) Focused Graph Mining (40 min) Refinement of Query Results (40 min)
|
|
- Osborne Simpson
- 5 years ago
- Views:
Transcription
1 Where we are Background (15 min) Graph models, subgraph isomorphism, subgraph mining, graph clustering Eploratory Graph Analysis (40 min) Focused Graph Mining (40 min) Refinement of Query Results (40 min) Machine Learning and Visualization (40 min) Challenges and discussion D. MOTTIN, E. MÜLLER 121
2 Interactivity An interactive approach allows the user to modify the state of the system Interactivity can be coupled with personalization: learning user preferences in an interactive way How do we learn on the way? Assume preferences follow a certain distribution and smoothness on neighbor nodes Try to minimize the number of interactions with the user D. MOTTIN, E. MÜLLER 122
3 Online learning (Dataset eploration) Main idea: Learn the items to show online as more points are acquired items Two ways of learning: passive and active Learn items Is t v or? v t Learn Passive Active D. MOTTIN, E. MÜLLER 123
4 Online learning taonomy Online learning Statistical Learning Conve optimization Game theory Passive Active Traditional Bayesian Selective Pool-based Kernel based Forests and trees First/secondorder learning Prediction with eperts Multi-armed bandits Sampling Uncertainty sampling Query by Committee Variance sampling Density based Cluster-based Multi-armed bandits S. C. H. Hoi, D. Sahoo, J. Lu, and P. Zhao, Online Learning: A Comprehensive Survey, arxiv, D. MOTTIN, E. MÜLLER 124
5 Passive learning application: MindReader Main idea: learn an implicit query from user eamples and optional scores Searching mildly overweighted patients Weight The doctor selects eamples by browsing patient database q The eamples have oblique correlation We can guess the implied query : good : very good Height Ishikawa, Y., Subramanya, R., & Faloutsos, C.. MindReader - Querying Databases Through Multiple Eamples. VLDB D. MOTTIN, E. MÜLLER 125
6 Learning an ellipsoid distance Euclidean q weighted Euclidean q generalized ellipsoid distance q ( œ, ß = œ ß æ(œ ß) Implicit query / Weighted distance matri ( œ, ß = o o Ç Éâ (œ É ß É )(œ â ß â ) É / â Learn the query minimizing the penalty = weighted sum of distances between query point and sample vectors ÇeîeÇe d o œ 0 ß æ(œ 0 ß) 0 F;i dkg Gc det æ = 1 Ishikawa, Y., Subramanya, R., & Faloutsos, C.. MindReader - Querying Databases Through Multiple Eamples. VLDB D. MOTTIN, E. MÜLLER 126
7 Learning the distance Query point is moved towards good eamples Rocchio formula in IR Q 0 : query point Q 1 : retrieved data : relevance judgments Q 1 : new query point Q 0 Learning can be done online!!! Ishikawa, Y., Subramanya, R., & Faloutsos, C.. MindReader - Querying Databases Through Multiple Eamples. VLDB D. MOTTIN, E. MÜLLER 127
8 Let s go back to SLQ Query: Prof., ~70 yrs? Candidate Match: ( ) Geoffrey Hinton (Professor, 1947) UT Google University of Toronto Google DNNresearch Features Node matching features: u 5, 5 = ] (5, 5 ) 0 Edge matching features: ~ d, d = Å É É (d, d ) É Matching Score y ep( q u v u 5, 5 + ~ v ~ (d, d )) D. MOTTIN, E. MÜLLER 128
9 Let s go back to SLQ Query: Prof., ~70 yrs? Candidate Match: ( ) Geoffrey Hinton (Professor, 1947) UT Google University of Toronto Google DNNresearch Features What if some answer is irrelevant? Node matching features: u 5, 5 = ] (5, 5 ) Edge matching features: ~ d, d = Å É É (d, d ) É Matching Score y ep( q u v u 5, 5 + ~ v ~ (d, d )) 0 D. MOTTIN, E. MÜLLER 129
10 Query-specific Tuning Find query-specific Ø that better aligned with the query using user feedback Ø = 1 ÿ = M t < <, Ø ÿ = M t < <, Ø M p M ä + f(ø, Ø ) User Feedback Regularization 0,1 D. MOTTIN, E. MÜLLER 130
11 Type Inference Infer the implicit type of each query node The types of the positive entities constitute a composite type for each query node Query Positive Feedback Candidate Nodes D. MOTTIN, E. MÜLLER 131
12 Multi-armed bandits Maimize the reward by successively playing gamble machines (the arms of the bandits) Invented in early 1950s by Robbins for decision making under uncertainty when the environment is unknown The lotteries are unknown ahead of time Credits: Pietro Lovato Reward + Reward - Reward Ä D. MOTTIN, E. MÜLLER 132
13 Multi-armed bandits Reward = random variable 0,/ ; 1 e fi, î 1 e = inde of the gambling machine î = number of plays fl 0 = epected reward of machine e. A policy, or allocation strategy is an algorithm that chooses the net machine to play based on the sequence of past plays and obtained rewards. D. MOTTIN, E. MÜLLER 133
14 A greedy algorithm - Naïve Choose the machine with current best epected reward Eploitation vs eploration dilemma: Should you eploit the information you ve learned or eplore new options in the hope of greater payoff? In the greedy case, the balance is completely towards eploitation D. MOTTIN, E. MÜLLER 134
15 Quality measure - Regret Total epected regret (after T plays): f = fl o fl É [ñ 0, ] 0 + fl : highest epected reward [ñ 0, ]: epected number of times machine i is played An algorithm solve the multi-armed problem if it matches the lower bound f = W(log ) [Think about binary search] D. MOTTIN, E. MÜLLER 135
16 Upper confidence bound (UCB) algorithm 1. Pull at each time the arm with the maimum probability of being the best / á 1 o î É,Á É Á log î É 2. Repeat until the budget (number of steps T) is depleted î É : number of times the arm j has been pulled Balance eploration and eploitation: The uncertainty diminishes asthe time passes D. MOTTIN, E. MÜLLER 136
17 Gaussian processes [Bishop et al., 2006] Model reward as a Gaussian Process A Gaussian Process (GP) is an infinite set of variables, any subset of this is Gaussian y Ë Σ, fl = 2 Σ + - ep( 1 2 Ë fl Σ ä+ (Ë fl)) Gaussian prior Specified only by mean and covariance / Given observations œ, È 0 + over an unknown function f drawn from a Gaussian prior, the posterior is Gaussian y Ë Í Î L y(ë, Ï, Í) D. MOTTIN, E. MÜLLER 137
18 Putting all together Active search Main idea: the system query the user to understand her preferences Get item System Ask user preference Learn unknown preferences and minimize the number of questions to the user Vanchinathan, H. P., et al. Discovering Valuable items from Massive Data. KDD D. MOTTIN, E. MÜLLER 138
19 Learning unknown preferences Problem: Find a set S that maimize the user preference within a budget (e.g., number of interactions) S (intended user set) arg ma o ÌKd4(5) User preferences q û subject to IcFG B i;l dg Cost for the set S Each item has a cost Vanchinathan, H. P., et al. Discovering Valuable items from Massive Data. KDD D. MOTTIN, E. MÜLLER 139
20 GP-Select Idea: Model the user preferences as a Gaussian Process Learn posterior Trades off eploration eploitation Ask user feedback Eploration: select items with high-variance Eploitation: select items with high-value Vanchinathan, H. P., et al. Discovering Valuable items from Massive Data. KDD D. MOTTIN, E. MÜLLER 140
21 GP-Select bound Similar to multi-armed bandits Trades off eploration-eploitation Ψ : Maimum information gain f = 4 œ ma Ñ + 4 œ Ñ Ψ î UCB- strategy. Halves the search space and the uncertainty Srinivas et al. Gaussian process optimization in the bandit setting: No regret and eperimental design. ICML 2009 D. MOTTIN, E. MÜLLER 141
22 Active learning on graphs which prior? Idea: Use the graph structure to infer the node classes Use graph Laplacian as prior Ù = (, A is the adjacency matri Laplacian: higher probability of having the same class if two nodes are connected Ma et al. Active Search and Bandits on Graphs using Sigma-Optimality. UAI D. MOTTIN, E. MÜLLER 142
23 Eplore-by-Eample: AIDE Relevance Feedback Relevant Samples Irrelevant Samples Data Classification User Model Samples User Model Query Formulation Sampling queries Space Eploration Data Etraction Query Dimitriadou et al. Eplore-by-eample - an automatic query steering framework for interactive data eploration. SIGMOD D. MOTTIN, E. MÜLLER 143
24 The AIDE algorithm 1. Divide the space into d-dimensional cubes 2. Find the sample points in the cubes (medoids) 3. Train the classifier 4. Refine the training sampling from neighbors of misclassified points 5. Boundary refinement Dimitriadou et al. Eplore-by-eample - an automatic query steering framework for interactive data eploration. SIGMOD D. MOTTIN, E. MÜLLER 144
25 Classification & Query Formulation Sample Red Green Relevant Object A Yes Object B No Object X Yes red red>14.82 red<=14.82 Irrelevant red red>=13.55 red<13.55 green Irrelevant green>13.74 green<=13.74 Decision Tree Classifier Irrelevant Relevant SELECT * FROM galay WHERE red<= AND red>= 13.5 AND green<=13.74 Dimitriadou et al. Eplore-by-eample - an automatic query steering framework for interactive data eploration. SIGMOD D. MOTTIN, E. MÜLLER 145
26 Misclassified Sample Eploitation Red wavelength Sampling Areas Green Wavelength Dimitriadou et al. Eplore-by-eample - an automatic query steering framework for interactive data eploration. SIGMOD D. MOTTIN, E. MÜLLER 146
27 Clustering-based Sampling Red wavelength Clusters- Sampling Areas Idea: Use a k-medoid approach to find sampling areas Green Wavelength Dimitriadou et al. Eplore-by-eample - an automatic query steering framework for interactive data eploration. SIGMOD D. MOTTIN, E. MÜLLER 147
An introduction to multi-armed bandits
An introduction to multi-armed bandits Henry WJ Reeve (Manchester) (henry.reeve@manchester.ac.uk) A joint work with Joe Mellor (Edinburgh) & Professor Gavin Brown (Manchester) Plan 1. An introduction to
More informationClustering. Supervised vs. Unsupervised Learning
Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now
More informationRecognition Tools: Support Vector Machines
CS 2770: Computer Vision Recognition Tools: Support Vector Machines Prof. Adriana Kovashka University of Pittsburgh January 12, 2017 Announcement TA office hours: Tuesday 4pm-6pm Wednesday 10am-12pm Matlab
More informationMachine Learning: k-nearest Neighbors. Lecture 08. Razvan C. Bunescu School of Electrical Engineering and Computer Science
Machine Learning: k-nearest Neighbors Lecture 08 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Nonparametric Methods: k-nearest Neighbors Input: A training dataset
More informationGraph Exploration: Taking the User into the Loop
Graph Exploration: Taking the User into the Loop Davide Mottin, Anja Jentzsch, Emmanuel Müller Hasso Plattner Institute, Potsdam, Germany 2016/10/24 CIKM2016, Indianapolis, US Where we are Background (5
More information6.867 Machine learning
6.867 Machine learning Final eam December 3, 24 Your name and MIT ID: J. D. (Optional) The grade ou would give to ourself + a brief justification. A... wh not? Problem 5 4.5 4 3.5 3 2.5 2.5 + () + (2)
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More information6.867 Machine learning
6.867 Machine learning Final eam December 3, 24 Your name and MIT ID: J. D. (Optional) The grade ou would give to ourself + a brief justification. A... wh not? Cite as: Tommi Jaakkola, course materials
More informationColoring 3-Colorable Graphs
Coloring -Colorable Graphs Charles Jin April, 015 1 Introduction Graph coloring in general is an etremely easy-to-understand yet powerful tool. It has wide-ranging applications from register allocation
More informationData mining. Classification k-nn Classifier. Piotr Paszek. (Piotr Paszek) Data mining k-nn 1 / 20
Data mining Piotr Paszek Classification k-nn Classifier (Piotr Paszek) Data mining k-nn 1 / 20 Plan of the lecture 1 Lazy Learner 2 k-nearest Neighbor Classifier 1 Distance (metric) 2 How to Determine
More informationCLASSIFICATION OF C4.5 AND CART ALGORITHMS USING DECISION TREE METHOD
CLASSIFICATION OF C4.5 AND CART ALGORITHMS USING DECISION TREE METHOD Khin Lay Myint 1, Aye Aye Cho 2, Aye Mon Win 3 1 Lecturer, Faculty of Information Science, University of Computer Studies, Hinthada,
More informationDS504/CS586: Big Data Analytics Big Data Clustering Prof. Yanhua Li
Welcome to DS504/CS586: Big Data Analytics Big Data Clustering Prof. Yanhua Li Time: 6:00pm 8:50pm Thu Location: AK 232 Fall 2016 High Dimensional Data v Given a cloud of data points we want to understand
More informationCS246: Mining Massive Datasets Jure Leskovec, Stanford University
CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu Web advertising We discussed how to match advertisers to queries in real-time But we did not discuss how to estimate
More informationFall 2017 ECEN Special Topics in Data Mining and Analysis
Fall 2017 ECEN 689-600 Special Topics in Data Mining and Analysis Nick Duffield Department of Electrical & Computer Engineering Teas A&M University Organization Organization Instructor: Nick Duffield,
More informationThis file contains an excerpt from the character code tables and list of character names for The Unicode Standard, Version 3.0.
Range: This file contains an excerpt from the character code tables and list of character names for The Unicode Standard, Version.. isclaimer The shapes of the reference glyphs used in these code charts
More informationThorsten Joachims Then: Universität Dortmund, Germany Now: Cornell University, USA
Retrospective ICML99 Transductive Inference for Text Classification using Support Vector Machines Thorsten Joachims Then: Universität Dortmund, Germany Now: Cornell University, USA Outline The paper in
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationSlides for Data Mining by I. H. Witten and E. Frank
Slides for Data Mining by I. H. Witten and E. Frank 7 Engineering the input and output Attribute selection Scheme-independent, scheme-specific Attribute discretization Unsupervised, supervised, error-
More informationCS246: Mining Massive Datasets Jure Leskovec, Stanford University
CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu [Kumar et al. 99] 2/13/2013 Jure Leskovec, Stanford CS246: Mining Massive Datasets, http://cs246.stanford.edu
More information3. Data Preprocessing. 3.1 Introduction
3. Data Preprocessing Contents of this Chapter 3.1 Introduction 3.2 Data cleaning 3.3 Data integration 3.4 Data transformation 3.5 Data reduction SFU, CMPT 740, 03-3, Martin Ester 84 3.1 Introduction Motivation
More information2. Data Preprocessing
2. Data Preprocessing Contents of this Chapter 2.1 Introduction 2.2 Data cleaning 2.3 Data integration 2.4 Data transformation 2.5 Data reduction Reference: [Han and Kamber 2006, Chapter 2] SFU, CMPT 459
More informationUnsupervised Learning
Unsupervised Learning Pierre Gaillard ENS Paris September 28, 2018 1 Supervised vs unsupervised learning Two main categories of machine learning algorithms: - Supervised learning: predict output Y from
More informationContent-based image and video analysis. Machine learning
Content-based image and video analysis Machine learning for multimedia retrieval 04.05.2009 What is machine learning? Some problems are very hard to solve by writing a computer program by hand Almost all
More informationMixture models and clustering
1 Lecture topics: Miture models and clustering, k-means Distance and clustering Miture models and clustering We have so far used miture models as fleible ays of constructing probability models for prediction
More informationECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 22, 2016 Course Information Website: http://www.stat.ucdavis.edu/~chohsieh/teaching/ ECS289G_Fall2016/main.html My office: Mathematical Sciences
More informationMining Data Streams. From Data-Streams Management System Queries to Knowledge Discovery from continuous and fast-evolving Data Records.
DATA STREAMS MINING Mining Data Streams From Data-Streams Management System Queries to Knowledge Discovery from continuous and fast-evolving Data Records. Hammad Haleem Xavier Plantaz APPLICATIONS Sensors
More informationDECISION-TREE-BASED MULTICLASS SUPPORT VECTOR MACHINES. Fumitake Takahashi, Shigeo Abe
DECISION-TREE-BASED MULTICLASS SUPPORT VECTOR MACHINES Fumitake Takahashi, Shigeo Abe Graduate School of Science and Technology, Kobe University, Kobe, Japan (E-mail: abe@eedept.kobe-u.ac.jp) ABSTRACT
More informationGaussian Processes for Robotics. McGill COMP 765 Oct 24 th, 2017
Gaussian Processes for Robotics McGill COMP 765 Oct 24 th, 2017 A robot must learn Modeling the environment is sometimes an end goal: Space exploration Disaster recovery Environmental monitoring Other
More informationUnlabeled Data Classification by Support Vector Machines
Unlabeled Data Classification by Support Vector Machines Glenn Fung & Olvi L. Mangasarian University of Wisconsin Madison www.cs.wisc.edu/ olvi www.cs.wisc.edu/ gfung The General Problem Given: Points
More informationUnsupervised Learning
Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised
More informationDiscussion: Clustering Random Curves Under Spatial Dependence
Discussion: Clustering Random Curves Under Spatial Dependence Gareth M. James, Wenguang Sun and Xinghao Qiao Abstract We discuss the advantages and disadvantages of a functional approach to clustering
More informationSolution Guide II-D. Classification. HALCON Progress
Solution Guide II-D Classification HALCON 17.12 Progress How to use classification, Version 17.12 All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted
More informationMachine Learning Techniques for Data Mining
Machine Learning Techniques for Data Mining Eibe Frank University of Waikato New Zealand 10/25/2000 1 PART VII Moving on: Engineering the input and output 10/25/2000 2 Applying a learner is not all Already
More informationPSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D.
PSU Student Research Symposium 2017 Bayesian Optimization for Refining Object Proposals, with an Application to Pedestrian Detection Anthony D. Rhodes 5/10/17 What is Machine Learning? Machine learning
More informationCS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample
CS 1675 Introduction to Machine Learning Lecture 18 Clustering Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square Clustering Groups together similar instances in the data sample Basic clustering problem:
More informationInternational Journal of Scientific Research & Engineering Trends Volume 4, Issue 6, Nov-Dec-2018, ISSN (Online): X
Analysis about Classification Techniques on Categorical Data in Data Mining Assistant Professor P. Meena Department of Computer Science Adhiyaman Arts and Science College for Women Uthangarai, Krishnagiri,
More informationCS 2750 Machine Learning. Lecture 19. Clustering. CS 2750 Machine Learning. Clustering. Groups together similar instances in the data sample
Lecture 9 Clustering Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square Clustering Groups together similar instances in the data sample Basic clustering problem: distribute data into k different groups
More informationCS6375: Machine Learning Gautam Kunapuli. Mid-Term Review
Gautam Kunapuli Machine Learning Data is identically and independently distributed Goal is to learn a function that maps to Data is generated using an unknown function Learn a hypothesis that minimizes
More informationCS6716 Pattern Recognition
CS6716 Pattern Recognition Prototype Methods Aaron Bobick School of Interactive Computing Administrivia Problem 2b was extended to March 25. Done? PS3 will be out this real soon (tonight) due April 10.
More informationCME323 Report: Distributed Multi-Armed Bandits
CME323 Report: Distributed Multi-Armed Bandits Milind Rao milind@stanford.edu 1 Introduction Consider the multi-armed bandit (MAB) problem. In this sequential optimization problem, a player gets to pull
More informationGenerative and discriminative classification techniques
Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14
More information6.867 Machine Learning
6.867 Machine Learning Problem set - solutions Thursday, October What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove. Do not
More informationSupervised Learning for Image Segmentation
Supervised Learning for Image Segmentation Raphael Meier 06.10.2016 Raphael Meier MIA 2016 06.10.2016 1 / 52 References A. Ng, Machine Learning lecture, Stanford University. A. Criminisi, J. Shotton, E.
More informationTopics in Machine Learning-EE 5359 Model Assessment and Selection
Topics in Machine Learning-EE 5359 Model Assessment and Selection Ioannis D. Schizas Electrical Engineering Department University of Texas at Arlington 1 Training and Generalization Training stage: Utilizing
More informationPV211: Introduction to Information Retrieval
PV211: Introduction to Information Retrieval http://www.fi.muni.cz/~sojka/pv211 IIR 15-1: Support Vector Machines Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk University,
More informationKDD 10 Tutorial: Recommender Problems for Web Applications. Deepak Agarwal and Bee-Chung Chen Yahoo! Research
KDD 10 Tutorial: Recommender Problems for Web Applications Deepak Agarwal and Bee-Chung Chen Yahoo! Research Agenda Focus: Recommender problems for dynamic, time-sensitive applications Content Optimization
More informationClustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford
Department of Engineering Science University of Oxford January 27, 2017 Many datasets consist of multiple heterogeneous subsets. Cluster analysis: Given an unlabelled data, want algorithms that automatically
More informationTight Clustering: a method for extracting stable and tight patterns in expression profiles
Statistical issues in microarra analsis Tight Clustering: a method for etracting stable and tight patterns in epression profiles Eperimental design Image analsis Normalization George C. Tseng Dept. of
More informationConstruction and Querying of Large-scale Knowledge Bases. Part II: Schema-agnostic Knowledge Base Querying
Construction and Querying of Large-scale Knowledge Bases Part II: Schema-agnostic Knowledge Base Querying Transformation in Information Search Desktop search Mobile search Which hotel has a roller coaster
More informationData Preprocessing. Why Data Preprocessing? MIT-652 Data Mining Applications. Chapter 3: Data Preprocessing. Multi-Dimensional Measure of Data Quality
Why Data Preprocessing? Data in the real world is dirty incomplete: lacking attribute values, lacking certain attributes of interest, or containing only aggregate data e.g., occupation = noisy: containing
More informationAnnouncements, schedule. Lecture 8: Fitting. Weighted graph representation. Outline. Segmentation by Graph Cuts. Images as graphs
Announcements, schedule Lecture 8: Fitting Tuesday, Sept 25 Grad student etensions Due of term Data sets, suggestions Reminder: Midterm Tuesday 10/9 Problem set 2 out Thursday, due 10/11 Outline Review
More informationSemi-supervised Learning
Semi-supervised Learning Piyush Rai CS5350/6350: Machine Learning November 8, 2011 Semi-supervised Learning Supervised Learning models require labeled data Learning a reliable model usually requires plenty
More informationBased on Raymond J. Mooney s slides
Instance Based Learning Based on Raymond J. Mooney s slides University of Texas at Austin 1 Example 2 Instance-Based Learning Unlike other learning algorithms, does not involve construction of an explicit
More informationReal-world bandit applications: Bridging the gap between theory and practice
Real-world bandit applications: Bridging the gap between theory and practice Audrey Durand EWRL 2018 The bandit setting (Thompson, 1933; Robbins, 1952; Lai and Robbins, 1985) Action Agent Context Environment
More informationMachine Learning Lecture 3
Machine Learning Lecture 3 Probability Density Estimation II 19.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Announcements Exam dates We re in the process
More informationMachine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves
Machine Learning A 708.064 11W 1sst KU Exercises Problems marked with * are optional. 1 Conditional Independence I [2 P] a) [1 P] Give an example for a probability distribution P (A, B, C) that disproves
More informationPredicting Popular Xbox games based on Search Queries of Users
1 Predicting Popular Xbox games based on Search Queries of Users Chinmoy Mandayam and Saahil Shenoy I. INTRODUCTION This project is based on a completed Kaggle competition. Our goal is to predict which
More informationUVA CS 6316/4501 Fall 2016 Machine Learning. Lecture 15: K-nearest-neighbor Classifier / Bias-Variance Tradeoff. Dr. Yanjun Qi. University of Virginia
UVA CS 6316/4501 Fall 2016 Machine Learning Lecture 15: K-nearest-neighbor Classifier / Bias-Variance Tradeoff Dr. Yanjun Qi University of Virginia Department of Computer Science 11/9/16 1 Rough Plan HW5
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationSolution Guide II-D. Classification. Building Vision for Business. MVTec Software GmbH
Solution Guide II-D Classification MVTec Software GmbH Building Vision for Business How to use classification, Version 10.0.4 All rights reserved. No part of this publication may be reproduced, stored
More informationCS178: Machine Learning and Data Mining. Complexity & Nearest Neighbor Methods
+ CS78: Machine Learning and Data Mining Complexity & Nearest Neighbor Methods Prof. Erik Sudderth Some materials courtesy Alex Ihler & Sameer Singh Machine Learning Complexity and Overfitting Nearest
More informationThe exam is closed book, closed notes except your one-page (two-sided) cheat sheet.
CS 189 Spring 2015 Introduction to Machine Learning Final You have 2 hours 50 minutes for the exam. The exam is closed book, closed notes except your one-page (two-sided) cheat sheet. No calculators or
More informationClassification. 1 o Semestre 2007/2008
Classification Departamento de Engenharia Informática Instituto Superior Técnico 1 o Semestre 2007/2008 Slides baseados nos slides oficiais do livro Mining the Web c Soumen Chakrabarti. Outline 1 2 3 Single-Class
More informationUVA CS 4501: Machine Learning. Lecture 10: K-nearest-neighbor Classifier / Bias-Variance Tradeoff. Dr. Yanjun Qi. University of Virginia
UVA CS 4501: Machine Learning Lecture 10: K-nearest-neighbor Classifier / Bias-Variance Tradeoff Dr. Yanjun Qi University of Virginia Department of Computer Science 1 Where are we? è Five major secfons
More informationthe power of machine vision Solution Guide II-D Classification
the power of machine vision Solution Guide II-D Classification How to use classification, Version 12.0.2 All rights reserved. No part of this publication may be reproduced, stored in a retrieval system,
More informationMachine Learning. Supervised Learning. Manfred Huber
Machine Learning Supervised Learning Manfred Huber 2015 1 Supervised Learning Supervised learning is learning where the training data contains the target output of the learning system. Training data D
More informationThe K-modes and Laplacian K-modes algorithms for clustering
The K-modes and Laplacian K-modes algorithms for clustering Miguel Á. Carreira-Perpiñán Electrical Engineering and Computer Science University of California, Merced http://faculty.ucmerced.edu/mcarreira-perpinan
More information08 An Introduction to Dense Continuous Robotic Mapping
NAVARCH/EECS 568, ROB 530 - Winter 2018 08 An Introduction to Dense Continuous Robotic Mapping Maani Ghaffari March 14, 2018 Previously: Occupancy Grid Maps Pose SLAM graph and its associated dense occupancy
More informationContents. Preface to the Second Edition
Preface to the Second Edition v 1 Introduction 1 1.1 What Is Data Mining?....................... 4 1.2 Motivating Challenges....................... 5 1.3 The Origins of Data Mining....................
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationIntroduction to Machine Learning. Xiaojin Zhu
Introduction to Machine Learning Xiaojin Zhu jerryzhu@cs.wisc.edu Read Chapter 1 of this book: Xiaojin Zhu and Andrew B. Goldberg. Introduction to Semi- Supervised Learning. http://www.morganclaypool.com/doi/abs/10.2200/s00196ed1v01y200906aim006
More informationUsing Reinforcement Learning to Optimize Storage Decisions Ravi Khadiwala Cleversafe
Using Reinforcement Learning to Optimize Storage Decisions Ravi Khadiwala Cleversafe Topics What is Reinforcement Learning? Exploration vs. Exploitation The Multi-armed Bandit Optimizing read locations
More informationOutline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation
Advanced Digital Image Processing and Others Xiaojun Qi -- REU Site Program in CVIP (7 Summer) Outline Segmentation Strategies and Data Structures Algorithms Overview K-Means Algorithm Hidden Markov Model
More informationBasis Functions. Volker Tresp Summer 2017
Basis Functions Volker Tresp Summer 2017 1 Nonlinear Mappings and Nonlinear Classifiers Regression: Linearity is often a good assumption when many inputs influence the output Some natural laws are (approximately)
More informationHow and what do we see? Segmentation and Grouping. Fundamental Problems. Polyhedral objects. Reducing the combinatorics of pose estimation
Segmentation and Grouping Fundamental Problems ' Focus of attention, or grouping ' What subsets of piels do we consider as possible objects? ' All connected subsets? ' Representation ' How do we model
More informationSolution Guide II-D. Classification. Building Vision for Business. MVTec Software GmbH
Solution Guide II-D Classification MVTec Software GmbH Building Vision for Business Overview In a broad range of applications classification is suitable to find specific objects or detect defects in images.
More informationLearning Statistical Models From Relational Data
Slides taken from the presentation (subset only) Learning Statistical Models From Relational Data Lise Getoor University of Maryland, College Park Includes work done by: Nir Friedman, Hebrew U. Daphne
More informationComputing Gaussian Mixture Models with EM using Equivalence Constraints
Computing Gaussian Mixture Models with EM using Equivalence Constraints Noam Shental, Aharon Bar-Hillel, Tomer Hertz and Daphna Weinshall email: tomboy,fenoam,aharonbh,daphna@cs.huji.ac.il School of Computer
More informationBatch Bayesian Optimization via Local Penalization
Batch Bayesian Optimization via Local Penalization Javier González j.h.gonzalez@sheffield.ac.uk Philipp Henning MPI for Intelligent Systems Tubingen, Germany phennig@tuebingen.mpg.de Zhenwen Dai z.dai@sheffield.ac.uk
More informationUnsupervised Learning. Supervised learning vs. unsupervised learning. What is Cluster Analysis? Applications of Cluster Analysis
7 Supervised learning vs unsupervised learning Unsupervised Learning Supervised learning: discover patterns in the data that relate data attributes with a target (class) attribute These patterns are then
More informationAn Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm
Proceedings of the National Conference on Recent Trends in Mathematical Computing NCRTMC 13 427 An Effective Performance of Feature Selection with Classification of Data Mining Using SVM Algorithm A.Veeraswamy
More information1 Case study of SVM (Rob)
DRAFT a final version will be posted shortly COS 424: Interacting with Data Lecturer: Rob Schapire and David Blei Lecture # 8 Scribe: Indraneel Mukherjee March 1, 2007 In the previous lecture we saw how
More informationClassification: Linear Discriminant Functions
Classification: Linear Discriminant Functions CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Discriminant functions Linear Discriminant functions
More informationAdaptive Dropout Training for SVMs
Department of Computer Science and Technology Adaptive Dropout Training for SVMs Jun Zhu Joint with Ning Chen, Jingwei Zhuo, Jianfei Chen, Bo Zhang Tsinghua University ShanghaiTech Symposium on Data Science,
More informationDS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University
DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University January 24 2019 Logistics HW 1 is due on Friday 01/25 Project proposal: due Feb 21 1 page description
More informationInformation Filtering for arxiv.org
Information Filtering for arxiv.org Bandits, Exploration vs. Exploitation, & the Cold Start Problem Peter Frazier School of Operations Research & Information Engineering Cornell University with Xiaoting
More informationMachine Learning : Clustering, Self-Organizing Maps
Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps Clustering The task: partition a set of objects into meaningful subsets (clusters). The
More informationMonte Carlo Tree Search PAH 2015
Monte Carlo Tree Search PAH 2015 MCTS animation and RAVE slides by Michèle Sebag and Romaric Gaudel Markov Decision Processes (MDPs) main formal model Π = S, A, D, T, R states finite set of states of the
More informationMachine Learning Lecture 3
Many slides adapted from B. Schiele Machine Learning Lecture 3 Probability Density Estimation II 26.04.2016 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course
More informationLocality Preserving Projections (LPP) Abstract
Locality Preserving Projections (LPP) Xiaofei He Partha Niyogi Computer Science Department Computer Science Department The University of Chicago The University of Chicago Chicago, IL 60615 Chicago, IL
More informationSupport Vector Machines + Classification for IR
Support Vector Machines + Classification for IR Pierre Lison University of Oslo, Dep. of Informatics INF3800: Søketeknologi April 30, 2014 Outline of the lecture Recap of last week Support Vector Machines
More informationA General Greedy Approximation Algorithm with Applications
A General Greedy Approximation Algorithm with Applications Tong Zhang IBM T.J. Watson Research Center Yorktown Heights, NY 10598 tzhang@watson.ibm.com Abstract Greedy approximation algorithms have been
More informationJeff Howbert Introduction to Machine Learning Winter
Collaborative Filtering Nearest es Neighbor Approach Jeff Howbert Introduction to Machine Learning Winter 2012 1 Bad news Netflix Prize data no longer available to public. Just after contest t ended d
More informationAnalysis: TextonBoost and Semantic Texton Forests. Daniel Munoz Februrary 9, 2009
Analysis: TextonBoost and Semantic Texton Forests Daniel Munoz 16-721 Februrary 9, 2009 Papers [shotton-eccv-06] J. Shotton, J. Winn, C. Rother, A. Criminisi, TextonBoost: Joint Appearance, Shape and Context
More information4 Linear Programming (LP) E. Amaldi -- Foundations of Operations Research -- Politecnico di Milano 1
4 Linear Programming (LP) E. Amaldi -- Foundations of Operations Research -- Politecnico di Milano 1 Definition: A Linear Programming (LP) problem is an optimization problem: where min f () s.t. X n the
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationCS4445 Data Mining and Knowledge Discovery in Databases. A Term 2008 Exam 2 October 14, 2008
CS4445 Data Mining and Knowledge Discovery in Databases. A Term 2008 Exam 2 October 14, 2008 Prof. Carolina Ruiz Department of Computer Science Worcester Polytechnic Institute NAME: Prof. Ruiz Problem
More informationFeature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule.
CS 188: Artificial Intelligence Fall 2008 Lecture 24: Perceptrons II 11/24/2008 Dan Klein UC Berkeley Feature Extractors A feature extractor maps inputs to feature vectors Dear Sir. First, I must solicit
More informationChapter 5: Outlier Detection
Ludwig-Maximilians-Universität München Institut für Informatik Lehr- und Forschungseinheit für Datenbanksysteme Knowledge Discovery in Databases SS 2016 Chapter 5: Outlier Detection Lecture: Prof. Dr.
More information