Computer Vision Group Prof. Daniel Cremers. 6. Boosting
|
|
- Marilyn Green
- 6 years ago
- Views:
Transcription
1 Prof. Daniel Cremers 6. Boosting
2 Repetition: Regression We start with a set of basis functions (x) =( 0 (x), 1(x),..., M 1(x)) x 2 í d The goal is to fit a model into the data y(x, w) =w T (x) To do this, we need to find an error function, e.g.: E(w) = 1 2 NX (w T (x i ) t i ) 2 i=1 To find the optimal parameters, we derived E with respect to w and set the derivative to zero. 2
3 Some Questions 1.Can we do the same for classification? As a special case we consider two classes: t i 2 { 1, 1} 8i =1,...,N 2.Can we use a different (better?) error function? 3.Can we learn the basis functions together with the model parameters? 4.Can we do the learning sequentially, i.e. one basis function after another? Answer to all questions: Yes, using Boosting! 3
4 The Loss Function Definition: a real-valued function L(t, y(x)), where t is a target value and y is a model, is called a loss function. Examples: 01-loss: L 01 (t, y(x)) = 0 if t = y(x) 1 else squared error loss: L sqe (t, y(x)) = (t y(x)) 2 exponential loss: L exp (t, y(x)) = exp( ty(x)) 4
5 Loss Functions 01-loss is not differentiable squared error loss has only one optimum 5
6 Sequential Fitting of Basis Functions Idea: We start with a basis function : 0(x) y 0 (x,w 0 )=w 0 0 (x) w 0 =1 Then, at iteration m, we add a new basis function m(x) to the model: y m (x,w 0,...,w m )=y m 1 (x,w 0,...,w m 1 )+w m m (x) Two questions need to be answered: 1.How do we find a good new basis function? 2.How can we determine a good value for w m? Idea: Minimize the exponential loss function 6
7 Minimizing the Exponential Loss Aim: find w m and m so that (w m, m) = arg min w, NX L(t i,y m 1 (x i )+w (x i )) i=1 where L(t, y) =exp( ty) 7
8 Minimizing the Exponential Loss Aim: find w m and m so that (w m, m) = arg min w, NX i=1 L(t i,y m 1 (x i )+w (x i )) where L(t, y) =exp( ty) Solution: m = arg min NX v i,m I(t i 6= (x i )) i=1 8
9 Minimizing the Exponential Loss Aim: find w m and m so that (w m, m) = arg min w, NX i=1 L(t i,y m 1 (x i )+w (x i )) where L(t, y) =exp( ty) Solution: w m = 1 2 log 1 m = arg min err m err m NX i=1 v i,m I(t i 6= (x i )) 9
10 Minimizing the Exponential Loss Aim: find w m and m so that (w m, m) = arg min w, NX i=1 L(t i,y m 1 (x i )+w (x i )) where L(t, y) =exp( ty) NX Solution: m = arg min i=1 v i,m I(t i 6= (x i )) w m = 1 2 log 1 err m err m v i,m+1 = v i,m exp(2w m I(t i 6= m (x i )) 10
11 Minimizing the Exponential Loss Aim: find w m and m so that (w m, m) = arg min w, NX i=1 L(t i,y m 1 (x i )+w (x i )) where L(t, y) =exp( ty) Condition: Must have training error less than 0.5! NX Solution: m = arg min i=1 v i,m I(t i 6= (x i )) w m = 1 2 log 1 err m err m v i,m+1 = v i,m exp(2w m I(t i 6= m (x i )) 11
12 1.For : The AdaBoost Algorithm i =1,...,N v i 1/N 2.For m =1,...,M Fit a classifier ( basis function ) Compute NX i=1 err m = v i I(t i 6= m (x i )) m P N i=1 v ii(t i 6= m (x i )) P N i=1 v i that minimizes and m := 2w m m = log 1 err m err m Update the weights: v i v i exp( m I(t i 6= m (x i ))) 3.Use the resulting classifier: y(x) = sgn MX m=1 12 m m (x)
13 The Basis Functions Can be any classifier that can deal with weighted data Most importantly: if these base classifiers provide a training error that is at most as bad as a random classifier would give (i.e. it is a weak classifier), then AdaBoost can return an arbitrarily small training error (i.e. AdaBoost is a strong classifier) Many possibilities for weak classifiers exist, e.g.: Decision stumps Decision trees 13
14 Decision Stumps Decision Stumps are a kind of very simple weak classifiers. Goal: Find an axis-aligned hyperplane that minimizes the class. error This can be done for each feature (i.e. for each dimension in feature space) It can be shown that the classif. error is always better than 0.5 (random guessing) Idea: apply many weak classifiers, where each is trained on the misclassified examples of the previous. 14
15 Classification Example 15
16 Classification Example 16
17 Classification Example 17
18 Classification Example 18
19 Classification Example 19
20 Classification Example 20
21 Decision Trees A more general version of decision stumps are decision trees: 10 X 1 t 1 9 X 2 t 2 R 1 X 1 t 3 X 2 t 4 R 2 R R 4 R 5 At every node, a decision is made Can be used for classification and for regression (Classification And Regression Trees CART) 21
22 Decision Trees for Classification blue color red 4,0 shape ellipse other other size < 10 yes no 1,1 0,2 4,0 0,5 Stores the distribution over class labels in each leaf (number of positives and negatives) With these, we can class label probabilities, e.g. p(y =1 x) =1/2 if we have a red ellipse 22
23 Growing a Decision Tree Finding the optimal partition of the data is an NP-complete problem! Instead: use a greedy strategy: function fittree(node, D, depth): 1. node.prediction = class label distribution 2. (j,t, D L, D R )=split(d) 3. if not worth splitting then return node 4. node.test x j <t D L 5. node.left = fittree(node,, depth +1) 6. node.right = fittree(node, D R, depth +1) 23
24 Growing a Decision Tree The Split-function finds an optimal feature and an optimal value for that feature For classification, it finds a split that minimizes some cost function, e.g. misclassification A decision stump is a decision tree with depth 1 Stopping criteria for growing the tree are: reduction of cost too small? maximum depth reached? is the distribution in the sub-trees homogenous? is the number of samples in the sub-trees too small? 24
25 Tree Pruning versicolor setosa SL < 5.45 W < 2.8 SW >= 2.8 SL < 6.15 SL >= 6.15 SW < 3.45 SW >= 3.45 SL < 7.05 SL >= 7.05 SL < 5.75 SL >= 5.75 SW < 2.4 SW >= 2.4 setosa SW < 3.1 SW >= 3.1 SL < 6.95 SL >= 6.95 versicolor versicolor SW < 2.95 SW >= 2.95 SW < 3.15 SW >= 3.15 versicolor versicolor versicolor virginica SL >= 5.45 virginica SL < 6.45 SL >= 6.45 SW < 2.65 SW >= 2.65 virginica versicolor SW < 2.85 SW >= 2.85 SW < 2.9 SW >= 2.9 versicolor virginica virginica versicolor SL < 6.55 SL >= 6.55 SW < 2.95 SW SL >= < SL >= 6.65 versicolor virginica virginica Cost (misclassification error) Cross validation Training set Min + 1 std. err. Best choice Number of terminal nodes If the tree grows too large, the algorithm overfits Simply stopping to grow can lead to situations where the tree is not expressive enough Idea: Build first full tree and then prune it Pruning can be done using cross-validation 25
26 Random Forests To reduce the variance of the classification estimate, we can train several trees on randomly sampled subsets of the data ( bagging ) However, this can result in correlated classifiers, limiting the reduction in variance Idea: chose data subset and variable (feature) subset randomly The resulting algorithm is known as Random Forests Random Forests have very good accuracy and are widely used, e.g. body pose recognition 26
27 Back to Boosting AdaBoost has been shown to perform very well, especially when using decision trees as weak classifiers However: the exponential loss weighs misclassified examples very high! 27
28 Using the Log-Loss The log-loss is defined as: L(t, y(x)) = log 2 (1 + exp( 2ty(x)) It penalizes misclassifications only linearly 28
29 The LogitBoost Algorithm 1.For : i =1,...,N v i 1/N i 1/2 2.For Compute the working response Compute the weights Find Update m =1,...,M m that minimizes NX i=1 3.Use the resulting classifier: z i = t i i i (1 i ) v i = i (1 i ) v i (z i (x i )) 2 y(x) y(x)+ 1 2 m (x) i y(x) = sgn 29 MX m=1 and m(x) exp( 2y(x i ))
30 Weighted Least-Squares Regression Instead of a weak classifier, LogitBoost uses weighted least-squares regression This is very similar to standard least-squares regression: E(w) = 1 2 NX i=1 v i (w T (x i ) t i ) 2 This results in a matrix ˆ = V 1/2 where The solution is V 1/2 = diag( p v 1,..., p v N ) w =(ˆ T ˆ) 1 ˆ T t 30
31 GentleBoost Numerically more stable than LogitBoost Tends to perform better than AdaBoost and LogitBoost 31
32 Generalization: Gradient Boost Initialize for m =1,...,M Compute the gradient residual Use the weak learner to compute that minimizes Update NX f 0 (x) = arg min L(t i, (x i, )) r im = i=1 i i ) f(x i )=f m 1 (x i ) NX (r im (x i ; m)) 2 i=1 f m (x) =f m 1 (x)+ (x, ) Return f(x) =f M (x) 32
33 Application of AdaBoost: Face Detection The biggest impact of AdaBoost was made in face detection Idea: extract features ( Haar-like features ) and train AdaBoost, use a cascade of classifiers Features can be computed very efficiently Weak classifiers can be decision stumps or decision trees As inference in AdaBoost is fast, the face detector can run in real-time! 33
34 Haar-like Features Defined as difference of rectangular integral area: The sum of the pixels which lie within the white rectangles are subtracted from the sum of pixels in the grey rectangles. One feature defined as: Feature type: A,B,C or D Feature position and size 34
35 The integral image Defined as : Integral on rectangle D can be computed in 4 access to I int : Very efficient way to compute features 35
36 Weak Classifiers Used A weak classifier has 3 attributes: A feature f j (type, size and position) A threshold θ j A comparison operator op j = < or > The resulting weak classifier is: x is a 24x24 pixels window in the image 36
37 Two First Classifiers Selected by AdaBoost A classifier with only this two features can be trained to recognise 100% of the faces, with 40% of false positives 37
38 The Inference Algorithm scale = 24x24 Do { For each position in the image { Try classifying the part of the image starting at this position, with the current scale, using the classifier selected by AdaBoost } Scale = Scale x 1.5 } until maximum scale 38
39 Another Improvement: the Cascade Basic idea: It is easy to detect that something is not a face Tune(boost) classifier to be very reliable at saying NO (i.e. very low false negative) Stop evaluating the cascade of classifier if one classifier says NO 39
40 Advantage of the Cascade Faster processing Quick elimination of useless windows Each individual classifier is trained to deal only with the example that the previous ones could not process Very specialised The deeper in the cascade, the more complex (the more features) in the classifiers. 40
41 Results (1) 41
42 Results (2) 42 PD Dr. Rudolph Triebel
43 Summary Boosting is a method to use a weak classifier and turn it into a strong one (arbitrarily small training error!) AdaBoost minimizes the exponential loss To be more robust against outliers, we can use LogitBoost or GentleBoost Weak learners can be decision stumps or decision trees Face detection can be solved with Boosting 43
Computer Vision Group Prof. Daniel Cremers. 8. Boosting and Bagging
Prof. Daniel Cremers 8. Boosting and Bagging Repetition: Regression We start with a set of basis functions (x) =( 0 (x), 1(x),..., M 1(x)) x 2 í d The goal is to fit a model into the data y(x, w) =w T
More information7. Boosting and Bagging Bagging
Group Prof. Daniel Cremers 7. Boosting and Bagging Bagging Bagging So far: Boosting as an ensemble learning method, i.e.: a combination of (weak) learners A different way to combine classifiers is known
More informationLearning to Detect Faces. A Large-Scale Application of Machine Learning
Learning to Detect Faces A Large-Scale Application of Machine Learning (This material is not in the text: for further information see the paper by P. Viola and M. Jones, International Journal of Computer
More informationEnsemble Methods, Decision Trees
CS 1675: Intro to Machine Learning Ensemble Methods, Decision Trees Prof. Adriana Kovashka University of Pittsburgh November 13, 2018 Plan for This Lecture Ensemble methods: introduction Boosting Algorithm
More informationData Mining Lecture 8: Decision Trees
Data Mining Lecture 8: Decision Trees Jo Houghton ECS Southampton March 8, 2019 1 / 30 Decision Trees - Introduction A decision tree is like a flow chart. E. g. I need to buy a new car Can I afford it?
More informationCSC411 Fall 2014 Machine Learning & Data Mining. Ensemble Methods. Slides by Rich Zemel
CSC411 Fall 2014 Machine Learning & Data Mining Ensemble Methods Slides by Rich Zemel Ensemble methods Typical application: classi.ication Ensemble of classi.iers is a set of classi.iers whose individual
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 12 Combining
More informationCS228: Project Report Boosted Decision Stumps for Object Recognition
CS228: Project Report Boosted Decision Stumps for Object Recognition Mark Woodward May 5, 2011 1 Introduction This project is in support of my primary research focus on human-robot interaction. In order
More informationMachine Learning for Signal Processing Detecting faces (& other objects) in images
Machine Learning for Signal Processing Detecting faces (& other objects) in images Class 8. 27 Sep 2016 11755/18979 1 Last Lecture: How to describe a face The typical face A typical face that captures
More informationRandom Forest A. Fornaser
Random Forest A. Fornaser alberto.fornaser@unitn.it Sources Lecture 15: decision trees, information theory and random forests, Dr. Richard E. Turner Trees and Random Forests, Adele Cutler, Utah State University
More informationThree Embedded Methods
Embedded Methods Review Wrappers Evaluation via a classifier, many search procedures possible. Slow. Often overfits. Filters Use statistics of the data. Fast but potentially naive. Embedded Methods A
More informationDecision Trees. This week: Next week: Algorithms for constructing DT. Pruning DT Ensemble methods. Random Forest. Intro to ML
Decision Trees This week: Algorithms for constructing DT Next week: Pruning DT Ensemble methods Random Forest 2 Decision Trees - Boolean x 1 0 1 +1 x 6 0 1 +1-1 3 Decision Trees - Continuous Decision stumps
More informationClassification: Decision Trees
Classification: Decision Trees IST557 Data Mining: Techniques and Applications Jessie Li, Penn State University 1 Decision Tree Example Will a pa)ent have high-risk based on the ini)al 24-hour observa)on?
More informationEnsemble Learning: An Introduction. Adapted from Slides by Tan, Steinbach, Kumar
Ensemble Learning: An Introduction Adapted from Slides by Tan, Steinbach, Kumar 1 General Idea D Original Training data Step 1: Create Multiple Data Sets... D 1 D 2 D t-1 D t Step 2: Build Multiple Classifiers
More informationAn introduction to random forests
An introduction to random forests Eric Debreuve / Team Morpheme Institutions: University Nice Sophia Antipolis / CNRS / Inria Labs: I3S / Inria CRI SA-M / ibv Outline Machine learning Decision tree Random
More informationCS 559: Machine Learning Fundamentals and Applications 10 th Set of Notes
1 CS 559: Machine Learning Fundamentals and Applications 10 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationAutomatic Initialization of the TLD Object Tracker: Milestone Update
Automatic Initialization of the TLD Object Tracker: Milestone Update Louis Buck May 08, 2012 1 Background TLD is a long-term, real-time tracker designed to be robust to partial and complete occlusions
More informationLecture 4 Face Detection and Classification. Lin ZHANG, PhD School of Software Engineering Tongji University Spring 2018
Lecture 4 Face Detection and Classification Lin ZHANG, PhD School of Software Engineering Tongji University Spring 2018 Any faces contained in the image? Who are they? Outline Overview Face detection Introduction
More informationCS6716 Pattern Recognition. Ensembles and Boosting (1)
CS6716 Pattern Recognition Ensembles and Boosting (1) Aaron Bobick School of Interactive Computing Administrivia Chapter 10 of the Hastie book. Slides brought to you by Aarti Singh, Peter Orbanz, and friends.
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationObject recognition. Methods for classification and image representation
Object recognition Methods for classification and image representation Credits Slides by Pete Barnum Slides by FeiFei Li Paul Viola, Michael Jones, Robust Realtime Object Detection, IJCV 04 Navneet Dalal
More informationClassifier Case Study: Viola-Jones Face Detector
Classifier Case Study: Viola-Jones Face Detector P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. CVPR 2001. P. Viola and M. Jones. Robust real-time face detection.
More informationDecision Trees Dr. G. Bharadwaja Kumar VIT Chennai
Decision Trees Decision Tree Decision Trees (DTs) are a nonparametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target
More informationDetecting Faces in Images. Detecting Faces in Images. Finding faces in an image. Finding faces in an image. Finding faces in an image
Detecting Faces in Images Detecting Faces in Images 37 Finding face like patterns How do we find if a picture has faces in it Where are the faces? A simple solution: Define a typical face Find the typical
More informationLars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany
Syllabus Fri. 27.10. (1) 0. Introduction A. Supervised Learning: Linear Models & Fundamentals Fri. 3.11. (2) A.1 Linear Regression Fri. 10.11. (3) A.2 Linear Classification Fri. 17.11. (4) A.3 Regularization
More informationMachine Learning Lecture 11
Machine Learning Lecture 11 Random Forests 23.11.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory Probability
More informationPredictive Analytics: Demystifying Current and Emerging Methodologies. Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA
Predictive Analytics: Demystifying Current and Emerging Methodologies Tom Kolde, FCAS, MAAA Linda Brobeck, FCAS, MAAA May 18, 2017 About the Presenters Tom Kolde, FCAS, MAAA Consulting Actuary Chicago,
More informationComputer Vision Group Prof. Daniel Cremers. 13. Online Learning
Prof. Daniel Cremers 13. Motivation So far, we have handled offline learning methods, but very often data is observed in an ongoing process Therefore, it would be good to adapt the learned models with
More informationThe Curse of Dimensionality
The Curse of Dimensionality ACAS 2002 p1/66 Curse of Dimensionality The basic idea of the curse of dimensionality is that high dimensional data is difficult to work with for several reasons: Adding more
More informationMachine Learning. A. Supervised Learning A.7. Decision Trees. Lars Schmidt-Thieme
Machine Learning A. Supervised Learning A.7. Decision Trees Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim, Germany 1 /
More informationSkin and Face Detection
Skin and Face Detection Linda Shapiro EE/CSE 576 1 What s Coming 1. Review of Bakic flesh detector 2. Fleck and Forsyth flesh detector 3. Details of Rowley face detector 4. Review of the basic AdaBoost
More informationFace detection and recognition. Detection Recognition Sally
Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification
More informationA Systematic Overview of Data Mining Algorithms. Sargur Srihari University at Buffalo The State University of New York
A Systematic Overview of Data Mining Algorithms Sargur Srihari University at Buffalo The State University of New York 1 Topics Data Mining Algorithm Definition Example of CART Classification Iris, Wine
More informationCSE 151 Machine Learning. Instructor: Kamalika Chaudhuri
CSE 151 Machine Learning Instructor: Kamalika Chaudhuri Announcements Midterm on Monday May 21 (decision trees, kernels, perceptron, and comparison to knns) Review session on Friday (enter time on Piazza)
More informationFace Tracking in Video
Face Tracking in Video Hamidreza Khazaei and Pegah Tootoonchi Afshar Stanford University 350 Serra Mall Stanford, CA 94305, USA I. INTRODUCTION Object tracking is a hot area of research, and has many practical
More informationFace detection and recognition. Many slides adapted from K. Grauman and D. Lowe
Face detection and recognition Many slides adapted from K. Grauman and D. Lowe Face detection and recognition Detection Recognition Sally History Early face recognition systems: based on features and distances
More informationSubject-Oriented Image Classification based on Face Detection and Recognition
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationMachine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013
Machine Learning Topic 5: Linear Discriminants Bryan Pardo, EECS 349 Machine Learning, 2013 Thanks to Mark Cartwright for his extensive contributions to these slides Thanks to Alpaydin, Bishop, and Duda/Hart/Stork
More informationPattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition
Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant
More informationSupervised Learning for Image Segmentation
Supervised Learning for Image Segmentation Raphael Meier 06.10.2016 Raphael Meier MIA 2016 06.10.2016 1 / 52 References A. Ng, Machine Learning lecture, Stanford University. A. Criminisi, J. Shotton, E.
More informationStatistical Methods for Data Mining
Statistical Methods for Data Mining Kuangnan Fang Xiamen University Email: xmufkn@xmu.edu.cn Tree-based Methods Here we describe tree-based methods for regression and classification. These involve stratifying
More informationGeneric Object Class Detection using Feature Maps
Generic Object Class Detection using Feature Maps Oscar Danielsson and Stefan Carlsson CVAP/CSC, KTH, Teknikringen 4, S- 44 Stockholm, Sweden {osda2,stefanc}@csc.kth.se Abstract. In this paper we describe
More informationAdvanced Video Content Analysis and Video Compression (5LSH0), Module 8B
Advanced Video Content Analysis and Video Compression (5LSH0), Module 8B 1 Supervised learning Catogarized / labeled data Objects in a picture: chair, desk, person, 2 Classification Fons van der Sommen
More informationClassification/Regression Trees and Random Forests
Classification/Regression Trees and Random Forests Fabio G. Cozman - fgcozman@usp.br November 6, 2018 Classification tree Consider binary class variable Y and features X 1,..., X n. Decide Ŷ after a series
More informationViola Jones Face Detection. Shahid Nabi Hiader Raiz Muhammad Murtaz
Viola Jones Face Detection Shahid Nabi Hiader Raiz Muhammad Murtaz Face Detection Train The Classifier Use facial and non facial images Train the classifier Find the threshold value Test the classifier
More informationActive learning for visual object recognition
Active learning for visual object recognition Written by Yotam Abramson and Yoav Freund Presented by Ben Laxton Outline Motivation and procedure How this works: adaboost and feature details Why this works:
More informationBoosting Simple Model Selection Cross Validation Regularization
Boosting: (Linked from class website) Schapire 01 Boosting Simple Model Selection Cross Validation Regularization Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University February 8 th,
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 189 Fall 2015 Introduction to Machine Learning Final Please do not turn over the page before you are instructed to do so. You have 2 hours and 50 minutes. Please write your initials on the top-right
More informationThe exam is closed book, closed notes except your one-page (two-sided) cheat sheet.
CS 189 Spring 2015 Introduction to Machine Learning Final You have 2 hours 50 minutes for the exam. The exam is closed book, closed notes except your one-page (two-sided) cheat sheet. No calculators or
More informationCS294-1 Final Project. Algorithms Comparison
CS294-1 Final Project Algorithms Comparison Deep Learning Neural Network AdaBoost Random Forest Prepared By: Shuang Bi (24094630) Wenchang Zhang (24094623) 2013-05-15 1 INTRODUCTION In this project, we
More informationBig Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1
Big Data Methods Chapter 5: Machine learning Big Data Methods, Chapter 5, Slide 1 5.1 Introduction to machine learning What is machine learning? Concerned with the study and development of algorithms that
More informationInformation Management course
Università degli Studi di Milano Master Degree in Computer Science Information Management course Teacher: Alberto Ceselli Lecture 20: 10/12/2015 Data Mining: Concepts and Techniques (3 rd ed.) Chapter
More informationFace Detection and Alignment. Prof. Xin Yang HUST
Face Detection and Alignment Prof. Xin Yang HUST Many slides adapted from P. Viola Face detection Face detection Basic idea: slide a window across image and evaluate a face model at every location Challenges
More informationChapter 7: Numerical Prediction
Ludwig-Maximilians-Universität München Institut für Informatik Lehr- und Forschungseinheit für Datenbanksysteme Knowledge Discovery in Databases SS 2016 Chapter 7: Numerical Prediction Lecture: Prof. Dr.
More informationSlides for Data Mining by I. H. Witten and E. Frank
Slides for Data Mining by I. H. Witten and E. Frank 7 Engineering the input and output Attribute selection Scheme-independent, scheme-specific Attribute discretization Unsupervised, supervised, error-
More informationObject Category Detection: Sliding Windows
03/18/10 Object Category Detection: Sliding Windows Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Goal: Detect all instances of objects Influential Works in Detection Sung-Poggio
More informationDetecting and Reading Text in Natural Scenes
October 19, 2004 X. Chen, A. L. Yuille Outline Outline Goals Example Main Ideas Results Goals Outline Goals Example Main Ideas Results Given an image of an outdoor scene, goals are to: Identify regions
More informationLogical Rhythm - Class 3. August 27, 2018
Logical Rhythm - Class 3 August 27, 2018 In this Class Neural Networks (Intro To Deep Learning) Decision Trees Ensemble Methods(Random Forest) Hyperparameter Optimisation and Bias Variance Tradeoff Biological
More informationEnsemble Learning. Another approach is to leverage the algorithms we have via ensemble methods
Ensemble Learning Ensemble Learning So far we have seen learning algorithms that take a training set and output a classifier What if we want more accuracy than current algorithms afford? Develop new learning
More informationFace Detection on OpenCV using Raspberry Pi
Face Detection on OpenCV using Raspberry Pi Narayan V. Naik Aadhrasa Venunadan Kumara K R Department of ECE Department of ECE Department of ECE GSIT, Karwar, Karnataka GSIT, Karwar, Karnataka GSIT, Karwar,
More informationSharing Visual Features for Multiclass and Multiview Object Detection
Sharing Visual Features for Multiclass and Multiview Object Detection A. Torralba, K. Murphy and W. Freeman IEEE TPAMI 2007 presented by J. Silva, Duke University Sharing Visual Features for Multiclass
More informationIntroduction. How? Rapid Object Detection using a Boosted Cascade of Simple Features. Features. By Paul Viola & Michael Jones
Rapid Object Detection using a Boosted Cascade of Simple Features By Paul Viola & Michael Jones Introduction The Problem we solve face/object detection What's new: Fast! 384X288 pixel images can be processed
More informationDecision Tree CE-717 : Machine Learning Sharif University of Technology
Decision Tree CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adapted from: Prof. Tom Mitchell Decision tree Approximating functions of usually discrete
More informationClassification. Instructor: Wei Ding
Classification Decision Tree Instructor: Wei Ding Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1 Preliminaries Each data record is characterized by a tuple (x, y), where x is the attribute
More informationBoosting Simple Model Selection Cross Validation Regularization. October 3 rd, 2007 Carlos Guestrin [Schapire, 1989]
Boosting Simple Model Selection Cross Validation Regularization Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University October 3 rd, 2007 1 Boosting [Schapire, 1989] Idea: given a weak
More informationFor Monday. Read chapter 18, sections Homework:
For Monday Read chapter 18, sections 10-12 The material in section 8 and 9 is interesting, but we won t take time to cover it this semester Homework: Chapter 18, exercise 25 a-b Program 4 Model Neuron
More informationDecision Trees Oct
Decision Trees Oct - 7-2009 Previously We learned two different classifiers Perceptron: LTU KNN: complex decision boundary If you are a novice in this field, given a classification application, are these
More informationSemi-supervised learning and active learning
Semi-supervised learning and active learning Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Combining classifiers Ensemble learning: a machine learning paradigm where multiple learners
More informationCS381V Experiment Presentation. Chun-Chen Kuo
CS381V Experiment Presentation Chun-Chen Kuo The Paper Indoor Segmentation and Support Inference from RGBD Images. N. Silberman, D. Hoiem, P. Kohli, and R. Fergus. ECCV 2012. 50 100 150 200 250 300 350
More informationObject Detection Design challenges
Object Detection Design challenges How to efficiently search for likely objects Even simple models require searching hundreds of thousands of positions and scales Feature design and scoring How should
More informationPartitioning Data. IRDS: Evaluation, Debugging, and Diagnostics. Cross-Validation. Cross-Validation for parameter tuning
Partitioning Data IRDS: Evaluation, Debugging, and Diagnostics Charles Sutton University of Edinburgh Training Validation Test Training : Running learning algorithms Validation : Tuning parameters of learning
More informationBusiness Club. Decision Trees
Business Club Decision Trees Business Club Analytics Team December 2017 Index 1. Motivation- A Case Study 2. The Trees a. What is a decision tree b. Representation 3. Regression v/s Classification 4. Building
More informationCS 584 Data Mining. Classification 1
CS 584 Data Mining Classification 1 Classification: Definition Given a collection of records (training set ) Each record contains a set of attributes, one of the attributes is the class. Find a model for
More information8. Tree-based approaches
Foundations of Machine Learning École Centrale Paris Fall 2015 8. Tree-based approaches Chloé-Agathe Azencott Centre for Computational Biology, Mines ParisTech chloe agathe.azencott@mines paristech.fr
More informationMIT Samberg Center Cambridge, MA, USA. May 30 th June 2 nd, by C. Rea, R.S. Granetz MIT Plasma Science and Fusion Center, Cambridge, MA, USA
Exploratory Machine Learning studies for disruption prediction on DIII-D by C. Rea, R.S. Granetz MIT Plasma Science and Fusion Center, Cambridge, MA, USA Presented at the 2 nd IAEA Technical Meeting on
More informationSupervised Learning of Classifiers
Supervised Learning of Classifiers Carlo Tomasi Supervised learning is the problem of computing a function from a feature (or input) space X to an output space Y from a training set T of feature-output
More informationCS229 Lecture notes. Raphael John Lamarre Townshend
CS229 Lecture notes Raphael John Lamarre Townshend Decision Trees We now turn our attention to decision trees, a simple yet flexible class of algorithms. We will first consider the non-linear, region-based
More informationDecision Trees. This week: Next week: constructing DT. Pruning DT. Ensemble methods. Greedy Algorithm Potential Function.
Decision Trees This week: constructing DT Greedy Algorithm Potential Function upper bounds the error Pruning DT Next week: Ensemble methods Random Forest 2 Decision Trees - Boolean x 0 + x 6 0 + - 3 Decision
More informationUnivariate and Multivariate Decision Trees
Univariate and Multivariate Decision Trees Olcay Taner Yıldız and Ethem Alpaydın Department of Computer Engineering Boğaziçi University İstanbul 80815 Turkey Abstract. Univariate decision trees at each
More informationModel Selection and Assessment
Model Selection and Assessment CS4780/5780 Machine Learning Fall 2014 Thorsten Joachims Cornell University Reading: Mitchell Chapter 5 Dietterich, T. G., (1998). Approximate Statistical Tests for Comparing
More informationii(i,j) = k i,l j efficiently computed in a single image pass using recurrences
4.2 Traditional image data structures 9 INTEGRAL IMAGE is a matrix representation that holds global image information [Viola and Jones 01] valuesii(i,j)... sums of all original image pixel-values left
More informationMachine Learning : Clustering, Self-Organizing Maps
Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps Clustering The task: partition a set of objects into meaningful subsets (clusters). The
More informationGlobally Induced Forest: A Prepruning Compression Scheme
Globally Induced Forest: A Prepruning Compression Scheme Jean-Michel Begon, Arnaud Joly, Pierre Geurts Systems and Modeling, Dept. of EE and CS, University of Liege, Belgium ICML 2017 Goal and motivation
More informationTraffic Sign Localization and Classification Methods: An Overview
Traffic Sign Localization and Classification Methods: An Overview Ivan Filković University of Zagreb Faculty of Electrical Engineering and Computing Department of Electronics, Microelectronics, Computer
More informationA Hybrid Face Detection System using combination of Appearance-based and Feature-based methods
IJCSNS International Journal of Computer Science and Network Security, VOL.9 No.5, May 2009 181 A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods Zahra Sadri
More informationIntroduction to Artificial Intelligence
Introduction to Artificial Intelligence COMP307 Machine Learning 2: 3-K Techniques Yi Mei yi.mei@ecs.vuw.ac.nz 1 Outline K-Nearest Neighbour method Classification (Supervised learning) Basic NN (1-NN)
More informationMachine Learning Techniques for Data Mining
Machine Learning Techniques for Data Mining Eibe Frank University of Waikato New Zealand 10/25/2000 1 PART VII Moving on: Engineering the input and output 10/25/2000 2 Applying a learner is not all Already
More informationNominal Data. May not have a numerical representation Distance measures might not make sense. PR and ANN
NonMetric Data Nominal Data So far we consider patterns to be represented by feature vectors of real or integer values Easy to come up with a distance (similarity) measure by using a variety of mathematical
More informationMondrian Forests: Efficient Online Random Forests
Mondrian Forests: Efficient Online Random Forests Balaji Lakshminarayanan (Gatsby Unit, UCL) Daniel M. Roy (Cambridge Toronto) Yee Whye Teh (Oxford) September 4, 2014 1 Outline Background and Motivation
More informationCS Machine Learning
CS 60050 Machine Learning Decision Tree Classifier Slides taken from course materials of Tan, Steinbach, Kumar 10 10 Illustrating Classification Task Tid Attrib1 Attrib2 Attrib3 Class 1 Yes Large 125K
More informationFace Detection Using Look-Up Table Based Gentle AdaBoost
Face Detection Using Look-Up Table Based Gentle AdaBoost Cem Demirkır and Bülent Sankur Boğaziçi University, Electrical-Electronic Engineering Department, 885 Bebek, İstanbul {cemd,sankur}@boun.edu.tr
More informationINF 4300 Classification III Anne Solberg The agenda today:
INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15
More informationLocating Salient Object Features
Locating Salient Object Features K.N.Walker, T.F.Cootes and C.J.Taylor Dept. Medical Biophysics, Manchester University, UK knw@sv1.smb.man.ac.uk Abstract We present a method for locating salient object
More informationArtificial Neural Networks (Feedforward Nets)
Artificial Neural Networks (Feedforward Nets) y w 03-1 w 13 y 1 w 23 y 2 w 01 w 21 w 22 w 02-1 w 11 w 12-1 x 1 x 2 6.034 - Spring 1 Single Perceptron Unit y w 0 w 1 w n w 2 w 3 x 0 =1 x 1 x 2 x 3... x
More informationMachine Learning under a Modern Optimization Lens
Machine Learning under a Modern Optimization Lens Dimitris Bertsimas Operations Research Center Massachusetts Institute of Technology January 2018 Bertsimas (MIT) ML and Modern Optimization January 2018
More informationApplying Supervised Learning
Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains
More informationFACE DETECTION BY HAAR CASCADE CLASSIFIER WITH SIMPLE AND COMPLEX BACKGROUNDS IMAGES USING OPENCV IMPLEMENTATION
FACE DETECTION BY HAAR CASCADE CLASSIFIER WITH SIMPLE AND COMPLEX BACKGROUNDS IMAGES USING OPENCV IMPLEMENTATION Vandna Singh 1, Dr. Vinod Shokeen 2, Bhupendra Singh 3 1 PG Student, Amity School of Engineering
More informationIntroduction to Machine Learning
Introduction to Machine Learning Eric Medvet 16/3/2017 1/77 Outline Machine Learning: what and why? Motivating example Tree-based methods Regression trees Trees aggregation 2/77 Teachers Eric Medvet Dipartimento
More informationBayesian model ensembling using meta-trained recurrent neural networks
Bayesian model ensembling using meta-trained recurrent neural networks Luca Ambrogioni l.ambrogioni@donders.ru.nl Umut Güçlü u.guclu@donders.ru.nl Yağmur Güçlütürk y.gucluturk@donders.ru.nl Julia Berezutskaya
More information