MATH 567: Mathematical Techniques in Data Science Support vector machines
|
|
- Carmel Leonard
- 5 years ago
- Views:
Transcription
1 1/8 MATH 567: Mathematical Techniques in Data Science Support vector machines Dominique Guillot Departments of Mathematical Sciences University of Delaware March 13, 2016
2 Hyperplanes 2/8 Recall: A hyperplane H in V = R n is a subspace of V of dimension n 1 (i.e., a subspace of codimension 1). Each hyperplane is determined by a nonzero vector β R n via H = {x R n : β T x = 0} = span(β).
3 Hyperplanes 2/8 Recall: A hyperplane H in V = R n is a subspace of V of dimension n 1 (i.e., a subspace of codimension 1). Each hyperplane is determined by a nonzero vector β R n via H = {x R n : β T x = 0} = span(β). An ane hyperplane H in R n is a subset of the form where β 0 R, β R n. H = {x R n : β 0 + β T x = 0}
4 Hyperplanes Recall: A hyperplane H in V = R n is a subspace of V of dimension n 1 (i.e., a subspace of codimension 1). Each hyperplane is determined by a nonzero vector β R n via H = {x R n : β T x = 0} = span(β). An ane hyperplane H in R n is a subset of the form where β 0 R, β R n. H = {x R n : β 0 + β T x = 0} We often use the term hyperplane for ane hyperplane. 2/8
5 Hyperplanes (cont.) 3/8 Let H = {x R n : β 0 + β T x = 0}.
6 Hyperplanes (cont.) Let H = {x R n : β 0 + β T x = 0}. Note that for x 0, x 1 H, β T (x 0 x 1 ) = 0. Thus β is perpendicular to H. It follows that for x R n, d(x, H) = βt β (x x 0) = β 0 + β T x. β 3/8
7 4/8 Separating hyperplane Suppose we have binary data with labels {+1, 1}. We want to separate data using an (ane) hyperplane. ESL, Figure (Orange = least-squares)
8 4/8 Separating hyperplane Suppose we have binary data with labels {+1, 1}. We want to separate data using an (ane) hyperplane. ESL, Figure (Orange = least-squares) Classify using G(x) = sgn(x T β + β 0 ).
9 4/8 Separating hyperplane Suppose we have binary data with labels {+1, 1}. We want to separate data using an (ane) hyperplane. ESL, Figure (Orange = least-squares) Classify using G(x) = sgn(x T β + β 0 ). Separating hyperplane may not be unique. Separating hyperplane may not exist (i.e., data may not be separable).
10 Margins 5/8 Uniqueness problem: when the data is separable, choose the hyperplane to maximize the margin (the no man's land).
11 Margins 5/8 Uniqueness problem: when the data is separable, choose the hyperplane to maximize the margin (the no man's land). Data: (y i, x i ) {+1, 1} R p (i = 1,..., n). Suppose β 0 + β T x is a separating hyperplane with β = 1.
12 Margins 5/8 Uniqueness problem: when the data is separable, choose the hyperplane to maximize the margin (the no man's land). Data: (y i, x i ) {+1, 1} R p (i = 1,..., n). Suppose β 0 + β T x is a separating hyperplane with β = 1. Note that: y i (x T i β + β 0 ) > 0 Correct classication y i (x T i β + β 0 ) < 0 Incorrect classication
13 Margins Uniqueness problem: when the data is separable, choose the hyperplane to maximize the margin (the no man's land). Data: (y i, x i ) {+1, 1} R p (i = 1,..., n). Suppose β 0 + β T x is a separating hyperplane with β = 1. Note that: y i (x T i β + β 0 ) > 0 Correct classication y i (x T i β + β 0 ) < 0 Incorrect classication Also, y i (x T i β + β 0) = distance between x and hyperplane (since β = 1). 5/8
14 Margins (cont.) 6/8 Thus, if the data is separable, we can solve max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M (i = 1,..., n).
15 Margins (cont.) 6/8 Thus, if the data is separable, we can solve max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M (i = 1,..., n). We will transform the problem into a usual form used in convex optimization.
16 6/8 Margins (cont.) Thus, if the data is separable, we can solve max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M (i = 1,..., n). We will transform the problem into a usual form used in convex optimization. We can remove β = 1 by replacing the constraint by 1 β y i(x T i β+β 0 ) M, or equivalently, y i (x T i β+β 0 ) M β.
17 Margins (cont.) 6/8 Thus, if the data is separable, we can solve max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M (i = 1,..., n). We will transform the problem into a usual form used in convex optimization. We can remove β = 1 by replacing the constraint by 1 β y i(x T i β+β 0 ) M, or equivalently, y i (x T i β+β 0 ) M β. We can always rescale (β, β 0 ) so that β = 1/M. Our problem is therefore equivalent to 1 min β 0,β R p 2 β 2 subject to y i (x T i β + β 0 ) 1 (i = 1,..., n).
18 Margins (cont.) Thus, if the data is separable, we can solve max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M (i = 1,..., n). We will transform the problem into a usual form used in convex optimization. We can remove β = 1 by replacing the constraint by 1 β y i(x T i β+β 0 ) M, or equivalently, y i (x T i β+β 0 ) M β. We can always rescale (β, β 0 ) so that β = 1/M. Our problem is therefore equivalent to 1 min β 0,β R p 2 β 2 subject to y i (x T i β + β 0 ) 1 (i = 1,..., n). We now recognize the problem as a convex optimization problem with a quadratic objective, and linear inequality constraints. 6/8
19 Support vector machines 7/8 The previous problem works well when the data is separable. What happens if there is no way to nd a margin?
20 Support vector machines 7/8 The previous problem works well when the data is separable. What happens if there is no way to nd a margin? We allow some points to be on the wrong side of the margin, but keep control on the error.
21 Support vector machines 7/8 The previous problem works well when the data is separable. What happens if there is no way to nd a margin? We allow some points to be on the wrong side of the margin, but keep control on the error. We replace y i (x T i β + β 0) M by y i (x T i β + β 0 ) M(1 ξ i ), ξ i 0, and add the constraint n ξ i C for some xed constant C > 0. i=1
22 Support vector machines 7/8 The previous problem works well when the data is separable. What happens if there is no way to nd a margin? We allow some points to be on the wrong side of the margin, but keep control on the error. We replace y i (x T i β + β 0) M by y i (x T i β + β 0 ) M(1 ξ i ), ξ i 0, and add the constraint n ξ i C for some xed constant C > 0. i=1 The problem becomes: max M β 0,β R p, β =1 subject to y i (x T i β + β 0 ) M(1 ξ i ) n ξ i 0, ξ i C. i=1
23 Support vector machines (cont.) 8/8 As before, we can transform the problem into its normal form: 1 min β 0,β 2 β 2 subject to y i (x T i β + β 0 ) 1 ξ i n ξ i 0, ξ i C. i=1 Problem can be solved using standard optimization packages.
SUPPORT VECTOR MACHINES
SUPPORT VECTOR MACHINES Today Reading AIMA 18.9 Goals (Naïve Bayes classifiers) Support vector machines 1 Support Vector Machines (SVMs) SVMs are probably the most popular off-the-shelf classifier! Software
More informationIntroduction to Machine Learning
Introduction to Machine Learning Maximum Margin Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574
More informationLecture 9: Support Vector Machines
Lecture 9: Support Vector Machines William Webber (william@williamwebber.com) COMP90042, 2014, Semester 1, Lecture 8 What we ll learn in this lecture Support Vector Machines (SVMs) a highly robust and
More informationMore on Classification: Support Vector Machine
More on Classification: Support Vector Machine The Support Vector Machine (SVM) is a classification method approach developed in the computer science field in the 1990s. It has shown good performance in
More informationSUPPORT VECTOR MACHINES
SUPPORT VECTOR MACHINES Today Reading AIMA 8.9 (SVMs) Goals Finish Backpropagation Support vector machines Backpropagation. Begin with randomly initialized weights 2. Apply the neural network to each training
More informationOptimal Separating Hyperplane and the Support Vector Machine. Volker Tresp Summer 2018
Optimal Separating Hyperplane and the Support Vector Machine Volker Tresp Summer 2018 1 (Vapnik s) Optimal Separating Hyperplane Let s consider a linear classifier with y i { 1, 1} If classes are linearly
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationSupport Vector Machines
Support Vector Machines About the Name... A Support Vector A training sample used to define classification boundaries in SVMs located near class boundaries Support Vector Machines Binary classifiers whose
More informationData Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)
Data Mining: Concepts and Techniques Chapter 9 Classification: Support Vector Machines 1 Support Vector Machines (SVMs) SVMs are a set of related supervised learning methods used for classification Based
More informationSupport vector machines
Support vector machines When the data is linearly separable, which of the many possible solutions should we prefer? SVM criterion: maximize the margin, or distance between the hyperplane and the closest
More informationTopics in Machine Learning
Topics in Machine Learning Gilad Lerman School of Mathematics University of Minnesota Text/slides stolen from G. James, D. Witten, T. Hastie, R. Tibshirani and A. Ng Machine Learning - Motivation Arthur
More informationMachine Learning for NLP
Machine Learning for NLP Support Vector Machines Aurélie Herbelot 2018 Centre for Mind/Brain Sciences University of Trento 1 Support Vector Machines: introduction 2 Support Vector Machines (SVMs) SVMs
More informationAn Introduction to Machine Learning
TRIPODS Summer Boorcamp: Topology and Machine Learning August 6, 2018 General Set-up Introduction Set-up and Goal Suppose we have X 1,X 2,...,X n data samples. Can we predict properites about any given
More informationAdvanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material
More informationDaily WeBWorK, #1. This means the two planes normal vectors must be multiples of each other.
Daily WeBWorK, #1 Consider the ellipsoid x 2 + 3y 2 + z 2 = 11. Find all the points where the tangent plane to this ellipsoid is parallel to the plane 2x + 3y + 2z = 0. In order for the plane tangent to
More informationLecture 7: Support Vector Machine
Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each
More informationNumerical Optimization
Convex Sets Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on Let x 1, x 2 R n, x 1 x 2. Line and line segment Line passing through x 1 and x 2 : {y
More informationLinear Models. Lecture Outline: Numeric Prediction: Linear Regression. Linear Classification. The Perceptron. Support Vector Machines
Linear Models Lecture Outline: Numeric Prediction: Linear Regression Linear Classification The Perceptron Support Vector Machines Reading: Chapter 4.6 Witten and Frank, 2nd ed. Chapter 4 of Mitchell Solving
More information.. Spring 2017 CSC 566 Advanced Data Mining Alexander Dekhtyar..
.. Spring 2017 CSC 566 Advanced Data Mining Alexander Dekhtyar.. Machine Learning: Support Vector Machines: Linear Kernel Support Vector Machines Extending Perceptron Classifiers. There are two ways to
More informationPerceptron Learning Algorithm (PLA)
Review: Lecture 4 Perceptron Learning Algorithm (PLA) Learning algorithm for linear threshold functions (LTF) (iterative) Energy function: PLA implements a stochastic gradient algorithm Novikoff s theorem
More informationChapter 4 Concepts from Geometry
Chapter 4 Concepts from Geometry An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Line Segments The line segment between two points and in R n is the set of points on the straight line joining
More information1. Show that the rectangle of maximum area that has a given perimeter p is a square.
Constrained Optimization - Examples - 1 Unit #23 : Goals: Lagrange Multipliers To study constrained optimization; that is, the maximizing or minimizing of a function subject to a constraint (or side condition).
More informationLOGISTIC REGRESSION FOR MULTIPLE CLASSES
Peter Orbanz Applied Data Mining Not examinable. 111 LOGISTIC REGRESSION FOR MULTIPLE CLASSES Bernoulli and multinomial distributions The mulitnomial distribution of N draws from K categories with parameter
More informationPerceptron Learning Algorithm
Perceptron Learning Algorithm An iterative learning algorithm that can find linear threshold function to partition linearly separable set of points. Assume zero threshold value. 1) w(0) = arbitrary, j=1,
More informationLP Geometry: outline. A general LP. minimize x c T x s.t. a T i. x b i, i 2 M 1 a T i x = b i, i 2 M 3 x j 0, j 2 N 1. where
LP Geometry: outline I Polyhedra I Extreme points, vertices, basic feasible solutions I Degeneracy I Existence of extreme points I Optimality of extreme points IOE 610: LP II, Fall 2013 Geometry of Linear
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing
More informationBounded, Closed, and Compact Sets
Bounded, Closed, and Compact Sets Definition Let D be a subset of R n. Then D is said to be bounded if there is a number M > 0 such that x < M for all x D. D is closed if it contains all the boundary points.
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions
More informationMATH 829: Introduction to Data Mining and Analysis Overview
1/13 MATH 829: Introduction to Data Mining and Analysis Overview Dominique Guillot Departments of Mathematical Sciences University of Delaware February 8, 2016 Supervised vs unsupervised learning 2/13
More informationChapter 3. Quadric hypersurfaces. 3.1 Quadric hypersurfaces Denition.
Chapter 3 Quadric hypersurfaces 3.1 Quadric hypersurfaces. 3.1.1 Denition. Denition 1. In an n-dimensional ane space A; given an ane frame fo;! e i g: A quadric hypersurface in A is a set S consisting
More informationClassification. 1 o Semestre 2007/2008
Classification Departamento de Engenharia Informática Instituto Superior Técnico 1 o Semestre 2007/2008 Slides baseados nos slides oficiais do livro Mining the Web c Soumen Chakrabarti. Outline 1 2 3 Single-Class
More informationNon-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines
Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007,
More informationDivision of the Humanities and Social Sciences. Convex Analysis and Economic Theory Winter Separation theorems
Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analysis and Economic Theory Winter 2018 Topic 8: Separation theorems 8.1 Hyperplanes and half spaces Recall that a hyperplane in
More informationMath 2331 Linear Algebra
4.2 Null Spaces, Column Spaces, & Linear Transformations Math 233 Linear Algebra 4.2 Null Spaces, Column Spaces, & Linear Transformations Jiwen He Department of Mathematics, University of Houston jiwenhe@math.uh.edu
More informationCOMS 4771 Support Vector Machines. Nakul Verma
COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron
More informationLecture Linear Support Vector Machines
Lecture 8 In this lecture we return to the task of classification. As seen earlier, examples include spam filters, letter recognition, or text classification. In this lecture we introduce a popular method
More informationLinear methods for supervised learning
Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes
More informationAlgebraic Geometry of Segmentation and Tracking
Ma191b Winter 2017 Geometry of Neuroscience Geometry of lines in 3-space and Segmentation and Tracking This lecture is based on the papers: Reference: Marco Pellegrini, Ray shooting and lines in space.
More informationData-driven Kernels for Support Vector Machines
Data-driven Kernels for Support Vector Machines by Xin Yao A research paper presented to the University of Waterloo in partial fulfillment of the requirement for the degree of Master of Mathematics in
More informationMATH 829: Introduction to Data Mining and Analysis Clustering I
1/13 MATH 829: Introduction to Data Mining and Analysis Clustering I Dominique Guillot Departments of Mathematical Sciences University of Delaware April 25, 2016 Supervised and unsupervised learning 2/13
More informationMachine Learning. Support Vector Machines. Fabio Vandin November 20, 2017
Machine Learning Support Vector Machines Fabio Vandin November 20, 2017 1 Classification and Margin Consider a classification problem with two classes: instance set X = R d label set Y = { 1, 1}. Training
More informationMaths for Signals and Systems Linear Algebra in Engineering. Some problems by Gilbert Strang
Maths for Signals and Systems Linear Algebra in Engineering Some problems by Gilbert Strang Problems. Consider u, v, w to be non-zero vectors in R 7. These vectors span a vector space. What are the possible
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationThe basics of rigidity
The basics of rigidity Lectures I and II Session on Granular Matter Institut Henri Poincaré R. Connelly Cornell University Department of Mathematics 1 What determines rigidity? 2 What determines rigidity?
More informationMathematics Parabolas
a place of mind F A C U L T Y O F E D U C A T I O N Department of Curriculum and Pedagog Mathematics Parabolas Science and Mathematics Education Research Group Supported b UBC Teaching and Learning Enhancement
More informationMore on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization
More on Learning Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization Neural Net Learning Motivated by studies of the brain. A network of artificial
More informationMath 5593 Linear Programming Lecture Notes
Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................
More informationSupport Vector Machines
Support Vector Machines SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions 6. Dealing
More informationPractical 7: Support vector machines
Practical 7: Support vector machines Support vector machines are implemented in several R packages, including e1071, caret and kernlab. We will use the e1071 package in this practical. install.packages('e1071')
More informationIntroduction to the h-principle
Introduction to the h-principle Kyler Siegel May 15, 2014 Contents 1 Introduction 1 2 Jet bundles and holonomic sections 2 3 Holonomic approximation 3 4 Applications 4 1 Introduction The goal of this talk
More informationCS446: Machine Learning Fall Problem Set 4. Handed Out: October 17, 2013 Due: October 31 th, w T x i w
CS446: Machine Learning Fall 2013 Problem Set 4 Handed Out: October 17, 2013 Due: October 31 th, 2013 Feel free to talk to other members of the class in doing the homework. I am more concerned that you
More informationTopic 4: Support Vector Machines
CS 4850/6850: Introduction to achine Learning Fall 2018 Topic 4: Support Vector achines Instructor: Daniel L Pimentel-Alarcón c Copyright 2018 41 Introduction Support vector machines (SVs) are considered
More informationModule 4. Non-linear machine learning econometrics: Support Vector Machine
Module 4. Non-linear machine learning econometrics: Support Vector Machine THE CONTRACTOR IS ACTING UNDER A FRAMEWORK CONTRACT CONCLUDED WITH THE COMMISSION Introduction When the assumption of linearity
More informationA fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song
More informationA Short SVM (Support Vector Machine) Tutorial
A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange
More informationCurve Construction via Local Fitting
Curve Construction via Local Fitting Suppose we are given points and tangents Q k, and T k (k = 0,..., n), and a fitting tolerance ε. We want to fit this data with the minimum (in some sense) number of
More informationCS 559: Machine Learning Fundamentals and Applications 9 th Set of Notes
1 CS 559: Machine Learning Fundamentals and Applications 9 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Overview
More informationAll lecture slides will be available at CSC2515_Winter15.html
CSC2515 Fall 2015 Introduc3on to Machine Learning Lecture 9: Support Vector Machines All lecture slides will be available at http://www.cs.toronto.edu/~urtasun/courses/csc2515/ CSC2515_Winter15.html Many
More informationSupervised vs. Unsupervised Learning. Supervised vs. Unsupervised Learning. Supervised vs. Unsupervised Learning. Supervised vs. Unsupervised Learning
Overview T7 - SVM and s Christian Vögeli cvoegeli@inf.ethz.ch Supervised/ s Support Vector Machines Kernels Based on slides by P. Orbanz & J. Keuchel Task: Apply some machine learning method to data from
More informationLecture 5: Linear Classification
Lecture 5: Linear Classification CS 194-10, Fall 2011 Laurent El Ghaoui EECS Department UC Berkeley September 8, 2011 Outline Outline Data We are given a training data set: Feature vectors: data points
More informationLecture 19: Convex Non-Smooth Optimization. April 2, 2007
: Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials
More informationLecture 12 March 4th
Math 239: Discrete Mathematics for the Life Sciences Spring 2008 Lecture 12 March 4th Lecturer: Lior Pachter Scribe/ Editor: Wenjing Zheng/ Shaowei Lin 12.1 Alignment Polytopes Recall that the alignment
More informationDM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini
DM545 Linear and Integer Programming Lecture 2 The Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 3. 4. Standard Form Basic Feasible Solutions
More informationConvexity I: Sets and Functions
Convexity I: Sets and Functions Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 See supplements for reviews of basic real analysis basic multivariate calculus basic
More informationMathematical Themes in Economics, Machine Learning, and Bioinformatics
Western Kentucky University From the SelectedWorks of Matt Bogard 2010 Mathematical Themes in Economics, Machine Learning, and Bioinformatics Matt Bogard, Western Kentucky University Available at: https://works.bepress.com/matt_bogard/7/
More informationRobot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning
Robot Learning 1 General Pipeline 1. Data acquisition (e.g., from 3D sensors) 2. Feature extraction and representation construction 3. Robot learning: e.g., classification (recognition) or clustering (knowledge
More informationRadical Functions. Attendance Problems. Identify the domain and range of each function.
Page 1 of 12 Radical Functions Attendance Problems. Identify the domain and range of each function. 1. f ( x) = x 2 + 2 2. f ( x) = 3x 3 Use the description to write the quadratic function g based on the
More informationChoosing the kernel parameters for SVMs by the inter-cluster distance in the feature space Authors: Kuo-Ping Wu, Sheng-De Wang Published 2008
Choosing the kernel parameters for SVMs by the inter-cluster distance in the feature space Authors: Kuo-Ping Wu, Sheng-De Wang Published 2008 Presented by: Nandini Deka UH Mathematics Spring 2014 Workshop
More information6 Randomized rounding of semidefinite programs
6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can
More informationSupport vector machines. Dominik Wisniewski Wojciech Wawrzyniak
Support vector machines Dominik Wisniewski Wojciech Wawrzyniak Outline 1. A brief history of SVM. 2. What is SVM and how does it work? 3. How would you classify this data? 4. Are all the separating lines
More information12 Classification using Support Vector Machines
160 Bioinformatics I, WS 14/15, D. Huson, January 28, 2015 12 Classification using Support Vector Machines This lecture is based on the following sources, which are all recommended reading: F. Markowetz.
More informationSupport Vector Machines
Support Vector Machines Chapter 9 Chapter 9 1 / 50 1 91 Maximal margin classifier 2 92 Support vector classifiers 3 93 Support vector machines 4 94 SVMs with more than two classes 5 95 Relationshiop to
More informationSupport Vector Machines
Support Vector Machines . Importance of SVM SVM is a discriminative method that brings together:. computational learning theory. previously known methods in linear discriminant functions 3. optimization
More informationMath 414 Lecture 2 Everyone have a laptop?
Math 44 Lecture 2 Everyone have a laptop? THEOREM. Let v,...,v k be k vectors in an n-dimensional space and A = [v ;...; v k ] v,..., v k independent v,..., v k span the space v,..., v k a basis v,...,
More informationApplied Lagrange Duality for Constrained Optimization
Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity
More informationAssignment 1 - Binary Linear Learning
Assignment 1 - Binary Linear Learning Xiang Zhang, John Langford and Yann LeCun February 26, 2013 Instructions Deadline: March 12, 2013 before midnight. You must answer the questions by yourself, but you
More informationLecture 2 - Introduction to Polytopes
Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.
More informationDavid G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer
David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms
More informationL-CONVEX-CONCAVE SETS IN REAL PROJECTIVE SPACE AND L-DUALITY
MOSCOW MATHEMATICAL JOURNAL Volume 3, Number 3, July September 2003, Pages 1013 1037 L-CONVEX-CONCAVE SETS IN REAL PROJECTIVE SPACE AND L-DUALITY A. KHOVANSKII AND D. NOVIKOV Dedicated to Vladimir Igorevich
More informationIntroduction to Modern Control Systems
Introduction to Modern Control Systems Convex Optimization, Duality and Linear Matrix Inequalities Kostas Margellos University of Oxford AIMS CDT 2016-17 Introduction to Modern Control Systems November
More informationKernel Methods. Chapter 9 of A Course in Machine Learning by Hal Daumé III. Conversion to beamer by Fabrizio Riguzzi
Kernel Methods Chapter 9 of A Course in Machine Learning by Hal Daumé III http://ciml.info Conversion to beamer by Fabrizio Riguzzi Kernel Methods 1 / 66 Kernel Methods Linear models are great because
More informationSupport vector machines
Support vector machines Cavan Reilly October 24, 2018 Table of contents K-nearest neighbor classification Support vector machines K-nearest neighbor classification Suppose we have a collection of measurements
More information1. What is the VC dimension of the family of finite unions of closed intervals over the real line?
Mehryar Mohri Foundations of Machine Learning Courant Institute of Mathematical Sciences Homework assignment 2 March 06, 2011 Due: March 22, 2011 A. VC Dimension 1. What is the VC dimension of the family
More informationSupport Vector Machines
Support Vector Machines Xiaojin Zhu jerryzhu@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [ Based on slides from Andrew Moore http://www.cs.cmu.edu/~awm/tutorials] slide 1
More informationCommon core standards from Grade 8 Math: General categories/domain:
Common core standards from Grade 8 Math: General categories/domain: 1. Ratio and Proportional Relationship (5 %) 2. Then Number System (5 %) 3. Expressions and Equations (25%) 4. (25 %) 5. Geometry (20
More informationarxiv: v1 [math.co] 24 Aug 2009
SMOOTH FANO POLYTOPES ARISING FROM FINITE PARTIALLY ORDERED SETS arxiv:0908.3404v1 [math.co] 24 Aug 2009 TAKAYUKI HIBI AND AKIHIRO HIGASHITANI Abstract. Gorenstein Fano polytopes arising from finite partially
More informationTHREE LECTURES ON BASIC TOPOLOGY. 1. Basic notions.
THREE LECTURES ON BASIC TOPOLOGY PHILIP FOTH 1. Basic notions. Let X be a set. To make a topological space out of X, one must specify a collection T of subsets of X, which are said to be open subsets of
More informationLagrange Multipliers
Lagrange Multipliers Christopher Croke University of Pennsylvania Math 115 How to deal with constrained optimization. How to deal with constrained optimization. Let s revisit the problem of finding the
More informationLocally convex topological vector spaces
Chapter 4 Locally convex topological vector spaces 4.1 Definition by neighbourhoods Let us start this section by briefly recalling some basic properties of convex subsets of a vector space over K (where
More informationModels for grids. Computer vision: models, learning and inference. Multi label Denoising. Binary Denoising. Denoising Goal.
Models for grids Computer vision: models, learning and inference Chapter 9 Graphical Models Consider models where one unknown world state at each pixel in the image takes the form of a grid. Loops in the
More informationConstraining strategies for Kaczmarz-like algorithms
Constraining strategies for Kaczmarz-like algorithms Ovidius University, Constanta, Romania Faculty of Mathematics and Computer Science Talk on April 7, 2008 University of Erlangen-Nürnberg The scanning
More informationSupport Vector Subset Scan for Spatial Pattern Detection
Support Vector Subset Scan for Spatial Pattern Detection Dylan Fitzpatrick, Yun Ni, and Daniel B. Neill Event and Pattern Detection Laboratory Carnegie Mellon University This work was partially supported
More informationConstraint Handling. Fernando Lobo. University of Algarve
Constraint Handling Fernando Lobo University of Algarve Outline Introduction Penalty methods Approach based on tournament selection Decoders Repair algorithms Constraint-preserving operators Introduction
More informationSupport Vector Machines for Face Recognition
Chapter 8 Support Vector Machines for Face Recognition 8.1 Introduction In chapter 7 we have investigated the credibility of different parameters introduced in the present work, viz., SSPD and ALR Feature
More informationSupport Vector Machines
Support Vector Machines RBF-networks Support Vector Machines Good Decision Boundary Optimization Problem Soft margin Hyperplane Non-linear Decision Boundary Kernel-Trick Approximation Accurancy Overtraining
More informationDirectional Derivatives and the Gradient Vector Part 2
Directional Derivatives and the Gradient Vector Part 2 Lecture 25 February 28, 2007 Recall Fact Recall Fact If f is a dierentiable function of x and y, then f has a directional derivative in the direction
More informationFast, Accurate Detection of 100,000 Object Classes on a Single Machine
Fast, Accurate Detection of 100,000 Object Classes on a Single Machine Thomas Dean etal. Google, Mountain View, CA CVPR 2013 best paper award Presented by: Zhenhua Wang 2013.12.10 Outline Background This
More informationConvex sets and convex functions
Convex sets and convex functions Convex optimization problems Convex sets and their examples Separating and supporting hyperplanes Projections on convex sets Convex functions, conjugate functions ECE 602,
More informationLECTURE 5: DUAL PROBLEMS AND KERNELS. * Most of the slides in this lecture are from
LECTURE 5: DUAL PROBLEMS AND KERNELS * Most of the slides in this lecture are from http://www.robots.ox.ac.uk/~az/lectures/ml Optimization Loss function Loss functions SVM review PRIMAL-DUAL PROBLEM Max-min
More informationLARGE MARGIN CLASSIFIERS
Admin Assignment 5 LARGE MARGIN CLASSIFIERS David Kauchak CS 451 Fall 2013 Midterm Download from course web page when you re ready to take it 2 hours to complete Must hand-in (or e-mail in) by 11:59pm
More information