Chap.12 Kernel methods [Book, Chap.7]

Similar documents
Support Vector Machines

Introduction to Support Vector Machines

All lecture slides will be available at CSC2515_Winter15.html

Support Vector Machines.

Support Vector Machines

Support Vector Machines

Support Vector Machines

DM6 Support Vector Machines

12 Classification using Support Vector Machines

SUPPORT VECTOR MACHINES

A Practical Guide to Support Vector Classification

Support Vector Machines

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Chakra Chennubhotla and David Koes

Software Documentation of the Potential Support Vector Machine

Data Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017

A Short SVM (Support Vector Machine) Tutorial

SUPPORT VECTOR MACHINES

Kernel Methods & Support Vector Machines

Learning from Data Linear Parameter Models

Basis Functions. Volker Tresp Summer 2017

Data Mining Practical Machine Learning Tools and Techniques. Slides for Chapter 6 of Data Mining by I. H. Witten and E. Frank

Machine Learning for NLP

Support Vector Machines (a brief introduction) Adrian Bevan.

Lab 2: Support vector machines

Support Vector Machines for Classification and Regression

A New Fuzzy Membership Computation Method for Fuzzy Support Vector Machines

5.6 Self-organizing maps (SOM) [Book, Sect. 10.3]

Feature scaling in support vector data description

Introduction to ANSYS DesignXplorer

Module 4. Non-linear machine learning econometrics: Support Vector Machine

kernlab A Kernel Methods Package

Support vector machines

Support Vector Machines

More on Classification: Support Vector Machine

5 Learning hypothesis classes (16 points)

Support Vector Machines.

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

SVM in Analysis of Cross-Sectional Epidemiological Data Dmitriy Fradkin. April 4, 2005 Dmitriy Fradkin, Rutgers University Page 1

LECTURE 5: DUAL PROBLEMS AND KERNELS. * Most of the slides in this lecture are from

Basis Functions. Volker Tresp Summer 2016

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

Function approximation using RBF network. 10 basis functions and 25 data points.

Model Answers to The Next Pixel Prediction Task

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)

Parallel & Scalable Machine Learning Introduction to Machine Learning Algorithms

Lecture 7: Support Vector Machine

CSE 417T: Introduction to Machine Learning. Lecture 22: The Kernel Trick. Henry Chai 11/15/18

Topic 4: Support Vector Machines

Classification by Support Vector Machines

Support Vector Machines

Lecture 10: Support Vector Machines and their Applications

HW2 due on Thursday. Face Recognition: Dimensionality Reduction. Biometrics CSE 190 Lecture 11. Perceptron Revisited: Linear Separators

CSE 5526: Introduction to Neural Networks Radial Basis Function (RBF) Networks

Kernel Methods. Chapter 9 of A Course in Machine Learning by Hal Daumé III. Conversion to beamer by Fabrizio Riguzzi

Kernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018

Support Vector Machines

Classification by Support Vector Machines

Radial Basis Function Networks: Algorithms

Support Vector Machines and their Applications

Learning from Data: Adaptive Basis Functions

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Embedding based Regression in Nonlinear System

10. Support Vector Machines

Neural Networks. Theory And Practice. Marco Del Vecchio 19/07/2017. Warwick Manufacturing Group University of Warwick

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

Support Vector Machines + Classification for IR

OBJECT CLASSIFICATION USING SUPPORT VECTOR MACHINES WITH KERNEL-BASED DATA PREPROCESSING

University of Cambridge Engineering Part IIB Paper 4F10: Statistical Pattern Processing Handout 11: Non-Parametric Techniques.

Support vector machines

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

CS 559: Machine Learning Fundamentals and Applications 9 th Set of Notes

University of Cambridge Engineering Part IIB Paper 4F10: Statistical Pattern Processing Handout 11: Non-Parametric Techniques

Kernel SVM. Course: Machine Learning MAHDI YAZDIAN-DEHKORDI FALL 2017

Support Vector Regression for Software Reliability Growth Modeling and Prediction

Evaluation of Performance Measures for SVR Hyperparameter Selection

model order p weights The solution to this optimization problem is obtained by solving the linear system

Radial Basis Function Networks

Optimal Separating Hyperplane and the Support Vector Machine. Volker Tresp Summer 2018

A Taxonomy of Semi-Supervised Learning Algorithms

Performance Measures

Natural Language Processing CS 6320 Lecture 6 Neural Language Models. Instructor: Sanda Harabagiu

3 Nonlinear Regression

Prediction of Dialysis Length. Adrian Loy, Antje Schubotz 2 February 2017

Bagging and Boosting Algorithms for Support Vector Machine Classifiers

Machine Learning Lecture 9

What is machine learning?

kernlab An R Package for Kernel Learning

SVM cont d. Applications face detection [IEEE INTELLIGENT SYSTEMS]

Deep Learning. Volker Tresp Summer 2014

Machine Learning. The Breadth of ML Neural Networks & Deep Learning. Marc Toussaint. Duy Nguyen-Tuong. University of Stuttgart

Bagging for One-Class Learning

More on Neural Networks. Read Chapter 5 in the text by Bishop, except omit Sections 5.3.3, 5.3.4, 5.4, 5.5.4, 5.5.5, 5.5.6, 5.5.7, and 5.

An Introduction to Machine Learning

Data mining with Support Vector Machine

Department of Electronics and Telecommunication Engineering 1 PG Student, JSPM s Imperial College of Engineering and Research, Pune (M.H.

Robustness of Selective Desensitization Perceptron Against Irrelevant and Partially Relevant Features in Pattern Classification

Application of Support Vector Machine In Bioinformatics

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Machine Learning Lecture 9

Transcription:

Chap.12 Kernel methods [Book, Chap.7] Neural network methods became popular in the mid to late 1980s, but by the mid to late 1990s, kernel methods have also become popular in machine learning. The first kernel methods were nonlinear classifiers called support vector machines (SVM), which were then generalized to nonlinear regression (support vector regression, SVR). Kernel methods were further developed to nonlinearly generalize principal component analysis (PCA), canonical correlation analysis (CCA), etc. The kernel method has also been extended to probabilistic models, e.g. Gaussian processes (GP). 1 / 25

12.1 From neural networks to kernel methods [Book, Sect. 7.1] Linear regression: y k = j w kj x j + b k, (1) with x the predictors and y the predictands or response variables. When a multi-layer perceptron (MLP) NN is used for nonlinear regression, the mapping does not proceed directly from x to y, but passses through an intermediate layer of variables h, i.e. the hidden neurons, ( ) h j = tanh w ji x i + b j, (2) where the mapping from x to h is nonlinear through an activation function like the hyperbolic tangent. i 2 / 25

The next stage is to map from h to y, and is most commonly done via a linear mapping y k = j w kj h j + b k. (3) Since (3) is formally identical to (1), one can think of the NN model as first mapping nonlinearly from the input space to a hidden space (containing the hidden neurons), i.e. φ : x h, then performing linear regression between the hidden space variables and the output variables. 3 / 25

Figure : Schematic diagram illustrating the effect of the nonlinear mapping φ from the input space to the hidden space, where a nonlinear relation between the input x and the output y becomes a linear relation (dashed line) between the hidden variables h and y. 4 / 25

If the relation between x and y is highly nonlinear, then one must use more hidden neurons, i.e. increase the dimension of the hidden space, before one can find a good linear regression fit between the hidden space variables and the output variables. As the parameters w ji, w kj, b j and b k are solved together, nonlinear optimation is involved, leading to local minima, which is the main disadvantage of the MLP NN. The kernel methods follow a somewhat similar procedure: In the first stage, a nonlinear function φ maps from the input space to a hidden space, called the feature space. In the second stage, one performs linear regression from the feature space to the output space. 5 / 25

Instead of linear regression, one can also perform linear classification, PCA, CCA, etc. during the second stage. Like the radial basis function NN with non-adaptive basis functions (Book, Sect. 4.6) (and unlike the MLP NN with adaptive basis functions), the optimization in stage two of a kernel method is independent of stage 1, so only linear optimization is involved, and there are no local minima a main advantage over the MLP. However, the feature space can be of very high (or even infinite) dimension. This disadvantage is eliminated by the use of a kernel trick, which manages to avoid the direct evaluation of the high dimensional function φ altogether. Hence many methods which were previously buried due to the curse of dimensionality (Book, p.97) have been revived in a kernel renaissance. 6 / 25

(Some kernel methods, e.g. Gaussian processes, use nonlinear optimization to find the hyperparameters, thus have the local minima problem.) 12.2 Advantages and disadvantages of kernel methods [Book, Sect. 7.5] Since the mathematical formulation of the kernel methods (Book, Sect. 7.2 7.4) is somewhat difficult, a summary of the main ideas of kernel methods and their advantages and disadvantages is given here. 7 / 25

First consider the simple linear regression problem with a single predictand variable y, y = i a i x i + a 0, (4) with x i the predictor variables, and a i and a 0 the regression parameters. In the MLP NN approach to nonlinear regression, nonlinear adaptive basis functions h j (also called hidden neurons) are introduced, so the linear regression is between y and h j, y = j a j h j (x; w) + a 0, (5) with typically h j (x; w) = tanh( i w ji x i + b j ). (6) 8 / 25

Since y depends on w nonlinearly (due to the nonlinear function tanh), the resulting optimization is nonlinear, with multiple minima in general. What happens if instead of adaptive basis functions h j (x; w), we use non-adaptive basis functions φ j (x), i.e. the basis functions do not have adjustable parameters w? In this situation, y = j a j φ j (x) + a 0. (7) e.g., Taylor series expansion with 2 predictors (x 1, x 2 ) would have {φ j } = x 1, x 2, x 2 1, x 1 x 2, x 2 2,.... The advantage of using non-adaptive basis functions is that y does not depend on any parameter nonlinearly, so only linear optimization is involved with no multiple minima problem. 9 / 25

The disadvantage is the curse of dimensionality (Book, p.97), i.e. one generally needs a huge number of non-adaptive basis functions compared to relatively few adaptive basis functions to model a nonlinear relation, hence the dominance of MLP NN models despite the local minima problem. The curse of dimensionality is finally solved with the kernel trick, i.e. although φ is a very high (or even infinite) dimensional vector function, as long as the solution of the problem can be formulated to involve only inner products like φ T (x )φ(x), then a kernel function K can be introduced K(x, x) = φ T (x )φ(x). (8) 10 / 25

The solution of the problem now only involves working with a very manageable kernel function K(x, x) instead of the unmanageable φ. From Book, Sect. 7.4, the kernel regression solution is of the form y = n α k K(x k, x), (9) k=1 where there are k = 1,..., n data points x k in the training dataset. The most commonly used kernel is the Gaussian kernel or radial basis function (RBF) kernel, ) K(x, x) = exp ( x x) 2, (10) 2σ 2 If a Gaussian kernel function is used, then y is simply a linear combination of Gaussian functions. 11 / 25

In summary, with the kernel method, one can analyze the structure in a high-dimensional feature space with only moderate computational costs as the kernel function gives the inner product in the feature space without having to evaluate the feature map φ directly. The kernel method is applicable to all pattern analysis algorithms expressed only in terms of inner products of the inputs. In the feature space, linear pattern analysis algorithms are applied, with no local minima problems since only linear optimization is involved. 12 / 25

Although only linear pattern analysis algorithms are used, the fully nonlinear patterns in the input space can be extracted due to the nonlinear feature map φ. The main disadvantage of the kernel method is the lack of an easy way to inversely map from the feature space back to the input data space a difficulty commonly referred to as the pre-image problem. This problem arises e.g. in kernel principal component analysis (kernel PCA, see Book, Sect. 10.4), where one wants to find the pattern in the input space corresponding to a principal component in the feature space. 13 / 25

In kernel PCA, where PCA is performed in the feature space, the eigenvectors must lie in the span of the data points in the feature space, i.e. n v = α i φ(x i ). (11) i=1 As the feature space F is generally a much higher dimensional space than the input space X and the mapping function φ is nonlinear, one may not be able to find an x (i.e. the pre-image ) in the input space, such that φ(x) = v. 14 / 25

Figure : Schematic diagram illustrating the pre-image problem in kernel methods. The input space X is mapped by φ to the grey area in the much larger feature space F. Two data points x 1 and x 2 are mapped to φ(x 1 ) and φ(x 2 ), respectively, in F. Although v is a linear combination of φ(x 1 ) and φ(x 2 ), it lies outside the grey area in F, hence there is no pre-image x in X, such that φ(x) = v. 15 / 25

12.3 Support vector regression (SVR) [Book, Sect. 9.1] Support vector regression is a widely used kernel method for nonlinear regression. In SVR, the objective function to be minimized is J = C N E[y(x n ) y dn ] + 1 2 w 2, (12) n=1 where C is the inverse weight penalty parameter, E is an error function and the second term is the weight penalty term. To retain the sparseness property of the SVM classifier, E is usually taken to be of the form E ɛ (z) = { z ɛ, if z > ɛ 0, otherwise. (13) This is an ɛ-insensitive error function, as it ignores errors of size smaller than ɛ. 16 / 25

Figure : The ɛ-insensitive error function E ɛ (z). Dashed line shows the mean absolute error (MAE) function. 17 / 25

Figure : A schematic diagram of support vector regression (SVR). Data points lying within the ɛ tube (i.e. between y ɛ and y + ɛ) are ignored by the ɛ-insensitive error function, so these points offer no support for the final solution. Hence data points lying on or outside the tube are called support vectors. For data points lying above and below the tube, their distances from the tube are given by the slack variables ξ and ξ. Data points lying inside (or right on) the tube have ξ = 0 = ξ. Those lying above the tube have ξ > 0 and ξ = 0, while those lying below have ξ = 0 and ξ > 0. 18 / 25

The regression is performed in the feature space, i.e. where φ is the feature map. y(x) = w T φ(x) + w 0, (14) With the kernel function K(x, x ) = φ T (x)φ(x ), and with a Lagrange multiplier approach, the solution can be written in the form N y(x) = (λ n λ n)k(x, x n ) + w 0. (15) n=1 All data points within the ɛ-tube have the Lagrange multipliers λ n = λ n = 0, so they fail to contribute to the summation in (15), thus leading to a sparse solution. 19 / 25

Support vectors are the data points which contribute in the above summation, i.e. either λ n 0 or λ n 0, meaning that the data point must lie either outside the ɛ-tube or exactly at the boundary of the tube. If K is the Gaussian or radial basis function (RBF) kernel (10), then (15) is simply a linear combination of radial basis functions centred at the training data points x n. SVR with the Gaussian kernel can be viewed as an extension of the RBF NN method (Book, Sect. 4.6). If the Gaussian or radial basis function (RBF) kernel (10) is used, then there are three hyperparameters in SVR, namely C, ɛ and σ (controlling the width of the Gaussian function in the kernel). 20 / 25

To find the optimal values of the hyperparameters, multiple models are trained for various values of the three hyperparameters, and upon evaluating their performance over validation data, the best estimates of the hyperparameters are obtained. The simplest way to implement this is a 3-D grid search for the 3 hyperparameters, but this can be computationally expensive. One can also use e.g. evolutionary computation to find the optimal hyperparameters. Inexpensive estimates of the hyperparameter values can be obtained following Cherkassky and Ma (2004). 21 / 25

2 (a) 2 (b) 1 1 y 0 y 0 1 1 2 2 0 1 2 x 0 1 2 x 2 (c) 2 (d) 1 1 y 0 y 0 1 1 2 2 0 1 2 x 0 1 2 x Figure : SVR applied to a test problem: (a) Optimal values of the hyperparameters C, ɛ and σ obtained from validation are used. The training data are the circles, the SVR solution is the solid curve and the 22 / 25

true relation is the dashed curve. (b) A larger C (i.e. less weight penalty) and a smaller σ (i.e. narrower Gaussian functions) result in overfitting. Underfitting occurs when (c) a larger ɛ (wider ɛ-tube) or (d) a smaller C (larger weight penalty) is used. The advantages of SVR over MLP NN are: (a) For given values of the hyperparameters, SVR trains significantly faster since it only solves sets of linear equations, hence is better suited for high-dimensional datasets, (b) SVR avoids the local minina problem from nonlinear optimization, and (c) the ɛ-insensitive error norm used by SVR is more robust to outliers in the training data than the MSE. 23 / 25

However, the 3-D grid search for the optimal hyperparameters in SVR can be computationally expensive. For nonlinear classification problems, the support vector machine (SVM) with Gaussian kernel has only 2 hyperparameters, namely C and σ (in contrast to SVR, where the 3 hyperparameters are C, σ and ɛ). SVM functions in Matlab SVM code for both classification and regression problems can be downloaded from LIBSVM (Chang and Lin, 2001). The code can be run on MATLAB, C++ and Java: www.csie.ntu.edu.tw/~cjlin/libsvm/ 24 / 25

MATLAB Bioinformatics toolbox has SVM for classification (not regression), called svmclassify: www.mathworks.com/help/toolbox/bioinfo/ref/ svmclassify.html References: Chang, C.-C. and Lin, C.-J. (2001). LIBSVM: A library for support vector machines. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvm. Cherkassky, V. and Ma, Y. Q. (2004). Practical selection of SVM parameters and noise estimation for SVM regression. Neural Networks, 17(1):113 26. 25 / 25