Support vector machines for classification of underwater targets in sidescan sonar imagery

Size: px
Start display at page:

Download "Support vector machines for classification of underwater targets in sidescan sonar imagery"

Transcription

1 Copy No. Defence Research and Development Canada Recherche et développement pour la défense Canada DEFENCE & DÉFENSE Support vector machines for classification of underwater targets in sidescan sonar imagery M. Couillard Defence R&D Canada CORA J. A. Fawcett Defence R&D Canada Atlantic V. L. Myers Defence R&D Canada Atlantic M. Davison University of Western Ontario Defence R&D Canada Atlantic Technical Memorandum DRDC Atlantic TM November 2008

2 This page intentionally left blank.

3 Support vector machines for classification of underwater targets in sidescan sonar imagery M. Couillard Defence R&D Canada CORA J. A. Fawcett Defence R&D Canada Atlantic V. L. Myers Defence R&D Canada Atlantic M. Davison University of Western Ontario Defence R&D Canada Atlantic Technical Memorandum DRDC Atlantic TM November 2008

4 Principal Author Original signed by M. Couillard M. Couillard Approved by Original signed by D. Hopkin D. Hopkin Head/Maritime Asset Protection Approved for release by Original signed by Ron Kuwahara for Calvin Hyatt Chair/Document Review Panel c Her Majesty the Queen in Right of Canada as represented by the Minister of National Defence, 2008 c Sa Majesté la Reine (en droit du Canada), telle que représentée par le ministre de la Défense nationale, 2008

5 Abstract High-frequency sidescan sonars have become a fundamental tool for modern mine hunting operations. They produce quality images of the seabed and a high probability of detection can be achieved. A crucial classification phase is then needed to accurately identify these contacts as harmless or as potential underwater mines. To ensure the security of the follow-on traffic, all mines have to be identified accurately. At the same time, to avoid delays in the minefield clearing operations, clutter should not be misclassified as mines. For this contact identification task, multiple classification tools are available. This technical memorandum focuses on a powerful classification tool: support vector machines. A comprehensive introduction to support vector machines is provided and their usefulness for the classification of underwater objects in sidescan sonar imagery is investigated. The database used in this study is made of real sidescan sonar images collected during the CITADEL sea trial held at the NATO Undersea Research Center in October It is shown that this classification tool yields excellent classification performances when shadow-based features are used. These performances increase significantly when highlight-based features are added. It is also shown that the Ridge regression approximation is faster than quadratic optimization for large dataset and yields a comparable performance. Résumé Les sonars hautes fréquences sont aujourd hui des outils essentiels pour les opérations modernes de chasse aux mines. Ils produisent des images de qualité du plancher océanique et permettent de réaliser des détections avec un haut degré de probabilité. Une phase de classification est alors nécessaire pour déterminer si les contacts obtenus sont inoffensifs ou s ils sont susceptibles d être des mines immergées. Afin que la sécurité des navires ultérieurs puisse être garantie, toutes les mines doivent être identifiées avec exactitude. D autre part, pour éviter de retarder les opérations de déminage, on ne doit pas confondre des objects inoffensifs, comme des pierres par exemple, avec une mine. Plusieurs outils de classification sont disponibles pour cette tâche d identification des contacts. Le présent document technique est consacré à un puissant outil de classification : les machines à vecteurs de support. Une introduction approfondie est présentée sur les machines à vecteurs de support ; le document examine également l utilité de celles-ci pour la classification des objets immergés dans les images de sonar à balayage latéral. La base de données utilisée dans la présente étude est constituée d authentiques images de sonar à balayage latéral recueillies lors de l essai marin CITADEL au NATO Undersea Research Centre, en octobre Il sera démontré que cet outil de classification présente une excellent capacité de classification lorsque sont utilisées les caractéristiques de classification axées sur l ombre acoustique. Cette performance est considérablement accrue lorsque DRDC Atlantic TM i

6 les caractéristiques de l écho acoustique sont également utilisées. Il sera également démontré que l approximation par méthode de régression Ridge est plus rapide que l optimisation quadratique lorsque les ensembles de données sont vastes, tandis que la performance de classification demeure comparable. ii DRDC Atlantic TM

7 Executive summary Support vector machines for classification of underwater targets in sidescan sonar imagery M. Couillard, J. A. Fawcett, V. L. Myers, M. Davison; DRDC Atlantic TM ; Defence R&D Canada Atlantic; November Background: High-frequency sidescan sonars have become a fundamental tool for modern mine hunting operations. They produce quality images of the seabed and a high number of contacts can be detected. A crucial classification phase is then needed to accurately identify these contacts as harmless or as potential underwater mines. For this contact identification task, multiple classification tools are available. Among these classification tools, a powerful one, support vector machines, has recently received a lot of attention in various fields such as data mining, face recognition, medical detection of micro calcifications and speech recognition. One of the objectives of this technical memorandum is to provide a comprehensive introduction to support vector machines. Also, their usefulness for the classification of underwater objects is investigated. Principal results: Support vector machines are used to discriminate mine shapes from clutter in sidescan sonar imagery. The database is made of real sidescan sonar images of mine shapes and various clutter collected during the 2005 CITADEL sea trial. These objects are first described only by shadow-based features and the impact of adding highlight-based features on the classification performance is then quantified. It was found that this classification tool yields excellent classification performances. It was also found that the Ridge regression approximation is faster than quadratic optimization for large dataset and yields a comparable performance. For our mines versus clutter classification problem, the best choices of support vector kernels are the Gaussian kernel with σ = 10 and the distance kernel with exponent 1 2. With a Gaussian kernel, the Ridge regression approximation and only the shadow based features, a hit rate of 70.0% is reached at a false alarm rate of 29.4%. Adding the highlight based features significantly increases the classification performance yielding a hit rate of 70.0% for a very small false alarm rate of 5.2%. With the mine shapes taken individually, the complete shadow and highlight feature sets produced outstanding classification results. At a 90.0% hit rate, we obtain false alarm rates of 10.0%, 12.5% and 19.0% respectively for the Manta, cylinder and Rockan shapes. Significance of results: This study provides a comprehensive introduction to support vector machines and shows that they are excellent classifiers that should be considered for computer-aided classification software designed for mine hunting. Support DRDC Atlantic TM iii

8 vectors are an efficient approach to large training sets as only a limited number of training points carrying all relevant information for the classification problem are used. The use of kernel functions also allows for non-linear decision boundaries in the original feature space by implicitly mapping the training data into a higherdimensional feature space. This classification performance study is more detailed and complete than shorter studies recently published and shows that highlight based features should be used in addition to the more traditionally accepted shadow-based features. Future work: In the future we would like to investigate the selection of feature sets and the classifier parameters which make the classification process more robust to limited training sets and different seabed conditions. For example, one would hope that the classifier would be able to correctly classify a cylindrical object which has slightly different dimensions than the cylinders/ in the training set. We also hope to apply kernel-based classification methods to other image and signal detection and classification problems. iv DRDC Atlantic TM

9 Sommaire Support vector machines for classification of underwater targets in sidescan sonar imagery M. Couillard, J. A. Fawcett, V. L. Myers, M. Davison ; DRDC Atlantic TM ; R & D pour la défense Canada Atlantique ; novembre Contexte : Les sonars hautes fréquences sont aujourd hui des outils essentiels pour les opérations modernes de chasse aux mines. Ils produisent des images de qualité du plancher océanique et permettent de réaliser des détections avec un haut degré de probabilité. Une phase de classification est alors nécessaire pour déterminer si les contacts obtenus sont inoffensifs ou s ils sont susceptibles d être des mines immergées. Plusieurs outils de classification sont disponibles pour cette tâche d identification des contacts. Les machines à vecteurs de support, qui comptent parmi les plus puissants de ces outils de classification, ont récemment reçu beaucoup d attention dans différents domaines tels que l exploration de données, la reconnaissance faciale, le dépistage médical de la microcalcification, ainsi que la reconnaissance de la parole. L un des objectifs de ce document technique est de présenter une introduction approfondie sur les machines à vecteurs de support. De plus, le document examine leur utilité pour la classification des objets immergés. Principaux résultats : Les machines à vecteurs de support sont utilisées pour établir une distinction entre les formes de mines et les objets inoffensifs (appelés clutters) dans les images de sonar à balayage latéral. La base de données est constituée d authentiques images de sonar à balayage latéral représentant des formes de mines et différents exemples de clutters. Ces images ont été recueillies lors de l essai marin CITADEL de Ces objets sont d abord décrits uniquement en termes de caractéristiques axées sur l ombre acoustique ; on examine ensuite l incidence de l inclusion des caractéristiques de l écho acoustique sur la performance de la classification. Il a été constaté que cet outil offre une excellente performance de classification. Il a également été constaté que l approximation par méthode de régression Ridge est plus rapide que l optimisation quadratique lorsque les ensembles de données sont vastes, tandis que la performance de classification demeure comparable. En appliquant un noyau gaussien, l approximation par méthode de régression Ridge et uniquement les caractéristiques axées sur l ombre acoustique, on obtient un taux de classification de 70,0%, pour un taux de fausse alarme de 29,4%. Si l on tient compte des caractéristiques de l écho acoustique, la performance de la classification est considérablement accrue : 70,0%, pour un très faible taux de fausse alarme : 5,2%. Si l on considère chacune des formes de mine individuellement, l ensemble complet de caractéristiques d ombre et d écho donne d excellents résultats de classification : pour DRDC Atlantic TM v

10 un taux de classification de 90.0%, les taux de fausse alarme sont respectivement de 10,0%, 12,5% et 19,0% pour les mines Manta, cylindriques et Rockan. Importance des résultats : L étude présente une introduction approfondie sur les machines à vecteurs de support et démontre que celles-ci constituent d excellents classificateurs qui devraient être envisagés pour les logiciels de classification assistée par ordinateur destinés à la chasse aux mines. L approche des vecteurs de support est efficace pour les vastes ensembles d apprentissage : elle n utilise que quelques points d apprentissage porteurs de l ensemble des renseignements pertinents pour la classification. L usage de fonctions noyaux rend également possible des limites décisionnelles non linéaires dans l espace des caractéristiques initiales grâce à une mise en correspondance implicite des données d apprentissage dans un espace de caractéristiques comportant davantage de dimensions. La présente étude sur la performance de classification est plus détaillée et plus complète que les études plus brèves publiées récemment ; elle indique que les caractéristiques d écho acoustique doivent être utilisées en plus des caractéristiques d ombre acoustique, dont l usage est traditionnellement plus répandu. Perspectives : Nous souhaiterions étudier ultérieurement la méthode de sélection des ensembles de caractéristiques et les paramètres du classificateur qui rendent le processus de classification plus robuste en présence d ensembles d apprentissage plus limités et dans des conditions de planchers océaniques différents. Par exemple, nous pourrions tenter d obtenir un classificateur qui est à même de classifier correctement un objet cylindrique dont les dimensions sont légèrement différentes de celles des cylindres de l ensemble d apprentissage. Nous espérons également appliquer les méthodes de classification fondées sur un noyau à d autres problèmes de détection et de classification d images et de signaux. vi DRDC Atlantic TM

11 Table of contents Abstract i Résumé i Executive summary Sommaire Table of contents List of figures iii v vii ix 1 Introduction Classification with support vector machines Support vector machines Linearly separable problems Non-linearly separable problems Kernels An example in one dimension Linear regression approximations Ridge regression Kernel partial least squares Fisher discriminant Classification of underwater targets Sonar images and feature extraction Quadratic optimization vs least square regression Performance of various kernels Classification performance for various underwater targets Summary DRDC Atlantic TM vii

12 References Annex A: Additional figures viii DRDC Atlantic TM

13 List of figures Figure 1: Hard margin decision boundary Figure 2: Soft margin decision boundary Figure 3: Discriminant function with Gaussian kernel Figure 4: Types of objects used for the Citadel trial Figure 5: High frequency sidescan sonar image Figure 6: Sonar images of cylinder shapes Figure 7: Quadratic optimization results for various values of C, Gaussian kernel Figure 8: Quadratic optimization vs Ridge regression, Gaussian kernel Figure 9: Results from various least square regression techniques, Gaussian kernel Figure 10: Gaussian kernel results for various values of σ Figure 11: Gaussian kernel vs distance kernels Figure 12: All mines vs clutter Figure 13: Cylinder shapes vs clutter Figure 14: Manta shapes vs clutter Figure 15: Rockan shapes vs clutter Figure A.1: Gaussian kernel results for various values of σ Figure A.2: Gaussian kernel vs homogeneous polynomial kernels Figure A.3: Gaussian kernel vs non homogeneous polynomial kernels Figure A.4: Gaussian kernel vs cosine and Minkowski kernels Figure A.5: Gaussian kernel vs other types of kernels DRDC Atlantic TM ix

14 This page intentionally left blank. x DRDC Atlantic TM

15 1 Introduction High-frequency sidescan sonars have become a fundamental tool for modern mine hunting operations. They produce quality images of the seabed and a high number of contacts can be detected. A crucial classification phase is then needed to accurately identify these contacts as harmless or as potential underwater mines. To ensure the security of the follow-on traffic, all mines have to be identified accurately. At the same time, to avoid delays in the minefield clearing operation, clutter should not be misclassified as mines. For this contact identification task, multiple classification tools are available [1]. This technical memorandum focuses on a powerful classification tool: support vector machines. This classification tool is already used for various tasks such as data mining [2], face recognition [3], medical detection of micro calcifications [4], and speech recognition [5]. One of the objectives of this technical memorandum is to provide a comprehensive introduction to support vector machines classifiers. Their usefulness for the classification of underwater objects in sidescan sonar imagery is also investigated. Other studies used this tool for computer-aided classification of underwater objects [6],[7] [8]. This memorandum provides a more detailed and complete classification performance analysis. Section 2 describes the fundamental principles of support vector machines and presents the main elements that made this classification tool so powerful. A 2-class problem is used to derive the support vector problem for the case of linearly separable classes and then for the more realistic case of non-linearly separable classes. The use of kernel functions for higher-dimensional mapping is also described. A simple onedimensional example is used to illustrate the discriminant function obtained with support vectors. Section 2 concludes with a description of linear regression approximation techniques used to decrease the time needed to obtain the support vector weights for large problems. In Section 3, the utility of support vector machines for the classification of underwater targets in sidescan sonar imagery is explored. The database is made of mine shapes and various clutter first described only by shadow-based features. Results obtained with quadratic optimization and linear regression are compared and the classification performance of various kernels is investigated. Finally, using support vector machines, the impact of adding highlight-based features on the classification performance is quantified. DRDC Atlantic TM

16 2 Classification with support vector machines In this section, we first derive the support vector problem for linearly separable classes and then for the more realistic case of non-linearly separable classes. The use of kernel functions for higher-dimensional mapping is also described. A simple onedimensional example is used to illustrate the discriminant function obtained with support vectors. We conclude this section with a description of linear regression approximation techniques used to decrease the time needed to obtain the support vector weights for large problems. 2.1 Support vector machines In this section, we use the following pattern classification terms. The feature space refers to the R n space (n-dimensional, real-valued number space) where each object to be classified is represented as a n-dimensional point. The dimension n is determined by the number of features used to characterize each object. Features are specific attributes of an object that describe it or set it apart from similar objects. For example, features could include quantities such as length, width or area. A label is an integer associated to each known object and identifying the class to which these objects belong. A discriminant function is a function taking the feature values as input and combining them to yield a classification label. A training set is a subset of all the available objects. The training set is used to create the discriminant function. A testing set is a subset of all the available objects which do not belong to the training set. The testing set is used to evaluate the classification performance of the discriminant function Linearly separable problems When faced with a 2-class classification problem, a common approach [9] [10] [11] is to find an hyperplane that separates the two classes in the feature space. Let {x 1,..., x n } be our training set with associated labels y i { 1, 1}, i =1...n, the Equation of this hyperplane is given by: w T x + b =0, w R n, b R (1) 2 DRDC Atlantic TM

17 Among all hyperplanes separating the classes, the optimal hyperplane is the one maximizing the distance, or margin, between the two classes. Such an hyperplane for linearly separable classes is called a hard margin hyperplane. Under the assumption of linearly separable classes, there exists an hyperplane separating the two classes perfectly. To find the optimal values of the parameters w and b, we can rescale them such that the points closest to the optimal hyperplane satisfy w T x i + b = ±1. The optimal hyperplane will be the one maximizing the distance between the hyperplane w T x + b = 1 and the hyperplane w T x + b = 1. The distance between these two hyperplanes is given by 2. Figure 1 illustrates these principles. w margin w T x + b = 1 ω 2 ω 1 w T x + b = 1 w T x + b = 0 Figure 1: Hard margin decision boundary. Therefore, to maximize the distance between the two classes, one needs to minimize w, orequivalently w 2. This can be written as a quadratic optimization problem: Minimize 1 2 w 2 (2) Subject to y i (w T x i + b) 1 i (3) The constraint can be written as: 1 y i (w T x i + b) 0 i (4) DRDC Atlantic TM

18 The solution vector w has an expansion w = s i=1 v ix i in terms of a subset of the training patterns falling on the hyperplanes w T x i + b = ±1 [11]. These training patterns carry all the information relevant to our classification problem and they are called the support vectors. This is one of the fundamental advantages of support vector machines: when faced with a large number of training points, the use of these support vectors can significantly reduce the complexity of the classification problem. Quadratic optimization problems such as the one defined by Equations 2 and 4 can easily be solved with commercially available solvers. Using w 2 = w T w,thelagrangian for this constrained optimization is: L (w,α,b)= 1 2 wt w + n ( α i 1 yi (w T x i + b) ) (5) i=1 Differentiating with respect to the primal variables w and b and setting to 0 we get: and w = n α i y i x i (6) i=1 n α i y i =0 (7) i=1 By substituting Equation 6 in the Lagrangian 5, we obtain after simplification: L = 1 2 n i=1 n α i α j y i y j x T i x j + j=1 n α i (8) Our new objective function is now in terms of α i only and it is this form of the optimization problem that we solve in S ection 3. This is known as the dual problem and it can be shown using Karush-Kuhn-Tucker theory [1] that instead of minimizing, we must instead maximize: i=1 Maximize W (α) = 1 2 n n α i α j y i y j x T i x j + i=1 j=1 n α i (9) i=1 4 DRDC Atlantic TM

19 Subject to n α i y i =0, α i 0 (10) i=1 The vector w can be recovered with: w = n α i y i x i (11) i=1 When testing a new data point z, the discriminant function f(z) isgivenby: Primal form: f(z) =w T z + b (12) Dual form: f(z) = x k S α k y k x T k z + b (13) Here S is the set of support vectors. The bias parameter b is obtained from: n b = y i y j α j x T j x i (14) j=1 Point z is classified as belonging to class one if f(z) 0andtoclasstwoiff(z) < Non-linearly separable problems In the previous section, we assumed our training points were linearly separable. For the majority of real life classification problems, such a perfect separation of the two classes is not possible. The optimal hyperplane separating the classes has to allow for misclassified points. To allow for the possibility of having training points violating constraint 3, we introduce slack variables ɛ i, i =1,..., n [10]. The hyperplane obtained with such an approach is called a soft margin hyperplane. Figure 2 illustrates these principles. The slack variables ɛ i can be computed by: DRDC Atlantic TM

20 x i ε i margin w T x + b = 1 ω 2 ω 1 w T x + b = 1 ε j x j w T x + b = 0 Figure 2: Soft margin decision boundary. y i ( w T x i + b ) 1 ɛ i, ɛ i 0 i (15) We also introduce a trade off parameter C>0 between the error and the margin. The optimization problem in this case becomes: Minimize 1 2 w 2 + C n ɛ i (16) i=1 Subject to y i ( w T x i + b ) 1 ɛ i, ɛ i 0 i (17) As in the previous section, the dual problem associated with this optimization is given by: Maximize W (α) = 1 2 n n α i α j y i y j x T i x j + i=1 j=1 n α i (18) i=1 6 DRDC Atlantic TM

21 The constraints are now: 0 α i C, n α i y i = 0 (19) i=1 As before, w can be recovered with Equation 11. The discriminant function and the bias terms are given by Equations 13 and 14. In Section 3, we use soft margins classifiers for our study of the classification of underwater targets. We also study the impact of the value of the trade off parameter C on the classification performance. 2.2 Kernels In the original feature space, separating the classes with an hyperplane might not yield the optimal classification performance as the true optimal decision boundaries might be non-linear. In Section 2.1, we presented the first fundamental principle of support vector machines: not all training points are used to define the discriminant function. In this section, we explore a second fundamental concept of support vector machines. To account for the possibility of non-linear decision boundaries in the original feature space, we map the training data into a higher-dimensional feature space using a non-linear transformation: Φ:R N F (20) x i Φ(x i ) We then find an optimal hyperplane in this higher-dimensional feature space [9] [12] [13]. In the optimization problem defined by Equations 18 and 19, the training points only appear as the inner products x T i x j with i, j =1...n. These inner products are now replaced by Φ(x i ) T Φ(x j ). Clearly, if the mapped feature space F is high-dimensional, the mapping and the computation of these inner products can be computationally expensive. To solve this problem, we introduce the notion of kernel function. A kernel is defined as: κ(x, y) =Φ(x) T Φ(y) (21) These kernel functions are quite useful as they allow us to compute the inner products of the mapped features in the original feature space. As long as we can do this, we do DRDC Atlantic TM

22 Table 1: Types of kernels Kernel type κ(x, y) Gaussian Homogeneous polynomial Non homogeneous polynomial Distance Homogeneous Cosine Minkowski All subsets exp( x y 2 ) 2σ 2 (x T y) ( d 1+x T y ) d x y d sign(x T y)(x T y) d 1 x T y x y ( x y d ) 1 d n i=1 (1 + x iy i ) not need the higher-dimensional mapping explicitly. Examples of kernels commonly used are given in table 1. Each of these kernel functions is associated with a specific mapping Φ. For example, with x =[x 1,x 2 ], the non homogeneous polynomial kernel of degree two corresponds to a mapping Φ(x) = ( 1, 2x 1, 2x 2, 2x 1 x 2,x 2 1,x2 2) : Φ(x) T Φ(y) = 1+2x 1 y 1 +2x 2 y 2 +2x 1 x 2 y 1 y 2 + x 2 1y1 2 + x 2 2y2 2 (22) = (1+x 1 y 1 + x 2 y 2 ) 2 = ( 1+x T y ) 2 We now need to update our optimal hyperplane problem to take into account the use of a kernel function. Replacing all inner products by κ (x i, x j ), the optimization problem 18, 19 becomes: Maximize W (α) = 1 2 n n α i α j y i y j κ (x i, x j )+ i=1 j=1 n α i (23) i=1 Subject to 0 α i C, n α i y i = 0 (24) i=1 8 DRDC Atlantic TM

23 The equation to obtain w becomes: w = n α i y i Φ(x i ) (25) i=1 Our new discriminant function f(z) is given by: f(z) =w T Φ(z)+b = x k S α k y k κ (x k, z)+b (26) Here S is the set of support vectors and the bias parameter b is obtained from: b = y i n y j α j κ(x i, x j ) (27) j=1 The choice of the best kernel to use for classification depends on the nature of the problem. In Section 3, we investigate different types of kernel for classification of underwater targets. 2.3 An example in one dimension To illustrate the utility of support vector machines for classification, we construct a simple one-dimensional example. Constructing an hyperplane using a kernel function leads to a non-linear decision boundary in the original feature space. The first two columns of table 2 contain the data points x i and their associated labels y i.weusea Gaussian kernel as defined in table 1 with σ = 3. Quadratic optimization is used to solve the system defined by Equations 23 and 24. We use C = 10. These parameter values for the optimization are determined by numerical experimentation. In general, one might wish to use a portion of the training set to determine the parameter values which minimize the classification error. The weights α i obtained are displayed in the third column of Table 2. Note that out of the six data points, only four are used as support vectors. These four points are used in Equation 26 to obtain a discriminant function for new test points z. The bias term is obtained with Equation 27. Values of this discriminant function with respect to z and the decision boundaries are shown in Figure 3. The decision boundaries lie where f(z) =0. Forf(z) 0, the data points are classified as having label y = 1 and for f(z) < 0, the associated label is y = 1. We see that all training points are correctly classified. DRDC Atlantic TM

24 Table 2: An example x i y i α i Discriminant function with Gaussian kernel ω 1 ω 2 ω 1 0 f(z) f(z) ω 1, y = 1 ω 2, y = z Figure 3: Discriminant function with Gaussian kernel. 2.4 Linear regression approximations When using large training sets with many features, the quadratic optimization of the support vector weights can take an excessive amount of time. In this section, we briefly describe three alternative techniques to approximate the support vector weights α i in the dual discriminant function: f(z) = n α i κ(x i, z)+b (28) i=1 For large problems, these regression techniques are faster than quadratic optimization. Their performance for the classification of underwater targets is discussed in Section DRDC Atlantic TM

25 2.4.1 Ridge regression A first technique to approximate the support vector weights is based on a least squares approach and is called Ridge regression [9]. Using κ as our kernel matrix and y as the column vector containing the training labels, the weights α i are obtained with [14]: α =(κ + λi n ) 1 y (29) Here λ is a regularization parameter used to ensure that the kernel matrix is not singular. Typically, a value of λ =0.01 is used Kernel partial least squares A second technique to estimate the weights α i in the discriminant function is the kernel partial least squares approach. A detailed description of partial least squares is outside the scope of this technical memorandum, but for more details the reader can refer to [9]. The fundamental principle of partial least square regression is to use the covariance of the training features with the associated labels to reduce, or deflate, the feature matrix and then perform least squares regression in this reduced feature space. This reduction of the feature space can be quite useful when we use a kernel function to map the features in a higher-dimensional space where the new coordinates can be correlated. Instead of using the original feature matrix, we use the kernel matrix and this matrix is deflated in the directions of maximum covariance. We first define κ 0 = κ and ŷ = y. Also, β j are the projection directions, c j are the output vectors and τ j are the scaled output vectors. For a fixed number of iterations j =1,..., n, the deflation of the kernel matrix is obtained with the following algorithm taken from [9]: β j = ŷ ŷ (30) τ j = κ j 1 β j (31) c j = ŷt τ j τ j 2 (32) ŷ =ŷ τ j c T j (33) DRDC Atlantic TM

26 κ j = ( I τ j τ T j / τ 2 j ) κj 1 ( I τj τ T j / τ 2 j ) (34) After completion of the n iterations, we can construct matrices B and Z: Finally, the weights α i are computed as: B =[β 1,..., β n ], Z =[τ 1,..., τ n ] (35) Fisher discriminant α = B(Z T κb) 1 Z T y (36) Our third technique to estimate the support vector weights is the Fisher discriminant approach. In the primal form of the discriminant function 12, this approach finds the direction w of the optimal hyperplane by maximizing the ratio [9]: J(w) = (μ+ w μ w )2 (σ + w) 2 +(σ w) 2 (37) Here μ + w is the mean of the projection of the training points with positive labels onto the direction w and μ w is the mean of the projection of the training points with negative labels. σ w + and σ w are the associated standard deviations. The use of this ratio ensures that the chosen direction w maximizes the separation of the means of the two classes scaled according to their variances in that direction. A more detailed description of the Fisher discriminant can be found in [1] and [9]. To find the weights α i for the dual form of the discriminant function 28, we first define n as the total number of training points, n + as the number of training points with associated label y i =1andn as the number of training points with associated label y i = 1. Of course, n = n + + n. To find α i, we define a diagonal matrix D, and two matrices C + and C : { 2n D ii = /n if y i =1 2n + /n if y i = 1 (38) 12 DRDC Atlantic TM

27 C + ij = { 2n /(nn + ) if y i = y j =1 0 otherwise (39) { 2n C ij = + /(nn ) if y i = y j = 1 0 otherwise (40) We then create a matrix B, givenbyb = D C + C. Finally, the weights α i can be obtained with : α =(Bκ + λi) 1 y (41) 3 Classification of underwater targets In this section, support vector machines are used for the classification of underwater targets in sidescan sonar imagery. Our database is made of bottom influence mine shapes and various harmless objects like rocks or seabed features (referred to as clutter) first described only by shadow-based features. We compare results obtained with quadratic optimization and linear regression and we investigate the classification performance of various kernels. Finally, using support vector machines, we quantify the impact of adding highlight-based features on the classification performance. 3.1 Sonar images and feature extraction Having introduced support vector machines in the previous section, we now explore their utility for the classification of underwater targets in sidescan sonar imagery. The data used in this study were gathered during the Citadel sea trial which took place during October 2005 at the NATO Undersea Research Center in La Spezia, Italy. Several targets designed to mimic various bottom influence mine types, as well as rocks, were deployed. Sonar images of these objects were obtained. The main asset used during this trial was the Dorado semi-submersible vehicle which towed a commercially available Klein 5500 sidescan sonar. This sonar operates at a centre frequency of 455 khz with a 20 khz bandwidth, giving an image with a nominal resolution of 0.10 m in the along-track by m in the across-track directions. Pictures of the deployed objects are shown in Figure 4. Figure 5 shows an example of a sonar image collected during the trial. The arrow in Figure 5 identifies a cylindrical mine shape. DRDC Atlantic TM

28 Cylinder shape Manta shape Rockan shape Rock Figure 4: Types of objects used for the Citadel trial. Figure 5: High frequency sidescan sonar image. The arrow identifies a cylindrical mine shape. 14 DRDC Atlantic TM

29 The objects are grouped into two classes. The first class is identified as clutter and includes 1080 images. This class includes unidentified natural clutter found while conducting the Citadel trial as well as 93 images of the rocks deployed during the trial. The second class is identified as mine shapes and includes 310 images. Among these mine images, we have 112 images of cylinder shapes, 99 images of Manta shapes and 99 images of Rockan shapes. Figure 6 shows examples of sonar images for the three mine shapes and also the rocks. Cylinder shape Manta shape Rockan shape Rock Figure 6: Sonar images of cylinder shapes. For our classification task, we are not using the sonar images directly. From these images, we extract multiple features. The first step of the feature extraction process is called the segmentation. Several methods exist for segmenting sonar images into highlight, shadow and background regions. Most of these segmentation algorithms employ some type of Markov Random Field, for instance those described by Reed [15] and Mignotte [16]. Our segmentation method is based mostly on work by Myers [17]. Myers algorithm employs a fuzzy decision function to determine whether a pixel belongs to the shadow or background region, where a pixel s class is flipped depending on its neighbours classes. This is sometimes called a relaxation method. The algorithm requires an initial image thresholded into shadow and background with which to perform the iterative algorithm. For our classification task, the shadow threshold was chosen at the 10th percentile of the cumulative image histogram while the highlight was thresholded at the 95th percentile. DRDC Atlantic TM

30 Once the shadow and highlight regions are obtained, we compute multiple features to represent each image in a multi-dimension feature space. The complete lists of shadow and highlight features used are given in tables 3 and 4. In Sections 3.2 and 3.3, we only use shadow based features. In Section 3.4, we include the highlight features to determine their impact on the classification performance. Table 3: Shadow Features Area Ratio area / best ellipse fit area Ratio perimeter of shadow / (perimeter of convex hull + 1) Sine of the orientation of the leading edge Sine of the orientation of the trailing edge Sine of the orientation of the best ellipse fit Estimated length of object from shadow (ellipse filtered) Estimated length of object from shadow Height profile Ratio standard deviation of height profile / mean height Maximum value of height profile Maximum value of height profile (ellipse filtered) Ratio major axis of ellipse / (minor axis of ellipse + 1) Speckle contrast (standard deviation of pixel amplitudes / mean of pixel amplitudes) Table 4: Highlight Features Area Ratio area / best ellipse fit area Estimated length of object from highlight Estimated length of object from highlight (ellipse filtered) Estimated across track extent of highlight (ellipse filtered) Sine of the orientation of the best ellipse fit Ratio major axis of ellipse / (minor axis of ellipse + 1) Speckle contrast (standard deviation of pixel amplitudes / mean of pixel amplitudes) 16 DRDC Atlantic TM

31 3.2 Quadratic optimization vs least square regression Now that we have defined our feature sets, we can proceed to the classification task. The clutter are identified with labels y i = 1 and the mine shapes are associated with labels y i = 1. The complete set of objects is randomly partitioned in a training set and a testing set. To ensure that the training will not be biased towards one of the classes, the same number of samples from each class is used in the training set. This number is chosen as 50% of the minimum between the number of samples of the first class and second class. All objects not selected in the training set are used in the testing set. The training set is used to obtain the support vector weights in the discriminant function. This discriminant function is then used to obtain predicted labels for the objects in the testing set. Relating these predicted labels to the true labels, we obtain two measures of performance: the hit rate and the false alarm rate. The hit rate is the fraction of mines correctly classified as mines and the false alarm rate is the fraction of clutter misclassified as mines. These classification performance measures are displayed as a receiver operating characteristic (ROC) curve where the hit rate is plotted as a function of the false alarm rate [1]. The standard threshold of the discriminant function is f(z) =T =0. Forf(z) T, the object is classified as belonging to the class 1 and for f(z) <T, the object is classified as belonging to class 1. This T = 0 threshold is but one point on our ROC curves. To construct a complete ROC curve, we increase the threshold value from negative values to positive values. Each point on the ROC curve is therefore associated with a given threshold value. To obtain smooth ROC curves, the random partitioning of the training/testing sets has to be repeated about 100 times and the hit and false alarm rates are obtained over all the testing sets realisations. We first use the soft margin approach as discussed in Section and we obtain the support vector weights with quadratic optimization. A Gaussian kernel with σ =10 is used. This choice of kernel is justified in Section 3.3. The value of the trade off parameter C in Equation 19 has an impact on the classification performance and Figure 7 shows the results obtained for various values of C. The best performance is achieved around C = 100. With this value of C, the hit rate reaches 70.0% for a false alarm rate of 29.5%. This indicates that our support vector classifier is producing satisfactory results when differentiating mines from clutter. One drawback of using quadratic optimization on a large dataset is the long time it takes to obtain the support vector weights. In addition, for our classification performance study, this optimization time is multiplied by the number of random training sets used. Section presented an alternative to approximate the support vector weights by using Ridge regression. Ridge regression is faster, but does it decrease the classification performance? Figure 8 compares the classification performance of the quadratic optimization approach with the performance of the Ridge regression approach. We see that the Ridge regression classifier yields a classification performance DRDC Atlantic TM

32 comparable to the quadratic optimization approach: a hit rate of 70.0% is obtained for a false alarm rate of 29.4%. We now examine the relative performance of the two linear regression techniques presented in Sections and Figure 9 compares the performance of the Ridge regression with the performances of the kernel partial least square and of the Fisher discriminant approaches. We see that the partial least square approach produces results comparable to the Ridge regression, and therefore to the quadratic optimization approach. However, the Fisher discriminant performance is lower reaching a hit rate of 70.0% at a false alarm rate of 37.0%. As the Ridge regression classifier is faster and produces satisfactory classification results, we will use this approach to obtain our ROC curves in Sections 3.3 and ROC Curve Quadratic Optimization Hit False Alarm C = 10 C = 50 C = 100 C = 200 Figure 7: Quadratic optimization results for various values of C, Gaussian kernel. 18 DRDC Atlantic TM

33 1 ROC Curve Quadratic optimization vs Ridge regression Hit False Alarm Quadratic optimization Ridge regression Figure 8: Quadratic optimization vs Ridge regression, Gaussian kernel. 1 ROC Curve SV regression Hit False Alarm PLS Fisher Ridge Figure 9: Results from various least square regression techniques, Gaussian kernel. DRDC Atlantic TM

34 3.3 Performance of various kernels Various types of kernel are defined in table 1, but which ones are the best for the classification of underwater targets? Using these kernels and various values of their parameters, we obtained a series of ROC curves showing the various classification performances for the case mine shapes versus clutter. Figure 10 shows the performance of a Gaussian kernel for various values of the parameter σ. The optimal classification performance is obtained around σ = 10. Performances of additional values of σ are shown in the appendix Figure A.1. Figure 11 and appendix Figures A.2 to A.5 show the results obtained for the other types of kernel. We see that for our mine classification problem, the best choices of kernels are the Gaussian kernel with σ =10andthe distance kernel with exponent 1. Some kernels like the non-homogeneous polynomial 2 kernel of degree 3 and the all subsets kernel yield very poor performances equivalent to random guessing. 1 ROC Curve Gaussian Kernel Hit False Alarm σ = 10 σ = 30 σ = 50 σ = 70 σ = 100 Figure 10: Gaussian kernel results for various values of σ. 20 DRDC Atlantic TM

35 1 ROC Curve Gaussian Kernel vs Distance Kernels Hit False Alarm Distance 1 Distance 2 Distance 0.5 Gaussian Figure 11: Gaussian kernel vs distance kernels. 3.4 Classification performance for various underwater targets Having identified the Gaussian kernel with σ = 10 as being one of the best kernels and knowing that the Ridge regression classifier produces satisfactory results, we now extend our analysis to include the highlight-based features. Figure 12 compares the ROC curve obtained with only the shadow based features, to the ROC curve obtained when adding the highlight based features to the shadow based ones. We see that adding the highlight features produces a significant increase in classification performance. With only the shadow features, a hit rate of 70.0% is reached at a false alarm rate of 29.4%, but with the highlight features, a hit rate of 70.0% is reached at a very small false alarm rate of 5.2%. Such a classification performance is excellent. In the previous analysis, the three mine shapes were grouped in one class. What kind of classification performance can we achieve for each type of mine shapes taken individually? Figures 13 to 15 summarize the results for the three classification problems: cylinders versus clutter, Mantas versus clutter and Rockans versus clutter. Once again, we see that using highlight based features in addition to the shadow based features increases the classification performance significantly. Using all features, we obtain outstanding classification performances with our support vectors approach. Looking this time at a 90.0% hit rate, we obtain false alarm rates of 10.0%, 12.5% and 19.0% respectively for the Mantas, cylinders and Rockans cases. DRDC Atlantic TM

36 1 ROC Curve Mine shapes vs Clutters Hit False Alarm Shadow only Shadow and highlight Figure 12: All mines vs clutter. 1 ROC Curve Cylinder shapes vs Clutters Hit False Alarm Shadow only Shadow and highlight Figure 13: Cylinder shapes vs clutter. 22 DRDC Atlantic TM

37 1 ROC Curve Manta shapes vs Clutters Hit False Alarm Shadow only Shadow and highlight Figure 14: Manta shapes vs clutter. 1 ROC Curve Rockan shapes vs Clutters Hit False Alarm Shadow only Shadow and highlight Figure 15: Rockan shapes vs clutter. DRDC Atlantic TM

38 4 Summary This technical memorandum investigated the utility of support vector machines for the classification of underwater targets. Support vectors are an efficient approach to large training sets as not all training points are used to define the discriminant function. Only the training points falling on the hyperplanes bounding the margin are used as they carry all relevant information for the classification problem. The use of kernel functions also allows for non-linear decision boundaries in the original feature space by mapping the training data into a higher-dimensional feature space where an hyperplane can be found. This mapping is not needed explicitly as kernel functions compute the inner products of the mapped features in the original feature space. Support vector machines were used to discriminate mine shapes from clutter in sidescan sonar imagery. This classification tool yielded excellent classification performances. It was found that the Ridge regression approximation is faster than quadratic optimization and yields a comparable performance. The partial least square approach also produces results comparable to the Ridge regression and quadratic optimization. The performance of the Fisher discriminant was however lower. For our mine classification problem, the best choices of kernels are the Gaussian kernel with σ =10and the distance kernel with exponent 1. With a Gaussian kernel, the Ridge regression 2 approximation and only the shadow based features, a hit rate of 70.0% is reached at a false alarm rate of 29.4%. Adding the highlight based features significantly increased the classification performance yielding a hit rate of 70.0% for a very small false alarm rate of 5.2%. With the mine shapes taken individually, the complete shadow and highlight feature sets produced outstanding classification results. At a 90.0% hit rate, we obtain false alarm rates of 10.0%, 12.5% and 19.0% respectively for the Manta, cylinder and Rockan shapes. Much work remains to be done in the field of computer aided-classification. Sidescan sonar images are currently widely used for mine countermeasures operations, but will certainly be replaced by the new synthetic aperture sonar images. These images have a much higher image resolution and should therefore yield a higher classification performance. If data from synthetic aperture sonars were available, it would be interesting to evaluate the level of classification performance achieved and contrast this performance to the one obtained in this memorandum. Furthermore, the usefulness of support vector machines is not limited to sidescan sonar classification problems, multiple other areas of interest could be investigated such as any other underwater acoustics detection/classification problem, studies of sphere properties and even tracking. 24 DRDC Atlantic TM

MULTI-VIEW TARGET CLASSIFICATION IN SYNTHETIC APERTURE SONAR IMAGERY

MULTI-VIEW TARGET CLASSIFICATION IN SYNTHETIC APERTURE SONAR IMAGERY MULTI-VIEW TARGET CLASSIFICATION IN SYNTHETIC APERTURE SONAR IMAGERY David Williams a, Johannes Groen b ab NATO Undersea Research Centre, Viale San Bartolomeo 400, 19126 La Spezia, Italy Contact Author:

More information

A Simple Model of a Rockan Mine in Processed Sidescan Sonar Imagery

A Simple Model of a Rockan Mine in Processed Sidescan Sonar Imagery Copy No. Defence Research and Development Canada Recherche et développement pour la défense Canada DEFENCE & DÉFENSE A Simple Model of a Rockan Mine in Processed Sidescan Sonar Imagery Anna Crawford Defence

More information

C-CORE Task #2 Report - Support for data processing and image analysis of the Fall 2012 through-wall field trials

C-CORE Task #2 Report - Support for data processing and image analysis of the Fall 2012 through-wall field trials C-CORE Task # Report - Support for data processing and image analysis of the Fall through-wall field trials B. Yue and J. Chamberland The scientific or technical validity of this Contract Report is entirely

More information

Kernel Methods & Support Vector Machines

Kernel Methods & Support Vector Machines & Support Vector Machines & Support Vector Machines Arvind Visvanathan CSCE 970 Pattern Recognition 1 & Support Vector Machines Question? Draw a single line to separate two classes? 2 & Support Vector

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing

More information

Support Vector Machines

Support Vector Machines Support Vector Machines . Importance of SVM SVM is a discriminative method that brings together:. computational learning theory. previously known methods in linear discriminant functions 3. optimization

More information

Data Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017

Data Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017 Data Analysis 3 Support Vector Machines Jan Platoš October 30, 2017 Department of Computer Science Faculty of Electrical Engineering and Computer Science VŠB - Technical University of Ostrava Table of

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines srihari@buffalo.edu SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions

More information

Module 4. Non-linear machine learning econometrics: Support Vector Machine

Module 4. Non-linear machine learning econometrics: Support Vector Machine Module 4. Non-linear machine learning econometrics: Support Vector Machine THE CONTRACTOR IS ACTING UNDER A FRAMEWORK CONTRACT CONCLUDED WITH THE COMMISSION Introduction When the assumption of linearity

More information

Support Vector Machines

Support Vector Machines Support Vector Machines SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions 6. Dealing

More information

Functional Blue Prints for the Development of a KMapper Prototype

Functional Blue Prints for the Development of a KMapper Prototype Functional Blue Prints for the Development of a KMapper Prototype SOFTWARE DESIGN DOCUMENT KMAPPER KNOWLEDGE INFERRING SERVICES And prepared by Martin Froment and Ludovic Tobin Fujitsu Consulting (Canada)

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Chapter 9 Chapter 9 1 / 50 1 91 Maximal margin classifier 2 92 Support vector classifiers 3 93 Support vector machines 4 94 SVMs with more than two classes 5 95 Relationshiop to

More information

Support Vector Machines + Classification for IR

Support Vector Machines + Classification for IR Support Vector Machines + Classification for IR Pierre Lison University of Oslo, Dep. of Informatics INF3800: Søketeknologi April 30, 2014 Outline of the lecture Recap of last week Support Vector Machines

More information

Support Vector Machines (a brief introduction) Adrian Bevan.

Support Vector Machines (a brief introduction) Adrian Bevan. Support Vector Machines (a brief introduction) Adrian Bevan email: a.j.bevan@qmul.ac.uk Outline! Overview:! Introduce the problem and review the various aspects that underpin the SVM concept.! Hard margin

More information

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2007 c 2007,

More information

Linear methods for supervised learning

Linear methods for supervised learning Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes

More information

All lecture slides will be available at CSC2515_Winter15.html

All lecture slides will be available at  CSC2515_Winter15.html CSC2515 Fall 2015 Introduc3on to Machine Learning Lecture 9: Support Vector Machines All lecture slides will be available at http://www.cs.toronto.edu/~urtasun/courses/csc2515/ CSC2515_Winter15.html Many

More information

CS 229 Midterm Review

CS 229 Midterm Review CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask

More information

A data set for change detection in port environments using sidescan sonar

A data set for change detection in port environments using sidescan sonar Copy No. Defence Research and Development Canada Recherche et développement pour la défense Canada DEFENCE & DÉFENSE A data set for change detection in port environments using sidescan sonar Vincent Myers

More information

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs)

Data Mining: Concepts and Techniques. Chapter 9 Classification: Support Vector Machines. Support Vector Machines (SVMs) Data Mining: Concepts and Techniques Chapter 9 Classification: Support Vector Machines 1 Support Vector Machines (SVMs) SVMs are a set of related supervised learning methods used for classification Based

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines CS 536: Machine Learning Littman (Wu, TA) Administration Slides borrowed from Martin Law (from the web). 1 Outline History of support vector machines (SVM) Two classes,

More information

Lab 2: Support vector machines

Lab 2: Support vector machines Artificial neural networks, advanced course, 2D1433 Lab 2: Support vector machines Martin Rehn For the course given in 2006 All files referenced below may be found in the following directory: /info/annfk06/labs/lab2

More information

DRDC Toronto No. CR Development and Documentation of the Software to Control the Noise Simulation Facility at DRDC Toronto

DRDC Toronto No. CR Development and Documentation of the Software to Control the Noise Simulation Facility at DRDC Toronto DRDC Toronto No. CR 2005-083 Development and Documentation of the Software to Control the Noise Simulation Facility at DRDC Toronto Élaboration et documentation du logiciel servant à contrôler l'installation

More information

Machine Learning for NLP

Machine Learning for NLP Machine Learning for NLP Support Vector Machines Aurélie Herbelot 2018 Centre for Mind/Brain Sciences University of Trento 1 Support Vector Machines: introduction 2 Support Vector Machines (SVMs) SVMs

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

Support Vector Machines

Support Vector Machines Support Vector Machines RBF-networks Support Vector Machines Good Decision Boundary Optimization Problem Soft margin Hyperplane Non-linear Decision Boundary Kernel-Trick Approximation Accurancy Overtraining

More information

9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives

9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives Foundations of Machine Learning École Centrale Paris Fall 25 9. Support Vector Machines Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech Learning objectives chloe agathe.azencott@mines

More information

A Model Based Approach to Mine Detection and Classification in Sidescan Sonar

A Model Based Approach to Mine Detection and Classification in Sidescan Sonar A Model Based Approach to Mine Detection and Classification in Sidescan Sonar S.Reed, Ocean Systems Lab, Heriot-Watt University Y.Petillot, Ocean Systems Lab, Heriot-Watt University J.Bell, Ocean Systems

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Maximum Margin Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574

More information

RMP Simulation User Guide

RMP Simulation User Guide Richard Sorensen Kihomac DRDC CORA CR 2011 099 October 2011 Defence R&D Canada Centre for Operational Research and Analysis National Defence Défense nationale Prepared By: Richard Sorensen Kihomac 5501

More information

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality

More information

Generative and discriminative classification techniques

Generative and discriminative classification techniques Generative and discriminative classification techniques Machine Learning and Category Representation 013-014 Jakob Verbeek, December 13+0, 013 Course website: http://lear.inrialpes.fr/~verbeek/mlcr.13.14

More information

Feature scaling in support vector data description

Feature scaling in support vector data description Feature scaling in support vector data description P. Juszczak, D.M.J. Tax, R.P.W. Duin Pattern Recognition Group, Department of Applied Physics, Faculty of Applied Sciences, Delft University of Technology,

More information

Data-driven Kernels for Support Vector Machines

Data-driven Kernels for Support Vector Machines Data-driven Kernels for Support Vector Machines by Xin Yao A research paper presented to the University of Waterloo in partial fulfillment of the requirement for the degree of Master of Mathematics in

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:

More information

Application of Support Vector Machine Algorithm in Spam Filtering

Application of Support Vector Machine Algorithm in  Spam Filtering Application of Support Vector Machine Algorithm in E-Mail Spam Filtering Julia Bluszcz, Daria Fitisova, Alexander Hamann, Alexey Trifonov, Advisor: Patrick Jähnichen Abstract The problem of spam classification

More information

Kernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018

Kernels + K-Means Introduction to Machine Learning. Matt Gormley Lecture 29 April 25, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Kernels + K-Means Matt Gormley Lecture 29 April 25, 2018 1 Reminders Homework 8:

More information

Canada s Energy Future:

Canada s Energy Future: Page 1 of 9 1DWLRQDO (QHUJ\ %RDUG 2IILFH QDWLRQDO GH OҋpQHUJLH Canada s Energy Future: ENERGY SUPPLY AND DEMAND PROJECTIONS TO 2035 Appendices AN ENERGY MARKET ASSESSMENT NOVEMBER 2011 Page 2 of 9 Canada

More information

DM6 Support Vector Machines

DM6 Support Vector Machines DM6 Support Vector Machines Outline Large margin linear classifier Linear separable Nonlinear separable Creating nonlinear classifiers: kernel trick Discussion on SVM Conclusion SVM: LARGE MARGIN LINEAR

More information

Well Analysis: Program psvm_welllogs

Well Analysis: Program psvm_welllogs Proximal Support Vector Machine Classification on Well Logs Overview Support vector machine (SVM) is a recent supervised machine learning technique that is widely used in text detection, image recognition

More information

VLANs. Commutation LAN et Wireless Chapitre 3

VLANs. Commutation LAN et Wireless Chapitre 3 VLANs Commutation LAN et Wireless Chapitre 3 ITE I Chapter 6 2006 Cisco Systems, Inc. All rights reserved. Cisco Public 1 Objectifs Expliquer le rôle des VLANs dans un réseau convergent. Expliquer le rôle

More information

Support Vector Machines

Support Vector Machines Support Vector Machines RBF-networks Support Vector Machines Good Decision Boundary Optimization Problem Soft margin Hyperplane Non-linear Decision Boundary Kernel-Trick Approximation Accurancy Overtraining

More information

User s Manual of Interactive Software for Predicting CPF Bow-Flare Impulsive Loads

User s Manual of Interactive Software for Predicting CPF Bow-Flare Impulsive Loads Copy No. Defence Research and Development Canada Recherche et développement pour la défense Canada DEFENCE & DÉFENSE User s Manual of Interactive Software for Predicting CPF Bow-Flare Impulsive Loads J.M.

More information

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013

Machine Learning. Topic 5: Linear Discriminants. Bryan Pardo, EECS 349 Machine Learning, 2013 Machine Learning Topic 5: Linear Discriminants Bryan Pardo, EECS 349 Machine Learning, 2013 Thanks to Mark Cartwright for his extensive contributions to these slides Thanks to Alpaydin, Bishop, and Duda/Hart/Stork

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

Support Vector Machines

Support Vector Machines Support Vector Machines About the Name... A Support Vector A training sample used to define classification boundaries in SVMs located near class boundaries Support Vector Machines Binary classifiers whose

More information

Lecture 7: Support Vector Machine

Lecture 7: Support Vector Machine Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each

More information

A Short SVM (Support Vector Machine) Tutorial

A Short SVM (Support Vector Machine) Tutorial A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange

More information

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018 Performance Estimation and Regularization Kasthuri Kannan, PhD. Machine Learning, Spring 2018 Bias- Variance Tradeoff Fundamental to machine learning approaches Bias- Variance Tradeoff Error due to Bias:

More information

Support vector machines

Support vector machines Support vector machines When the data is linearly separable, which of the many possible solutions should we prefer? SVM criterion: maximize the margin, or distance between the hyperplane and the closest

More information

About Transferring License Rights for. PL7 V4.5 and Unity Pro V2.3 SP1 Software

About Transferring License Rights for. PL7 V4.5 and Unity Pro V2.3 SP1 Software Page 1 of 38 Click here to access the English Cliquez ici pour accéder au Français Klicken Sie hier, um zum Deutschen zu gelangen Premete qui per accedere all' Italiano Pulse acquì para acceder al Español

More information

Mardi 3 avril Epreuve écrite sur un document en anglais

Mardi 3 avril Epreuve écrite sur un document en anglais C O L L E CONCOURS INTERNE ET EXTERNE DE TECHNICIEN DE CLASSE NORMALE DES SYSTEMES D INFORMATION ET DE COMMUNICATION Ne pas cacher le cadre d identité. Cette opération sera réalisée par l administration

More information

Dealing with Categorical Data Types in a Designed Experiment

Dealing with Categorical Data Types in a Designed Experiment Dealing with Categorical Data Types in a Designed Experiment Part II: Sizing a Designed Experiment When Using a Binary Response Best Practice Authored by: Francisco Ortiz, PhD STAT T&E COE The goal of

More information

Machine Learning: Think Big and Parallel

Machine Learning: Think Big and Parallel Day 1 Inderjit S. Dhillon Dept of Computer Science UT Austin CS395T: Topics in Multicore Programming Oct 1, 2013 Outline Scikit-learn: Machine Learning in Python Supervised Learning day1 Regression: Least

More information

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation.

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation. Equation to LaTeX Abhinav Rastogi, Sevy Harris {arastogi,sharris5}@stanford.edu I. Introduction Copying equations from a pdf file to a LaTeX document can be time consuming because there is no easy way

More information

HW2 due on Thursday. Face Recognition: Dimensionality Reduction. Biometrics CSE 190 Lecture 11. Perceptron Revisited: Linear Separators

HW2 due on Thursday. Face Recognition: Dimensionality Reduction. Biometrics CSE 190 Lecture 11. Perceptron Revisited: Linear Separators HW due on Thursday Face Recognition: Dimensionality Reduction Biometrics CSE 190 Lecture 11 CSE190, Winter 010 CSE190, Winter 010 Perceptron Revisited: Linear Separators Binary classification can be viewed

More information

Practice EXAM: SPRING 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

Practice EXAM: SPRING 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE Practice EXAM: SPRING 0 CS 6375 INSTRUCTOR: VIBHAV GOGATE The exam is closed book. You are allowed four pages of double sided cheat sheets. Answer the questions in the spaces provided on the question sheets.

More information

Fast checking of CMM geometry with a patented tool

Fast checking of CMM geometry with a patented tool 17 International Congress of Metrology, 13012 (2015) DOI: 10.1051/ metrolo gy/201513012 C Owned by the authors, published by EDP Sciences, 2015 Fast checking of CMM geometry with a patented tool Jean-François

More information

CSE 417T: Introduction to Machine Learning. Lecture 22: The Kernel Trick. Henry Chai 11/15/18

CSE 417T: Introduction to Machine Learning. Lecture 22: The Kernel Trick. Henry Chai 11/15/18 CSE 417T: Introduction to Machine Learning Lecture 22: The Kernel Trick Henry Chai 11/15/18 Linearly Inseparable Data What can we do if the data is not linearly separable? Accept some non-zero in-sample

More information

Multicollinearity and Validation CIVL 7012/8012

Multicollinearity and Validation CIVL 7012/8012 Multicollinearity and Validation CIVL 7012/8012 2 In Today s Class Recap Multicollinearity Model Validation MULTICOLLINEARITY 1. Perfect Multicollinearity 2. Consequences of Perfect Multicollinearity 3.

More information

A MODEL-BASED MULTI-VIEW IMAGE REGISTRATION METHOD FOR SAS IMAGES

A MODEL-BASED MULTI-VIEW IMAGE REGISTRATION METHOD FOR SAS IMAGES A MODEL-BASED MULTI-VIEW IMAGE REGISTRATION METHOD FOR SAS IMAGES Johannes Groen, David Williams, Warren Fox NATO Undersea Research Centre, Viale San Bartolomeo 400, 19126 La Spezia, Italy. Contact: J

More information

SVM in Analysis of Cross-Sectional Epidemiological Data Dmitriy Fradkin. April 4, 2005 Dmitriy Fradkin, Rutgers University Page 1

SVM in Analysis of Cross-Sectional Epidemiological Data Dmitriy Fradkin. April 4, 2005 Dmitriy Fradkin, Rutgers University Page 1 SVM in Analysis of Cross-Sectional Epidemiological Data Dmitriy Fradkin April 4, 2005 Dmitriy Fradkin, Rutgers University Page 1 Overview The goals of analyzing cross-sectional data Standard methods used

More information

Multi-Source Data Fusion for the Estimation of Mean Shipping Densities

Multi-Source Data Fusion for the Estimation of Mean Shipping Densities Multi-Source Data Fusion for the Estimation of Mean Shipping Densities Yvan Gauthier Maritime Forces Pacific Operational Research Team Peter Minter Co-op student, University of Victoria DRDC CORA TM 2005

More information

Applying Supervised Learning

Applying Supervised Learning Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains

More information

Face detection and recognition. Detection Recognition Sally

Face detection and recognition. Detection Recognition Sally Face detection and recognition Detection Recognition Sally Face detection & recognition Viola & Jones detector Available in open CV Face recognition Eigenfaces for face recognition Metric learning identification

More information

ELEC Dr Reji Mathew Electrical Engineering UNSW

ELEC Dr Reji Mathew Electrical Engineering UNSW ELEC 4622 Dr Reji Mathew Electrical Engineering UNSW Review of Motion Modelling and Estimation Introduction to Motion Modelling & Estimation Forward Motion Backward Motion Block Motion Estimation Motion

More information

5 Learning hypothesis classes (16 points)

5 Learning hypothesis classes (16 points) 5 Learning hypothesis classes (16 points) Consider a classification problem with two real valued inputs. For each of the following algorithms, specify all of the separators below that it could have generated

More information

Machine Learning in Biology

Machine Learning in Biology Università degli studi di Padova Machine Learning in Biology Luca Silvestrin (Dottorando, XXIII ciclo) Supervised learning Contents Class-conditional probability density Linear and quadratic discriminant

More information

Lecture 9: Support Vector Machines

Lecture 9: Support Vector Machines Lecture 9: Support Vector Machines William Webber (william@williamwebber.com) COMP90042, 2014, Semester 1, Lecture 8 What we ll learn in this lecture Support Vector Machines (SVMs) a highly robust and

More information

Statistical Methods in AI

Statistical Methods in AI Statistical Methods in AI Distance Based and Linear Classifiers Shrenik Lad, 200901097 INTRODUCTION : The aim of the project was to understand different types of classification algorithms by implementing

More information

Second Order SMO Improves SVM Online and Active Learning

Second Order SMO Improves SVM Online and Active Learning Second Order SMO Improves SVM Online and Active Learning Tobias Glasmachers and Christian Igel Institut für Neuroinformatik, Ruhr-Universität Bochum 4478 Bochum, Germany Abstract Iterative learning algorithms

More information

Divide and Conquer Kernel Ridge Regression

Divide and Conquer Kernel Ridge Regression Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem

More information

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Exploratory data analysis tasks Examine the data, in search of structures

More information

Analysis of Functional MRI Timeseries Data Using Signal Processing Techniques

Analysis of Functional MRI Timeseries Data Using Signal Processing Techniques Analysis of Functional MRI Timeseries Data Using Signal Processing Techniques Sea Chen Department of Biomedical Engineering Advisors: Dr. Charles A. Bouman and Dr. Mark J. Lowe S. Chen Final Exam October

More information

Classification by Support Vector Machines

Classification by Support Vector Machines Classification by Support Vector Machines Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Practical DNA Microarray Analysis 2003 1 Overview I II III

More information

AIS Indexer User Guide

AIS Indexer User Guide AIS Indexer User Guide Dan Radulescu Prepared by: OODA Technologies Inc. 4891 Av. Grosvenor, Montreal Qc, H3W 2M2 Project Manager: Anthony W. Isenor Contract Number: W7707-115137, Call Up 6, 4500959431

More information

Robust Shape Retrieval Using Maximum Likelihood Theory

Robust Shape Retrieval Using Maximum Likelihood Theory Robust Shape Retrieval Using Maximum Likelihood Theory Naif Alajlan 1, Paul Fieguth 2, and Mohamed Kamel 1 1 PAMI Lab, E & CE Dept., UW, Waterloo, ON, N2L 3G1, Canada. naif, mkamel@pami.uwaterloo.ca 2

More information

Classification by Support Vector Machines

Classification by Support Vector Machines Classification by Support Vector Machines Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Practical DNA Microarray Analysis 2003 1 Overview I II III

More information

Semi-supervised learning and active learning

Semi-supervised learning and active learning Semi-supervised learning and active learning Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Combining classifiers Ensemble learning: a machine learning paradigm where multiple learners

More information

Lab 2: Support Vector Machines

Lab 2: Support Vector Machines Articial neural networks, advanced course, 2D1433 Lab 2: Support Vector Machines March 13, 2007 1 Background Support vector machines, when used for classication, nd a hyperplane w, x + b = 0 that separates

More information

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response

Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response Categorical Data in a Designed Experiment Part 2: Sizing with a Binary Response Authored by: Francisco Ortiz, PhD Version 2: 19 July 2018 Revised 18 October 2018 The goal of the STAT COE is to assist in

More information

6 Model selection and kernels

6 Model selection and kernels 6. Bias-Variance Dilemma Esercizio 6. While you fit a Linear Model to your data set. You are thinking about changing the Linear Model to a Quadratic one (i.e., a Linear Model with quadratic features φ(x)

More information

Mobile Human Detection Systems based on Sliding Windows Approach-A Review

Mobile Human Detection Systems based on Sliding Windows Approach-A Review Mobile Human Detection Systems based on Sliding Windows Approach-A Review Seminar: Mobile Human detection systems Njieutcheu Tassi cedrique Rovile Department of Computer Engineering University of Heidelberg

More information

Topics in Machine Learning

Topics in Machine Learning Topics in Machine Learning Gilad Lerman School of Mathematics University of Minnesota Text/slides stolen from G. James, D. Witten, T. Hastie, R. Tibshirani and A. Ng Machine Learning - Motivation Arthur

More information

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule.

Feature Extractors. CS 188: Artificial Intelligence Fall Some (Vague) Biology. The Binary Perceptron. Binary Decision Rule. CS 188: Artificial Intelligence Fall 2008 Lecture 24: Perceptrons II 11/24/2008 Dan Klein UC Berkeley Feature Extractors A feature extractor maps inputs to feature vectors Dear Sir. First, I must solicit

More information

Feature-Based Facial Expression Recognition: Experiments With a Multi-Layer Perceptron

Feature-Based Facial Expression Recognition: Experiments With a Multi-Layer Perceptron Feature-Based Facial Expression Recognition: Experiments With a Multi-Layer Perceptron Zhengyou Zhang To cite this version: Zhengyou Zhang. Feature-Based Facial Expression Recognition: Experiments With

More information

Transductive Learning: Motivation, Model, Algorithms

Transductive Learning: Motivation, Model, Algorithms Transductive Learning: Motivation, Model, Algorithms Olivier Bousquet Centre de Mathématiques Appliquées Ecole Polytechnique, FRANCE olivier.bousquet@m4x.org University of New Mexico, January 2002 Goal

More information

SUPPORT VECTOR MACHINES

SUPPORT VECTOR MACHINES SUPPORT VECTOR MACHINES Today Reading AIMA 18.9 Goals (Naïve Bayes classifiers) Support vector machines 1 Support Vector Machines (SVMs) SVMs are probably the most popular off-the-shelf classifier! Software

More information

Large synthetic data sets to compare different data mining methods

Large synthetic data sets to compare different data mining methods Large synthetic data sets to compare different data mining methods Victoria Ivanova, Yaroslav Nalivajko Superviser: David Pfander, IPVS ivanova.informatics@gmail.com yaroslav.nalivayko@gmail.com June 3,

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation Learning 4 Supervised Learning 4 Unsupervised Learning 4

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Kernel SVM. Course: Machine Learning MAHDI YAZDIAN-DEHKORDI FALL 2017

Kernel SVM. Course: Machine Learning MAHDI YAZDIAN-DEHKORDI FALL 2017 Kernel SVM Course: MAHDI YAZDIAN-DEHKORDI FALL 2017 1 Outlines SVM Lagrangian Primal & Dual Problem Non-linear SVM & Kernel SVM SVM Advantages Toolboxes 2 SVM Lagrangian Primal/DualProblem 3 SVM LagrangianPrimalProblem

More information

Model and Data Management Tool for the Air Force Structure Analysis Model - Final Report

Model and Data Management Tool for the Air Force Structure Analysis Model - Final Report Air Force Structure Analysis Model - Final Report D.G. Hunter DRDC CORA Prepared By: CAE Integrated Enterprise Solutions - Canada 1135 Innovation Drive Ottawa, ON, K2K 3G7 Canada Telephone: 613-247-0342

More information

Classification Part 4

Classification Part 4 Classification Part 4 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville Model Evaluation Metrics for Performance Evaluation How to evaluate

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

IPv6 Protocol (RFC 2460 DS)

IPv6 Protocol (RFC 2460 DS) IPv6 Protocol (RFC 2460 DS) Copy Rights This slide set is the ownership of the 6DISS project via its partners The Powerpoint version of this material may be reused and modified only with written authorization

More information

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München

Evaluation Measures. Sebastian Pölsterl. April 28, Computer Aided Medical Procedures Technische Universität München Evaluation Measures Sebastian Pölsterl Computer Aided Medical Procedures Technische Universität München April 28, 2015 Outline 1 Classification 1. Confusion Matrix 2. Receiver operating characteristics

More information

Last time... Coryn Bailer-Jones. check and if appropriate remove outliers, errors etc. linear regression

Last time... Coryn Bailer-Jones. check and if appropriate remove outliers, errors etc. linear regression Machine learning, pattern recognition and statistical data modelling Lecture 3. Linear Methods (part 1) Coryn Bailer-Jones Last time... curse of dimensionality local methods quickly become nonlocal as

More information

Distortion-invariant Kernel Correlation Filters for General Object Recognition

Distortion-invariant Kernel Correlation Filters for General Object Recognition Distortion-invariant Kernel Correlation Filters for General Object Recognition Dissertation by Rohit Patnaik Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy

More information