Abstract R-splines are introduced as splines t with a polynomial null space plus the sum of radial basis functions. Thin plate splines are a special c

Size: px
Start display at page:

Download "Abstract R-splines are introduced as splines t with a polynomial null space plus the sum of radial basis functions. Thin plate splines are a special c"

Transcription

1 R-Splines for Response Surface Modeling July 12, 2000 Sarah W. Hardy North Carolina State University Douglas W. Nychka National Center for Atmospheric Research Raleigh, NC Boulder, CO

2 Abstract R-splines are introduced as splines t with a polynomial null space plus the sum of radial basis functions. Thin plate splines are a special case of R-splines. By this broader denition of an R-spline, however, it includes splines in which 2m? d 0 where the traditional roughness penalty is not guaranteed to be non-negative denite. This papers discusses a modication of the roughness penalty that allows for the tting of reduced polynomial null spaces. A series of examples are used to demonstrate the behavior of this modied roughness penalty. Keywords: Thin plate spline, nonparametric, response surfaces, roughness penalty, Demmler-Reinsch basis functions. 1 Introduction Thin plate splines are nonparametric method for tting response surfaces. A thin plate spline is a smooth surface constructed by adding a base, or null space, polynomial component, and a sum of radial basis functions. The thin plate smoothing splines are a solution to a minimization problem and can be interpreted as a natural generalization of the traditional polynomial model estimated by minimizing residual sum of squares. Unlike the traditional formulation, however, the thin plate spline minimization problem also involves a roughness penalty. This roughness penalty is a function of m and d where where m? 1 is the degree of the polynomial null space and d is the number of explanatory variables. To guarantee that the roughness penalty is positive and the function is a true thin plate spline, 2m? d > 0. Thus, the size of the polynomial null space grows very quickly as a function of the number of explanatory variables. Many typical industrial experiments do not have enough observations to t the number of parameters needed for the required null space. Specically, this becomes problematic when there is a large set of explanatory variables and it is desirable to use an all subsets approach to variable selection for thin 2

3 plate spline modeling. R-splines arose in the context of attempting to modify the thin plate spline to allow for a greater number of explanatory variables in the model. R-splines are splines t with a polynomial null space plus the sum of radial basis functions. Clearly, thin plate splines fall into this category. By this broader denition of an R-spline, however, it also includes other splines. Specically, it includes splines in which 2m? d 0. In other words, the polynomial null space is reduced or of a lower order than required by the thin plate spline restraint. When 2m? d 0, the seamless way in which the polynomial function and roughness penalty t together (i.e, the space spanned by the polynomial terms is the null space of the roughness penalty) no longer holds or the relationship is \broken." The term broken spline will be used to indicate this kind of spline where the penalty matrix is chosen to be similar to the thin plate spline. The term R-splines refers to the broader class of splines including both broken and thin plate splines. This paper begins by discussing the thin plate spline roughness penalty. It then presents the broken spline modication to the roughness penalty. The Demmler-Reinsch basis function representation of splines is used to show the role the eigenvalues of the roughness penalty matrix have in the spline solution. A series of four two and three dimensional examples from data sets available in S-PLUS and FUNFITS (Nychka, et. al., 1996), is used to compare the roughness penalties for the thin plate and broken spline. Another ve dimensional example using data collected at Becton Dickinson is also presented. 2 Thin plate spline roughness penalty The roughness functional below, J m (f), will increase in magnitude as a function departs from an m? 1 order polynomial. This roughness functional is: 3

4 Z X m! J m (f) = < d 1!::: d!! m :du 1 1 :::@u d d The sum in the integrand is taken over all non-negative integer vectors,, such that P d = m. Clearly, for m? 1 order polynomials J m (f) = 0, because all m th derivatives are 0. The thin plate spline estimator of f (Wahba, 1990) is the minimizer of the sum of the mean-squared error and the roughness penalty which is weighted by a smoothing parameter. Details on the form of the thin plate spline can be found in the appendix. In matrix form, the thin plate spline estimate is f(x i ) = T + M where T T = 0 and 2m? d > 0; where T is the design matrix of a polynomial model of order m? 1 and M k;i = E(k x i? x k k; m; d). For linear combinations of basis functions that satisfy the constraints, J m (f) = T M > 0. In other words, M is guaranteed to be positive denite. Dening the matrix W nn, a diagonal matrix proportional to the reciprocal variances of the errors, allows for the following matrix representation of the penalized sums of squares: S = 1 n (Y? T? M)T W (Y? T? M) + T M: (1) After taking partial derivatives of (1) with respect to and, a QR decomposition of T, F T T = R , is used to enforce the constraint T T = 0. F nn is an orthogonal matrix that can be partioned F = [F1 j F2] where F1 has columns that span the column space of T, i.e T = F1R, and F2 is orthogonal to the column space of T. Reparameterizing by letting = F2!2,!1, and! T = (!1;!2) T allows for the problem to be posed in ridge 4

5 regression form: X T W X! + H! = X T W Y; where X = T MF2 T and H = : 0 F2 T MF 2 Note that in the ridge regression formulation the roughness penalty is represented by the matrix H. 3 Broken spline roughness penalty The degree of the polynomial component implies a specic roughness penalty in the thin plate spline, and thus if T, the design matrix for the polynomial component does not span P m?1 the roughness penalty minimized in the ridge regression formulation will not be the same roughness penalty as the one used to derive the thin plate spline. We will use the term broken spline to describe splines where T does not span P m?1. For purposes of this comparative discussion the T matrix for a thin plate spline will be denoted T P and the T matrix for a broken spline will be denoted T R. Note that the column space of T R is a subset of that of T P. Partition F F = [F a F b F c ] where F a spans the space of T R, F b spans the columns of T P that are not in T R, and F c spans the space orthogonal to T P. Thus in the QR decomposition of T P, F2 = F c and in the QR decomposition of T R, F2 = [F b F c ]: 5

6 For the broken spline H, the roughness penalty matrix, becomes H R = rr Fb T MF b Fc T MF b 0 Fb T MF c Fc T MF c ; as compared to the H for the thin plate spline which is H t = rr t?rt?r Fc T MF c : Hence, H R for the broken spline is the roughness penalty for the thin plate spline, plus other terms, H R = H t F T b MF b F T c MF b 0 F T b MF c : With thin plate splines, H t is guaranteed to be non-negative denite. Recall T M > 0 where T T = 0 and 2m? d > 0, which guarantees that M is positive denite. F2 T MF 2, is a quadratic form of a positive denite matrix and as such is also positive denite. H t, having F2 T MF 2 in the lower right block and zeros elsewhere, is non-negative denite. With broken splines, however, H R is no longer a non-negative denite matrix. In the computational solution to the thin plate spline, UDU T is the singular value decomposition of BHB so H t = B?1 UDU T B?1. Now, BHB is no longer positive denite and thus its singular value decomposition is UDV T, where U 6= V, and H R = B?1 UDV T B?1. Let H R = B?1 UDU T B?1, which is a non-negative denite matrix because D is a diagonal matrix of singular values (which are non-negative by denition) and H is a quadratic form of D. If H R is replaced by H R, a non-negative denite roughness penalty is minimized. Because H R is a symmetric matrix, U and V will be the same except for sign changes in 6

7 columns corresponding to the negative eigenvalues. Hence, H R, is the matrix that results from forcing the negative eigenvalues of H R to be positive. How \close" the two matrices are depends on the magnitude of these eigenvalues. The negative eigenvalues that occur correspond to the \missing" polynomial terms in the null space. Thus, the roughness penalty is forced into penalizing roughness in the surface that might be modeled by these missing polynomial terms, tting them instead with the exible radial basis functions. 3.1 Demmler-Reinsch basis functions Denition Constructing a Demmler-Reinsch basis for the smoothing problem aids in the understanding of the roles the eigenvalues and eigenvectors of H have in the spline solution. Before dening the Demmler-Reinsch basis, two inner products are dened: < h ; h >1= X k h (x k )W k h (x k ) and < h ; h >2= Z < d X m! 1!::: m m 1 1 :::@x d 1 1 :::@x d d! dx The Demmler-Reinsch basis is denoted by fg g and is dened by the three following properties: 1. fg g for 1 N spans the same subspace as f j g and the radial basis functions. 2. < g ; g >1= 1 for = and 0 for 6=. 3. < g ; g >2= D for = and 0 for 6=. In other words, this basis consists of linear combinations of the polynomial and radial basis functions that are orthogonal with respect to the inner products by which the weighted 7

8 residual sums of squares and the roughness penalty can be expressed. By convention D are in ascending order Role in spline solution Because this basis spans the same space as the components of f(x) we can now express both f(x) and y as a linear combinations of these basis functions: NX NX f(x k ) = g (x k ) and y k = u g (x k ): (2) =1 =1 Using these expression it can be easily shown that u = P n k=1 g (x k )W k Y k : Moreover, under the assumption that the random errors are independent and distributed N(0; 2 I), the u 's are also independent and are distributed N(; 2 I): With the Demmler-Reinsch basis functions the residual sum of squares and the roughness penalty can be expressed: nx nx (Y k? f(x k )) 2 w k = (u? ) 2 k=1 =1 and nx J m (f) = D 2 : =1 Thus, it is apparent that minimizing S (f) is equivalent to nding The minimizing expression is: min 2< n nx nx (u? ) 2 + D 2 : (3) =1 =1 and so, c = u 1 + D ; ^f(x) = NX =1 c g (x) = 8 NX =1 u 1 + D g (x): (4)

9 In fact, the G = fg g =1;::;N we are seeking is (cleverly) the same G we found in the ridge regression formulation and property (2) is easily veried by noting that G T (X T W X)G = I: To determine if property (3) holds, observe that G T HG = D. As an important point of clarication recall that the upper r r block of H is 0 where r is the dimension of the null space. This implies D = 0 for = 1; ::; r where D are the eigenvalues of G T HG. More intuitively, by virtue of the way in which the basis functions are dened, the roughness of the basis functions is J m (g ) = D and r of these basis functions will be polynomials whose roughness is 0. With this formulation, it is apparent that if the D 's, which are the eigenvalues of H, and the g 's, which are the Demmler-Reinsch basis functions, are correlated for the thin plate and broken spline, then the two roughness penalties are similar and measuring the same features of the data. 4 Examples 4.1 Example datasets Four sample datasets that are available in S-Plus (StatSci, 1993), the stack and ethanol datasets, and FUNFITS, the mini-triathalon and BD2 datasets were used to compare these roughness penalties for ts to the data with a thin plate spline and a broken spline. The stack data is from operation of a plant for the oxidation of ammonia to nitric acid, measured on 21 consecutive days. The explanatory variables are air ow to the plant, cooling water inlet temperature, and acid concentration and the response variable is percent of ammonia lost. The ethanol data frame records 88 measurements an experiment in which ethanol was burned in a single cylinder automobile test engine. The explanatory variables are the compression ratio of the engine and the equivalence ratio at which the 9

10 engine was run (a measure of the richness of the air/ethanol mix). The response variable is the concentration of nitric oxide and nitrogen dioxide in engine exhaust, normalized by the work done by the engine. The mini-triathalon data contains swimming, biking, and running times from 110 entrants in a Cary, NC event. In the example we are using swimming and biking times to predict running times. The BD2 dataset comes from results of a sequence of RSM designs in a DNA amplication experiment performed at Becton Dickinson Technologies. The explanatory variables are potassium phosphate ( a buer), magnesium acetate (a salt), and dimethyl sulfoxide ( a solvent). The response variable is yield. In summary, the ethanol and mini-triathalon datasets each have 2 explanatory variables and the stack and BD2 datasets have 3. Thin plate splines were t with m = 2 or a linear null space and the broken splines were t with a m = 1 or a constant null space. Figure 1 shows the plots of the eigenvalues of H for the two types of splines. These plots demonstrate that these eigenvalues are nearly equivalent for all but a few points. Notice that the number of eigenvalues that dier signicantly equals the number of explanatory variables, which in turn equals the number of linear terms missing from the polynomial null space. These are the specic eigenvalues that correspond to polynomial terms in the null space of the thin plate spline, and hence are 0 for the thin plate spline and positive for the broken spline. 10

11 minitri H eigenvalues H eigenvalues for m=1 and stack.loss H eigenvalues m= m=1 ethanol H eigenvalues BD2 H eigenvalues m= m=1 Figure 1: Comparison of eigenvalues for m=1 (broken splines) and (thin plate splines). 11

12 Figure 2 shows image plots of the correlations of the thin plate spline and broken spline Demmler-Reinsch basis functions for these four data sets. Lines on the graphs indicate the division of the basis functions associated with the null space and roughness penalty. The areas in the regions greater than the lines on the graph indicate the regions where the basis functions are those associated with the null space. The area enclosed within the lines indicates the basis functions associated with the roughness penalty, which these plots show are highly correlated. As expected, the area outside the lines is not as highly correlated, as the polynomials in the null space are not the same for the thin plate and broken splines. There is, however, some correlation between the basis functions representing the null space in the thin plate spline and those representing the roughness penalty in the broken spline. This indicates that the polynomial terms missing from the broken spline are being accounted for in some way by the roughness penalty. There is some dierence between the two sets of basis functions, but overall when ordered in ascending order by eigenvalues the two sets of basis functions match closely beyond the rst few functions related to the null space. Clearly, more theoretical work is needed to explicitly quantify the relationship given these promising empirical results. Figure 3 demonstrates the results of tting the mini-triathalon dataset with broken and thin plate splines. The ts are approximately the same and nothing essential has been sacriced by tting the reduced null space, again evidence that the roughness penalties are similar. 12

13 Correlation of Demmler-Reinsch basis functions for m=1 and minitri stack.loss m=1 m=1 ethanol BD m=1 m=1 Figure 2: Correlation of basis functions for m=1 (broken splines) and (thin plate splines). 13

14 Predicted Values Observed Values R^2 = 69.56% Q^2 = 64.88% Predicted values Residuals m=1, Broken spline RMSE = Effective number of parameters Average Prediction Error Eff. df. = 8 Res. df. = 102 GCV min = GCV Predicted Values Observed Values R^2 = 70.23% Q^2 = 64.66% Predicted values Residuals , Thin plate Spline RMSE = Effective number of parameters Average Prediction Error Eff. df. = 9.6 Res. df. = GCV min = GCV Figure 3: Diagnostic plots for the mini-triathalon dataset t with a broken and thin plate spline. 14

15 4.2 A larger example dataset example m= Figure 4: Comparison of H eigenvalues for the example dataset t with a broken and thin plate spline. The next example arose from research being conducted at Becton Dickinson Technologies and because of the proprietary nature of the data, the details can not be given. This example has 5 explanatory variables, and so m for a thin plate spline must be 3. In other words, for a thin plate spline a full quadratic null space with 21 terms must be t. The broken spline was t with m = 2, or a linear null space with 6 terms. Figure 4 is a scatterplot of the 291 eigenvalues of H for the broken versus the thin plate spline. This plot shows that the majority of eigenvalues are very closely related, except for 15 of the eigenvalues related to the terms in the null space of the thin plate spline that are not in the null space of the broken spline. Another approximately 15 eigenvalues are smaller for 15

16 the broken spline than for the thin plate spline. example m= Figure 5: Correlation of basis functions for the example dataset t with a broken and thin plate spline. Figure 5 shows that, again, the basis functions are highly correlated for the thin plate and broken spline. They are less correlated at the higher end than the previous examples because of the higher dimensionality. The diagnostic plots, Figure 6, indicate that in this 16

17 case the broken spline is doing a slightly better job of tting the data. Predicted Values Observed Values R^2 = 45.12% Q^2 = 10.18% Predicted values Residuals , Broken spline RMSE = Pure Error = Effective number of parameters Average Prediction Error Eff. df. = 46.3 Res. df. = GCV min = GCV Predicted Values Observed Values R^2 = 37.5% Q^2 = 6.3% Predicted values Residuals m=3, Thin plate Spline RMSE = Pure Error = Effective number of parameters Average Prediction Error Eff. df. = 46.3 Res. df. = GCV min = GCV Figure 6: Diagnostic plots for the example dataset t with a broken and thin plate spline. 5 Conclusions The evidence indicates that broken splines are a feasible model tting tool. Like the thin plate spline, the broken spline is the sum of polynomial terms and radial basis functions and as such it is a smooth function with considerably more exibility than the traditional polynomial model. Also like the thin plate spline, it is the solution to a minimization problem in which the mean-squared error and a roughness component are minimized. 17

18 It can be shown that the thin plate spline solution is the unique and optimal solution to the posed minimization problem. The broken spline is also a unique solution to the minimization problem given a xed roughness penalty matrix as described. In the thin plate spline, the roughness penalty is explicitly dened and makes intuitive sense from a physical perspective as the bending energy of a thin plate. In fact, positive denite matrices arise in many applications involving energy, and clearly only make sense for a roughness penalty that is to be minimized. With the broken spline, although the eective roughness penalty can not be explicitly dened, it is closely related to the \bending energy" roughness penalty and empirically gives good results. The new insight that it is possible to achieve good results using the sum of polynomial terms and radial basis functions to estimate smooth functions without the restrictions imposed for thin plate splines opens the door for experimentation using this modeling technique for many more types of datasets. A Denition of a thin plate spline The thin plate spline estimator of f (Wahba, 1990) is the minimizer of the following penalized sums of squares for a d-dimensional explanatory variable x: nx S (f) = 1 w i (y i? f(x i )) 2 + J m (f) for > 0: (5) n i=1 Thus, an ^f minimizing (5) will result in f with some level of smoothness dictated by. The function that minimizes this expression has the form tx NX f(x i ) = j (x i ) j + E(k x i? x k k; m; d) k (6) j=1 k=1 where tx j (x i ) j = 0; 1 j t and 2m? d > 0: (7) j=1 18

19 In this formulation, the j (x i ) are a set of t polynomial functions (of order m? 1) and E(k x i? x k k; m; d) are a set of N radial basis functions. It is assumed that j is estimable. The radial basis functions are explicitly dened as below: E(r; m; d) = 8 >< >: a md k r k (2m?d) log(k r k) a md k r k (2m?d) d even d odd where a md depends only on m and d. One standard way of determining is by generalized cross-validation (GCV) (Bates et al, 1987). B References Bates, D.M., Lindstrom, M.J., Wahba, G. and Yandell, B.S. (1987). GCVPACK - Routines for generalized cross validation. Comm. Stat. Sim. Comp. 16, Berlin, pp Nychka, D., B. Bailey, S. Ellner, P. Haaland, and M. O'Connell (1996). FunFits data analysis and statistical tools for estimating functions. Software and paper available from statlib. StatSci (1993) S-PLUS Reference Manual Vol. 2, Version 3.2 MathSoft, Inc., Seattle, WA. Wahba, G. (1990). Spline Models for Observational Data. Society for Industrial Applied Mathematics. Philadelphia PA. 19

2 <Author Name(s)> and Green, 1994). This research was motivated by a specic application with twelve potential independent variables. If the functiona

2 <Author Name(s)> and Green, 1994). This research was motivated by a specic application with twelve potential independent variables. If the functiona Variable Selection for Response Surface Modeling Using R-Splines Sarah W. HARDY Douglas W. NYCHKA North Carolina State University National Center for Atmospheric Research Raleigh, NC 27695-8203 Boulder,

More information

Linear Methods for Regression and Shrinkage Methods

Linear Methods for Regression and Shrinkage Methods Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors

More information

Soft Threshold Estimation for Varying{coecient Models 2 ations of certain basis functions (e.g. wavelets). These functions are assumed to be smooth an

Soft Threshold Estimation for Varying{coecient Models 2 ations of certain basis functions (e.g. wavelets). These functions are assumed to be smooth an Soft Threshold Estimation for Varying{coecient Models Artur Klinger, Universitat Munchen ABSTRACT: An alternative penalized likelihood estimator for varying{coecient regression in generalized linear models

More information

Spline Models. Introduction to CS and NCS. Regression splines. Smoothing splines

Spline Models. Introduction to CS and NCS. Regression splines. Smoothing splines Spline Models Introduction to CS and NCS Regression splines Smoothing splines 3 Cubic Splines a knots: a< 1 < 2 < < m

More information

Adaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented wh

Adaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented wh Adaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented which, for a large-dimensional exponential family G,

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression John Fox Department of Sociology McMaster University 1280 Main Street West Hamilton, Ontario Canada L8S 4M4 jfox@mcmaster.ca February 2004 Abstract Nonparametric regression analysis

More information

Approximate Smoothing Spline Methods for Large Data Sets in the Binary Case Dong Xiang, SAS Institute Inc. Grace Wahba, University of Wisconsin at Mad

Approximate Smoothing Spline Methods for Large Data Sets in the Binary Case Dong Xiang, SAS Institute Inc. Grace Wahba, University of Wisconsin at Mad DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 982 September 30, 1997 Approximate Smoothing Spline Methods for Large Data Sets in the Binary

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Last time... Bias-Variance decomposition. This week

Last time... Bias-Variance decomposition. This week Machine learning, pattern recognition and statistical data modelling Lecture 4. Going nonlinear: basis expansions and splines Last time... Coryn Bailer-Jones linear regression methods for high dimensional

More information

Generalized Additive Model

Generalized Additive Model Generalized Additive Model by Huimin Liu Department of Mathematics and Statistics University of Minnesota Duluth, Duluth, MN 55812 December 2008 Table of Contents Abstract... 2 Chapter 1 Introduction 1.1

More information

APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES

APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES A. Likas, K. Blekas and A. Stafylopatis National Technical University of Athens Department

More information

Lecture 17: Smoothing splines, Local Regression, and GAMs

Lecture 17: Smoothing splines, Local Regression, and GAMs Lecture 17: Smoothing splines, Local Regression, and GAMs Reading: Sections 7.5-7 STATS 202: Data mining and analysis November 6, 2017 1 / 24 Cubic splines Define a set of knots ξ 1 < ξ 2 < < ξ K. We want

More information

Package rgcvpack. February 20, Index 6. Fitting Thin Plate Smoothing Spline. Fit thin plate splines of any order with user specified knots

Package rgcvpack. February 20, Index 6. Fitting Thin Plate Smoothing Spline. Fit thin plate splines of any order with user specified knots Version 0.1-4 Date 2013/10/25 Title R Interface for GCVPACK Fortran Package Author Xianhong Xie Package rgcvpack February 20, 2015 Maintainer Xianhong Xie

More information

Moving Beyond Linearity

Moving Beyond Linearity Moving Beyond Linearity The truth is never linear! 1/23 Moving Beyond Linearity The truth is never linear! r almost never! 1/23 Moving Beyond Linearity The truth is never linear! r almost never! But often

More information

Splines and penalized regression

Splines and penalized regression Splines and penalized regression November 23 Introduction We are discussing ways to estimate the regression function f, where E(y x) = f(x) One approach is of course to assume that f has a certain shape,

More information

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview Chapter 888 Introduction This procedure generates D-optimal designs for multi-factor experiments with both quantitative and qualitative factors. The factors can have a mixed number of levels. For example,

More information

6 Randomized rounding of semidefinite programs

6 Randomized rounding of semidefinite programs 6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can

More information

Lecture 16: High-dimensional regression, non-linear regression

Lecture 16: High-dimensional regression, non-linear regression Lecture 16: High-dimensional regression, non-linear regression Reading: Sections 6.4, 7.1 STATS 202: Data mining and analysis November 3, 2017 1 / 17 High-dimensional regression Most of the methods we

More information

Regression for non-linear data

Regression for non-linear data Regression for non-linear data don t let students go model shopping NCTM National Conference Boston MA April 2015 Julie Graves graves@ncssm.edu The North Carolina School of Science and Mathematics Model

More information

Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica

Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica Boston-Keio Workshop 2016. Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica... Mihoko Minami Keio University, Japan August 15, 2016 Joint work with

More information

Feature selection. Term 2011/2012 LSI - FIB. Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/ / 22

Feature selection. Term 2011/2012 LSI - FIB. Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/ / 22 Feature selection Javier Béjar cbea LSI - FIB Term 2011/2012 Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/2012 1 / 22 Outline 1 Dimensionality reduction 2 Projections 3 Attribute selection

More information

Spatial Distributions of Precipitation Events from Regional Climate Models

Spatial Distributions of Precipitation Events from Regional Climate Models Spatial Distributions of Precipitation Events from Regional Climate Models N. Lenssen September 2, 2010 1 Scientific Reason The Institute of Mathematics Applied to Geosciences (IMAGe) and the National

More information

EE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov.

EE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. EE613 Machine Learning for Engineers LINEAR REGRESSION Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. 4, 2015 1 Outline Multivariate ordinary least squares Singular value

More information

STAT 705 Introduction to generalized additive models

STAT 705 Introduction to generalized additive models STAT 705 Introduction to generalized additive models Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 22 Generalized additive models Consider a linear

More information

EE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov.

EE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. EE613 Machine Learning for Engineers LINEAR REGRESSION Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. 9, 2017 1 Outline Multivariate ordinary least squares Matlab code:

More information

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2010 1 / 26 Additive predictors

More information

Telecommunication and Informatics University of North Carolina, Technical University of Gdansk Charlotte, NC 28223, USA

Telecommunication and Informatics University of North Carolina, Technical University of Gdansk Charlotte, NC 28223, USA A Decoder-based Evolutionary Algorithm for Constrained Parameter Optimization Problems S lawomir Kozie l 1 and Zbigniew Michalewicz 2 1 Department of Electronics, 2 Department of Computer Science, Telecommunication

More information

Splines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric)

Splines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric) Splines Patrick Breheny November 20 Patrick Breheny STA 621: Nonparametric Statistics 1/46 Introduction Introduction Problems with polynomial bases We are discussing ways to estimate the regression function

More information

Generalized Additive Models

Generalized Additive Models :p Texts in Statistical Science Generalized Additive Models An Introduction with R Simon N. Wood Contents Preface XV 1 Linear Models 1 1.1 A simple linear model 2 Simple least squares estimation 3 1.1.1

More information

(Creating Arrays & Matrices) Applied Linear Algebra in Geoscience Using MATLAB

(Creating Arrays & Matrices) Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB (Creating Arrays & Matrices) Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional

More information

Lecture 3: Camera Calibration, DLT, SVD

Lecture 3: Camera Calibration, DLT, SVD Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P

More information

Improved smoothing spline regression by combining estimates of dierent smoothness

Improved smoothing spline regression by combining estimates of dierent smoothness Available online at www.sciencedirect.com Statistics & Probability Letters 67 (2004) 133 140 Improved smoothing spline regression by combining estimates of dierent smoothness Thomas C.M. Lee Department

More information

Hands-On Standards Deluxe Grades: 7, 8 States: California Content Standards

Hands-On Standards Deluxe Grades: 7, 8 States: California Content Standards Hands-On Standards Deluxe Grades: 7, 8 States: Hands-On Standards Deluxe Edition Kit, Grades 5-6: Algebra Summary: This resource guide meets 5 and 6 math curriculum standards by matching activities to

More information

Chapter 9 Robust Regression Examples

Chapter 9 Robust Regression Examples Chapter 9 Robust Regression Examples Chapter Table of Contents OVERVIEW...177 FlowChartforLMS,LTS,andMVE...179 EXAMPLES USING LMS AND LTS REGRESSION...180 Example 9.1 LMS and LTS with Substantial Leverage

More information

Generalized Additive Models

Generalized Additive Models Generalized Additive Models Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Additive Models GAMs are one approach to non-parametric regression in the multiple predictor setting.

More information

An Approach to Identify the Number of Clusters

An Approach to Identify the Number of Clusters An Approach to Identify the Number of Clusters Katelyn Gao Heather Hardeman Edward Lim Cristian Potter Carl Meyer Ralph Abbey July 11, 212 Abstract In this technological age, vast amounts of data are generated.

More information

Localization in Graphs. Richardson, TX Azriel Rosenfeld. Center for Automation Research. College Park, MD

Localization in Graphs. Richardson, TX Azriel Rosenfeld. Center for Automation Research. College Park, MD CAR-TR-728 CS-TR-3326 UMIACS-TR-94-92 Samir Khuller Department of Computer Science Institute for Advanced Computer Studies University of Maryland College Park, MD 20742-3255 Localization in Graphs Azriel

More information

What is machine learning?

What is machine learning? Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship

More information

Introduction to ANSYS DesignXplorer

Introduction to ANSYS DesignXplorer Lecture 4 14. 5 Release Introduction to ANSYS DesignXplorer 1 2013 ANSYS, Inc. September 27, 2013 s are functions of different nature where the output parameters are described in terms of the input parameters

More information

( ) = Y ˆ. Calibration Definition A model is calibrated if its predictions are right on average: ave(response Predicted value) = Predicted value.

( ) = Y ˆ. Calibration Definition A model is calibrated if its predictions are right on average: ave(response Predicted value) = Predicted value. Calibration OVERVIEW... 2 INTRODUCTION... 2 CALIBRATION... 3 ANOTHER REASON FOR CALIBRATION... 4 CHECKING THE CALIBRATION OF A REGRESSION... 5 CALIBRATION IN SIMPLE REGRESSION (DISPLAY.JMP)... 5 TESTING

More information

CS 229 Midterm Review

CS 229 Midterm Review CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask

More information

Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu

Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu 1 Introduction Yelp Dataset Challenge provides a large number of user, business and review data which can be used for

More information

I How does the formulation (5) serve the purpose of the composite parameterization

I How does the formulation (5) serve the purpose of the composite parameterization Supplemental Material to Identifying Alzheimer s Disease-Related Brain Regions from Multi-Modality Neuroimaging Data using Sparse Composite Linear Discrimination Analysis I How does the formulation (5)

More information

In a two-way contingency table, the null hypothesis of quasi-independence. (QI) usually arises for two main reasons: 1) some cells involve structural

In a two-way contingency table, the null hypothesis of quasi-independence. (QI) usually arises for two main reasons: 1) some cells involve structural Simulate and Reject Monte Carlo Exact Conditional Tests for Quasi-independence Peter W. F. Smith and John W. McDonald Department of Social Statistics, University of Southampton, Southampton, SO17 1BJ,

More information

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines A popular method for moving beyond linearity 2. Basis expansion and regularization 1 Idea: Augment the vector inputs x with additional variables which are transformation of x use linear models in this

More information

Radial Basis Function Networks: Algorithms

Radial Basis Function Networks: Algorithms Radial Basis Function Networks: Algorithms Neural Computation : Lecture 14 John A. Bullinaria, 2015 1. The RBF Mapping 2. The RBF Network Architecture 3. Computational Power of RBF Networks 4. Training

More information

NONPARAMETRIC REGRESSION SPLINES FOR GENERALIZED LINEAR MODELS IN THE PRESENCE OF MEASUREMENT ERROR

NONPARAMETRIC REGRESSION SPLINES FOR GENERALIZED LINEAR MODELS IN THE PRESENCE OF MEASUREMENT ERROR NONPARAMETRIC REGRESSION SPLINES FOR GENERALIZED LINEAR MODELS IN THE PRESENCE OF MEASUREMENT ERROR J. D. Maca July 1, 1997 Abstract The purpose of this manual is to demonstrate the usage of software for

More information

Feature Selection Using Modified-MCA Based Scoring Metric for Classification

Feature Selection Using Modified-MCA Based Scoring Metric for Classification 2011 International Conference on Information Communication and Management IPCSIT vol.16 (2011) (2011) IACSIT Press, Singapore Feature Selection Using Modified-MCA Based Scoring Metric for Classification

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 2015 MODULE 4 : Modelling experimental data Time allowed: Three hours Candidates should answer FIVE questions. All questions carry equal

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 13 UNSUPERVISED LEARNING If you have access to labeled training data, you know what to do. This is the supervised setting, in which you have a teacher telling

More information

Model Answers to The Next Pixel Prediction Task

Model Answers to The Next Pixel Prediction Task Model Answers to The Next Pixel Prediction Task December 2, 25. (Data preprocessing and visualization, 8 marks) (a) Solution. In Algorithm we are told that the data was discretized to 64 grey scale values,...,

More information

Equations and Functions, Variables and Expressions

Equations and Functions, Variables and Expressions Equations and Functions, Variables and Expressions Equations and functions are ubiquitous components of mathematical language. Success in mathematics beyond basic arithmetic depends on having a solid working

More information

Nonparametric Classification Methods

Nonparametric Classification Methods Nonparametric Classification Methods We now examine some modern, computationally intensive methods for regression and classification. Recall that the LDA approach constructs a line (or plane or hyperplane)

More information

J. Weston, A. Gammerman, M. Stitson, V. Vapnik, V. Vovk, C. Watkins. Technical Report. February 5, 1998

J. Weston, A. Gammerman, M. Stitson, V. Vapnik, V. Vovk, C. Watkins. Technical Report. February 5, 1998 Density Estimation using Support Vector Machines J. Weston, A. Gammerman, M. Stitson, V. Vapnik, V. Vovk, C. Watkins. Technical Report CSD-TR-97-3 February 5, 998!()+, -./ 3456 Department of Computer Science

More information

davidr Cornell University

davidr Cornell University 1 NONPARAMETRIC RANDOM EFFECTS MODELS AND LIKELIHOOD RATIO TESTS Oct 11, 2002 David Ruppert Cornell University www.orie.cornell.edu/ davidr (These transparencies and preprints available link to Recent

More information

USING PRINCIPAL COMPONENTS ANALYSIS FOR AGGREGATING JUDGMENTS IN THE ANALYTIC HIERARCHY PROCESS

USING PRINCIPAL COMPONENTS ANALYSIS FOR AGGREGATING JUDGMENTS IN THE ANALYTIC HIERARCHY PROCESS Analytic Hierarchy To Be Submitted to the the Analytic Hierarchy 2014, Washington D.C., U.S.A. USING PRINCIPAL COMPONENTS ANALYSIS FOR AGGREGATING JUDGMENTS IN THE ANALYTIC HIERARCHY PROCESS Natalie M.

More information

Geometric Transformations and Image Warping

Geometric Transformations and Image Warping Geometric Transformations and Image Warping Ross Whitaker SCI Institute, School of Computing University of Utah Univ of Utah, CS6640 2009 1 Geometric Transformations Greyscale transformations -> operate

More information

Minitab 17 commands Prepared by Jeffrey S. Simonoff

Minitab 17 commands Prepared by Jeffrey S. Simonoff Minitab 17 commands Prepared by Jeffrey S. Simonoff Data entry and manipulation To enter data by hand, click on the Worksheet window, and enter the values in as you would in any spreadsheet. To then save

More information

A review of spline function selection procedures in R

A review of spline function selection procedures in R Matthias Schmid Department of Medical Biometry, Informatics and Epidemiology University of Bonn joint work with Aris Perperoglou on behalf of TG2 of the STRATOS Initiative September 1, 2016 Introduction

More information

Minoru SASAKI and Kenji KITA. Department of Information Science & Intelligent Systems. Faculty of Engineering, Tokushima University

Minoru SASAKI and Kenji KITA. Department of Information Science & Intelligent Systems. Faculty of Engineering, Tokushima University Information Retrieval System Using Concept Projection Based on PDDP algorithm Minoru SASAKI and Kenji KITA Department of Information Science & Intelligent Systems Faculty of Engineering, Tokushima University

More information

Nonparametric Risk Attribution for Factor Models of Portfolios. October 3, 2017 Kellie Ottoboni

Nonparametric Risk Attribution for Factor Models of Portfolios. October 3, 2017 Kellie Ottoboni Nonparametric Risk Attribution for Factor Models of Portfolios October 3, 2017 Kellie Ottoboni Outline The problem Page 3 Additive model of returns Page 7 Euler s formula for risk decomposition Page 11

More information

Operator-Based Backward Motion Estimation. Aria Nosratinia and Michael T. Orchard. Beckman Institute for Advanced Science and Technology.

Operator-Based Backward Motion Estimation. Aria Nosratinia and Michael T. Orchard. Beckman Institute for Advanced Science and Technology. Operator-Based Backward Motion Estimation Aria Nosratinia and Michael T. Orchard Beckman Institute for Advanced Science and Technology University of Illinois at Urbana-Champaign 405 N. Mathews Ave., Urbana,

More information

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important

More information

Divide and Conquer Kernel Ridge Regression

Divide and Conquer Kernel Ridge Regression Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem

More information

Robust Regression. Robust Data Mining Techniques By Boonyakorn Jantaranuson

Robust Regression. Robust Data Mining Techniques By Boonyakorn Jantaranuson Robust Regression Robust Data Mining Techniques By Boonyakorn Jantaranuson Outline Introduction OLS and important terminology Least Median of Squares (LMedS) M-estimator Penalized least squares What is

More information

Stat 8053, Fall 2013: Additive Models

Stat 8053, Fall 2013: Additive Models Stat 853, Fall 213: Additive Models We will only use the package mgcv for fitting additive and later generalized additive models. The best reference is S. N. Wood (26), Generalized Additive Models, An

More information

Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling

Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling Session #176, Beatini Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling NCTM

More information

Incorporating Geospatial Data in House Price Indexes: A Hedonic Imputation Approach with Splines. Robert J. Hill and Michael Scholz

Incorporating Geospatial Data in House Price Indexes: A Hedonic Imputation Approach with Splines. Robert J. Hill and Michael Scholz Incorporating Geospatial Data in House Price Indexes: A Hedonic Imputation Approach with Splines Robert J. Hill and Michael Scholz Department of Economics University of Graz, Austria OeNB Workshop Vienna,

More information

Nonparametric regression using kernel and spline methods

Nonparametric regression using kernel and spline methods Nonparametric regression using kernel and spline methods Jean D. Opsomer F. Jay Breidt March 3, 016 1 The statistical model When applying nonparametric regression methods, the researcher is interested

More information

Fast Orthogonal Search For Training Radial Basis. Function Neural Networks. By Wahid Ahmed. Thesis Advisor: Donald M. Hummels, Ph.D.

Fast Orthogonal Search For Training Radial Basis. Function Neural Networks. By Wahid Ahmed. Thesis Advisor: Donald M. Hummels, Ph.D. Fast Orthogonal Search For Training Radial Basis Function Neural Networks By Wahid Ahmed Thesis Advisor: Donald M. Hummels, Ph.D. An Abstract of the Thesis Presented in Partial Fulllment of the Requirements

More information

All 0-1 Polytopes are. Abstract. We study the facial structure of two important permutation polytopes

All 0-1 Polytopes are. Abstract. We study the facial structure of two important permutation polytopes All 0-1 Polytopes are Traveling Salesman Polytopes L.J. Billera and A. Sarangarajan y Abstract We study the facial structure of two important permutation polytopes in R n2, the Birkho or assignment polytope

More information

An Overview MARS is an adaptive procedure for regression, and is well suited for high dimensional problems.

An Overview MARS is an adaptive procedure for regression, and is well suited for high dimensional problems. Data mining and Machine Learning, Mar 24, 2008 MARS: Multivariate Adaptive Regression Splines (Friedman 1991, Friedman and Silverman 1989) An Overview MARS is an adaptive procedure for regression, and

More information

North Carolina Standard Course of Study, 2003, grade 8 [NC] PH Course 3 Lesson

North Carolina Standard Course of Study, 2003, grade 8 [NC] PH Course 3 Lesson [NC] North Carolina Standard Course of Study, 2003, grade 8 PH Course 3 Lesson COMPETENCY GOAL 1: The learner will understand and compute with real 1.01 Develop number sense for the real numbers. 1-2,

More information

An Estimated General Cross Validation Function for Periodic Control Theoretic. smoothing splines

An Estimated General Cross Validation Function for Periodic Control Theoretic. smoothing splines An Estimated General Cross Validation Function for Periodic Control Theoretic Smoothing Splines M. Karasalo X. Hu and C. F. Martin Abstract In this paper, a method is developed for estimating the optimal

More information

and therefore the system throughput in a distributed database system [, 1]. Vertical fragmentation further enhances the performance of database transa

and therefore the system throughput in a distributed database system [, 1]. Vertical fragmentation further enhances the performance of database transa Vertical Fragmentation and Allocation in Distributed Deductive Database Systems Seung-Jin Lim Yiu-Kai Ng Department of Computer Science Brigham Young University Provo, Utah 80, U.S.A. Email: fsjlim,ngg@cs.byu.edu

More information

Y7 Learning Stage 1. Y7 Learning Stage 2. Y7 Learning Stage 3

Y7 Learning Stage 1. Y7 Learning Stage 2. Y7 Learning Stage 3 Y7 Learning Stage 1 Y7 Learning Stage 2 Y7 Learning Stage 3 Understand simple algebraic notation. Collect like terms to simplify algebraic expressions. Use coordinates in the first quadrant. Make a comparative

More information

SGN (4 cr) Chapter 11

SGN (4 cr) Chapter 11 SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter

More information

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material

More information

(Preliminary Version 2 ) Jai-Hoon Kim Nitin H. Vaidya. Department of Computer Science. Texas A&M University. College Station, TX

(Preliminary Version 2 ) Jai-Hoon Kim Nitin H. Vaidya. Department of Computer Science. Texas A&M University. College Station, TX Towards an Adaptive Distributed Shared Memory (Preliminary Version ) Jai-Hoon Kim Nitin H. Vaidya Department of Computer Science Texas A&M University College Station, TX 77843-3 E-mail: fjhkim,vaidyag@cs.tamu.edu

More information

1. In this problem we use Monte Carlo integration to approximate the area of a circle of radius, R = 1.

1. In this problem we use Monte Carlo integration to approximate the area of a circle of radius, R = 1. 1. In this problem we use Monte Carlo integration to approximate the area of a circle of radius, R = 1. A. For method 1, the idea is to conduct a binomial experiment using the random number generator:

More information

Linear penalized spline model estimation using ranked set sampling technique

Linear penalized spline model estimation using ranked set sampling technique Hacettepe Journal of Mathematics and Statistics Volume 46 (4) (2017), 669 683 Linear penalized spline model estimation using ranked set sampling technique Al Kadiri M A Abstract Benets of using Ranked

More information

The theory of the linear model 41. Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that

The theory of the linear model 41. Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that The theory of the linear model 41 Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that E(Y X) =X 0 b 0 0 the F-test statistic follows an F-distribution with (p p 0, n p) degrees

More information

Graph drawing in spectral layout

Graph drawing in spectral layout Graph drawing in spectral layout Maureen Gallagher Colleen Tygh John Urschel Ludmil Zikatanov Beginning: July 8, 203; Today is: October 2, 203 Introduction Our research focuses on the use of spectral graph

More information

Applied Regression Modeling: A Business Approach

Applied Regression Modeling: A Business Approach i Applied Regression Modeling: A Business Approach Computer software help: SAS SAS (originally Statistical Analysis Software ) is a commercial statistical software package based on a powerful programming

More information

Evaluation of Loop Subdivision Surfaces

Evaluation of Loop Subdivision Surfaces Evaluation of Loop Subdivision Surfaces Jos Stam Alias wavefront, Inc. 8 Third Ave, 8th Floor, Seattle, WA 980, U.S.A. jstam@aw.sgi.com Abstract This paper describes a technique to evaluate Loop subdivision

More information

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1 Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches

More information

Principal Component Analysis

Principal Component Analysis Copyright 2004, Casa Software Ltd. All Rights Reserved. 1 of 16 Principal Component Analysis Introduction XPS is a technique that provides chemical information about a sample that sets it apart from other

More information

Machine Learning / Jan 27, 2010

Machine Learning / Jan 27, 2010 Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,

More information

Institute for Advanced Computer Studies. Department of Computer Science. Direction of Arrival and The Rank-Revealing. E. C. Boman y. M. F.

Institute for Advanced Computer Studies. Department of Computer Science. Direction of Arrival and The Rank-Revealing. E. C. Boman y. M. F. University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{91{166 TR{2813 Direction of Arrival and The Rank-Revealing URV Decomposition E. C. Boman y

More information

Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet. Y. C. Pati R. Rezaiifar and P. S.

Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet. Y. C. Pati R. Rezaiifar and P. S. / To appear in Proc. of the 27 th Annual Asilomar Conference on Signals Systems and Computers, Nov. {3, 993 / Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet

More information

An Efficient Clustering for Crime Analysis

An Efficient Clustering for Crime Analysis An Efficient Clustering for Crime Analysis Malarvizhi S 1, Siddique Ibrahim 2 1 UG Scholar, Department of Computer Science and Engineering, Kumaraguru College Of Technology, Coimbatore, Tamilnadu, India

More information

A Robust Optimum Response Surface Methodology based On MM-estimator

A Robust Optimum Response Surface Methodology based On MM-estimator A Robust Optimum Response Surface Methodology based On MM-estimator, 2 HABSHAH MIDI,, 2 MOHD SHAFIE MUSTAFA,, 2 ANWAR FITRIANTO Department of Mathematics, Faculty Science, University Putra Malaysia, 434,

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

New user profile learning for extremely sparse data sets

New user profile learning for extremely sparse data sets New user profile learning for extremely sparse data sets Tomasz Hoffmann, Tadeusz Janasiewicz, and Andrzej Szwabe Institute of Control and Information Engineering, Poznan University of Technology, pl.

More information

Goals of the Lecture. SOC6078 Advanced Statistics: 9. Generalized Additive Models. Limitations of the Multiple Nonparametric Models (2)

Goals of the Lecture. SOC6078 Advanced Statistics: 9. Generalized Additive Models. Limitations of the Multiple Nonparametric Models (2) SOC6078 Advanced Statistics: 9. Generalized Additive Models Robert Andersen Department of Sociology University of Toronto Goals of the Lecture Introduce Additive Models Explain how they extend from simple

More information

Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension

Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension Claudia Sannelli, Mikio Braun, Michael Tangermann, Klaus-Robert Müller, Machine Learning Laboratory, Dept. Computer

More information

3. Data Analysis and Statistics

3. Data Analysis and Statistics 3. Data Analysis and Statistics 3.1 Visual Analysis of Data 3.2.1 Basic Statistics Examples 3.2.2 Basic Statistical Theory 3.3 Normal Distributions 3.4 Bivariate Data 3.1 Visual Analysis of Data Visual

More information

then present results of directly applying the SFFS feature selection algorithms for mammographic mass detection, discussing how noise and the sie of t

then present results of directly applying the SFFS feature selection algorithms for mammographic mass detection, discussing how noise and the sie of t The role of feature selection in building pattern recogniers for computer-aided diagnosis Clay Spence and Paul Sajda National Information Display Laboratory Sarno Corporation Princeton, NJ 08543-5300,

More information

Graphical Analysis of Data using Microsoft Excel [2016 Version]

Graphical Analysis of Data using Microsoft Excel [2016 Version] Graphical Analysis of Data using Microsoft Excel [2016 Version] Introduction In several upcoming labs, a primary goal will be to determine the mathematical relationship between two variable physical parameters.

More information

8. MINITAB COMMANDS WEEK-BY-WEEK

8. MINITAB COMMANDS WEEK-BY-WEEK 8. MINITAB COMMANDS WEEK-BY-WEEK In this section of the Study Guide, we give brief information about the Minitab commands that are needed to apply the statistical methods in each week s study. They are

More information