Abstract R-splines are introduced as splines t with a polynomial null space plus the sum of radial basis functions. Thin plate splines are a special c
|
|
- Katherine King
- 5 years ago
- Views:
Transcription
1 R-Splines for Response Surface Modeling July 12, 2000 Sarah W. Hardy North Carolina State University Douglas W. Nychka National Center for Atmospheric Research Raleigh, NC Boulder, CO
2 Abstract R-splines are introduced as splines t with a polynomial null space plus the sum of radial basis functions. Thin plate splines are a special case of R-splines. By this broader denition of an R-spline, however, it includes splines in which 2m? d 0 where the traditional roughness penalty is not guaranteed to be non-negative denite. This papers discusses a modication of the roughness penalty that allows for the tting of reduced polynomial null spaces. A series of examples are used to demonstrate the behavior of this modied roughness penalty. Keywords: Thin plate spline, nonparametric, response surfaces, roughness penalty, Demmler-Reinsch basis functions. 1 Introduction Thin plate splines are nonparametric method for tting response surfaces. A thin plate spline is a smooth surface constructed by adding a base, or null space, polynomial component, and a sum of radial basis functions. The thin plate smoothing splines are a solution to a minimization problem and can be interpreted as a natural generalization of the traditional polynomial model estimated by minimizing residual sum of squares. Unlike the traditional formulation, however, the thin plate spline minimization problem also involves a roughness penalty. This roughness penalty is a function of m and d where where m? 1 is the degree of the polynomial null space and d is the number of explanatory variables. To guarantee that the roughness penalty is positive and the function is a true thin plate spline, 2m? d > 0. Thus, the size of the polynomial null space grows very quickly as a function of the number of explanatory variables. Many typical industrial experiments do not have enough observations to t the number of parameters needed for the required null space. Specically, this becomes problematic when there is a large set of explanatory variables and it is desirable to use an all subsets approach to variable selection for thin 2
3 plate spline modeling. R-splines arose in the context of attempting to modify the thin plate spline to allow for a greater number of explanatory variables in the model. R-splines are splines t with a polynomial null space plus the sum of radial basis functions. Clearly, thin plate splines fall into this category. By this broader denition of an R-spline, however, it also includes other splines. Specically, it includes splines in which 2m? d 0. In other words, the polynomial null space is reduced or of a lower order than required by the thin plate spline restraint. When 2m? d 0, the seamless way in which the polynomial function and roughness penalty t together (i.e, the space spanned by the polynomial terms is the null space of the roughness penalty) no longer holds or the relationship is \broken." The term broken spline will be used to indicate this kind of spline where the penalty matrix is chosen to be similar to the thin plate spline. The term R-splines refers to the broader class of splines including both broken and thin plate splines. This paper begins by discussing the thin plate spline roughness penalty. It then presents the broken spline modication to the roughness penalty. The Demmler-Reinsch basis function representation of splines is used to show the role the eigenvalues of the roughness penalty matrix have in the spline solution. A series of four two and three dimensional examples from data sets available in S-PLUS and FUNFITS (Nychka, et. al., 1996), is used to compare the roughness penalties for the thin plate and broken spline. Another ve dimensional example using data collected at Becton Dickinson is also presented. 2 Thin plate spline roughness penalty The roughness functional below, J m (f), will increase in magnitude as a function departs from an m? 1 order polynomial. This roughness functional is: 3
4 Z X m! J m (f) = < d 1!::: d!! m :du 1 1 :::@u d d The sum in the integrand is taken over all non-negative integer vectors,, such that P d = m. Clearly, for m? 1 order polynomials J m (f) = 0, because all m th derivatives are 0. The thin plate spline estimator of f (Wahba, 1990) is the minimizer of the sum of the mean-squared error and the roughness penalty which is weighted by a smoothing parameter. Details on the form of the thin plate spline can be found in the appendix. In matrix form, the thin plate spline estimate is f(x i ) = T + M where T T = 0 and 2m? d > 0; where T is the design matrix of a polynomial model of order m? 1 and M k;i = E(k x i? x k k; m; d). For linear combinations of basis functions that satisfy the constraints, J m (f) = T M > 0. In other words, M is guaranteed to be positive denite. Dening the matrix W nn, a diagonal matrix proportional to the reciprocal variances of the errors, allows for the following matrix representation of the penalized sums of squares: S = 1 n (Y? T? M)T W (Y? T? M) + T M: (1) After taking partial derivatives of (1) with respect to and, a QR decomposition of T, F T T = R , is used to enforce the constraint T T = 0. F nn is an orthogonal matrix that can be partioned F = [F1 j F2] where F1 has columns that span the column space of T, i.e T = F1R, and F2 is orthogonal to the column space of T. Reparameterizing by letting = F2!2,!1, and! T = (!1;!2) T allows for the problem to be posed in ridge 4
5 regression form: X T W X! + H! = X T W Y; where X = T MF2 T and H = : 0 F2 T MF 2 Note that in the ridge regression formulation the roughness penalty is represented by the matrix H. 3 Broken spline roughness penalty The degree of the polynomial component implies a specic roughness penalty in the thin plate spline, and thus if T, the design matrix for the polynomial component does not span P m?1 the roughness penalty minimized in the ridge regression formulation will not be the same roughness penalty as the one used to derive the thin plate spline. We will use the term broken spline to describe splines where T does not span P m?1. For purposes of this comparative discussion the T matrix for a thin plate spline will be denoted T P and the T matrix for a broken spline will be denoted T R. Note that the column space of T R is a subset of that of T P. Partition F F = [F a F b F c ] where F a spans the space of T R, F b spans the columns of T P that are not in T R, and F c spans the space orthogonal to T P. Thus in the QR decomposition of T P, F2 = F c and in the QR decomposition of T R, F2 = [F b F c ]: 5
6 For the broken spline H, the roughness penalty matrix, becomes H R = rr Fb T MF b Fc T MF b 0 Fb T MF c Fc T MF c ; as compared to the H for the thin plate spline which is H t = rr t?rt?r Fc T MF c : Hence, H R for the broken spline is the roughness penalty for the thin plate spline, plus other terms, H R = H t F T b MF b F T c MF b 0 F T b MF c : With thin plate splines, H t is guaranteed to be non-negative denite. Recall T M > 0 where T T = 0 and 2m? d > 0, which guarantees that M is positive denite. F2 T MF 2, is a quadratic form of a positive denite matrix and as such is also positive denite. H t, having F2 T MF 2 in the lower right block and zeros elsewhere, is non-negative denite. With broken splines, however, H R is no longer a non-negative denite matrix. In the computational solution to the thin plate spline, UDU T is the singular value decomposition of BHB so H t = B?1 UDU T B?1. Now, BHB is no longer positive denite and thus its singular value decomposition is UDV T, where U 6= V, and H R = B?1 UDV T B?1. Let H R = B?1 UDU T B?1, which is a non-negative denite matrix because D is a diagonal matrix of singular values (which are non-negative by denition) and H is a quadratic form of D. If H R is replaced by H R, a non-negative denite roughness penalty is minimized. Because H R is a symmetric matrix, U and V will be the same except for sign changes in 6
7 columns corresponding to the negative eigenvalues. Hence, H R, is the matrix that results from forcing the negative eigenvalues of H R to be positive. How \close" the two matrices are depends on the magnitude of these eigenvalues. The negative eigenvalues that occur correspond to the \missing" polynomial terms in the null space. Thus, the roughness penalty is forced into penalizing roughness in the surface that might be modeled by these missing polynomial terms, tting them instead with the exible radial basis functions. 3.1 Demmler-Reinsch basis functions Denition Constructing a Demmler-Reinsch basis for the smoothing problem aids in the understanding of the roles the eigenvalues and eigenvectors of H have in the spline solution. Before dening the Demmler-Reinsch basis, two inner products are dened: < h ; h >1= X k h (x k )W k h (x k ) and < h ; h >2= Z < d X m! 1!::: m m 1 1 :::@x d 1 1 :::@x d d! dx The Demmler-Reinsch basis is denoted by fg g and is dened by the three following properties: 1. fg g for 1 N spans the same subspace as f j g and the radial basis functions. 2. < g ; g >1= 1 for = and 0 for 6=. 3. < g ; g >2= D for = and 0 for 6=. In other words, this basis consists of linear combinations of the polynomial and radial basis functions that are orthogonal with respect to the inner products by which the weighted 7
8 residual sums of squares and the roughness penalty can be expressed. By convention D are in ascending order Role in spline solution Because this basis spans the same space as the components of f(x) we can now express both f(x) and y as a linear combinations of these basis functions: NX NX f(x k ) = g (x k ) and y k = u g (x k ): (2) =1 =1 Using these expression it can be easily shown that u = P n k=1 g (x k )W k Y k : Moreover, under the assumption that the random errors are independent and distributed N(0; 2 I), the u 's are also independent and are distributed N(; 2 I): With the Demmler-Reinsch basis functions the residual sum of squares and the roughness penalty can be expressed: nx nx (Y k? f(x k )) 2 w k = (u? ) 2 k=1 =1 and nx J m (f) = D 2 : =1 Thus, it is apparent that minimizing S (f) is equivalent to nding The minimizing expression is: min 2< n nx nx (u? ) 2 + D 2 : (3) =1 =1 and so, c = u 1 + D ; ^f(x) = NX =1 c g (x) = 8 NX =1 u 1 + D g (x): (4)
9 In fact, the G = fg g =1;::;N we are seeking is (cleverly) the same G we found in the ridge regression formulation and property (2) is easily veried by noting that G T (X T W X)G = I: To determine if property (3) holds, observe that G T HG = D. As an important point of clarication recall that the upper r r block of H is 0 where r is the dimension of the null space. This implies D = 0 for = 1; ::; r where D are the eigenvalues of G T HG. More intuitively, by virtue of the way in which the basis functions are dened, the roughness of the basis functions is J m (g ) = D and r of these basis functions will be polynomials whose roughness is 0. With this formulation, it is apparent that if the D 's, which are the eigenvalues of H, and the g 's, which are the Demmler-Reinsch basis functions, are correlated for the thin plate and broken spline, then the two roughness penalties are similar and measuring the same features of the data. 4 Examples 4.1 Example datasets Four sample datasets that are available in S-Plus (StatSci, 1993), the stack and ethanol datasets, and FUNFITS, the mini-triathalon and BD2 datasets were used to compare these roughness penalties for ts to the data with a thin plate spline and a broken spline. The stack data is from operation of a plant for the oxidation of ammonia to nitric acid, measured on 21 consecutive days. The explanatory variables are air ow to the plant, cooling water inlet temperature, and acid concentration and the response variable is percent of ammonia lost. The ethanol data frame records 88 measurements an experiment in which ethanol was burned in a single cylinder automobile test engine. The explanatory variables are the compression ratio of the engine and the equivalence ratio at which the 9
10 engine was run (a measure of the richness of the air/ethanol mix). The response variable is the concentration of nitric oxide and nitrogen dioxide in engine exhaust, normalized by the work done by the engine. The mini-triathalon data contains swimming, biking, and running times from 110 entrants in a Cary, NC event. In the example we are using swimming and biking times to predict running times. The BD2 dataset comes from results of a sequence of RSM designs in a DNA amplication experiment performed at Becton Dickinson Technologies. The explanatory variables are potassium phosphate ( a buer), magnesium acetate (a salt), and dimethyl sulfoxide ( a solvent). The response variable is yield. In summary, the ethanol and mini-triathalon datasets each have 2 explanatory variables and the stack and BD2 datasets have 3. Thin plate splines were t with m = 2 or a linear null space and the broken splines were t with a m = 1 or a constant null space. Figure 1 shows the plots of the eigenvalues of H for the two types of splines. These plots demonstrate that these eigenvalues are nearly equivalent for all but a few points. Notice that the number of eigenvalues that dier signicantly equals the number of explanatory variables, which in turn equals the number of linear terms missing from the polynomial null space. These are the specic eigenvalues that correspond to polynomial terms in the null space of the thin plate spline, and hence are 0 for the thin plate spline and positive for the broken spline. 10
11 minitri H eigenvalues H eigenvalues for m=1 and stack.loss H eigenvalues m= m=1 ethanol H eigenvalues BD2 H eigenvalues m= m=1 Figure 1: Comparison of eigenvalues for m=1 (broken splines) and (thin plate splines). 11
12 Figure 2 shows image plots of the correlations of the thin plate spline and broken spline Demmler-Reinsch basis functions for these four data sets. Lines on the graphs indicate the division of the basis functions associated with the null space and roughness penalty. The areas in the regions greater than the lines on the graph indicate the regions where the basis functions are those associated with the null space. The area enclosed within the lines indicates the basis functions associated with the roughness penalty, which these plots show are highly correlated. As expected, the area outside the lines is not as highly correlated, as the polynomials in the null space are not the same for the thin plate and broken splines. There is, however, some correlation between the basis functions representing the null space in the thin plate spline and those representing the roughness penalty in the broken spline. This indicates that the polynomial terms missing from the broken spline are being accounted for in some way by the roughness penalty. There is some dierence between the two sets of basis functions, but overall when ordered in ascending order by eigenvalues the two sets of basis functions match closely beyond the rst few functions related to the null space. Clearly, more theoretical work is needed to explicitly quantify the relationship given these promising empirical results. Figure 3 demonstrates the results of tting the mini-triathalon dataset with broken and thin plate splines. The ts are approximately the same and nothing essential has been sacriced by tting the reduced null space, again evidence that the roughness penalties are similar. 12
13 Correlation of Demmler-Reinsch basis functions for m=1 and minitri stack.loss m=1 m=1 ethanol BD m=1 m=1 Figure 2: Correlation of basis functions for m=1 (broken splines) and (thin plate splines). 13
14 Predicted Values Observed Values R^2 = 69.56% Q^2 = 64.88% Predicted values Residuals m=1, Broken spline RMSE = Effective number of parameters Average Prediction Error Eff. df. = 8 Res. df. = 102 GCV min = GCV Predicted Values Observed Values R^2 = 70.23% Q^2 = 64.66% Predicted values Residuals , Thin plate Spline RMSE = Effective number of parameters Average Prediction Error Eff. df. = 9.6 Res. df. = GCV min = GCV Figure 3: Diagnostic plots for the mini-triathalon dataset t with a broken and thin plate spline. 14
15 4.2 A larger example dataset example m= Figure 4: Comparison of H eigenvalues for the example dataset t with a broken and thin plate spline. The next example arose from research being conducted at Becton Dickinson Technologies and because of the proprietary nature of the data, the details can not be given. This example has 5 explanatory variables, and so m for a thin plate spline must be 3. In other words, for a thin plate spline a full quadratic null space with 21 terms must be t. The broken spline was t with m = 2, or a linear null space with 6 terms. Figure 4 is a scatterplot of the 291 eigenvalues of H for the broken versus the thin plate spline. This plot shows that the majority of eigenvalues are very closely related, except for 15 of the eigenvalues related to the terms in the null space of the thin plate spline that are not in the null space of the broken spline. Another approximately 15 eigenvalues are smaller for 15
16 the broken spline than for the thin plate spline. example m= Figure 5: Correlation of basis functions for the example dataset t with a broken and thin plate spline. Figure 5 shows that, again, the basis functions are highly correlated for the thin plate and broken spline. They are less correlated at the higher end than the previous examples because of the higher dimensionality. The diagnostic plots, Figure 6, indicate that in this 16
17 case the broken spline is doing a slightly better job of tting the data. Predicted Values Observed Values R^2 = 45.12% Q^2 = 10.18% Predicted values Residuals , Broken spline RMSE = Pure Error = Effective number of parameters Average Prediction Error Eff. df. = 46.3 Res. df. = GCV min = GCV Predicted Values Observed Values R^2 = 37.5% Q^2 = 6.3% Predicted values Residuals m=3, Thin plate Spline RMSE = Pure Error = Effective number of parameters Average Prediction Error Eff. df. = 46.3 Res. df. = GCV min = GCV Figure 6: Diagnostic plots for the example dataset t with a broken and thin plate spline. 5 Conclusions The evidence indicates that broken splines are a feasible model tting tool. Like the thin plate spline, the broken spline is the sum of polynomial terms and radial basis functions and as such it is a smooth function with considerably more exibility than the traditional polynomial model. Also like the thin plate spline, it is the solution to a minimization problem in which the mean-squared error and a roughness component are minimized. 17
18 It can be shown that the thin plate spline solution is the unique and optimal solution to the posed minimization problem. The broken spline is also a unique solution to the minimization problem given a xed roughness penalty matrix as described. In the thin plate spline, the roughness penalty is explicitly dened and makes intuitive sense from a physical perspective as the bending energy of a thin plate. In fact, positive denite matrices arise in many applications involving energy, and clearly only make sense for a roughness penalty that is to be minimized. With the broken spline, although the eective roughness penalty can not be explicitly dened, it is closely related to the \bending energy" roughness penalty and empirically gives good results. The new insight that it is possible to achieve good results using the sum of polynomial terms and radial basis functions to estimate smooth functions without the restrictions imposed for thin plate splines opens the door for experimentation using this modeling technique for many more types of datasets. A Denition of a thin plate spline The thin plate spline estimator of f (Wahba, 1990) is the minimizer of the following penalized sums of squares for a d-dimensional explanatory variable x: nx S (f) = 1 w i (y i? f(x i )) 2 + J m (f) for > 0: (5) n i=1 Thus, an ^f minimizing (5) will result in f with some level of smoothness dictated by. The function that minimizes this expression has the form tx NX f(x i ) = j (x i ) j + E(k x i? x k k; m; d) k (6) j=1 k=1 where tx j (x i ) j = 0; 1 j t and 2m? d > 0: (7) j=1 18
19 In this formulation, the j (x i ) are a set of t polynomial functions (of order m? 1) and E(k x i? x k k; m; d) are a set of N radial basis functions. It is assumed that j is estimable. The radial basis functions are explicitly dened as below: E(r; m; d) = 8 >< >: a md k r k (2m?d) log(k r k) a md k r k (2m?d) d even d odd where a md depends only on m and d. One standard way of determining is by generalized cross-validation (GCV) (Bates et al, 1987). B References Bates, D.M., Lindstrom, M.J., Wahba, G. and Yandell, B.S. (1987). GCVPACK - Routines for generalized cross validation. Comm. Stat. Sim. Comp. 16, Berlin, pp Nychka, D., B. Bailey, S. Ellner, P. Haaland, and M. O'Connell (1996). FunFits data analysis and statistical tools for estimating functions. Software and paper available from statlib. StatSci (1993) S-PLUS Reference Manual Vol. 2, Version 3.2 MathSoft, Inc., Seattle, WA. Wahba, G. (1990). Spline Models for Observational Data. Society for Industrial Applied Mathematics. Philadelphia PA. 19
2 <Author Name(s)> and Green, 1994). This research was motivated by a specic application with twelve potential independent variables. If the functiona
Variable Selection for Response Surface Modeling Using R-Splines Sarah W. HARDY Douglas W. NYCHKA North Carolina State University National Center for Atmospheric Research Raleigh, NC 27695-8203 Boulder,
More informationLinear Methods for Regression and Shrinkage Methods
Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors
More informationSoft Threshold Estimation for Varying{coecient Models 2 ations of certain basis functions (e.g. wavelets). These functions are assumed to be smooth an
Soft Threshold Estimation for Varying{coecient Models Artur Klinger, Universitat Munchen ABSTRACT: An alternative penalized likelihood estimator for varying{coecient regression in generalized linear models
More informationSpline Models. Introduction to CS and NCS. Regression splines. Smoothing splines
Spline Models Introduction to CS and NCS Regression splines Smoothing splines 3 Cubic Splines a knots: a< 1 < 2 < < m
More informationAdaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented wh
Adaptive Estimation of Distributions using Exponential Sub-Families Alan Gous Stanford University December 1996 Abstract: An algorithm is presented which, for a large-dimensional exponential family G,
More informationNonparametric Regression
Nonparametric Regression John Fox Department of Sociology McMaster University 1280 Main Street West Hamilton, Ontario Canada L8S 4M4 jfox@mcmaster.ca February 2004 Abstract Nonparametric regression analysis
More informationApproximate Smoothing Spline Methods for Large Data Sets in the Binary Case Dong Xiang, SAS Institute Inc. Grace Wahba, University of Wisconsin at Mad
DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 982 September 30, 1997 Approximate Smoothing Spline Methods for Large Data Sets in the Binary
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationLast time... Bias-Variance decomposition. This week
Machine learning, pattern recognition and statistical data modelling Lecture 4. Going nonlinear: basis expansions and splines Last time... Coryn Bailer-Jones linear regression methods for high dimensional
More informationGeneralized Additive Model
Generalized Additive Model by Huimin Liu Department of Mathematics and Statistics University of Minnesota Duluth, Duluth, MN 55812 December 2008 Table of Contents Abstract... 2 Chapter 1 Introduction 1.1
More informationAPPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES
APPLICATION OF THE FUZZY MIN-MAX NEURAL NETWORK CLASSIFIER TO PROBLEMS WITH CONTINUOUS AND DISCRETE ATTRIBUTES A. Likas, K. Blekas and A. Stafylopatis National Technical University of Athens Department
More informationLecture 17: Smoothing splines, Local Regression, and GAMs
Lecture 17: Smoothing splines, Local Regression, and GAMs Reading: Sections 7.5-7 STATS 202: Data mining and analysis November 6, 2017 1 / 24 Cubic splines Define a set of knots ξ 1 < ξ 2 < < ξ K. We want
More informationPackage rgcvpack. February 20, Index 6. Fitting Thin Plate Smoothing Spline. Fit thin plate splines of any order with user specified knots
Version 0.1-4 Date 2013/10/25 Title R Interface for GCVPACK Fortran Package Author Xianhong Xie Package rgcvpack February 20, 2015 Maintainer Xianhong Xie
More informationMoving Beyond Linearity
Moving Beyond Linearity The truth is never linear! 1/23 Moving Beyond Linearity The truth is never linear! r almost never! 1/23 Moving Beyond Linearity The truth is never linear! r almost never! But often
More informationSplines and penalized regression
Splines and penalized regression November 23 Introduction We are discussing ways to estimate the regression function f, where E(y x) = f(x) One approach is of course to assume that f has a certain shape,
More informationD-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview
Chapter 888 Introduction This procedure generates D-optimal designs for multi-factor experiments with both quantitative and qualitative factors. The factors can have a mixed number of levels. For example,
More information6 Randomized rounding of semidefinite programs
6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can
More informationLecture 16: High-dimensional regression, non-linear regression
Lecture 16: High-dimensional regression, non-linear regression Reading: Sections 6.4, 7.1 STATS 202: Data mining and analysis November 3, 2017 1 / 17 High-dimensional regression Most of the methods we
More informationRegression for non-linear data
Regression for non-linear data don t let students go model shopping NCTM National Conference Boston MA April 2015 Julie Graves graves@ncssm.edu The North Carolina School of Science and Mathematics Model
More informationDoubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica
Boston-Keio Workshop 2016. Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica... Mihoko Minami Keio University, Japan August 15, 2016 Joint work with
More informationFeature selection. Term 2011/2012 LSI - FIB. Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/ / 22
Feature selection Javier Béjar cbea LSI - FIB Term 2011/2012 Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/2012 1 / 22 Outline 1 Dimensionality reduction 2 Projections 3 Attribute selection
More informationSpatial Distributions of Precipitation Events from Regional Climate Models
Spatial Distributions of Precipitation Events from Regional Climate Models N. Lenssen September 2, 2010 1 Scientific Reason The Institute of Mathematics Applied to Geosciences (IMAGe) and the National
More informationEE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov.
EE613 Machine Learning for Engineers LINEAR REGRESSION Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. 4, 2015 1 Outline Multivariate ordinary least squares Singular value
More informationSTAT 705 Introduction to generalized additive models
STAT 705 Introduction to generalized additive models Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 22 Generalized additive models Consider a linear
More informationEE613 Machine Learning for Engineers LINEAR REGRESSION. Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov.
EE613 Machine Learning for Engineers LINEAR REGRESSION Sylvain Calinon Robot Learning & Interaction Group Idiap Research Institute Nov. 9, 2017 1 Outline Multivariate ordinary least squares Matlab code:
More informationLecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010
Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2010 1 / 26 Additive predictors
More informationTelecommunication and Informatics University of North Carolina, Technical University of Gdansk Charlotte, NC 28223, USA
A Decoder-based Evolutionary Algorithm for Constrained Parameter Optimization Problems S lawomir Kozie l 1 and Zbigniew Michalewicz 2 1 Department of Electronics, 2 Department of Computer Science, Telecommunication
More informationSplines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric)
Splines Patrick Breheny November 20 Patrick Breheny STA 621: Nonparametric Statistics 1/46 Introduction Introduction Problems with polynomial bases We are discussing ways to estimate the regression function
More informationGeneralized Additive Models
:p Texts in Statistical Science Generalized Additive Models An Introduction with R Simon N. Wood Contents Preface XV 1 Linear Models 1 1.1 A simple linear model 2 Simple least squares estimation 3 1.1.1
More information(Creating Arrays & Matrices) Applied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB (Creating Arrays & Matrices) Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional
More informationLecture 3: Camera Calibration, DLT, SVD
Computer Vision Lecture 3 23--28 Lecture 3: Camera Calibration, DL, SVD he Inner Parameters In this section we will introduce the inner parameters of the cameras Recall from the camera equations λx = P
More informationImproved smoothing spline regression by combining estimates of dierent smoothness
Available online at www.sciencedirect.com Statistics & Probability Letters 67 (2004) 133 140 Improved smoothing spline regression by combining estimates of dierent smoothness Thomas C.M. Lee Department
More informationHands-On Standards Deluxe Grades: 7, 8 States: California Content Standards
Hands-On Standards Deluxe Grades: 7, 8 States: Hands-On Standards Deluxe Edition Kit, Grades 5-6: Algebra Summary: This resource guide meets 5 and 6 math curriculum standards by matching activities to
More informationChapter 9 Robust Regression Examples
Chapter 9 Robust Regression Examples Chapter Table of Contents OVERVIEW...177 FlowChartforLMS,LTS,andMVE...179 EXAMPLES USING LMS AND LTS REGRESSION...180 Example 9.1 LMS and LTS with Substantial Leverage
More informationGeneralized Additive Models
Generalized Additive Models Statistics 135 Autumn 2005 Copyright c 2005 by Mark E. Irwin Generalized Additive Models GAMs are one approach to non-parametric regression in the multiple predictor setting.
More informationAn Approach to Identify the Number of Clusters
An Approach to Identify the Number of Clusters Katelyn Gao Heather Hardeman Edward Lim Cristian Potter Carl Meyer Ralph Abbey July 11, 212 Abstract In this technological age, vast amounts of data are generated.
More informationLocalization in Graphs. Richardson, TX Azriel Rosenfeld. Center for Automation Research. College Park, MD
CAR-TR-728 CS-TR-3326 UMIACS-TR-94-92 Samir Khuller Department of Computer Science Institute for Advanced Computer Studies University of Maryland College Park, MD 20742-3255 Localization in Graphs Azriel
More informationWhat is machine learning?
Machine learning, pattern recognition and statistical data modelling Lecture 12. The last lecture Coryn Bailer-Jones 1 What is machine learning? Data description and interpretation finding simpler relationship
More informationIntroduction to ANSYS DesignXplorer
Lecture 4 14. 5 Release Introduction to ANSYS DesignXplorer 1 2013 ANSYS, Inc. September 27, 2013 s are functions of different nature where the output parameters are described in terms of the input parameters
More information( ) = Y ˆ. Calibration Definition A model is calibrated if its predictions are right on average: ave(response Predicted value) = Predicted value.
Calibration OVERVIEW... 2 INTRODUCTION... 2 CALIBRATION... 3 ANOTHER REASON FOR CALIBRATION... 4 CHECKING THE CALIBRATION OF A REGRESSION... 5 CALIBRATION IN SIMPLE REGRESSION (DISPLAY.JMP)... 5 TESTING
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationRecommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu
Recommendation System Using Yelp Data CS 229 Machine Learning Jia Le Xu, Yingran Xu 1 Introduction Yelp Dataset Challenge provides a large number of user, business and review data which can be used for
More informationI How does the formulation (5) serve the purpose of the composite parameterization
Supplemental Material to Identifying Alzheimer s Disease-Related Brain Regions from Multi-Modality Neuroimaging Data using Sparse Composite Linear Discrimination Analysis I How does the formulation (5)
More informationIn a two-way contingency table, the null hypothesis of quasi-independence. (QI) usually arises for two main reasons: 1) some cells involve structural
Simulate and Reject Monte Carlo Exact Conditional Tests for Quasi-independence Peter W. F. Smith and John W. McDonald Department of Social Statistics, University of Southampton, Southampton, SO17 1BJ,
More informationA popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines
A popular method for moving beyond linearity 2. Basis expansion and regularization 1 Idea: Augment the vector inputs x with additional variables which are transformation of x use linear models in this
More informationRadial Basis Function Networks: Algorithms
Radial Basis Function Networks: Algorithms Neural Computation : Lecture 14 John A. Bullinaria, 2015 1. The RBF Mapping 2. The RBF Network Architecture 3. Computational Power of RBF Networks 4. Training
More informationNONPARAMETRIC REGRESSION SPLINES FOR GENERALIZED LINEAR MODELS IN THE PRESENCE OF MEASUREMENT ERROR
NONPARAMETRIC REGRESSION SPLINES FOR GENERALIZED LINEAR MODELS IN THE PRESENCE OF MEASUREMENT ERROR J. D. Maca July 1, 1997 Abstract The purpose of this manual is to demonstrate the usage of software for
More informationFeature Selection Using Modified-MCA Based Scoring Metric for Classification
2011 International Conference on Information Communication and Management IPCSIT vol.16 (2011) (2011) IACSIT Press, Singapore Feature Selection Using Modified-MCA Based Scoring Metric for Classification
More informationEXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 2015 MODULE 4 : Modelling experimental data Time allowed: Three hours Candidates should answer FIVE questions. All questions carry equal
More informationA Course in Machine Learning
A Course in Machine Learning Hal Daumé III 13 UNSUPERVISED LEARNING If you have access to labeled training data, you know what to do. This is the supervised setting, in which you have a teacher telling
More informationModel Answers to The Next Pixel Prediction Task
Model Answers to The Next Pixel Prediction Task December 2, 25. (Data preprocessing and visualization, 8 marks) (a) Solution. In Algorithm we are told that the data was discretized to 64 grey scale values,...,
More informationEquations and Functions, Variables and Expressions
Equations and Functions, Variables and Expressions Equations and functions are ubiquitous components of mathematical language. Success in mathematics beyond basic arithmetic depends on having a solid working
More informationNonparametric Classification Methods
Nonparametric Classification Methods We now examine some modern, computationally intensive methods for regression and classification. Recall that the LDA approach constructs a line (or plane or hyperplane)
More informationJ. Weston, A. Gammerman, M. Stitson, V. Vapnik, V. Vovk, C. Watkins. Technical Report. February 5, 1998
Density Estimation using Support Vector Machines J. Weston, A. Gammerman, M. Stitson, V. Vapnik, V. Vovk, C. Watkins. Technical Report CSD-TR-97-3 February 5, 998!()+, -./ 3456 Department of Computer Science
More informationdavidr Cornell University
1 NONPARAMETRIC RANDOM EFFECTS MODELS AND LIKELIHOOD RATIO TESTS Oct 11, 2002 David Ruppert Cornell University www.orie.cornell.edu/ davidr (These transparencies and preprints available link to Recent
More informationUSING PRINCIPAL COMPONENTS ANALYSIS FOR AGGREGATING JUDGMENTS IN THE ANALYTIC HIERARCHY PROCESS
Analytic Hierarchy To Be Submitted to the the Analytic Hierarchy 2014, Washington D.C., U.S.A. USING PRINCIPAL COMPONENTS ANALYSIS FOR AGGREGATING JUDGMENTS IN THE ANALYTIC HIERARCHY PROCESS Natalie M.
More informationGeometric Transformations and Image Warping
Geometric Transformations and Image Warping Ross Whitaker SCI Institute, School of Computing University of Utah Univ of Utah, CS6640 2009 1 Geometric Transformations Greyscale transformations -> operate
More informationMinitab 17 commands Prepared by Jeffrey S. Simonoff
Minitab 17 commands Prepared by Jeffrey S. Simonoff Data entry and manipulation To enter data by hand, click on the Worksheet window, and enter the values in as you would in any spreadsheet. To then save
More informationA review of spline function selection procedures in R
Matthias Schmid Department of Medical Biometry, Informatics and Epidemiology University of Bonn joint work with Aris Perperoglou on behalf of TG2 of the STRATOS Initiative September 1, 2016 Introduction
More informationMinoru SASAKI and Kenji KITA. Department of Information Science & Intelligent Systems. Faculty of Engineering, Tokushima University
Information Retrieval System Using Concept Projection Based on PDDP algorithm Minoru SASAKI and Kenji KITA Department of Information Science & Intelligent Systems Faculty of Engineering, Tokushima University
More informationNonparametric Risk Attribution for Factor Models of Portfolios. October 3, 2017 Kellie Ottoboni
Nonparametric Risk Attribution for Factor Models of Portfolios October 3, 2017 Kellie Ottoboni Outline The problem Page 3 Additive model of returns Page 7 Euler s formula for risk decomposition Page 11
More informationOperator-Based Backward Motion Estimation. Aria Nosratinia and Michael T. Orchard. Beckman Institute for Advanced Science and Technology.
Operator-Based Backward Motion Estimation Aria Nosratinia and Michael T. Orchard Beckman Institute for Advanced Science and Technology University of Illinois at Urbana-Champaign 405 N. Mathews Ave., Urbana,
More informationLocally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling
Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important
More informationDivide and Conquer Kernel Ridge Regression
Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem
More informationRobust Regression. Robust Data Mining Techniques By Boonyakorn Jantaranuson
Robust Regression Robust Data Mining Techniques By Boonyakorn Jantaranuson Outline Introduction OLS and important terminology Least Median of Squares (LMedS) M-estimator Penalized least squares What is
More informationStat 8053, Fall 2013: Additive Models
Stat 853, Fall 213: Additive Models We will only use the package mgcv for fitting additive and later generalized additive models. The best reference is S. N. Wood (26), Generalized Additive Models, An
More informationHands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling
Session #176, Beatini Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling Hands-On Activities + Technology = Mathematical Understanding Through Authentic Modeling NCTM
More informationIncorporating Geospatial Data in House Price Indexes: A Hedonic Imputation Approach with Splines. Robert J. Hill and Michael Scholz
Incorporating Geospatial Data in House Price Indexes: A Hedonic Imputation Approach with Splines Robert J. Hill and Michael Scholz Department of Economics University of Graz, Austria OeNB Workshop Vienna,
More informationNonparametric regression using kernel and spline methods
Nonparametric regression using kernel and spline methods Jean D. Opsomer F. Jay Breidt March 3, 016 1 The statistical model When applying nonparametric regression methods, the researcher is interested
More informationFast Orthogonal Search For Training Radial Basis. Function Neural Networks. By Wahid Ahmed. Thesis Advisor: Donald M. Hummels, Ph.D.
Fast Orthogonal Search For Training Radial Basis Function Neural Networks By Wahid Ahmed Thesis Advisor: Donald M. Hummels, Ph.D. An Abstract of the Thesis Presented in Partial Fulllment of the Requirements
More informationAll 0-1 Polytopes are. Abstract. We study the facial structure of two important permutation polytopes
All 0-1 Polytopes are Traveling Salesman Polytopes L.J. Billera and A. Sarangarajan y Abstract We study the facial structure of two important permutation polytopes in R n2, the Birkho or assignment polytope
More informationAn Overview MARS is an adaptive procedure for regression, and is well suited for high dimensional problems.
Data mining and Machine Learning, Mar 24, 2008 MARS: Multivariate Adaptive Regression Splines (Friedman 1991, Friedman and Silverman 1989) An Overview MARS is an adaptive procedure for regression, and
More informationNorth Carolina Standard Course of Study, 2003, grade 8 [NC] PH Course 3 Lesson
[NC] North Carolina Standard Course of Study, 2003, grade 8 PH Course 3 Lesson COMPETENCY GOAL 1: The learner will understand and compute with real 1.01 Develop number sense for the real numbers. 1-2,
More informationAn Estimated General Cross Validation Function for Periodic Control Theoretic. smoothing splines
An Estimated General Cross Validation Function for Periodic Control Theoretic Smoothing Splines M. Karasalo X. Hu and C. F. Martin Abstract In this paper, a method is developed for estimating the optimal
More informationand therefore the system throughput in a distributed database system [, 1]. Vertical fragmentation further enhances the performance of database transa
Vertical Fragmentation and Allocation in Distributed Deductive Database Systems Seung-Jin Lim Yiu-Kai Ng Department of Computer Science Brigham Young University Provo, Utah 80, U.S.A. Email: fsjlim,ngg@cs.byu.edu
More informationY7 Learning Stage 1. Y7 Learning Stage 2. Y7 Learning Stage 3
Y7 Learning Stage 1 Y7 Learning Stage 2 Y7 Learning Stage 3 Understand simple algebraic notation. Collect like terms to simplify algebraic expressions. Use coordinates in the first quadrant. Make a comparative
More informationSGN (4 cr) Chapter 11
SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter
More informationAdvanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material
More information(Preliminary Version 2 ) Jai-Hoon Kim Nitin H. Vaidya. Department of Computer Science. Texas A&M University. College Station, TX
Towards an Adaptive Distributed Shared Memory (Preliminary Version ) Jai-Hoon Kim Nitin H. Vaidya Department of Computer Science Texas A&M University College Station, TX 77843-3 E-mail: fjhkim,vaidyag@cs.tamu.edu
More information1. In this problem we use Monte Carlo integration to approximate the area of a circle of radius, R = 1.
1. In this problem we use Monte Carlo integration to approximate the area of a circle of radius, R = 1. A. For method 1, the idea is to conduct a binomial experiment using the random number generator:
More informationLinear penalized spline model estimation using ranked set sampling technique
Hacettepe Journal of Mathematics and Statistics Volume 46 (4) (2017), 669 683 Linear penalized spline model estimation using ranked set sampling technique Al Kadiri M A Abstract Benets of using Ranked
More informationThe theory of the linear model 41. Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that
The theory of the linear model 41 Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that E(Y X) =X 0 b 0 0 the F-test statistic follows an F-distribution with (p p 0, n p) degrees
More informationGraph drawing in spectral layout
Graph drawing in spectral layout Maureen Gallagher Colleen Tygh John Urschel Ludmil Zikatanov Beginning: July 8, 203; Today is: October 2, 203 Introduction Our research focuses on the use of spectral graph
More informationApplied Regression Modeling: A Business Approach
i Applied Regression Modeling: A Business Approach Computer software help: SAS SAS (originally Statistical Analysis Software ) is a commercial statistical software package based on a powerful programming
More informationEvaluation of Loop Subdivision Surfaces
Evaluation of Loop Subdivision Surfaces Jos Stam Alias wavefront, Inc. 8 Third Ave, 8th Floor, Seattle, WA 980, U.S.A. jstam@aw.sgi.com Abstract This paper describes a technique to evaluate Loop subdivision
More informationPreface to the Second Edition. Preface to the First Edition. 1 Introduction 1
Preface to the Second Edition Preface to the First Edition vii xi 1 Introduction 1 2 Overview of Supervised Learning 9 2.1 Introduction... 9 2.2 Variable Types and Terminology... 9 2.3 Two Simple Approaches
More informationPrincipal Component Analysis
Copyright 2004, Casa Software Ltd. All Rights Reserved. 1 of 16 Principal Component Analysis Introduction XPS is a technique that provides chemical information about a sample that sets it apart from other
More informationMachine Learning / Jan 27, 2010
Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,
More informationInstitute for Advanced Computer Studies. Department of Computer Science. Direction of Arrival and The Rank-Revealing. E. C. Boman y. M. F.
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{91{166 TR{2813 Direction of Arrival and The Rank-Revealing URV Decomposition E. C. Boman y
More informationOrthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet. Y. C. Pati R. Rezaiifar and P. S.
/ To appear in Proc. of the 27 th Annual Asilomar Conference on Signals Systems and Computers, Nov. {3, 993 / Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet
More informationAn Efficient Clustering for Crime Analysis
An Efficient Clustering for Crime Analysis Malarvizhi S 1, Siddique Ibrahim 2 1 UG Scholar, Department of Computer Science and Engineering, Kumaraguru College Of Technology, Coimbatore, Tamilnadu, India
More informationA Robust Optimum Response Surface Methodology based On MM-estimator
A Robust Optimum Response Surface Methodology based On MM-estimator, 2 HABSHAH MIDI,, 2 MOHD SHAFIE MUSTAFA,, 2 ANWAR FITRIANTO Department of Mathematics, Faculty Science, University Putra Malaysia, 434,
More informationDetecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationNew user profile learning for extremely sparse data sets
New user profile learning for extremely sparse data sets Tomasz Hoffmann, Tadeusz Janasiewicz, and Andrzej Szwabe Institute of Control and Information Engineering, Poznan University of Technology, pl.
More informationGoals of the Lecture. SOC6078 Advanced Statistics: 9. Generalized Additive Models. Limitations of the Multiple Nonparametric Models (2)
SOC6078 Advanced Statistics: 9. Generalized Additive Models Robert Andersen Department of Sociology University of Toronto Goals of the Lecture Introduce Additive Models Explain how they extend from simple
More informationEstimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension
Estimating Noise and Dimensionality in BCI Data Sets: Towards Illiteracy Comprehension Claudia Sannelli, Mikio Braun, Michael Tangermann, Klaus-Robert Müller, Machine Learning Laboratory, Dept. Computer
More information3. Data Analysis and Statistics
3. Data Analysis and Statistics 3.1 Visual Analysis of Data 3.2.1 Basic Statistics Examples 3.2.2 Basic Statistical Theory 3.3 Normal Distributions 3.4 Bivariate Data 3.1 Visual Analysis of Data Visual
More informationthen present results of directly applying the SFFS feature selection algorithms for mammographic mass detection, discussing how noise and the sie of t
The role of feature selection in building pattern recogniers for computer-aided diagnosis Clay Spence and Paul Sajda National Information Display Laboratory Sarno Corporation Princeton, NJ 08543-5300,
More informationGraphical Analysis of Data using Microsoft Excel [2016 Version]
Graphical Analysis of Data using Microsoft Excel [2016 Version] Introduction In several upcoming labs, a primary goal will be to determine the mathematical relationship between two variable physical parameters.
More information8. MINITAB COMMANDS WEEK-BY-WEEK
8. MINITAB COMMANDS WEEK-BY-WEEK In this section of the Study Guide, we give brief information about the Minitab commands that are needed to apply the statistical methods in each week s study. They are
More information