1D Regression. i.i.d. with mean 0. Univariate Linear Regression: fit by least squares. Minimize: to get. The set of all possible functions is...

Similar documents
Moving Beyond Linearity

Nonparametric Mixed-Effects Models for Longitudinal Data

Last time... Bias-Variance decomposition. This week

A popular method for moving beyond linearity. 2. Basis expansion and regularization 1. Examples of transformations. Piecewise-polynomials and splines

lecture 10: B-Splines

Splines and penalized regression

Curve fitting using linear models

Lecture 17: Smoothing splines, Local Regression, and GAMs

Doubly Cyclic Smoothing Splines and Analysis of Seasonal Daily Pattern of CO2 Concentration in Antarctica

Spline Models. Introduction to CS and NCS. Regression splines. Smoothing splines

This is called a linear basis expansion, and h m is the mth basis function For example if X is one-dimensional: f (X) = β 0 + β 1 X + β 2 X 2, or

Nonparametric Approaches to Regression

Generalized Additive Models

Lecture 16: High-dimensional regression, non-linear regression

Splines. Patrick Breheny. November 20. Introduction Regression splines (parametric) Smoothing splines (nonparametric)

Interactive Graphics. Lecture 9: Introduction to Spline Curves. Interactive Graphics Lecture 9: Slide 1

Design considerations

A comparison of spline methods in R for building explanatory models

Generalized Additive Model

Linear Methods for Regression and Shrinkage Methods

EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines

What is machine learning?

Lecture 9: Introduction to Spline Curves

Curve and Surface Fitting with Splines. PAUL DIERCKX Professor, Computer Science Department, Katholieke Universiteit Leuven, Belgium

Nonparametric regression using kernel and spline methods

The pspline Package. August 4, Author S original by Jim Ramsey R port by Brian Ripley

GAMs semi-parametric GLMs. Simon Wood Mathematical Sciences, University of Bath, U.K.

davidr Cornell University

In this course we will need a set of techniques to represent curves and surfaces in 2-d and 3-d. Some reasons for this include

BASIC LOESS, PBSPLINE & SPLINE

Knowledge Discovery and Data Mining

Assessing the Quality of the Natural Cubic Spline Approximation

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

Lecture 7: Splines and Generalized Additive Models

Consider functions such that then satisfies these properties: So is represented by the cubic polynomials on on and on.

The theory of the linear model 41. Theorem 2.5. Under the strong assumptions A3 and A5 and the hypothesis that

Machine Learning / Jan 27, 2010

Generalized additive models I

Analyzing Longitudinal Data Using Regression Splines

Moving Beyond Linearity

STAT 705 Introduction to generalized additive models

Computational Physics PHYS 420

Multiple-Choice Test Spline Method Interpolation COMPLETE SOLUTION SET

CS 450 Numerical Analysis. Chapter 7: Interpolation

Package freeknotsplines

CS130 : Computer Graphics Curves. Tamar Shinar Computer Science & Engineering UC Riverside

Improved smoothing spline regression by combining estimates of dierent smoothness

Statistical techniques for data analysis in Cosmology

CPSC 340: Machine Learning and Data Mining. More Regularization Fall 2017

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

CSE446: Linear Regression. Spring 2017

Nonparametric Risk Attribution for Factor Models of Portfolios. October 3, 2017 Kellie Ottoboni

Curves and Surfaces. CS475 / 675, Fall Siddhartha Chaudhuri

Chapter 5: Basis Expansion and Regularization

Four equations are necessary to evaluate these coefficients. Eqn

Fitting to a set of data. Lecture on fitting

Robust Regression. Robust Data Mining Techniques By Boonyakorn Jantaranuson

CS130 : Computer Graphics Curves (cont.) Tamar Shinar Computer Science & Engineering UC Riverside

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010

TECHNICAL REPORT NO December 11, 2001

Cubic spline interpolation

Lecture IV Bézier Curves

CSC 411: Lecture 02: Linear Regression

B-spline Curves. Smoother than other curve forms

See the course website for important information about collaboration and late policies, as well as where and when to turn in assignments.

Bezier Curves, B-Splines, NURBS

Fall CSCI 420: Computer Graphics. 4.2 Splines. Hao Li.

Need for Parametric Equations

Interpolating and approximating scattered 3D-data with hierarchical tensor product B-splines

Approximation of 3D-Parametric Functions by Bicubic B-spline Functions

GENREG DID THAT? Clay Barker Research Statistician Developer JMP Division, SAS Institute

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

A Curve Tutorial for Introductory Computer Graphics

3 Nonlinear Regression

Economics Nonparametric Econometrics

Applied Statistics : Practical 9

(Spline, Bezier, B-Spline)

Lecture VIII. Global Approximation Methods: I

A review of spline function selection procedures in R

Smoothing Dissimilarities for Cluster Analysis: Binary Data and Functional Data

An introduction to interpolation and splines

Nonparametric Methods Recap

P-spline ANOVA-type interaction models for spatio-temporal smoothing

2D Spline Curves. CS 4620 Lecture 18

Know it. Control points. B Spline surfaces. Implicit surfaces

Bayesian Time-Stratified-Petersen estimators for abundance. Sampling Protocol. Simon J. Bonner (UBC) Carl James Schwarz (SFU)

Splines. Parameterization of a Curve. Curve Representations. Roller coaster. What Do We Need From Curves in Computer Graphics? Modeling Complex Shapes

COMP3421. Global Lighting Part 2: Radiosity

Polynomial Splines over locally refined box-partitions

2D Spline Curves. CS 4620 Lecture 13

Mastery. PRECALCULUS Student Learning Targets

Knot-Placement to Avoid Over Fitting in B-Spline Scedastic Smoothing. Hirokazu Yanagihara* and Megu Ohtaki**

Natural Quartic Spline

Handout 4 - Interpolation Examples

Introduction to Computer Graphics

CSE446: Linear Regression. Spring 2017

Homework #6 Brief Solutions 2012

MS in Applied Statistics: Study Guide for the Data Science concentration Comprehensive Examination. 1. MAT 456 Applied Regression Analysis

CS-184: Computer Graphics. Today

HW 10 STAT 472, Spring 2018

Transcription:

1D Regression i.i.d. with mean 0. Univariate Linear Regression: fit by least squares. Minimize: to get. The set of all possible functions is... 1

Non-linear problems What if the underlying function is not linear? Try: fit non-linear function from a bag of functions Problem: which bag? The space of all functions is HUGE Another problem: We only have SOME data: want to find the underlying function but avoid noise Need to be selective in choosing possible non-linear functions 2

+ Basis expansion: polynomial terms Univariate LS has two basis functions:! "# $ % '& The resulting fit is a linear combination of $ ( ) *) $ )! : One way: add non-linear functions of to the bag. Polynomial terms seem as good as any:,+ -./102010354 Construct matrix, with: + + and fit linear regression with 4 " terms 3

Global vs Local fits One problem with polynomial regression: global fit Must find very good global basis for global fit: unlikely to find the true one Other way: fit locally with simple functions Why it works: It is easier to find a suitable basis for a part of a function. Tradeoff: in each part we only have a fraction of data to work with: must be extra-careful not to overfit. 4

Polynomial Splines Flebility: fit low-order polynomials in small windows of the support of. Most popular are order 4 (cubic) splines Must join the pieces somehow: with M-order splines we make sure derivatives up-to M-2 order match at knots Naive basis for cubic splines: 7 " 7 08 7 9 7 ;:7 but many coefficents constrained by macthing derivatives Truncated-power basis set: "< = ;:>? + @:AB&DC +FE $ equivalent to naive set plus constraints Procedure: 5

G G G choose knots,? + H- "#2010203JI populate matrix using truncated power basis set (in columns) each evaluated at all data points, (rows) Run linear regression with?? terms. Natural Cubic splines: linear beyond data: extra two constraints on each side The number of parameters (degrees of freedom) is now?

S Regularization Avoid knot-selection problem. Use all possible knots (unique s) But have over-parameterized regression (N+2 parameters, N data points) Need to regularize (shrink) coeficients: KL9MON subject to: SVU WBS X Y PRQ + T + + Without constraint we get usual least squares fit: here we get infinite number of them The constraint on T only allows those fits with certain T. W controls over-all smoothness of the final fit: W +[Z ]\^\ + \^\ Z `_a 7

c b This remarkably solves a general variational problem: KLJMaN PRQ c d e \^\ gf9 & _#f is in one-to-one correspondance with Y above. Solution: Natural Cubic Spline with knots at each. Benefit: Can get all fits h in O(N). 8

q( \ $ j i \ B-spline Basis Most smoothing splines computationally fitted using B-spline basis B-spline are a basis for polynomial splines on a closed interval. Each cubic B-spline spans at most 5 knots. Computationally, one sets up an i k< matrix of ordered, evaluated B-spline basis. Each column, -, is a - th B-spline, and its center moves from left-most to right-most point. has banded structure and so does c W, where: W lmh-n o \^\ pf9 `o \^\ + gfj `_#f One then solves a penalized regression problem: c W r \ts This is actually done using Choleski and 9

s Y q back-substitution to get O(N) running time. Conceptually, the function to be fitted is expanded into a B-spline basis set: + u + o + and fit obtained by constrained least-squares: ( KL9MON PRQ b vwv vwv subject to penalty: x H X Here x H is the familiar squared second-derivative functional: x H y \^\ gf9 & _#f and Y is one-to-one with c 10

U $ Equivalent DF Smoothing spline (and many other smoothing procedures) are usually called semi-parametric models Once you expand into basis set, it looks like any other regression BUT individual terms, n+ have no real meaning With penalties, one cannot count number of terms to get degrees of freedom Equivalent expression is needed for guidance and (approx) inference In regular regression: z { } L r U this is the trace of a hat, or projection matrix. All penalized regressions (including cubic 11

$ smoothing splines) are obtained by: (s ~ ~ U ~ c W r ~ U s s while is not a projection matrix, it has similar properties (it is a shrunk projection operator) Define: ƒ } L 12