Stat 4510/7510 Homework 6
|
|
- Garry Flowers
- 5 years ago
- Views:
Transcription
1 Stat 4510/7510 1/11. Stat 4510/7510 Homework 6 Instructions: Please list your name and student number clearly. In order to receive credit for a problem, your solution must show sufficient details so that the grader can determine how you obtained your answer. 1. The file CommunityCrime.csv is a dataset containing 319 observations on 123 variables. The observations are communities within the United States. The data combines socio-economic data from the 1990 US Census, law enforcement data from the 1990 US LEMAS survey, and crime data from the 1995 FBI Uniform Crime Reporting program. A detailed description of all variables is available at We seek to predict the variable ViolentCrimesPerPop, the total number of violent crimes per 100,000 people. Note: when asked to perform cross validation to select a tuning parameter, be sure to conduct this cross validation on the training data only, then see how well your cross validated tuning parameter does on the test data. (a). Set a seed of 1 and split the data into a 90% training set, and a 10% test set. crime=read.csv("communitycrime.csv") set.seed(1) train.obs=sample(1:nrow(crime),.9*nrow(crime),replace=false) crime.train = crime[train.obs,] crime.test = crime[-train.obs,] VC.test = crime.test$violentcrimesperpop (b). Fit a linear model using least squares on the training set. Report the test error obtained. crime.lm=lm(violentcrimesperpop~., data=crime.train) lm.pred=predict.lm(crime.lm,crime.test) mean((vc.test-lm.pred)^2) ## [1] The MSPE for the linear model fitted with least squares is (c). Fit a ridge regression model on the training set with λ chosen by cross-validation. Report the test error obtained. 1
2 Stat 4510/7510 2/11 library(glmnet) ## Loading required package: Matrix ## Loading required package: foreach ## Loaded glmnet X <- model.matrix(violentcrimesperpop~., data=crime.train) cv.out=cv.glmnet(x,y=crime.train$violentcrimesperpop,alpha=0) plot(cv.out) Mean Squared Error log(lambda) 2
3 Stat 4510/7510 3/11 cv.out$lambda.min ## [1] ridge.mod=glmnet(x = X,y=crime.train$ViolentCrimesPerPop,alpha=0, lambda = cv.out$lambda.min) X.new <- model.matrix(violentcrimesperpop~., data=crime.test) ridge.pred=predict(ridge.mod,newx=x.new,s=cv.out$lambda.min) mean((vc.test-ridge.pred)^2) ## [1] The MSPE for the linear model fitted with ridge regression is The optimal λ value was 0.31 (d). Fit a lasso model on the training set with λ chosen by cross-validation. Report the test error obtained, along with the number of non-zero coefficient estimates. X <- model.matrix(violentcrimesperpop~., data=crime.train) cv.out=cv.glmnet(x,y=crime.train$violentcrimesperpop,alpha=1) plot(cv.out) 3
4 Stat 4510/7510 4/ Mean Squared Error log(lambda) cv.out$lambda.min ## [1] lasso.mod=glmnet(x = X,y=crime.train$ViolentCrimesPerPop,alpha=1, lambda = cv.out$lambda.min) X.new <- model.matrix(violentcrimesperpop~., data=crime.test) lasso.pred=predict(lasso.mod,newx=x.new,s=cv.out$lambda.min) mean((vc.test-lasso.pred)^2) ## [1]
5 Stat 4510/7510 5/11 The MSPE for the linear model fitted with lasso is There are 16 non-zeo coefficient estimates. The optimal λ value is (e). Fit a PCR model on the training set with M chosen by crossvalidation. Report the test error obtained along with the value of M selected by cross-validation. library(pls) ## ## Attaching package: pls ## The following object is masked from package:stats : ## ## loadings pcr.mod=pcr(violentcrimesperpop~., data=crime.train,scale=false, validation="cv") validationplot(pcr.mod,val.type="msep",xlim=c(0,40),ylim=c(.025,.03)) 5
6 Stat 4510/7510 6/11 ViolentCrimesPerPop MSEP number of components pcr.pred=predict(pcr.mod,newdata=x.new[,-1],ncomp=16) mean((vc.test-pcr.pred)^2) ## [1] The value of M that minimizes cross-validation error is 16. The MSPE for the linear model fitted with PCR is (f). Fit a PLS model on the training set with M chosen by cross-validation. Report the test error obtained, along with the value of M selected by cross-validation. 6
7 Stat 4510/7510 7/11 library(pls) pls.mod=plsr(violentcrimesperpop~., data=crime.train, scale=true, validation="cv") validationplot(pls.mod,val.type="msep",ylim=c(0.02,.04),xlim=c(0,10)) ViolentCrimesPerPop MSEP number of components pls.pred=predict(pls.mod,x.new[,-1],ncomp=3) mean((vc.test-pls.pred)^2) ## [1] The lowest cross-validation error occurs when only M = 3 partial least squares directions are used. The MSPE is again
8 Stat 4510/7510 8/11 (g). Comment on the above parts and how well you believe we can predict violent crime rate using these methods. 2. Each of the four approaches above outperforms the original least squares fitted linear model in terms of out-of-sample prediction. This is because the least squares method suffers from over-fitting due to the large number of possible predictor variables. Overall, we can predict violent crime rate marginally well as the predictive standard error is 0.17 (sqrt(0.032)) which is still somewhat large for data that is restrained to be between 0 and 1. Run the following R code: set.seed(1) n<-80 p<-50 X<-matrix(0,nrow = n,ncol=p) for(j in 1:p) { X[,j]<-runif(n = n,min = 0,max = 1) } beta0<-2 betas<-rep(0,p) betas[1:3]<-c(1,2,3) betas<-matrix(betas,ncol=1) Y<-beta0 + X %*% betas + rnorm(n,0,1) (a). Briefly describe what this code is doing. What are n and p? Which X j are related to Y? What are the true values of the β j? The code is generating a matrix of predicor variables, X, that is of dimension n p. Then, setting the intercept term to 2, the first three coefficients to 1, 2, and 3, and the remaining p 3 coefficients to 0, it generates the response variable from a normal distribution with mean β 0 + Xβ and variance 1. Here, n = 80, the number of observations, p = 50, the number of possible covariates. (b). Run a linear model with the lm() function and examine the coefficients. Do the estimates match your description in (a)? sim.lm=lm(y~x) No, the intercept is estimated at 1.19, and the first three coefficient estimtes are 2.98, 2.37, and They all are significantly different than zero but not very close to the true values. Also, five other covariates are found significant at the α = 0.10 level. 8
9 Stat 4510/7510 9/11 (c). Conduct lasso with a grid of λ ranging from 10 2 to 1 and construct a traceplot of coefficient value vs λ for all 50 variables. Color the first 3 X variables (which are related to Y ) red and all others black. Describe your findings. lasso.mod<-glmnet(x,y,alpha=1,lambda = 10^seq(from=-2,to=.1,length=100)) plot(y=coef(lasso.mod)[2,],x=lasso.mod$lambda,type="l",ylim=c(-1,3),col=2) lines(y=coef(lasso.mod)[3,],x=lasso.mod$lambda,type="l",col=2) lines(y=coef(lasso.mod)[4,],x=lasso.mod$lambda,type="l",col=2) for(j in 5:(p+1)) {lines(y=coef(lasso.mod)[j,],x=lasso.mod$lambda,type="l",col=1)} coef(lasso.mod)[2, ] lasso.mod$lambda The lasso model has correctly given much large coefficient estimates to the three signal predictors than the other noise predictors in the model. For values of λ > 0.3, the coefficients of the noise predictors all are shrunk to 0. 9
10 Stat 4510/ /11 (d). Repeat (c) with ridge regression and report the differences. You may need to change the range for λ to get an informative plot. ridge.mod<-glmnet(x,y,alpha=0,lambda = 10^seq(from=-2,to=.1,length=100)) plot(y=coef(ridge.mod)[2,],x=ridge.mod$lambda,type="l",ylim=c(-1,3),col=2) lines(y=coef(ridge.mod)[3,],x=ridge.mod$lambda,type="l",col=2) lines(y=coef(ridge.mod)[4,],x=ridge.mod$lambda,type="l",col=2) for(j in 5:(p+1)) {lines(y=coef(ridge.mod)[j,],x=ridge.mod$lambda,type="l",col=1)} coef(ridge.mod)[2, ] ridge.mod$lambda Similarly to the lasso, the ridge regression approach also finds the first three coefficients the most significant. The remaining noise covariates have much smaller coefficient estiamtes. (e). Conduct a lasso again but use cross validation to select the best λ. Report this value as well 10
11 Stat 4510/ /11 as the results of the model selection it performs. Did it find all 3 of our truly significant variables significant? Any false positives? cv.out=cv.glmnet(x,y,alpha=1) cv.out$lambda.min ## [1] lasso.mod=glmnet(x,y,alpha=1,lambda = cv.out$lambda.min) lasso.coef=predict(lasso.mod,type="coefficients",s=cv.out$lambda.min) The chosen value of λ is The intercept term and first three coefficients were found significant. There were 14 false positive covariates. 11
14. League: A factor with levels A and N indicating player s league at the end of 1986
PENALIZED REGRESSION Ridge and The LASSO Note: The example contained herein was copied from the lab exercise in Chapter 6 of Introduction to Statistical Learning by. For this exercise, we ll use some baseball
More informationLasso. November 14, 2017
Lasso November 14, 2017 Contents 1 Case Study: Least Absolute Shrinkage and Selection Operator (LASSO) 1 1.1 The Lasso Estimator.................................... 1 1.2 Computation of the Lasso Solution............................
More informationPenalized regression Statistical Learning, 2011
Penalized regression Statistical Learning, 2011 Niels Richard Hansen September 19, 2011 Penalized regression is implemented in several different R packages. Ridge regression can, in principle, be carried
More information1 StatLearn Practical exercise 5
1 StatLearn Practical exercise 5 Exercise 1.1. Download the LA ozone data set from the book homepage. We will be regressing the cube root of the ozone concentration on the other variables. Divide the data
More informationLinear Model Selection and Regularization. especially usefull in high dimensions p>>100.
Linear Model Selection and Regularization especially usefull in high dimensions p>>100. 1 Why Linear Model Regularization? Linear models are simple, BUT consider p>>n, we have more features than data records
More informationLecture 13: Model selection and regularization
Lecture 13: Model selection and regularization Reading: Sections 6.1-6.2.1 STATS 202: Data mining and analysis October 23, 2017 1 / 17 What do we know so far In linear regression, adding predictors always
More informationPackage parcor. February 20, 2015
Type Package Package parcor February 20, 2015 Title Regularized estimation of partial correlation matrices Version 0.2-6 Date 2014-09-04 Depends MASS, glmnet, ppls, Epi, GeneNet Author, Juliane Schaefer
More informationRegularization Methods. Business Analytics Practice Winter Term 2015/16 Stefan Feuerriegel
Regularization Methods Business Analytics Practice Winter Term 2015/16 Stefan Feuerriegel Today s Lecture Objectives 1 Avoiding overfitting and improving model interpretability with the help of regularization
More informationLecture 16: High-dimensional regression, non-linear regression
Lecture 16: High-dimensional regression, non-linear regression Reading: Sections 6.4, 7.1 STATS 202: Data mining and analysis November 3, 2017 1 / 17 High-dimensional regression Most of the methods we
More informationHW 10 STAT 472, Spring 2018
HW 10 STAT 472, Spring 2018 1) (0 points) Do parts (a), (b), (c), and (e) of Exercise 2 on p. 298 of ISL. 2) (0 points) Do Exercise 3 on p. 298 of ISL. 3) For this problem, you can merely submit the things
More informationPackage msgps. February 20, 2015
Type Package Package msgps February 20, 2015 Title Degrees of freedom of elastic net, adaptive lasso and generalized elastic net Version 1.3 Date 2012-5-17 Author Kei Hirose Maintainer Kei Hirose
More informationPerformance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018
Performance Estimation and Regularization Kasthuri Kannan, PhD. Machine Learning, Spring 2018 Bias- Variance Tradeoff Fundamental to machine learning approaches Bias- Variance Tradeoff Error due to Bias:
More informationCS294-1 Assignment 2 Report
CS294-1 Assignment 2 Report Keling Chen and Huasha Zhao February 24, 2012 1 Introduction The goal of this homework is to predict a users numeric rating for a book from the text of the user s review. The
More informationLecture on Modeling Tools for Clustering & Regression
Lecture on Modeling Tools for Clustering & Regression CS 590.21 Analysis and Modeling of Brain Networks Department of Computer Science University of Crete Data Clustering Overview Organizing data into
More informationPackage LINselect. June 9, 2018
Title Selection of Linear Estimators Version 1.1 Date 2017-04-20 Package LINselect June 9, 2018 Author Yannick Baraud, Christophe Giraud, Sylvie Huet Maintainer ORPHANED Description Estimate the mean of
More informationStat 4510/7510 Homework 4
Stat 45/75 1/7. Stat 45/75 Homework 4 Instructions: Please list your name and student number clearly. In order to receive credit for a problem, your solution must show sufficient details so that the grader
More informationSTAT432 Mini-Midterm Exam I (green) University of Illinois Urbana-Champaign February 25 (Monday), :00 10:50a SOLUTIONS
STAT432 Mini-Midterm Exam I (green) University of Illinois Urbana-Champaign February 25 (Monday), 2019 10:00 10:50a SOLUTIONS Question 1 (6.5 points) 2 (3.5 points) 3 (10 points) extra-credit (1 point)
More informationHW 10 STAT 672, Summer 2018
HW 10 STAT 672, Summer 2018 1) (0 points) Do parts (a), (b), (c), and (e) of Exercise 2 on p. 298 of ISL. 2) (0 points) Do Exercise 3 on p. 298 of ISL. 3) For this problem, try to use the 64 bit version
More informationPackage flam. April 6, 2018
Type Package Package flam April 6, 2018 Title Fits Piecewise Constant Models with Data-Adaptive Knots Version 3.2 Date 2018-04-05 Author Ashley Petersen Maintainer Ashley Petersen
More informationLecture 7: Linear Regression (continued)
Lecture 7: Linear Regression (continued) Reading: Chapter 3 STATS 2: Data mining and analysis Jonathan Taylor, 10/8 Slide credits: Sergio Bacallado 1 / 14 Potential issues in linear regression 1. Interactions
More informationDS Machine Learning and Data Mining I. Alina Oprea Associate Professor, CCIS Northeastern University
DS 4400 Machine Learning and Data Mining I Alina Oprea Associate Professor, CCIS Northeastern University September 20 2018 Review Solution for multiple linear regression can be computed in closed form
More informationMachine Learning: An Applied Econometric Approach Online Appendix
Machine Learning: An Applied Econometric Approach Online Appendix Sendhil Mullainathan mullain@fas.harvard.edu Jann Spiess jspiess@fas.harvard.edu April 2017 A How We Predict In this section, we detail
More informationMultiple Regression White paper
+44 (0) 333 666 7366 Multiple Regression White paper A tool to determine the impact in analysing the effectiveness of advertising spend. Multiple Regression In order to establish if the advertising mechanisms
More informationPackage intccr. September 12, 2017
Type Package Package intccr September 12, 2017 Title Semiparametric Competing Risks Regression under Interval Censoring Version 0.2.0 Author Giorgos Bakoyannis , Jun Park
More informationLecture 27: Review. Reading: All chapters in ISLR. STATS 202: Data mining and analysis. December 6, 2017
Lecture 27: Review Reading: All chapters in ISLR. STATS 202: Data mining and analysis December 6, 2017 1 / 16 Final exam: Announcements Tuesday, December 12, 8:30-11:30 am, in the following rooms: Last
More informationMACHINE LEARNING TOOLBOX. Random forests and wine
MACHINE LEARNING TOOLBOX Random forests and wine Random forests Popular type of machine learning model Good for beginners Robust to overfitting Yield very accurate, non-linear models Random forests Unlike
More informationPackage sparsenetgls
Type Package Package sparsenetgls March 18, 2019 Title Using Gaussian graphical structue learning estimation in generalized least squared regression for multivariate normal regression Version 1.0.1 Maintainer
More informationLecture 22 The Generalized Lasso
Lecture 22 The Generalized Lasso 07 December 2015 Taylor B. Arnold Yale Statistics STAT 312/612 Class Notes Midterm II - Due today Problem Set 7 - Available now, please hand in by the 16th Motivation Today
More informationThe Data. Math 158, Spring 2016 Jo Hardin Shrinkage Methods R code Ridge Regression & LASSO
Math 158, Spring 2016 Jo Hardin Shrinkage Methods R code Ridge Regression & LASSO The Data The following dataset is from Hastie, Tibshirani and Friedman (2009), from a studyby Stamey et al. (1989) of prostate
More informationLecture 25: Review I
Lecture 25: Review I Reading: Up to chapter 5 in ISLR. STATS 202: Data mining and analysis Jonathan Taylor 1 / 18 Unsupervised learning In unsupervised learning, all the variables are on equal standing,
More informationLECTURE 12: LINEAR MODEL SELECTION PT. 3. October 23, 2017 SDS 293: Machine Learning
LECTURE 12: LINEAR MODEL SELECTION PT. 3 October 23, 2017 SDS 293: Machine Learning Announcements 1/2 Presentation of the CS Major & Minors TODAY @ lunch Ford 240 FREE FOOD! Announcements 2/2 CS Internship
More informationAssignment 5: Collaborative Filtering
Assignment 5: Collaborative Filtering Arash Vahdat Fall 2015 Readings You are highly recommended to check the following readings before/while doing this assignment: Slope One Algorithm: https://en.wikipedia.org/wiki/slope_one.
More informationChapter 6: Linear Model Selection and Regularization
Chapter 6: Linear Model Selection and Regularization As p (the number of predictors) comes close to or exceeds n (the sample size) standard linear regression is faced with problems. The variance of the
More informationPackage EBglmnet. January 30, 2016
Type Package Package EBglmnet January 30, 2016 Title Empirical Bayesian Lasso and Elastic Net Methods for Generalized Linear Models Version 4.1 Date 2016-01-15 Author Anhui Huang, Dianting Liu Maintainer
More informationApplied Statistics : Practical 9
Applied Statistics : Practical 9 This practical explores nonparametric regression and shows how to fit a simple additive model. The first item introduces the necessary R commands for nonparametric regression
More informationLecture 26: Missing data
Lecture 26: Missing data Reading: ESL 9.6 STATS 202: Data mining and analysis December 1, 2017 1 / 10 Missing data is everywhere Survey data: nonresponse. 2 / 10 Missing data is everywhere Survey data:
More informationPackage StVAR. February 11, 2017
Type Package Title Student's t Vector Autoregression (StVAR) Version 1.1 Date 2017-02-10 Author Niraj Poudyal Maintainer Niraj Poudyal Package StVAR February 11, 2017 Description Estimation
More informationLeveling Up as a Data Scientist. ds/2014/10/level-up-ds.jpg
Model Optimization Leveling Up as a Data Scientist http://shorelinechurch.org/wp-content/uploa ds/2014/10/level-up-ds.jpg Bias and Variance Error = (expected loss of accuracy) 2 + flexibility of model
More informationPractical OmicsFusion
Practical OmicsFusion Introduction In this practical, we will analyse data, from an experiment which aim was to identify the most important metabolites that are related to potato flesh colour, from an
More informationGetting started with simulating data in R: some helpful functions and how to use them Ariel Muldoon August 28, 2018
Getting started with simulating data in R: some helpful functions and how to use them Ariel Muldoon August 28, 2018 Contents Overview 2 Generating random numbers 2 rnorm() to generate random numbers from
More informationLinear Methods for Regression and Shrinkage Methods
Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors
More informationMoving Beyond Linearity
Moving Beyond Linearity Basic non-linear models one input feature: polynomial regression step functions splines smoothing splines local regression. more features: generalized additive models. Polynomial
More informationPackage RAMP. May 25, 2017
Type Package Package RAMP May 25, 2017 Title Regularized Generalized Linear Models with Interaction Effects Version 2.0.1 Date 2017-05-24 Author Yang Feng, Ning Hao and Hao Helen Zhang Maintainer Yang
More informationPackage ordinalnet. December 5, 2017
Type Package Title Penalized Ordinal Regression Version 2.4 Package ordinalnet December 5, 2017 Fits ordinal regression models with elastic net penalty. Supported model families include cumulative probability,
More informationPS 6: Regularization. PART A: (Source: HTF page 95) The Ridge regression problem is:
Economics 1660: Big Data PS 6: Regularization Prof. Daniel Björkegren PART A: (Source: HTF page 95) The Ridge regression problem is: : β "#$%& = argmin (y # β 2 x #4 β 4 ) 6 6 + λ β 4 #89 Consider the
More informationModel selection and validation 1: Cross-validation
Model selection and validation 1: Cross-validation Ryan Tibshirani Data Mining: 36-462/36-662 March 26 2013 Optional reading: ISL 2.2, 5.1, ESL 7.4, 7.10 1 Reminder: modern regression techniques Over the
More informationHomework 3: Solutions
Homework 3: Solutions Statistics 43 Fall 207 Data Analysis: Note: All data analysis results are provided by Yixin Chen STAT43 - HW3 Yixin Chen Data Analysis in R Pipeline (a) Since the purpose is to find
More informationCSE446: Linear Regression. Spring 2017
CSE446: Linear Regression Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Prediction of continuous variables Billionaire says: Wait, that s not what I meant! You say: Chill
More informationLasso.jl Documentation
Lasso.jl Documentation Release 0.0.1 Simon Kornblith Jan 07, 2018 Contents 1 Lasso paths 3 2 Fused Lasso and trend filtering 7 3 Indices and tables 9 i ii Lasso.jl Documentation, Release 0.0.1 Contents:
More informationCSE Data Mining Concepts and Techniques STATISTICAL METHODS (REGRESSION) Professor- Anita Wasilewska. Team 13
CSE 634 - Data Mining Concepts and Techniques STATISTICAL METHODS Professor- Anita Wasilewska (REGRESSION) Team 13 Contents Linear Regression Logistic Regression Bias and Variance in Regression Model Fit
More informationPackage bayescl. April 14, 2017
Package bayescl April 14, 2017 Version 0.0.1 Date 2017-04-10 Title Bayesian Inference on a GPU using OpenCL Author Rok Cesnovar, Erik Strumbelj Maintainer Rok Cesnovar Description
More informationFMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu
FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)
More informationAmong those 14 potential explanatory variables,non-dummy variables are:
Among those 14 potential explanatory variables,non-dummy variables are: Size: 2nd column in the dataset Land: 14th column in the dataset Bed.Rooms: 5th column in the dataset Fireplace: 7th column in the
More informationRegression on SAT Scores of 374 High Schools and K-means on Clustering Schools
Regression on SAT Scores of 374 High Schools and K-means on Clustering Schools Abstract In this project, we study 374 public high schools in New York City. The project seeks to use regression techniques
More informationLecture 17: Smoothing splines, Local Regression, and GAMs
Lecture 17: Smoothing splines, Local Regression, and GAMs Reading: Sections 7.5-7 STATS 202: Data mining and analysis November 6, 2017 1 / 24 Cubic splines Define a set of knots ξ 1 < ξ 2 < < ξ K. We want
More informationPackage sgpls. R topics documented: August 29, Type Package Title Sparse Group Partial Least Square Methods Version 1.
Type Package Title Sparse Group Partial Least Square Methods Version 1.7 Date 2017-08-29 Package sgpls August 29, 2017 Author Benoit Liquet and Pierre Lafaye de Micheaux and Camilo Broc Maintainer Benoit
More informationContents Cont Hypothesis testing
Lecture 5 STATS/CME 195 Contents Hypothesis testing Hypothesis testing Exploratory vs. confirmatory data analysis Two approaches of statistics to analyze data sets: Exploratory: use plotting, transformations
More informationThis is a simple example of how the lasso regression model works.
1 of 29 5/25/2016 11:26 AM This is a simple example of how the lasso regression model works. save.image("backup.rdata") rm(list=ls()) library("glmnet") ## Loading required package: Matrix ## ## Attaching
More informationThe grplasso Package
The grplasso Package June 27, 2007 Type Package Title Fitting user specified models with Group Lasso penalty Version 0.2-1 Date 2007-06-27 Author Lukas Meier Maintainer Lukas Meier
More informationPackage TANDEM. R topics documented: June 15, Type Package
Type Package Package TANDEM June 15, 2017 Title A Two-Stage Approach to Maximize Interpretability of Drug Response Models Based on Multiple Molecular Data Types Version 1.0.2 Date 2017-04-07 Author Nanne
More informationPackage lspls. July 27, 2018
Title LS-PLS Models Version 0.2-2 Date 2018-07-26 Author [aut, cre] Package lspls July 27, 2018 Maintainer Encoding UTF-8 Depends pls (>= 2.2.0) Imports grdevices, graphics, methods, stats
More informationSupport Vector Machines
Support Vector Machines Chapter 9 Chapter 9 1 / 50 1 91 Maximal margin classifier 2 92 Support vector classifiers 3 93 Support vector machines 4 94 SVMs with more than two classes 5 95 Relationshiop to
More informationPackage polywog. April 20, 2018
Package polywog April 20, 2018 Title Bootstrapped Basis Regression with Oracle Model Selection Version 0.4-1 Date 2018-04-03 Author Maintainer Brenton Kenkel Routines for flexible
More informationPackage SSLASSO. August 28, 2018
Package SSLASSO August 28, 2018 Version 1.2-1 Date 2018-08-28 Title The Spike-and-Slab LASSO Author Veronika Rockova [aut,cre], Gemma Moran [aut] Maintainer Gemma Moran Description
More informationSTAT 311 (3 CREDITS) VARIANCE AND REGRESSION ANALYSIS ELECTIVE: ALL STUDENTS. CONTENT Introduction to Computer application of variance and regression
STAT 311 (3 CREDITS) VARIANCE AND REGRESSION ANALYSIS ELECTIVE: ALL STUDENTS. CONTENT Introduction to Computer application of variance and regression analysis. Analysis of Variance: one way classification,
More informationIntroduction to Automated Text Analysis. bit.ly/poir599
Introduction to Automated Text Analysis Pablo Barberá School of International Relations University of Southern California pablobarbera.com Lecture materials: bit.ly/poir599 Today 1. Solutions for last
More informationComparison of Linear Regression with K-Nearest Neighbors
Comparison of Linear Regression with K-Nearest Neighbors Rebecca C. Steorts, Duke University STA 325, Chapter 3.5 ISL Agenda Intro to KNN Comparison of KNN and Linear Regression K-Nearest Neighbors vs
More informationRegression Analysis and Linear Regression Models
Regression Analysis and Linear Regression Models University of Trento - FBK 2 March, 2015 (UNITN-FBK) Regression Analysis and Linear Regression Models 2 March, 2015 1 / 33 Relationship between numerical
More informationStandard Errors in OLS Luke Sonnet
Standard Errors in OLS Luke Sonnet Contents Variance-Covariance of ˆβ 1 Standard Estimation (Spherical Errors) 2 Robust Estimation (Heteroskedasticity Constistent Errors) 4 Cluster Robust Estimation 7
More informationMachine Learning / Jan 27, 2010
Revisiting Logistic Regression & Naïve Bayes Aarti Singh Machine Learning 10-701/15-781 Jan 27, 2010 Generative and Discriminative Classifiers Training classifiers involves learning a mapping f: X -> Y,
More informationPackage TVsMiss. April 5, 2018
Type Package Title Variable Selection for Missing Data Version 0.1.1 Date 2018-04-05 Author Jiwei Zhao, Yang Yang, and Ning Yang Maintainer Yang Yang Package TVsMiss April 5, 2018
More information2017 ITRON EFG Meeting. Abdul Razack. Specialist, Load Forecasting NV Energy
2017 ITRON EFG Meeting Abdul Razack Specialist, Load Forecasting NV Energy Topics 1. Concepts 2. Model (Variable) Selection Methods 3. Cross- Validation 4. Cross-Validation: Time Series 5. Example 1 6.
More informationCSE446: Linear Regression. Spring 2017
CSE446: Linear Regression Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Prediction of continuous variables Billionaire says: Wait, that s not what I meant! You say: Chill
More informationSimulation studies. Patrick Breheny. September 8. Monte Carlo simulation Example: Ridge vs. Lasso vs. Subset
Simulation studies Patrick Breheny September 8 Patrick Breheny BST 764: Applied Statistical Modeling 1/17 Introduction In statistics, we are often interested in properties of various estimation and model
More informationNonparametric Methods Recap
Nonparametric Methods Recap Aarti Singh Machine Learning 10-701/15-781 Oct 4, 2010 Nonparametric Methods Kernel Density estimate (also Histogram) Weighted frequency Classification - K-NN Classifier Majority
More informationOverfitting. Machine Learning CSE546 Carlos Guestrin University of Washington. October 2, Bias-Variance Tradeoff
Overfitting Machine Learning CSE546 Carlos Guestrin University of Washington October 2, 2013 1 Bias-Variance Tradeoff Choice of hypothesis class introduces learning bias More complex class less bias More
More informationModel Inference and Averaging. Baging, Stacking, Random Forest, Boosting
Model Inference and Averaging Baging, Stacking, Random Forest, Boosting Bagging Bootstrap Aggregating Bootstrap Repeatedly select n data samples with replacement Each dataset b=1:b is slightly different
More informationGlmnet Vignette. Introduction. Trevor Hastie and Junyang Qian
Glmnet Vignette Trevor Hastie and Junyang Qian Stanford September 13, 2016 Introduction Installation Quick Start Linear Regression Logistic Regression Poisson Models Cox Models Sparse Matrices Appendix
More informationEE 511 Linear Regression
EE 511 Linear Regression Instructor: Hanna Hajishirzi hannaneh@washington.edu Slides adapted from Ali Farhadi, Mari Ostendorf, Pedro Domingos, Carlos Guestrin, and Luke Zettelmoyer, Announcements Hw1 due
More informationIntroducing Categorical Data/Variables (pp )
Notation: Means pencil-and-paper QUIZ Means coding QUIZ Definition: Feature Engineering (FE) = the process of transforming the data to an optimal representation for a given application. Scaling (see Chs.
More informationChapter 7: Numerical Prediction
Ludwig-Maximilians-Universität München Institut für Informatik Lehr- und Forschungseinheit für Datenbanksysteme Knowledge Discovery in Databases SS 2016 Chapter 7: Numerical Prediction Lecture: Prof. Dr.
More informationSPSS INSTRUCTION CHAPTER 9
SPSS INSTRUCTION CHAPTER 9 Chapter 9 does no more than introduce the repeated-measures ANOVA, the MANOVA, and the ANCOVA, and discriminant analysis. But, you can likely envision how complicated it can
More informationUniversity of California, Berkeley
University of California, Berkeley U.C. Berkeley Division of Biostatistics Working Paper Series Year 2014 Paper 321 A Scalable Supervised Subsemble Prediction Algorithm Stephanie Sapp Mark J. van der Laan
More information6 Model selection and kernels
6. Bias-Variance Dilemma Esercizio 6. While you fit a Linear Model to your data set. You are thinking about changing the Linear Model to a Quadratic one (i.e., a Linear Model with quadratic features φ(x)
More informationAssignment 6 - Model Building
Assignment 6 - Model Building your name goes here Due: Wednesday, March 7, 2018, noon, to Sakai Summary Primarily from the topics in Chapter 9 of your text, this homework assignment gives you practice
More informationCSE 158 Lecture 2. Web Mining and Recommender Systems. Supervised learning Regression
CSE 158 Lecture 2 Web Mining and Recommender Systems Supervised learning Regression Supervised versus unsupervised learning Learning approaches attempt to model data in order to solve a problem Unsupervised
More informationMultiple Linear Regression
Multiple Linear Regression Rebecca C. Steorts, Duke University STA 325, Chapter 3 ISL 1 / 49 Agenda How to extend beyond a SLR Multiple Linear Regression (MLR) Relationship Between the Response and Predictors
More informationScatterplot: The Bridge from Correlation to Regression
Scatterplot: The Bridge from Correlation to Regression We have already seen how a histogram is a useful technique for graphing the distribution of one variable. Here is the histogram depicting the distribution
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationCSE 446 Bias-Variance & Naïve Bayes
CSE 446 Bias-Variance & Naïve Bayes Administrative Homework 1 due next week on Friday Good to finish early Homework 2 is out on Monday Check the course calendar Start early (midterm is right before Homework
More informationLab 10 - Ridge Regression and the Lasso in Python
Lab 10 - Ridge Regression and the Lasso in Python March 9, 2016 This lab on Ridge Regression and the Lasso is a Python adaptation of p. 251-255 of Introduction to Statistical Learning with Applications
More informationMidterm Examination , Fall 2011
Midterm Examination 36-350, Fall 2011 Instructions: Provide all answers in your bluebook. Only work in the bluebook will be graded. Clearly indicate which problem each answer goes with. There are 100 points
More informationEconomics Nonparametric Econometrics
Economics 217 - Nonparametric Econometrics Topics covered in this lecture Introduction to the nonparametric model The role of bandwidth Choice of smoothing function R commands for nonparametric models
More informationNonparametric Regression and Cross-Validation Yen-Chi Chen 5/27/2017
Nonparametric Regression and Cross-Validation Yen-Chi Chen 5/27/2017 Nonparametric Regression In the regression analysis, we often observe a data consists of a response variable Y and a covariate (this
More informationLast time... Bias-Variance decomposition. This week
Machine learning, pattern recognition and statistical data modelling Lecture 4. Going nonlinear: basis expansions and splines Last time... Coryn Bailer-Jones linear regression methods for high dimensional
More informationPackage DTRlearn. April 6, 2018
Type Package Package DTRlearn April 6, 2018 Title Learning Algorithms for Dynamic Treatment Regimes Version 1.3 Date 2018-4-05 Author Ying Liu, Yuanjia Wang, Donglin Zeng Maintainer Ying Liu
More informationPackage gwrr. February 20, 2015
Type Package Package gwrr February 20, 2015 Title Fits geographically weighted regression models with diagnostic tools Version 0.2-1 Date 2013-06-11 Author David Wheeler Maintainer David Wheeler
More informationTopics in Machine Learning-EE 5359 Model Assessment and Selection
Topics in Machine Learning-EE 5359 Model Assessment and Selection Ioannis D. Schizas Electrical Engineering Department University of Texas at Arlington 1 Training and Generalization Training stage: Utilizing
More informationStudy Guide. Module 1. Key Terms
Study Guide Module 1 Key Terms general linear model dummy variable multiple regression model ANOVA model ANCOVA model confounding variable squared multiple correlation adjusted squared multiple correlation
More informationCSC 411: Lecture 02: Linear Regression
CSC 411: Lecture 02: Linear Regression Raquel Urtasun & Rich Zemel University of Toronto Sep 16, 2015 Urtasun & Zemel (UofT) CSC 411: 02-Regression Sep 16, 2015 1 / 16 Today Linear regression problem continuous
More information