A Note on Model Selection in Mixture Experiments

Size: px
Start display at page:

Download "A Note on Model Selection in Mixture Experiments"

Transcription

1 Journal of Mathematics Statistics (): 9-99, 007 ISSN Science Publications Note on Model Selection in Miture Eperiments Kadri Ulaş KY University of Marmara, Departments of Mathematics, Göztepe Kampüsü Kadıköy, 7 Istanbul Turkey bstract: In miture eperiments, determination of the best model for modeling the miture system is significant in both understing interpreting the system. For obtaining the best model in miture eperiments, different methods have been used. Most commonly used methods are the stepwise type methods. However, the models obtained with these methods are not always the best model depending on the chosen criteria. s the models obtained with these methods can be affected by collinearity, in this paper, an alternative approach is used for the determination of the models taken into account in the modeling of the miture surface, which is obtained on the eperimental region. This approach depends on the eamination of all possible subset regression models obtained for the miture model. To determine the best subset model, the condition numbers of models the model control graphs are also taken into account. Then, proposed approach has been investigated on flare data set, which is widely known in literature. Key words: Miture model, all possible subset selection, variable selection, model reduction, miture components, collinearity, condition number INTRODUCTION In miture eperiments, the measured response is assumed to depend only on the proportions of ingredients present in the miture not on the amount of miture. For eample, the response might be the tensile strength of stainless steel which is a miture of iron, nickel, copper chromium, or, it might be octane rating of a blend of gasoline. The purpose of miture eperiments is to build an appropriate model relating the response(s) to miture components. ll of the work on miture models has been based on response surface concepts. model is fitted to data by an eperimental design. Various miture models can be used in the analysis of miture eperiments. However, the determination of the best model for modeling the miture system is important in both understing interpreting the miture system since the fitted models are used to screen the components, predict the response(s), determine the effects of components on the response(s), or optimize the response(s) over the eperimental region. In general, computer-based methods for choosing the best subset regression have been suggested for determining the best model in miture eperiments. Cornell [], Piepel et al. [-] Martin et al. [5] used the stepwise regression for choosing the model in miture eperiments. In addition, Draper St. John [6] used the backward elimination regression procedure. Different from these methods, there are many various methods which eamine all possible subset regression models. One of these methods is RSQURE procedure in SS. This approach was used in miture eperiments, by Khuri [7] Cornell []. Khuri revised the work done by Cornell [] he gave some collinearity diagnostic measures for each p-parameter submodel with the highest R value. However, for each submodel with p-parameter, the models including different terms should not be ignored as an alternative to the model with the highest R value. Using the model with the highest R value may not be suitable for the interpretation of the miture model, as it can be affected by collinearity compared to other models. The purpose of this study was to get attention on the results obtained by using an alternative approach, in choosing a model in miture eperiments. This approach depends on the eamination of all possible subset regression models obtained for the miture model. By comparing all possible subsets, an investigator can not only determine the best reduced models according to the selected criteria such as R, but also identify alternatives to the best ones. In addition, etra criteria will also be taken into account Corresponding uthor: Kadri Ulaş kay, University of Marmara, Departments of Mathematics, Göztepe Kampüsü Kadıköy, 7 Istanbul Turkey 9

2 J. Math. & Stat. (): 9-99, 007 for the determination of alternative models. In this way, with the help of models including different interaction terms, the miture system can be interpreted much better the role of components in the system can be understood much easier. Miture eperiments: miture eperiment involves miing various proportions of two or more components to make different compositions of an end product. In a -components miture in which i represents the proportion of the ith components present in miture, 0 i i =,,..., i= i = () The composition space of the components takes the form of a regular ( ) dimensional simple. Physical, theoretical, or economic considerations often impose additional constraints on individual components 0 Li i Ui i =,,..., () where L i U i denote lower upper bounds, respectively. In general, restriction () reduce the constraint region given by () to an irregular ( ) dimensional hyperpolyhedron. When the component proportions are restricted by lower upper bounds, collinearity appears all too freuently [8]. It is assumed that the response or property of interest, denoted by η, is to be epressed in terms of a suitable function f of the miture variables i, η = f (,,..., ) () typical model may thus be written yi = ηi + εi () where ε i is assumed that εi NID ( 0, σ ). Miture model forms most commonly used in fitting data are the canonical polynomials introduced by Scheffé [9] in the form ( ) (5) E Y = η = β + β i i ij i j i= i< j For modeling well-behaved systems, generally the Scheffé polynomials are adeuate. For some situations, however, there are better modeling forms than Scheffé polynomials which could be used. For eample, as an alternative to Scheffé miture models, models including inverse term are used in order to model an etreme change in the response behavior of one or more components, which are close to boundary of the simple region [6]. Following, uadratic model including an inverse term has been proposed by Draper St. John, ( ) = βi i + βij i j + β i i (6) E Y i= i< j i= Scheffé polynomial models fails to satisfy the modeling of additive effect of one component at the same time accommodate the curvilinear blending effects of the remaining components. To model these effects jointly, Becker has developed a set of miture models which are homogeneous of degree one [0]. They provide alternatives to the Scheffé polynomials. Becker s three second order models are of the form H: η = βii + βij min( i, j) β... min(,,..., ) i= i < j (7) i j... H: η = β + β β i i ij... i i j i = < + j ( ) H: η = βii + βij( i j) β... (... ) i= i < j In the H model, i j ( i j) 0 ( i + j) = 0. + = whenever s usual, we can represent the Scheffé canonical polynomial models, miture models with inverse terms Becker Homogenous models in matri form by Y = Xβ+ ε (8) where Y is n vector of observations on the response matri, where p is number of variable, X is n p( ) terms in the model, β is the p vector of parameters to be estimated ε is n vector of errors. It was assumed that the errors have the property E( ε ) = 0, E( εε ) = σ I n (9) where I n is identity matri σ is the error variance. Hence E ( Y ) = µ = Xβ where µ is column b XX. vector of all epected responses. The least suares estimator for β is b = ( XX ) Xy variancecovariance matri of b is ( ) ( ) var = σ comprehensive reference on the design analysis of miture data is given by Cornell [,]. Determination comparison of miture models: In miture eperiments, reduction of the model is as important as choosing between different miture model forms, since it is not a very good approach to add all the terms of the chosen model to itself. In a situation like this, the model may include correlated terms. It may also be hard to make comments on the miture system as the parameter values may be affected. In miture eperiments, determination of the best model for modeling the miture system is significant in 9

3 J. Math. & Stat. (): 9-99, 007 both understing interpreting the system. There are various methods for choosing a regression model such as forward selection, backward elimination stepwise regression when there are many cidate model terms. In addition, Cornell [] mentioned that the stepwise regression model can be investigated for various models in miture eperiments. The objective is to obtain a model form that not only contains an adeuate amount of information about the miture system under investigation but whose form also makes sense. There are serious problems with stepwise type methods since they do not give the best model (based on the selected criteria, for eample R ). This is because they hle variables one at a time. In addition, only one model is obtained with these methods. Therefore, there is a possibility of missing better models. Mitures problems are particularly prone to illconditioning (or collinearity) because of constraint (). Collinearity between the terms may lead to inconsistent or confusing conclusions when comparing the result of different stepwise regression procedures. This inconsistency makes variable selection a potentially misleading process when collinearity is present. Illconditioning may not be a problem if the goal is prediction within the eperimental region. However, illconditioning can be a serious problem if interpretation of the coefficients is the objective [8]. Stard variable selection methods do not perform well when the data is highly multi-collinear. For this reason, in order to obtain the best reduced model, all the possible subset regression models should be eamined. The seuential model fitting methods proposed by Draper St. John [6] for miture eperiments can be useful. But, if there are many terms, it can reuire too much labor. more preferable method than these methods is to fit all possible regression models, to evaluate these according to some criterion. In this way a number of best regression models can be selected. In order to find the best subset regression model RESERCH procedure on GENSTT was used []. While using this procedure, three criteria will be taken into account in determining the best models. First of all, linear miture terms (,,..., ) were kept in the model all possible combinations for the rest of the terms were added to the linear miture terms. The reason for keeping the linear miture terms in the model is that the model proves the hierarchy principle. Hierarchy principle is important for the euivalence of the models [7]. s an addition to linear miture terms, the number of models with the term t ( t p ) is p. Therefore, for different t values in the miture t system, as a total of ( p ) subset regression models will be obtained. For the models obtained, the terms which have p value smaller than 0.05 according to F- statistics are meaningful. However this situation can affect because of the collinearity therefore, some important terms for the miture system can be ignored. In this situation, instead of taking models with meaningful terms into account, the VIF values of the terms should be taken into account. The condition numbers of the models can also be used for comparing the reduced models. useful measure of collinearity is the condition number, κ, defined by λ ma κ = λmin where λ ma λ min denote the largest the smallest of the eigenvalues of XX (the columns of X have been scaled to unit length), respectively. Smaller condition numbers indicate more stability (better conditioning) in the least suares estimates than indicated by larger condition numbers. In this study, subset regression models with a condition number less than 0 will be taken into account. In some situations, the condition number of the model with the highest R can be greater than those for other models this affects the parameter values which may cause misinterpretations about the system. The condition number being less than 0 does not guarantee that VIF values of the terms are less than 00. For this reason, the models with condition numbers less than 0 the terms with VIF values less than 00 will be taken into account. Thirdly, in order to eamine which of the models are adeuate, model control graphs should be obtained. For the models whose model control graphs are adeuate, a final decision can be made by looking at R MSE values of the models. The proposed approach will be eamined in the following part over the flare data set. Flare eperiment: McLean nderson [] presented an eample to illustrate their etreme-vertices design. flare is manufactured by miing magnesium ( ),, sodium nitrate ( ), strontium nitrate ( ) 95

4 J. Math. & Stat. (): 9-99, 007 Table : Components proportions illumination response values for flare eperiment Blend No Component Proportions Illumination (000 cles) binder ( ) under the following constraints, The component proportions for design points as well as the measured illumination values are given in Table. Snee [5] used Homogenous miture models (7) for the modeling of the flare data set. Draper St. John [6] made a comparison of the miture models (6) (7) for the flare data set. On the other h, Piepel Cornell [6] gave a summary of the models proposed for the flare data set till now. When these models are eamined, it can be seen that they have three terms as an addition to linear miture terms they also have the highest R values. However, as the eperimental region is restricted, the parameter predictions are affected due to collinearity. For this reason, comparing the models obtained to their condition numbers using the models with small condition number are more accurate for interpreting the system. On the other h, Snee [5] Draper St. John [6] considered pseudocomponents for flare data set. In this paper, subset regression model for actual components will be given by using Scheffé, Homogenous H Models including inverse term. First of all, let s take the second degree Scheffé models into account. For Scheffé model, as an addition to linear miture terms with a condition number less than 0, models with one two terms are obtained. The models with one term have the terms,,. The condition numbers of these models are.6, , respectively. The models with two terms have the terms,,, their condition numbers being eual to each other, is. However, the model recommended for the second degree Scheffé model have the terms, [6]. The condition number of this model is.5. In addition, when the stepwise regression forward selection is used for the Scheffé model taken into account, the model with only is obtained but with backward elimination method, model with is obtained. The condition numbers for these models obtained are , respectively. lthough the model obtained with backward elimination method has the highest R value ( R = 7.59), this model is the most affected by the collinearity among other models with two terms. When the homogenous H model is eamined, five models with one term condition numbers less than 0 are obtained. These models include the terms ecept for ( + ). mong these models the model with only ( + ) term has the terms with VIF values less than 00. s some of the terms in the models with two terms have VIF values greater than 00, they are ignored. On the other h, for the homogenous H model, a model with ( + ) ( + ) is obtained by using backward elimination method. In contrast to this, with forward selection stepwise regression a model with ( + ) is obtained. The condition numbers for these models are 5..7, respectively. The model control graphs of the Scheffé Homogenous H 96

5 subset regression models obtained show that these models are not adeuate. Now let s take miture models including inverse term into account. If the model including inverse term is from the first degree, then the models with condition number less than 0 have, with condition numbers 6.8, , respectively. s the VIF value of is 06, this model was ignored. lthough the model with has the highest R value ( R = 78.), it is the model the most affected from the collinearity with a condition number of 7. The model control graphs of the models with are adeuate. The model with two terms, only has the condition number of this model is.. On the other h, with backward elimination stepwise regression, the model with is obtained. lthough this model is the model with the highest R, its condition number is With forward selection, the model with the terms, is obtained. This model with highest R value was also used by Piepel Cornell [6]. The VIF value of in the model is the condition number is 79.. Therefore, the models that can be recommended for the model including inverse term are only the models with J. Math. & Stat. (): 9-99, 007 from the models with one term. The summary statistics of these models are R = 6., MSE = 79 R = 60., MSE = 68, respectively. If the model including inverse term is from the second degree, there is a linear relation between the terms. For this reason, some of the terms in the model are ignored. In this eample, the terms are kept outside the model as they are the linear composition of other terms. In this situation, si models with two terms condition numbers less than 0 are obtained. These models include the term couples (, ), (, ), (, ), (, ), (, ) (, ). The condition numbers of these models are 5.6, 0.9, 0.6, 5.5, 0.6, 0.9, respectively. The model control graphs of all models ecept for the last model also show that the models are also adeuate. However, as some linear miture terms have VIF values close to or greater than 00 in models with (, ) (, ), they were ignored. VIF values are eual 99. for the term in the 97 first model, the for the second model. Therefore, the models that can be recommended for the second degree models including inverse term include the terms (, ), (, ) (, ). The summary statistics of these models are R = 60.7, MSE = 587 ; R = 60., MSE = 6 R = 55.8, MSE = 0, respectively. Fig. : Model control graphs of model including inverse terms For the models including inverse term from the second degree, si models can also be obtained with three terms condition numbers less than 0. The models with only adeuate control graph include the terms (,, ) (,, ). The condition numbers of these models are ,

6 respectively. However, as VIF values of the term in these models are , respectively, these models were ignored. In addition, although the model with three terms the highest R value include the,,, VIF values of these terms are terms ( )., 9. 9., respectively. The condition number of this model is also In models with three terms, the model with the smallest condition number (.) κ = includes the terms (,, ) J. Math. & Stat. (): 9-99, 007. On the other h, for the model including inverse term from the second degree, only a model with the couple (, ) is obtained by using backward elimination, forward selection stepwise regression. In this model, VIF value for is 6.6 condition number is 7.. Fig. : Miture surfaces obtained for model including inverse terms The models including inverse term are better for the interpretation of the miture system than Scheffé Homogenous H miture models due to their 98 condition numbers adeuate model control graphs. For this reason, the investigator can choose the best model among the models with one or two terms. For eample, the model control graph of the model with terms (, ) is given in Fig.. The miture surface for = 0.0 = 0.08 on the eperimental region for the model is shown respectively in Fig.. CONCLUSION In this study, comparisons of the results, which were obtained by different subset selection methods for the flare data set, were done. The models obtained by backward elimination, forward selection stepwise regression methods are one of the models obtained by all possible subset selection. By using all possible subset selection, we can obtain models according to criteria we want. On the other h, due to effect by collinearity in the models with the highest R value obtained by RSQURE procedure, etra criteria were taken into account in choosing the models that can be used in interpreting the miture system. These criteria are the comparison of the condition numbers the investigation of the model control graphs of the models with small condition numbers. s a result of comparing the condition numbers of the models, models with more consistent parameter values were obtained. Therefore the condition number of the each model should be taken into account when the all possible subset regression models for the determination of the miture model are investigated. s an addition, linear miture terms of the model were kept in the model to prove the hierarchy principle. Hierarchy principle in miture eperiments is essential for the models, obtained by pseudo-components actual components, to be euivalent. Epressing the components in terms of pseudo-components will also alleviate the problems due to the correlations among the coefficients. However, this property due to the structure of the models including inverse term that of homogenous H miture models has to be investigated for both the pseudo-components the actual components. In addition, in the presence of collinearity, ridge trace can be used as an alternative approach in choosing the model for miture eperiments.

7 J. Math. & Stat. (): 9-99, 007 REFERENCES. Cornell, J.., 000. Fitting a Slack-Variable Model to Miture Data: Some Questions Raised. J. Quality Tech.,, : -7. Piepel, G. T. Redgate, 997. Miture eperiment techniues for reducing the number of components applied for modeling waste glass sodium release. J. m. Ceramic Soc., 80: Piepel, G. T. Redgate, 998. miture eperiment analysis of the hald cement data. The m. Stat., 5, : -0.. Piepel, G. F., J.M. Szychowski J.M. Loeppky, 00. ugmenting Scheffé linear miture models with suared /or crossproduct terms. J. Quality Tech.,, : Martin, R. J., L.M. Platts,.B. Seddon E.C. Stillman, 00. The design analysis of a miture eperiment on glass durability. ustralian & New Zeal J. Stat., 5: Draper, N. R. R. C. St. John, 977. mitures model with inverse terms. Technometrics, 9: Khuri,.I Slack-variable models versus Scheffé s miture models. J. pplied Stat., 9: St. John, R. C., 98. Eperiments with Mitures, Ill-Conditioning, ridge regression. J. Quality Tech., 6, : Scheffé, H., 958. Eperiments with mitures. J. Royal Stat. Soc., B, 0: Becker, N. G., 968. Models for the response of a miture. J. Royal Statistical Soc., B, 0: Cornell, J.., 000. Developing Miture Models, re we done?. Journal of Statistical Computation Simulation, 66: 7-.. Cornell, J.., 00. Eperiments with mitures, rd. Ed. Wiley-Interscience.. GENSTT, Release 7., 00. The Guide to Genstat Release 7.: Part Statistics.. McLean, R.. V. L. nderson, 966. Etreme vertices design of miture eperiments. Technometrics, 8: Snee, R. D., 97. Techniues for the analysis of miture data. Technometrics, 5: Piepel, G.F. J.. Cornell, 99. Miture eperiment approaches: eamples, discussion, recommendations. J. Quality Tech., 6:

Selecting the Right Central Composite Design

Selecting the Right Central Composite Design International Journal of Statistics and Applications 1, (1): 1-3 DOI:.93/j.statistics.11. Selecting the Right Central Composite Design B. A. Oyejola 1, J. C. Nwanya,* 1 Department of Statistics, University

More information

2014 Stat-Ease, Inc. All Rights Reserved.

2014 Stat-Ease, Inc. All Rights Reserved. What s New in Design-Expert version 9 Factorial split plots (Two-Level, Multilevel, Optimal) Definitive Screening and Single Factor designs Journal Feature Design layout Graph Columns Design Evaluation

More information

Linear Model Selection and Regularization. especially usefull in high dimensions p>>100.

Linear Model Selection and Regularization. especially usefull in high dimensions p>>100. Linear Model Selection and Regularization especially usefull in high dimensions p>>100. 1 Why Linear Model Regularization? Linear models are simple, BUT consider p>>n, we have more features than data records

More information

Discussion: Clustering Random Curves Under Spatial Dependence

Discussion: Clustering Random Curves Under Spatial Dependence Discussion: Clustering Random Curves Under Spatial Dependence Gareth M. James, Wenguang Sun and Xinghao Qiao Abstract We discuss the advantages and disadvantages of a functional approach to clustering

More information

Tips on JMP ing into Mixture Experimentation

Tips on JMP ing into Mixture Experimentation Tips on JMP ing into Mixture Experimentation Daniell. Obermiller, The Dow Chemical Company, Midland, MI Abstract Mixture experimentation has unique challenges due to the fact that the proportion of the

More information

CDAA No. 4 - Part Two - Multiple Regression - Initial Data Screening

CDAA No. 4 - Part Two - Multiple Regression - Initial Data Screening CDAA No. 4 - Part Two - Multiple Regression - Initial Data Screening Variables Entered/Removed b Variables Entered GPA in other high school, test, Math test, GPA, High school math GPA a Variables Removed

More information

Adaptive Filtering using Steepest Descent and LMS Algorithm

Adaptive Filtering using Steepest Descent and LMS Algorithm IJSTE - International Journal of Science Technology & Engineering Volume 2 Issue 4 October 2015 ISSN (online): 2349-784X Adaptive Filtering using Steepest Descent and LMS Algorithm Akash Sawant Mukesh

More information

DECISION-TREE-BASED MULTICLASS SUPPORT VECTOR MACHINES. Fumitake Takahashi, Shigeo Abe

DECISION-TREE-BASED MULTICLASS SUPPORT VECTOR MACHINES. Fumitake Takahashi, Shigeo Abe DECISION-TREE-BASED MULTICLASS SUPPORT VECTOR MACHINES Fumitake Takahashi, Shigeo Abe Graduate School of Science and Technology, Kobe University, Kobe, Japan (E-mail: abe@eedept.kobe-u.ac.jp) ABSTRACT

More information

CHEE 392A Unit 5: Design of Experiments

CHEE 392A Unit 5: Design of Experiments PROJECT LABORATORY R.J. MUNZ CHEE 9A 9 Unit 5: Design of Eperiments References: ASPE, section., chapter Quote of the Unit No man ought to be discouraged if the eperiments he puts in practice answer not

More information

Module 4. Non-linear machine learning econometrics: Support Vector Machine

Module 4. Non-linear machine learning econometrics: Support Vector Machine Module 4. Non-linear machine learning econometrics: Support Vector Machine THE CONTRACTOR IS ACTING UNDER A FRAMEWORK CONTRACT CONCLUDED WITH THE COMMISSION Introduction When the assumption of linearity

More information

Stresstest procedures for feature selection algorithms

Stresstest procedures for feature selection algorithms Stresstest procedures for feature selection algorithms A. M. Katrutsa 1,2 and V. V. Strijov 1 1 Moscow Institute of Physics and Technology, Institutskiy lane 9, Dolgoprudny city, 1417, Russian Federation

More information

Detecting and Circumventing Collinearity or Ill-Conditioning Problems

Detecting and Circumventing Collinearity or Ill-Conditioning Problems Chapter 8 Detecting and Circumventing Collinearity or Ill-Conditioning Problems Section 8.1 Introduction Multicollinearity/Collinearity/Ill-Conditioning The terms multicollinearity, collinearity, and ill-conditioning

More information

Multiresponse Sparse Regression with Application to Multidimensional Scaling

Multiresponse Sparse Regression with Application to Multidimensional Scaling Multiresponse Sparse Regression with Application to Multidimensional Scaling Timo Similä and Jarkko Tikka Helsinki University of Technology, Laboratory of Computer and Information Science P.O. Box 54,

More information

Living with Collinearity in Local Regression Models

Living with Collinearity in Local Regression Models Living with Collinearity in Local Regression Models Chris Brunsdon 1, Martin Charlton 2, Paul Harris 2 1 People Space and Place, Roxby Building, University of Liverpool,L69 7ZT, UK Tel. +44 151 794 2837

More information

Basics of Multivariate Modelling and Data Analysis

Basics of Multivariate Modelling and Data Analysis Basics of Multivariate Modelling and Data Analysis Kurt-Erik Häggblom 9. Linear regression with latent variables 9.1 Principal component regression (PCR) 9.2 Partial least-squares regression (PLS) [ mostly

More information

Outline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation

Outline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation Advanced Digital Image Processing and Others Xiaojun Qi -- REU Site Program in CVIP (7 Summer) Outline Segmentation Strategies and Data Structures Algorithms Overview K-Means Algorithm Hidden Markov Model

More information

Linear Methods for Regression and Shrinkage Methods

Linear Methods for Regression and Shrinkage Methods Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors

More information

Lab 07: Multiple Linear Regression: Variable Selection

Lab 07: Multiple Linear Regression: Variable Selection Lab 07: Multiple Linear Regression: Variable Selection OBJECTIVES 1.Use PROC REG to fit multiple regression models. 2.Learn how to find the best reduced model. 3.Variable diagnostics and influential statistics

More information

A Robust Optimum Response Surface Methodology based On MM-estimator

A Robust Optimum Response Surface Methodology based On MM-estimator A Robust Optimum Response Surface Methodology based On MM-estimator, 2 HABSHAH MIDI,, 2 MOHD SHAFIE MUSTAFA,, 2 ANWAR FITRIANTO Department of Mathematics, Faculty Science, University Putra Malaysia, 434,

More information

Lecture 13: Model selection and regularization

Lecture 13: Model selection and regularization Lecture 13: Model selection and regularization Reading: Sections 6.1-6.2.1 STATS 202: Data mining and analysis October 23, 2017 1 / 17 What do we know so far In linear regression, adding predictors always

More information

Title. Author(s)B. ŠĆEPANOVIĆ; M. KNEŽEVIĆ; D. LUČIĆ; O. MIJUŠKOVIĆ. Issue Date Doc URL. Type. Note. File Information

Title. Author(s)B. ŠĆEPANOVIĆ; M. KNEŽEVIĆ; D. LUČIĆ; O. MIJUŠKOVIĆ. Issue Date Doc URL. Type. Note. File Information Title FORECAST OF COLLAPSE MODE IN ECCENTRICALLY PATCH LOA ARTIFICIAL NEURAL NETWORKS Author(s)B. ŠĆEPANOVIĆ; M. KNEŽEVIĆ; D. LUČIĆ; O. MIJUŠKOVIĆ Issue Date 2013-09-11 Doc URL http://hdl.handle.net/2115/54224

More information

Lecture 27: Review. Reading: All chapters in ISLR. STATS 202: Data mining and analysis. December 6, 2017

Lecture 27: Review. Reading: All chapters in ISLR. STATS 202: Data mining and analysis. December 6, 2017 Lecture 27: Review Reading: All chapters in ISLR. STATS 202: Data mining and analysis December 6, 2017 1 / 16 Final exam: Announcements Tuesday, December 12, 8:30-11:30 am, in the following rooms: Last

More information

Math 214 Introductory Statistics Summer Class Notes Sections 3.2, : 1-21 odd 3.3: 7-13, Measures of Central Tendency

Math 214 Introductory Statistics Summer Class Notes Sections 3.2, : 1-21 odd 3.3: 7-13, Measures of Central Tendency Math 14 Introductory Statistics Summer 008 6-9-08 Class Notes Sections 3, 33 3: 1-1 odd 33: 7-13, 35-39 Measures of Central Tendency odd Notation: Let N be the size of the population, n the size of the

More information

Optimal Partially Replicated Cube, Star and Center Runs in Face-centered Central Composite Designs

Optimal Partially Replicated Cube, Star and Center Runs in Face-centered Central Composite Designs International Journal of Statistics and Probability; Vol. 4, No. 4; 25 ISSN 927-732 E-ISSN 927-74 Published by anadian enter of Science and Education Optimal Partially Replicated ube, Star and enter Runs

More information

Systems of Linear Equations and their Graphical Solution

Systems of Linear Equations and their Graphical Solution Proceedings of the World Congress on Engineering and Computer Science Vol I WCECS, - October,, San Francisco, USA Systems of Linear Equations and their Graphical Solution ISBN: 98-988-95-- ISSN: 8-958

More information

7. Collinearity and Model Selection

7. Collinearity and Model Selection Sociology 740 John Fox Lecture Notes 7. Collinearity and Model Selection Copyright 2014 by John Fox Collinearity and Model Selection 1 1. Introduction I When there is a perfect linear relationship among

More information

Principal Component Analysis of Lack of Cohesion in Methods (LCOM) metrics

Principal Component Analysis of Lack of Cohesion in Methods (LCOM) metrics Principal Component Analysis of Lack of Cohesion in Methods (LCOM) metrics Anuradha Lakshminarayana Timothy S.Newman Department of Computer Science University of Alabama in Huntsville Abstract In this

More information

Information Criteria Methods in SAS for Multiple Linear Regression Models

Information Criteria Methods in SAS for Multiple Linear Regression Models Paper SA5 Information Criteria Methods in SAS for Multiple Linear Regression Models Dennis J. Beal, Science Applications International Corporation, Oak Ridge, TN ABSTRACT SAS 9.1 calculates Akaike s Information

More information

Gradient LASSO algoithm

Gradient LASSO algoithm Gradient LASSO algoithm Yongdai Kim Seoul National University, Korea jointly with Yuwon Kim University of Minnesota, USA and Jinseog Kim Statistical Research Center for Complex Systems, Korea Contents

More information

Linear Programming has been used to:

Linear Programming has been used to: Linear Programming Linear programming became important during World War II: used to solve logistics problems for the military. Linear Programming (LP) was the first widely used form of optimization in

More information

The problem we have now is called variable selection or perhaps model selection. There are several objectives.

The problem we have now is called variable selection or perhaps model selection. There are several objectives. STAT-UB.0103 NOTES for Wednesday 01.APR.04 One of the clues on the library data comes through the VIF values. These VIFs tell you to what extent a predictor is linearly dependent on other predictors. We

More information

SPSS INSTRUCTION CHAPTER 9

SPSS INSTRUCTION CHAPTER 9 SPSS INSTRUCTION CHAPTER 9 Chapter 9 does no more than introduce the repeated-measures ANOVA, the MANOVA, and the ANCOVA, and discriminant analysis. But, you can likely envision how complicated it can

More information

Feature Selection Using Principal Feature Analysis

Feature Selection Using Principal Feature Analysis Feature Selection Using Principal Feature Analysis Ira Cohen Qi Tian Xiang Sean Zhou Thomas S. Huang Beckman Institute for Advanced Science and Technology University of Illinois at Urbana-Champaign Urbana,

More information

Globally Stabilized 3L Curve Fitting

Globally Stabilized 3L Curve Fitting Globally Stabilized 3L Curve Fitting Turker Sahin and Mustafa Unel Department of Computer Engineering, Gebze Institute of Technology Cayirova Campus 44 Gebze/Kocaeli Turkey {htsahin,munel}@bilmuh.gyte.edu.tr

More information

Based on Regression Diagnostics

Based on Regression Diagnostics Automatic Detection of Region-Mura Defects in TFT-LCD Based on Regression Diagnostics Yu-Chiang Chuang 1 and Shu-Kai S. Fan 2 Department of Industrial Engineering and Management, Yuan Ze University, Tao

More information

PS 6: Regularization. PART A: (Source: HTF page 95) The Ridge regression problem is:

PS 6: Regularization. PART A: (Source: HTF page 95) The Ridge regression problem is: Economics 1660: Big Data PS 6: Regularization Prof. Daniel Björkegren PART A: (Source: HTF page 95) The Ridge regression problem is: : β "#$%& = argmin (y # β 2 x #4 β 4 ) 6 6 + λ β 4 #89 Consider the

More information

Exploring Econometric Model Selection Using Sensitivity Analysis

Exploring Econometric Model Selection Using Sensitivity Analysis Exploring Econometric Model Selection Using Sensitivity Analysis William Becker Paolo Paruolo Andrea Saltelli Nice, 2 nd July 2013 Outline What is the problem we are addressing? Past approaches Hoover

More information

Mixture models and clustering

Mixture models and clustering 1 Lecture topics: Miture models and clustering, k-means Distance and clustering Miture models and clustering We have so far used miture models as fleible ays of constructing probability models for prediction

More information

Divide and Conquer Kernel Ridge Regression

Divide and Conquer Kernel Ridge Regression Divide and Conquer Kernel Ridge Regression Yuchen Zhang John Duchi Martin Wainwright University of California, Berkeley COLT 2013 Yuchen Zhang (UC Berkeley) Divide and Conquer KRR COLT 2013 1 / 15 Problem

More information

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi Unsupervised Learning Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi Content Motivation Introduction Applications Types of clustering Clustering criterion functions Distance functions Normalization Which

More information

Preprocessing and Visualization. Jonathan Diehl

Preprocessing and Visualization. Jonathan Diehl RWTH Aachen University Chair of Computer Science VI Prof. Dr.-Ing. Hermann Ney Seminar Data Mining WS 2003/2004 Preprocessing and Visualization Jonathan Diehl January 19, 2004 onathan Diehl Preprocessing

More information

Sparse Unmixing using an approximate L 0 Regularization Yang Guo 1,a, Tai Gao 1,b, Chengzhi Deng 2,c, Shengqian Wang 2,d and JianPing Xiao 1,e

Sparse Unmixing using an approximate L 0 Regularization Yang Guo 1,a, Tai Gao 1,b, Chengzhi Deng 2,c, Shengqian Wang 2,d and JianPing Xiao 1,e International Conference on Information Sciences, Machinery, Materials and Energy (ICISMME 5) Sparse Unmiing using an approimate L Regularization Yang Guo,a, ai Gao,b, Chengzhi Deng,c, Shengqian Wang,d

More information

Trees, Dendrograms and Sensitivity

Trees, Dendrograms and Sensitivity Trees, Dendrograms and Sensitivity R.D. Braddock Cooperative Research Centre for Catchment Hydrology, Griffith University, Nathan, Qld 4111, Australia (r.braddock@mailbo.gu.edu.au) Abstract: Dendrograms

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Pierre Gaillard ENS Paris September 28, 2018 1 Supervised vs unsupervised learning Two main categories of machine learning algorithms: - Supervised learning: predict output Y from

More information

Find the Relationship: An Exercise in Graphical Analysis

Find the Relationship: An Exercise in Graphical Analysis Find the Relationship: An Eercise in Graphical Analsis In several laborator investigations ou do this ear, a primar purpose will be to find the mathematical relationship between two variables. For eample,

More information

Genotype x Environmental Analysis with R for Windows

Genotype x Environmental Analysis with R for Windows Genotype x Environmental Analysis with R for Windows Biometrics and Statistics Unit Angela Pacheco CIMMYT,Int. 23-24 Junio 2015 About GEI In agricultural experimentation, a large number of genotypes are

More information

Image Contrast Enhancement in Wavelet Domain

Image Contrast Enhancement in Wavelet Domain Advances in Computational Sciences and Technology ISSN 0973-6107 Volume 10, Number 6 (2017) pp. 1915-1922 Research India Publications http://www.ripublication.com Image Contrast Enhancement in Wavelet

More information

Advanced Topics in Experimental Design. Jason Anderson 6/11/2015

Advanced Topics in Experimental Design. Jason Anderson 6/11/2015 Advanced Topics in Experimental Design Jason Anderson 6/11/2015 Table of Contents Introduction... 2 Chapter 1: Mixture Designs... 3 Introduction... 3 Brief History... 3 Design Space... 3 Model... 4 Simplex-Lattice

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2012 http://ce.sharif.edu/courses/90-91/2/ce725-1/ Agenda Features and Patterns The Curse of Size and

More information

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018

Performance Estimation and Regularization. Kasthuri Kannan, PhD. Machine Learning, Spring 2018 Performance Estimation and Regularization Kasthuri Kannan, PhD. Machine Learning, Spring 2018 Bias- Variance Tradeoff Fundamental to machine learning approaches Bias- Variance Tradeoff Error due to Bias:

More information

Intro to Linear Programming. The problem that we desire to address in this course is loosely stated below.

Intro to Linear Programming. The problem that we desire to address in this course is loosely stated below. . Introduction Intro to Linear Programming The problem that we desire to address in this course is loosely stated below. Given a number of generators make price-quantity offers to sell (each provides their

More information

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes.

Resources for statistical assistance. Quantitative covariates and regression analysis. Methods for predicting continuous outcomes. Resources for statistical assistance Quantitative covariates and regression analysis Carolyn Taylor Applied Statistics and Data Science Group (ASDa) Department of Statistics, UBC January 24, 2017 Department

More information

Where s Waldo: Visualizing Collinearity Diagnostics

Where s Waldo: Visualizing Collinearity Diagnostics In press, The American Statistician Where s Waldo: Visualizing Collinearity Diagnostics Michael Friendly Ernest Kwan Abstract Collinearity diagnostics are widely used, but the typical tabular output used

More information

CUTTING EXPERIMENTAL DESIGNS INTO BLOCKS. Nam-Ky Nguyen 1. Design Computing. Summary

CUTTING EXPERIMENTAL DESIGNS INTO BLOCKS. Nam-Ky Nguyen 1. Design Computing. Summary Aust. N. Z. J. Stat. 43(3), 2001, 367 374 CUTTING EXPERIMENTAL DESIGNS INTO BLOCKS Nam-Ky Nguyen 1 Design Computing Summary Most experimental material in agriculture and industry is heterogeneous in nature

More information

Variable selection is intended to select the best subset of predictors. But why bother?

Variable selection is intended to select the best subset of predictors. But why bother? Chapter 10 Variable Selection Variable selection is intended to select the best subset of predictors. But why bother? 1. We want to explain the data in the simplest way redundant predictors should be removed.

More information

These notes are in two parts: this part has topics 1-3 above.

These notes are in two parts: this part has topics 1-3 above. IEEM 0: Linear Programming and Its Applications Outline of this series of lectures:. How can we model a problem so that it can be solved to give the required solution 2. Motivation: eamples of typical

More information

Predictor Selection Algorithm for Bayesian Lasso

Predictor Selection Algorithm for Bayesian Lasso Predictor Selection Algorithm for Baesian Lasso Quan Zhang Ma 16, 2014 1 Introduction The Lasso [1] is a method in regression model for coefficients shrinkage and model selection. It is often used in the

More information

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate

More information

VALIDATION AND VERIFICATION OF HULL RESISTANCE COMPONENTS USING A COMMERCIAL CFD CODE SUMMARY

VALIDATION AND VERIFICATION OF HULL RESISTANCE COMPONENTS USING A COMMERCIAL CFD CODE SUMMARY VALIDATION AND VERIFICATION OF HULL RESISTANCE COMPONENTS USING A COMMERCIAL CFD CODE C.A. Perez G, University of Southampton, UK. Universidad Pontificia Bolivariana, Colombia, M. Tan and P.A. Wilson University

More information

Numerical Computing: An Introduction

Numerical Computing: An Introduction Numerical Computing: An Introduction Gyula Horváth Horvath@inf.u-szeged.hu Tom Verhoeff T.Verhoeff@TUE.NL University of Szeged Hungary Eindhoven University of Technology The Netherlands Numerical Computing

More information

Complexity Challenges to the Discovery of Relationships in Eddy Current Non-destructive Test Data

Complexity Challenges to the Discovery of Relationships in Eddy Current Non-destructive Test Data Complexity Challenges to the Discovery of Relationships in Eddy Current Non-destructive Test Data CPT John R. Brence United States Military Academy Donald E. Brown, PhD University of Virginia Outline Background

More information

Introduction to Mixed Models: Multivariate Regression

Introduction to Mixed Models: Multivariate Regression Introduction to Mixed Models: Multivariate Regression EPSY 905: Multivariate Analysis Spring 2016 Lecture #9 March 30, 2016 EPSY 905: Multivariate Regression via Path Analysis Today s Lecture Multivariate

More information

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview Chapter 888 Introduction This procedure generates D-optimal designs for multi-factor experiments with both quantitative and qualitative factors. The factors can have a mixed number of levels. For example,

More information

THE GRAPH OF FRACTAL DIMENSIONS OF JULIA SETS Bünyamin Demir 1, Yunus Özdemir2, Mustafa Saltan 3. Anadolu University Eskişehir, TURKEY

THE GRAPH OF FRACTAL DIMENSIONS OF JULIA SETS Bünyamin Demir 1, Yunus Özdemir2, Mustafa Saltan 3. Anadolu University Eskişehir, TURKEY International Journal of Pure and Applied Mathematics Volume 70 No. 3 2011, 401-409 THE GRAPH OF FRACTAL DIMENSIONS OF JULIA SETS Bünyamin Demir 1, Yunus Özdemir2, Mustafa Saltan 3 1,2,3 Department of

More information

Homework set 4 - Solutions

Homework set 4 - Solutions Homework set 4 - Solutions Math 3200 Renato Feres 1. (Eercise 4.12, page 153) This requires importing the data set for Eercise 4.12. You may, if you wish, type the data points into a vector. (a) Calculate

More information

Discrete Optimization. Lecture Notes 2

Discrete Optimization. Lecture Notes 2 Discrete Optimization. Lecture Notes 2 Disjunctive Constraints Defining variables and formulating linear constraints can be straightforward or more sophisticated, depending on the problem structure. The

More information

Estimating basis functions for spectral sensitivity of digital cameras

Estimating basis functions for spectral sensitivity of digital cameras (MIRU2009) 2009 7 Estimating basis functions for spectral sensitivity of digital cameras Abstract Hongxun ZHAO, Rei KAWAKAMI, Robby T.TAN, and Katsushi IKEUCHI Institute of Industrial Science, The University

More information

Assessing the Quality of the Natural Cubic Spline Approximation

Assessing the Quality of the Natural Cubic Spline Approximation Assessing the Quality of the Natural Cubic Spline Approximation AHMET SEZER ANADOLU UNIVERSITY Department of Statisticss Yunus Emre Kampusu Eskisehir TURKEY ahsst12@yahoo.com Abstract: In large samples,

More information

For more information, see the Math Notes box in Lesson of the Core Connections, Course 1 text.

For more information, see the Math Notes box in Lesson of the Core Connections, Course 1 text. Number TYPES OF NUMBERS When two or more integers are multiplied together, each number is a factor of the product. Nonnegative integers that have eactly two factors, namely, one and itself, are called

More information

NAG Toolbox for MATLAB. g02ef.1

NAG Toolbox for MATLAB. g02ef.1 G2 Correlation and Regression Analysis g2ef Purpose NAG Toolbox for MATLAB g2ef g2ef calculates a full stepwise selection from p variables by using Clarke s sweep algorithm on the correlation matrix of

More information

A Dendrogram. Bioinformatics (Lec 17)

A Dendrogram. Bioinformatics (Lec 17) A Dendrogram 3/15/05 1 Hierarchical Clustering [Johnson, SC, 1967] Given n points in R d, compute the distance between every pair of points While (not done) Pick closest pair of points s i and s j and

More information

DECISION SCIENCES INSTITUTE. Exponentially Derived Antithetic Random Numbers. (Full paper submission)

DECISION SCIENCES INSTITUTE. Exponentially Derived Antithetic Random Numbers. (Full paper submission) DECISION SCIENCES INSTITUTE (Full paper submission) Dennis Ridley, Ph.D. SBI, Florida A&M University and Scientific Computing, Florida State University dridley@fsu.edu Pierre Ngnepieba, Ph.D. Department

More information

Computer Experiments: Space Filling Design and Gaussian Process Modeling

Computer Experiments: Space Filling Design and Gaussian Process Modeling Computer Experiments: Space Filling Design and Gaussian Process Modeling Best Practice Authored by: Cory Natoli Sarah Burke, Ph.D. 30 March 2018 The goal of the STAT COE is to assist in developing rigorous,

More information

This lesson combines vertical translations and dilations in several quadratic and inverse variation modeling applications.

This lesson combines vertical translations and dilations in several quadratic and inverse variation modeling applications. Learning Objectives Combined Vertical Transformations Algebra ; Pre-Calculus Time required: 90 min. This lesson combines vertical translations and dilations in several quadratic and inverse variation modeling

More information

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY 23 CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY 3.1 DESIGN OF EXPERIMENTS Design of experiments is a systematic approach for investigation of a system or process. A series

More information

CHAPTER 3 MAINTENANCE STRATEGY SELECTION USING AHP AND FAHP

CHAPTER 3 MAINTENANCE STRATEGY SELECTION USING AHP AND FAHP 31 CHAPTER 3 MAINTENANCE STRATEGY SELECTION USING AHP AND FAHP 3.1 INTRODUCTION Evaluation of maintenance strategies is a complex task. The typical factors that influence the selection of maintenance strategy

More information

The Ohio State University Columbus, Ohio, USA Universidad Autónoma de Nuevo León San Nicolás de los Garza, Nuevo León, México, 66450

The Ohio State University Columbus, Ohio, USA Universidad Autónoma de Nuevo León San Nicolás de los Garza, Nuevo León, México, 66450 Optimization and Analysis of Variability in High Precision Injection Molding Carlos E. Castro 1, Blaine Lilly 1, José M. Castro 1, and Mauricio Cabrera Ríos 2 1 Department of Industrial, Welding & Systems

More information

Normalized cuts and image segmentation

Normalized cuts and image segmentation Normalized cuts and image segmentation Department of EE University of Washington Yeping Su Xiaodan Song Normalized Cuts and Image Segmentation, IEEE Trans. PAMI, August 2000 5/20/2003 1 Outline 1. Image

More information

TRICAP Chania. Typical rank when arrays have symmetric slices, and the Carroll & Chang conjecture of equal CP components.

TRICAP Chania. Typical rank when arrays have symmetric slices, and the Carroll & Chang conjecture of equal CP components. TRICAP 26 Chania Typical rank when arrays have symmetric slices, and the Carroll & Chang conjecture of equal CP components Jos ten Berge Research with Henk Kiers, Nikos Sidiropoulos, Roberto Rocci, and

More information

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions ENEE 739Q: STATISTICAL AND NEURAL PATTERN RECOGNITION Spring 2002 Assignment 2 Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions Aravind Sundaresan

More information

found to be robust to outliers in the response variable, but turned out not to be resistant to outliers in the explanatory variables, called leverage

found to be robust to outliers in the response variable, but turned out not to be resistant to outliers in the explanatory variables, called leverage 3rd International Conference on Multimedia echnology(icm 013) Robust Estimation of Parameters for Lucas-Kanade Algorithm Yih-Lon Lin 1 Abstract he object tracking problem is an important research topic

More information

General Factorial Models

General Factorial Models In Chapter 8 in Oehlert STAT:5201 Week 9 - Lecture 2 1 / 34 It is possible to have many factors in a factorial experiment. In DDD we saw an example of a 3-factor study with ball size, height, and surface

More information

Quality Improvement in the Multi-response Problem by Using Clustering Characteristic

Quality Improvement in the Multi-response Problem by Using Clustering Characteristic Proceedings of the 2007 WSEAS International Conference on Computer Engineering and Applications, Gold Coast, Australia, January 17-19, 2007 325 Quality Improvement in the Multi-response Problem by Using

More information

A Simulation Based Comparative Study of Normalization Procedures in Multiattribute Decision Making

A Simulation Based Comparative Study of Normalization Procedures in Multiattribute Decision Making Proceedings of the 6th WSEAS Int. Conf. on Artificial Intelligence, Knowledge Engineering and Data Bases, Corfu Island, Greece, February 16-19, 2007 102 A Simulation Based Comparative Study of Normalization

More information

Business Club. Decision Trees

Business Club. Decision Trees Business Club Decision Trees Business Club Analytics Team December 2017 Index 1. Motivation- A Case Study 2. The Trees a. What is a decision tree b. Representation 3. Regression v/s Classification 4. Building

More information

This is called a linear basis expansion, and h m is the mth basis function For example if X is one-dimensional: f (X) = β 0 + β 1 X + β 2 X 2, or

This is called a linear basis expansion, and h m is the mth basis function For example if X is one-dimensional: f (X) = β 0 + β 1 X + β 2 X 2, or STA 450/4000 S: February 2 2005 Flexible modelling using basis expansions (Chapter 5) Linear regression: y = Xβ + ɛ, ɛ (0, σ 2 ) Smooth regression: y = f (X) + ɛ: f (X) = E(Y X) to be specified Flexible

More information

Statistical Matching using Fractional Imputation

Statistical Matching using Fractional Imputation Statistical Matching using Fractional Imputation Jae-Kwang Kim 1 Iowa State University 1 Joint work with Emily Berg and Taesung Park 1 Introduction 2 Classical Approaches 3 Proposed method 4 Application:

More information

The Evaluation of Useful Method of Effort Estimation in Software Projects

The Evaluation of Useful Method of Effort Estimation in Software Projects The Evaluation of Useful Method of Effort Estimation in Software Projects Abstract Amin Moradbeiky, Vahid Khatibi Bardsiri Kerman Branch, Islamic Azad University, Kerman, Iran moradbeigi@csri.ac.i Kerman

More information

Lasso. November 14, 2017

Lasso. November 14, 2017 Lasso November 14, 2017 Contents 1 Case Study: Least Absolute Shrinkage and Selection Operator (LASSO) 1 1.1 The Lasso Estimator.................................... 1 1.2 Computation of the Lasso Solution............................

More information

Machine Learning. Topic 4: Linear Regression Models

Machine Learning. Topic 4: Linear Regression Models Machine Learning Topic 4: Linear Regression Models (contains ideas and a few images from wikipedia and books by Alpaydin, Duda/Hart/ Stork, and Bishop. Updated Fall 205) Regression Learning Task There

More information

Position Error Reduction of Kinematic Mechanisms Using Tolerance Analysis and Cost Function

Position Error Reduction of Kinematic Mechanisms Using Tolerance Analysis and Cost Function Position Error Reduction of Kinematic Mechanisms Using Tolerance Analysis and Cost Function B.Moetakef-Imani, M.Pour Department of Mechanical Engineering, Faculty of Engineering, Ferdowsi University of

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Features and Feature Selection Hamid R. Rabiee Jafar Muhammadi Spring 2013 http://ce.sharif.edu/courses/91-92/2/ce725-1/ Agenda Features and Patterns The Curse of Size and

More information

I How does the formulation (5) serve the purpose of the composite parameterization

I How does the formulation (5) serve the purpose of the composite parameterization Supplemental Material to Identifying Alzheimer s Disease-Related Brain Regions from Multi-Modality Neuroimaging Data using Sparse Composite Linear Discrimination Analysis I How does the formulation (5)

More information

Division of the Humanities and Social Sciences. Convex Analysis and Economic Theory Winter Separation theorems

Division of the Humanities and Social Sciences. Convex Analysis and Economic Theory Winter Separation theorems Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analysis and Economic Theory Winter 2018 Topic 8: Separation theorems 8.1 Hyperplanes and half spaces Recall that a hyperplane in

More information

Development of a tool for the easy determination of control factor interaction in the Design of Experiments and the Taguchi Methods

Development of a tool for the easy determination of control factor interaction in the Design of Experiments and the Taguchi Methods Development of a tool for the easy determination of control factor interaction in the Design of Experiments and the Taguchi Methods IKUO TANABE Department of Mechanical Engineering, Nagaoka University

More information

Chapter 6: Linear Model Selection and Regularization

Chapter 6: Linear Model Selection and Regularization Chapter 6: Linear Model Selection and Regularization As p (the number of predictors) comes close to or exceeds n (the sample size) standard linear regression is faced with problems. The variance of the

More information

DM6 Support Vector Machines

DM6 Support Vector Machines DM6 Support Vector Machines Outline Large margin linear classifier Linear separable Nonlinear separable Creating nonlinear classifiers: kernel trick Discussion on SVM Conclusion SVM: LARGE MARGIN LINEAR

More information

Missing Data Analysis for the Employee Dataset

Missing Data Analysis for the Employee Dataset Missing Data Analysis for the Employee Dataset 67% of the observations have missing values! Modeling Setup For our analysis goals we would like to do: Y X N (X, 2 I) and then interpret the coefficients

More information

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will Lectures Recent advances in Metamodel of Optimal Prognosis Thomas Most & Johannes Will presented at the Weimar Optimization and Stochastic Days 2010 Source: www.dynardo.de/en/library Recent advances in

More information

Linear Programming. Revised Simplex Method, Duality of LP problems and Sensitivity analysis

Linear Programming. Revised Simplex Method, Duality of LP problems and Sensitivity analysis Linear Programming Revised Simple Method, Dualit of LP problems and Sensitivit analsis Introduction Revised simple method is an improvement over simple method. It is computationall more efficient and accurate.

More information