Module 4 : Solving Linear Algebraic Equations Section 11 Appendix C: Steepest Descent / Gradient Search Method

Similar documents
APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING

f( x ), or a solution to the equation f( x) 0. You are already familiar with ways of solving

Deccan Education Society s

Modern Methods of Data Analysis - WS 07/08

COURSE: NUMERICAL ANALYSIS. LESSON: Methods for Solving Non-Linear Equations

3.7. Vertex and tangent

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms

Today. Golden section, discussion of error Newton s method. Newton s method, steepest descent, conjugate gradient

Today s class. Roots of equation Finish up incremental search Open methods. Numerical Methods, Fall 2011 Lecture 5. Prof. Jinbo Bi CSE, UConn

Optimization Methods: Optimization using Calculus Kuhn-Tucker Conditions 1. Module - 2 Lecture Notes 5. Kuhn-Tucker Conditions

Convergence of C 2 Deficient Quartic Spline Interpolation

Directional Derivatives. Directional Derivatives. Directional Derivatives. Directional Derivatives. Directional Derivatives. Directional Derivatives

Search direction improvement for gradient-based optimization problems

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Maximizing an interpolating quadratic

Ordinary Differential Equations

Local and Global Minimum

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

February 23 Math 2335 sec 51 Spring 2016

Contents. Hilary Term. Summary of Numerical Analysis for this term. Sources of error in numerical calculation. Solving Problems

Adaptive Filtering using Steepest Descent and LMS Algorithm

Introduction to optimization methods and line search

Chapter 3 Numerical Methods

Deficient Quartic Spline Interpolation

Introduction to Optimization Problems and Methods

A Brief Look at Optimization

Constrained Optimization with Calculus. Background Three Big Problems Setup and Vocabulary

Visualizing Images. Lecture 2: Intensity Surfaces and Gradients. Images as Surfaces. Bridging the Gap. Examples. Examples

Lecture 12: Feasible direction methods

99 International Journal of Engineering, Science and Mathematics

Machine Learning for Signal Processing Lecture 4: Optimization

Robert Collins CSE486, Penn State. Lecture 2: Intensity Surfaces and Gradients

14.6 Directional Derivatives and the Gradient Vector

100 Myung Hwan Na log-hazard function. The discussion section of Abrahamowicz, et al.(1992) contains a good review of many of the papers on the use of

Experimental Data and Training

REGRESSION ANALYSIS : LINEAR BY MAUAJAMA FIRDAUS & TULIKA SAHA

8 Piecewise Polynomial Interpolation

Nonlinear Programming

Numerical Analysis Fall. Numerical Differentiation

Linear Separability. Linear Separability. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons. Capabilities of Threshold Neurons

LECTURE NOTES Non-Linear Programming

Interpolation. TANA09 Lecture 7. Error analysis for linear interpolation. Linear Interpolation. Suppose we have a table x x 1 x 2...

Curve and Surface Basics

Programming, numerics and optimization

In other words, we want to find the domain points that yield the maximum or minimum values (extrema) of the function.

Numerical Optimization

Introduction to Design Optimization: Search Methods

A Derivative-Free Approximate Gradient Sampling Algorithm for Finite Minimax Problems

Chapter Multidimensional Gradient Method

In this course we will need a set of techniques to represent curves and surfaces in 2-d and 3-d. Some reasons for this include

INTERPOLATION BY QUARTIC SPLINE

Functions of Several Variables

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Using MATLAB, SIMULINK and Control System Toolbox

Rendering Algebraic Surfaces CS348B Final Project

The Rectangular Coordinate System and Equations of Lines. College Algebra

Surfaces and Partial Derivatives

1. Assumptions. 1. Introduction. 2. Terminology

10.6 Conjugate Gradient Methods in Multidimensions

Constrained and Unconstrained Optimization

The Bisection Method versus Newton s Method in Maple (Classic Version for Windows)

Solve Non-Linear Parabolic Partial Differential Equation by Spline Collocation Method

Recent Developments in Model-based Derivative-free Optimization

Introduction to Design Optimization

Sung-Eui Yoon ( 윤성의 )

Energy Minimization -Non-Derivative Methods -First Derivative Methods. Background Image Courtesy: 3dciencia.com visual life sciences

MATH Harrell. Which way is up? Lecture 9. Copyright 2008 by Evans M. Harrell II.

(Part - 1) P. Sam Johnson. April 14, Numerical Solution of. Ordinary Differential Equations. (Part - 1) Sam Johnson. NIT Karnataka.

Multidimensional scaling

Natasha S. Sharma, PhD

Mathematics 6 12 Section 26

International Association of Scientific Innovation and Research (IASIR) (An Association Unifying the Sciences, Engineering, and Applied Research)

CS 450 Numerical Analysis. Chapter 7: Interpolation

Driven Cavity Example

Grade 9 Math Terminology

Mathematics for chemical engineers

Various Methods for Medical Image Segmentation

Dynamic Analysis of Structures Using Neural Networks

Optimization. Industrial AI Lab.

CHAPTER 6 Parametric Spline Curves

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Chapter 3 Limits and Derivative Concepts

2.3. Graphing Calculators; Solving Equations and Inequalities Graphically

13.1. Functions of Several Variables. Introduction to Functions of Several Variables. Functions of Several Variables. Objectives. Example 1 Solution

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

3.9 LINEAR APPROXIMATION AND THE DERIVATIVE

Drawing fractals in a few lines of Matlab

MATHEMATICAL ANALYSIS, MODELING AND OPTIMIZATION OF COMPLEX HEAT TRANSFER PROCESSES

Handout 4 - Interpolation Examples

AC : MATHEMATICAL MODELING AND SIMULATION US- ING LABVIEW AND LABVIEW MATHSCRIPT

Natural Quartic Spline

Graphical Approach to Solve the Transcendental Equations Salim Akhtar 1 Ms. Manisha Dawra 2

Themes in the Texas CCRS - Mathematics

MATH3016: OPTIMIZATION

Iterative Methods for Solving Linear Problems

Data-Driven Modeling. Scientific Computation J. NATHAN KUTZ OXPORD. Methods for Complex Systems & Big Data

EARLY INTERIOR-POINT METHODS

1.2 Numerical Solutions of Flow Problems

On the Parameter Estimation of the Generalized Exponential Distribution Under Progressive Type-I Interval Censoring Scheme

Transcription:

Module 4 : Solving Linear Algebraic Equations Section 11 Appendix C: Steepest Descent / Gradient Search Method 11 Appendix C: Steepest Descent / Gradient Search Method In the module on Problem Discretization using Approximation Theory, we derived the necessary condition for a point to be an extreme point of a twice differentiable scalar function and sufficient conditions for the extreme point to qualify either as a minimum or a maximum. The question that needs to be answered in practice is, given how does one locate an extreme point A direct approach can be to solve for simultaneous equations -------(242) in unknowns. When these equations are bell behaved nonlinear functions of an iterative scheme, such as Newton-Raphson method, can be employed. Alternatively, iterative search schemes can be derived using evaluations of and its derivatives. The steepest descent / gradient method is the simplest iterative search scheme in the unconstrained optimization and forms a basis for developing many sophisticated optimization algorithms. In this section, we discuss details of this numerical optimization approach. 11.1 Gradient / Steepest Descent / Cauchy's method Set of vectors such that where is a constant, is called the level surface of for value. By tracing out one by one level surfaces we obtain contour plot (see Figure steepdescent). Suppose is a point lying on one of the level surfaces. If is continuous and differentiable then, using Taylor series expansion in a neighborhood of we can write If we neglect the second and higher order terms, we obtained -------(243) -------(244 & 245) This is equation of the plane tangent to surface at point. The equation of level surface through is -------(246) Combining above two equations, the equation of tangent surface at -------(247) is obtained as Thus, gradient at is perpendicular to the level surface passing through (See Figure steepdescent).

Figure 2: Contour plots (level surfaces) of = (x) and the local tangent and the steepest descent direction at point (n2,n1) We will now show that it points in the direction in which the function increases most rapidly, and in fact, is the direction of maximum slope. If then ------(248) and is called as descent direction. Suppose we fix our selves to unit sphere in the neighborhood of i.e. set of all such that and want to find direction such that algebraically minimum. Using Cauchy-Schwartz inequality together with we have ------(249) This implies ------(250) and minimum value can attain when is restricted within the unit ball equals In fact, the equality will hold if and only if is chosen colinear with. Let denote unit vector along the -ve of the gradient direction, i.e. ------(251) Table 7: Gradient Search Algorithm INITIALIZE: WHILE

END WHILE Then, is the direction of steepest or maximum descent in which reduces at the maximum rate. While these arguments provides us with the local direction for the steepest descent, they does not give any clue on how much we should move in that direction so that the function continues to decresae. To arrive at an optimal step size, the subsequent guess is constructed as follows ------(252) and the step size is determined by solving the following one dimensional minimization (line search) problem There are several numerical approaches available for solving this one dimensional optimization problem (ref Rao). An approach based on cubic polynomial interpolation is presented in the next sub-section. The iterative gradient based search algorithm can be summarized as in Table GradientSearch. Alternate criteria which can be used for termination of iterations are as follows The Method of steepest descent may appear to be the best unconstrained minimization method. However, due to the fact that steepest descent is a local property, this method is not effective in many problems. If the objective function are distorted, then the method can be hopelessly slow. 11.2 Line Search: One Dimensional Optimization In any multi-dimensional minimization problem, at each iteration, we have to solve a one dimensional minimization problem of the form ------(253) where is the descent (search) direction. (In gradient search method, i.e. -ve of the gradient direction.) There are many ways of solving this problem numerically (Rao). In this subsection, we discuss the cubic interpolation method, which is one of the popular techniques for performing the line search. The first step in the line search is to find bounds on the optimal step size by finding two points, say and, such that the slope These are established

------(254 & 255) has opposite signs at these points. We know that at ------(256) as is assumed to be a descent direction. Thus, we take corresponding to and try to find a point such that The point can be taken as the first value out of for which where is some pre-assigned initial step size. As changes sign in the interval the optimum is bounded in the interval The next step is to approximate over interval by a cubic interpolating polynomial for the form The parameters and be computed as ------(257) The parameters and can be computed by solving i.e. by solving The application of necessary condition for optimality yields i.e. ------(258) ------(259) One of the two values correspond to the minimum. The sufficiency condition for minimum requires ------(260) The fact that has opposite sign at and ensures that the equation Q does not have imaginary roots. Table 8: Line Search using Cubic Interpolation Algorithm INITIALIZE:

Step 1: Find WHILE END WHILE Step 2: Solve for and using and Step 3: Find using sufficient condition for optimality References 1. Gupta, S. K.; Numerical Methods for Engineers. Wiley Eastern, New Delhi, 1995. 2. Gourdin, A. and M Boumhrat; Applied Numerical Methods. Prentice Hall India, New Delhi. 3. Strang, G.; Linear Algebra and Its Applications. Harcourt Brace Jevanovich College Publisher, New York, 1988. 4. Kelley, C.T.; Iterative Methods for Linear and Nonlinear Equations, Philadelphia : SIAM, 1995. 5. Demidovich, B. P. and I. A. Maron; Computational Mathematics. Mir Publishers, Moskow, 1976. 6. Atkinson, K. E.; An Introduction to Numerical Analysis, John Wiley, 2001. 7. Linfield, G. and J. Penny; Numerical Methods Using Matlab, Prentice Hall, 1999. 8. Phillips, G. M. and P. J. Taylor; Theory and Applications of Numerical Analysis, Academic Press, 1996. 9. Rao, S. S., Optimization: Theory and Applications, Wiley Eastern, New Delhi, 1978. 10. Bazara, M.S., Sherali, H. D., Shetty, C. M., Nonlinear Programming, John Wiley, 1979.