Unconstrained Optimization

Similar documents
(1) Given the following system of linear equations, which depends on a parameter a R, 3x y + 5z = 2 4x + y + (a 2 14)z = a + 2

MTAEA Convexity and Quasiconvexity

5 Day 5: Maxima and minima for n variables.

Convexity and Optimization

Convexity and Optimization

Minima, Maxima, Saddle points

22. LECTURE 22. I can define critical points. I know the difference between local and absolute minimums/maximums.

Lecture 2 Optimization with equality constraints

3.3 Optimizing Functions of Several Variables 3.4 Lagrange Multipliers

LECTURE 18 - OPTIMIZATION

Lecture 25 Nonlinear Programming. November 9, 2009

Applied Lagrange Duality for Constrained Optimization

Bob Brown Math 251 Calculus 1 Chapter 4, Section 4 Completed 1 CCBC Dundalk

Lecture 5: Properties of convex sets

EC422 Mathematical Economics 2

Key points. Assume (except for point 4) f : R n R is twice continuously differentiable. strictly above the graph of f except at x

Section 4.2 selected answers Math 131 Multivariate Calculus D Joyce, Spring 2014

EC5555 Economics Masters Refresher Course in Mathematics September Lecture 6 Optimization with equality constraints Francesco Feri

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS

EC 521 MATHEMATICAL METHODS FOR ECONOMICS. Lecture 2: Convex Sets

Lecture 2 September 3

2 Second Derivatives. As we have seen, a function f (x, y) of two variables has four different partial derivatives: f xx. f yx. f x y.

FAQs on Convex Optimization

Chapter 4 Concepts from Geometry

MATH2111 Higher Several Variable Calculus Lagrange Multipliers

MS&E 213 / CS 269O : Chapter 1 Introduction to Introduction to Optimization Theory

Convexity. 1 X i is convex. = b is a hyperplane in R n, and is denoted H(p, b) i.e.,

Lecture 5: Duality Theory

Lecture 2 - Introduction to Polytopes

Chapter 8. Voronoi Diagrams. 8.1 Post Oce Problem

AM 221: Advanced Optimization Spring 2016

Math 21a Homework 22 Solutions Spring, 2014

Mathematical Programming and Research Methods (Part II)

Convexity Theory and Gradient Methods

Simplex Algorithm in 1 Slide

THREE LECTURES ON BASIC TOPOLOGY. 1. Basic notions.

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

The complement of PATH is in NL

Inverse and Implicit functions

for Approximating the Analytic Center of a Polytope Abstract. The analytic center of a polytope P + = fx 0 : Ax = b e T x =1g log x j

Tangent Planes/Critical Points

M3P1/M4P1 (2005) Dr M Ruzhansky Metric and Topological Spaces Summary of the course: definitions, examples, statements.

Local and Global Minimum

POLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if

Bounded, Closed, and Compact Sets

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Convex Sets (cont.) Convex Functions

Chapter 3 Numerical Methods

Increasing/Decreasing Behavior

control polytope. These points are manipulated by a descent method to compute a candidate global minimizer. The second method is described in Section

Lecture 4: Convexity

Linear Programming. Larry Blume. Cornell University & The Santa Fe Institute & IHS

CS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm. Instructor: Shaddin Dughmi

Math 443/543 Graph Theory Notes 11: Graph minors and Kuratowski s Theorem

Point-Set Topology 1. TOPOLOGICAL SPACES AND CONTINUOUS FUNCTIONS

A metric space is a set S with a given distance (or metric) function d(x, y) which satisfies the conditions

Lecture 15: The subspace topology, Closed sets

Constrained Optimization

Introduction to optimization

Chapter 15 Introduction to Linear Programming

Chapter 3. Convex Sets and Concave Functions

Finding local maxima and minima for functions z=f(x,y) [the cookbook]

.Math 0450 Honors intro to analysis Spring, 2009 Notes #4 corrected (as of Monday evening, 1/12) some changes on page 6, as in .

Increasing/Decreasing Behavior

Polynomial Functions Graphing Investigation Unit 3 Part B Day 1. Graph 1: y = (x 1) Graph 2: y = (x 1)(x + 2) Graph 3: y =(x 1)(x + 2)(x 3)

QEM Optimization, WS 2017/18 Part 4. Constrained optimization

Lecture-12: Closed Sets

Properties of a Function s Graph

Section 13. Basis for a Topology

Math 414 Lecture 2 Everyone have a laptop?

Numerical Optimization

Compact Sets. James K. Peterson. September 15, Department of Biological Sciences and Department of Mathematical Sciences Clemson University

Lecture 2: August 31

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.

Hw 4 Due Feb 22. D(fg) x y z (

QUADRATIC FUNCTIONS: MINIMUM/MAXIMUM POINTS, USE OF SYMMETRY. 7.1 Minimum/Maximum, Recall: Completing the square

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Lecture 1: Introduction

A small review, Second Midterm, Calculus 3, Prof. Montero 3450: , Fall 2008

Open and Closed Sets

Second Midterm Exam Math 212 Fall 2010

Math 734 Aug 22, Differential Geometry Fall 2002, USC

Lesson 4: Gradient Vectors, Level Curves, Maximums/Minimums/Saddle Points

Lecture 2: August 29, 2018

Math 5593 Linear Programming Lecture Notes

1 Linear programming relaxation

Worksheet 2.7: Critical Points, Local Extrema, and the Second Derivative Test

Convex sets and convex functions

CPSC 320 Sample Solution, Playing with Graphs!

ISM206 Lecture, April 26, 2005 Optimization of Nonlinear Objectives, with Non-Linear Constraints

Multivariate Calculus Review Problems for Examination Two

Lec13p1, ORF363/COS323

Section 4.3: How Derivatives Affect the Shape of the Graph

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

2017 SOLUTIONS (PRELIMINARY VERSION)

CSE 460. Today we will look at" Classes of Optimization Problems" Linear Programming" The Simplex Algorithm"

1. Gradient and Chain Rule. First a reminder about the gradient f of a smooth function f. In R n, the gradient operator or grad has the form

This lecture: Convex optimization Convex sets Convex functions Convex optimization problems Why convex optimization? Why so early in the course?

CMU-Q Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization. Teacher: Gianni A. Di Caro

2 ATTILA FAZEKAS The tracking model of the robot car The schematic picture of the robot car can be seen on Fig.1. Figure 1. The main controlling task

Transcription:

Unconstrained Optimization Joshua Wilde, revised by Isabel Tecu, Takeshi Suzuki and María José Boccardi August 13, 2013 1 Denitions Economics is a science of optima We maximize utility functions, minimize cost functions, and nd optimal allocations In order to study optimization, we must rst dene what maxima and minima are Let f : X Y be a function Then 1 We say x X is a local maximum of f on X if there is r > 0 such that f(x f(y for all y X B(x, r If the inequality is strict, then we have a strict local maximum 2 We say x X is a local minimum of f on X if there is r > 0 such that f(x f(y for all y X B(x, r If the inequality is strict, then we have a strict local minimum 3 We say x X is a global maximum of f on X if f(x f(y for all y X If the inequality is strict, then we have a strict global maximum 4 We say x X is a global minimum of f on X if f(x f(y for all y X If the inequality is strict, then we have a strict global minimum Optimization problems are often written in the form max f(x x X In this notation max refers to the global maximum of f on X The point at which the maximum is achieved is called the maximizer of f on X and usually denoted x or argmax x X f(x If there are multiple global maxima of f on X, then argmax x X f(x denotes the whole set of them In the following we seek conditions whereby we can tell whether maxima or minima exist, and if a point x X is a local maximum or minimum 2 Weierstrass Theorem The Weierstrass Theorem is one of the most important in economics It states conditions under which we are guaranteed to nd a global maximum We state it here without proof Weierstrass Theorem Let D R N be a compact set, and f : D R a continuous function Then f attains a (global maximum and a (global minimum on D, ie z 1, z 2 D such that f(z 1 f(x f(z 2 x D The crucial part of this theorem is that the set D has to be compact, that is, bounded and closed f attains its maximum either at the boundary of D or in the interior The following discussion will 1

2 Unconstrained Optimization focus on extreme points to be found in the interior of a set (in fact, we will usually let the domain of our function be an open set But we should not forget that if a function's domain is compact, the extreme points can also be attained on the boundary of the domain These extreme points cannot be found with the rst order conditions that apply for open domains 3 First Order Conditions Let F : U R 1 be a continuously dierentiable function dened on U an open subset of R n If x is a local maximum or minimum of F in U, then DF (x = 0 Notice that the converse is not true, namely "if DF (x = 0, then x is a local maximum or a local minimum" For example, consider the function f(x = x 3 on R 1 Then Df = 3x 2, which implies that when x = 0, Df(0 = 0 However, x = 0 is not a local maximum or minimum since if x > 0 in any ɛ-ball about 0 then f(x > f(0, and if x < 0 in any ɛ-ball about 0 then f(x < f(0 A condition which only goes in the direction such as this is called a necessary condition A condition which only goes in the direction is called a sucient condition Therefore, if a condition goes in both directions, we say it is a necessary and sucient condition Note that our rst order condition for maxima or minima is a necessary condition, but not sucient Examples 1 Let f : R R, f(x = 2x 3 3x 2 Then Df(x = 6x 2 6x = 6x(x 1, which implies that the only candidates for a maximum or minimum are x = 0 and x = 1 Without further conditions, however, we cannot say whether these are actual maxima or minima 2 Let F : R 2 R, F (x, y = x 3 y 3 + 9xy Then DF (x = (3x 2 + 9y, 3y 2 + 9x, which implies that the only candidates for a maximum or minimum are when 3x 2 + 9y = 0 and 3y 2 + 9x = 0 Solving the rst equation for y yields y = 1 3 x2 Plugging this into the other equation we have: This equation can be re-written as: ( 0 = 3y 2 + 9x = 3 1 2 3 x2 + 9x = 1 3 x4 + 9x 1 3 x4 + 9x = 27x x 4 = x(27 x 3, which implies that x = 0 and x = 3 are possible solutions Plugging the x solutions into the equation for y gives y = 0 and y = 3 respectively Therefore, the only possible optima are at (x, y = (0, 0 and (x, y = (3, 3 Without further conditions, however, we cannot say whether these are actual maxima or minima 4 Second Order Conditions The following theorem provides a sucient condition for nding local maxima or minima: Let F : U R be twice continuously dierentiable, where U is an open subset of R n, and the rst order condition holds for some x U:

Math Camp 3 1 If the Hessian matrix D 2 F (x is a negative denite matrix, then x is a strict local maximum of F 2 If the Hessian matrix D 2 F (x is a positive denite matrix, then x is a strict local minimum of F 3 If the Hessian matrix D 2 F (x is an indenite matrix, then x is neither a local maximum nor a local minimum of F In this case x is called a saddle point Notice again, however, that this proof does not go both ways For example, it is not true that all local minima have positive denite Hessian matrices For example, take the function f(x = x 4, which has a local minimum at x = 0, but its Hessian at x = 0 is D 2 f(x = 0, which is not positive denite The following theorem provides weaker necessary conditions on the Hessian for a local maximum or minimum: 1 Let F : U R be twice continuously dierentiable, where U is an open subset of R n, and x is a local maximum of F on U Then DF (x = 0, and D 2 F (x is negative semidenite 2 Let F : U R be twice continuously dierentiable, where U is an open subset of R n, and x is a local minimum of F on U Then DF (x = 0, and D 2 F (x is positive semidenite According to the weaker necessary conditions, if we can nd x such that DF (x = 0 and D 2 F (x = 0 is either negative (or positive semidenite, then that x is a candidate for a local maximum (or minimum However, we cannot know for sure without further inspection Examples 1 Recall the function f : R R, f(x = 2x 3 3x 2 has DF (x = 0 when x = 0 or x = 1 We can calculate that D 2 F (x = 12x 6 When x = 0, then D 2 F (x = 6 which is negative denite, so we can be sure that x = 0 is a local maximum However, when x = 1, then D 2 F (x = 6 which is positive denite, so we can be sure that x = 1 is a local minimum 2 Recall the function F : R 2 R, F (x, y = x 3 y 3 + 9xy has DF (x, y = 0 when (x, y = (0, 0 or (x, y = (3, 3 We can calculate that D 2 F (x = ( 6x 9 9 6y When (x, y = (0, 0, then D 2 F (x = ( 0 9 9 0 In this case, the rst order leading principal minor (the determinant of the matrix left after we delete the last row and column, or the determinant of the top left element is 0, and the second order principal minor (the determinant of the whole matrix is 81 Therefore, this matrix is indenite, and (x, y = (0, 0 is neither a maximum or minimum When (x, y = (3, 3, then D 2 F (x = ( 18 9 9 18 In this case, the rst order leading principal minor is 18, and the second order principal minor is 243 Therefore, this matrix is positive denite, and (x, y = (0, 0 is a strict local minimum

4 Unconstrained Optimization 5 Concavity, Convexity, and Global Optima Let F : U R be twice continuously dierentiable, and U an open set Then the function F is concave i D 2 F (x is negative semidenite for all x U, and is convex i D 2 F (x is positive semidenite for all x U If F is a concave function and DF (x = 0 for some x U, then x is a global maximum of F on U If F is a convex function and DF (x = 0 for some x U, then x is a global minimum of F on U Example 1 Inspect the function F (x, y = x 2 + y 2, and nd any maxima or minima Can you tell whether they are local or global? The Hessian is DF (x, y = (2x, 2y DF (x, y = 0 for x = 0 and y = 0 D 2 F (x, y = ( 2 0 0 2 Since this is a diagonal matrix with only strictly positive entries on the diagonal, it is positive denite, independent of x and y Therefore, it is positive denite at the critical point (0, 0, and (0, 0 is a strict local min (second order condition But we can say even more: The Hessian is positive denite at all points, hence also positive semi-denite (a weaker condition at all points Therefore F (x, y is convex By the theorem above, the point (0, 0 is a strict global minimum 2 Consider the function f(x = x 4 Find the critical values and classify them f (x = 4x 3 so the only critical point is x = 0 The Hessian is f (x = 12x 2 Evaluated at x, the Hessian is f (x = 0, so the second order condition does not tell us whether x is a maximum or a minimum However, looking at the Hessian for all points in the domain gives us more information: f (x = 12x 2 0 for all x, so the Hessian is positive semi-denite on the whole domain and f is convex Therefore, x has to be a global minimum 3 Consider F (x, y = x 2 y 2 Find the critical values and classify them DF (x, y = (2xy 2, 2x 2 y DF (x, y = 0 for x = 0 or y = 0 We have innitely many critical points which are of three dierent forms: The Hessian is z 1 = (x, 0 with x 0, z 2 = (0, y with y 0, z 3 = (0, 0 D 2 F (x, y = ( 2y 2 4xy 4xy 2x 2 First of all, the Hessian is not always positive semidenite or always negative denite (rst oder principal minors are 0, second order principal minor is 0, so F is neither concave nor convex Let's determine the deniteness of D 2 F (x, y at critical points of the form (x, 0 with x 0 In this case the Hessian is NOT negative semidenite, so these points cannot be local maxima The Hessian is positive semidenite, but that is not sucient to conclude anything else Similarly, at critical points of the form (0, y with y 0 the Hessian is NOT negative denite, so these points cannot be local maxima either, but we can't say more At the critical point (0, 0, the Hessian is positive semidenite and negative semidenite, so we can't say anything This is an example where the conditions on the Hessian do not provide a lot of information But we can use "common sense" instead: F (x, y 0 for all x, y, so points at which F (x, y = 0 have to be global minima Therefore, all the critical points are global minima

Math Camp 5 6 Homework 1 For each function, determine whether the it denitely has a maximum, denitively does not have a maximum, or that there is not enough information to tell, using the Weierstrass Theorem If it denitely has a maximum, prove that this is the case (a f : R R, f(x = x (b f : [ 1, 1] R, f(x = x (c f : ( 1, 1 R, f(x = x { 0 if x = 1 (d f : [ 1, 1] R, f(x = x otherwise { 1 if x = 5 (e f : R ++ R, f(x = 0 otherwise 2 Consider the standard utility maximization problem max U(x, where B(p, I = {x x B(p,I Rn + p x I} Prove a solution exists for any U(x continuous, I > 0 and p R n ++ Show a solution may not necessarily exist if p R n + 3 Search for local maxima and minima in the following functions More specically, nd the points where DF (x = 0, and then classify then as a local maximum, a local minimum, denitely not a maximum or minimum, or can't tell Also, check whether the functions are concave, convex, or neither The answers (except for the concavity/convexity part are found in the back of Simon and Blume, Exercises 171-172 (a F (x, y = x 4 + x 2 6xy + 3y 2 (b F (x, y = x 2 6xy + 2y 2 + 10x + 2y 5 (c F (x, y = xy 2 + x 3 y xy (d F (x, y = 3x 4 + 3x 2 y y 3 (e F (x, y, z = x 2 + 6xy + y 2 3yz + 4z 2 10x 5y 21z (f F (x, y, z = ( x 2 + 2y 2 + 3z 2 e (x2 +y 2 +z 2