Size: px
Start display at page:

Download ""

Transcription

1 Numerical experience with a derivative-free trust-funnel method for nonlinear optimization problems with general nonlinear constraints by Ph. R. Sampaio and Ph. L. Toint Report naxys-3-25 August 2, 25 Log 2 Scaled Performance Profile on Subset of CUTEst ρs() Subbasis Min. Frobenius norm Min. l. 2 norm Regression University of Namur, 6, rue de Bruxelles, B5 Namur (Belgium)

2 Numerical experience with a derivative-free trust-funnel method for nonlinear optimization problems with general nonlinear constraints Ph. R. Sampaio and Ph. L. Toint August 2, 25 Abstract A trust-funnel method is proposed for solving nonlinear optimization problems with general nonlinear constraints. It extends the one presented by Gould and Toint (Math. Prog., 22():55-96, 2), originally proposed for equality-constrained optimization problems only, to problems with both equality and inequality constraints and where simple bounds are also considered. As the original one, our method maes use of neither filter nor penalty functions and considers the objective function and the constraints as independently as possible. To handle the bounds, an activeset approach is employed. We then exploit techniques developed for derivative-free optimization to obtain a method that can also be used to solve problems where the derivatives are unavailable or are available at a prohibitive cost. The resulting approach extends the DEFT-FUNNEL algorithm presented by Sampaio and Toint (Comput. Optim. Appl., 6():25-49, 25), which implements a derivative-free trust-funnel method for equality-constrained problems. Numerical experiments with the extended algorithm show that our approach compares favorably to other well-nown modelbased algorithms for derivative-free optimization. Keywords: Constrained nonlinear optimization, trust-region method, trust funnel, derivative-free optimization. Introduction We consider the solution of the nonlinear optimization problem min f(x) x s.t.: l s c(x) u s, l x x u x, (.) where we assume that f : IR n IR and c : IR n IR m are twice continuously differentiable, and that f is bounded below on the feasible domain. The vectors l s and u s are lower and upper bounds, respectively, on the constraints values c(x), while l x and u x are bounds on the x variables, with l s (IR ) m, u s (IR ) m, l x (IR ) n and u x (IR ) n. Namur Center for Complex Systems (naxys) and Department of Mathematics, University of Namur, 6, rue de Bruxelles, B-5 Namur, Belgium. phillipece@gmail.com Namur Center for Complex Systems (naxys) and Department of Mathematics, University of Namur, 6, rue de Bruxelles, B-5 Namur, Belgium. philippe.toint@unamur.be

3 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 2 By ining f(x, s) = f(x) and c(x, s) = c(x) s, the problem above may be rewritten as the following equality-constrained optimization problem with simple bounds min f(x, s) (x,s) s.t.: c(x, s) =, (.2) l s s u s, l x x u x, which is the one we address throughout this paper. The trust-funnel method was first introduced by Gould and Toint [] (see also []) as a SQP (Sequential Quadratic Programming) algorithm for equality-constrained optimization problems whose convergence is driven by an adaptive bound imposed on the allowed infeasibility at each iteration. Such bound is monotonically decreased as the algorithm progresses, assuring its global convergence whilst seeing optimality; hence the name trust funnel. It belongs to the class of trust-region methods and maes use of a composite-step approach to calculate a new direction at each iteration: a normal step is computed first in the hope of reducing the infeasibility measure ensuing from the constraint functions values, and a tangent step is subsequently calculated with the aim of improving optimality of the iterates with regard to the objective function. These computations are carried out with the use of two different trust regions, one for each step component. The main idea is to consider the objective function and the constraints as independently as possible. The method is noticeable among others for constrained problems as a parameter-free alternative, for neither filter nor penalties are needed, freeing the user from common difficulties encountered when choosing the initial penalty parameter, for instance. An extension to problems with both equalities and inequalities was developed recently by Curtis et al. [7], who presented an interior-point trust-funnel algorithm for solving large-scale problems that may be characterized as a barrier-sqp method. Although many methods have been designed for unconstrained DFO (Derivative-Free Optimization), the range of possibilities for the constrained case remains quite open. A derivative-free trustfunnel method called DEFT-FUNNEL (DErivative-Free Trust FUNNEL), proposed by Sampaio and Toint [22] for equality-constrained problems, has given good results in comparison with a filter-sqp method introduced by Colson [4] and a trust-region method by Powell named COBYLA (Constrained Optimization BY Linear Approximations) [9, 2], all centred on local polynomial interpolation models. Recently, another trust-region method by Powell [2] entitled LINCOA has been developed for linearly-constrained optimization without derivatives and maes use of quadratic interpolation-based models for the objective function. Many other methods that use direct search instead of trust-region techniques combined to interpolation-based models also have been developed for the constrained case (e.g., Lewis and Torczon [4, 5, 6], Audet and Dennis [], Lucidi et al. [7], Yu and Li [24]). The primary contribution of this paper is the extension of both the original trust-funnel method [] and its derivative-free adaptation (DEFT-FUNNEL) [22] to problems with general nonlinear constraints as well as the presentation of numerical experience with the resulting algorithm. Particularly for the derivative-free algorithm, a subspace minimization approach proposed by Gratton et al. [2] that considers the poisedness of the interpolation set is employed in order to avoid its degeneration. The final method described here features four main steps to solve problems with general nonlinear constraints, namely: a subspace minimization approach to handle the bounds on the x variables, which maes the DEFT-FUNNEL an active-set method, a bounded linear least-squares solver to calculate the normal step, a projected gradient method to calculate the tangent step and the control of the permitted infeasibility of the iterates through the funnel bound. The reason behind the choice of exploring subspaces ined by the active bounds is dual. Besides the fact that we aim to avoid treating the bounds on the x variables as general inequality constraints, the reduction of the dimension of the problem after having identified the active bounds helps to thwart a possible degeneration of

4 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 3 the interpolation set when the sample points become close to each other and thus affinely dependent, which happens often as the optimal solution is approached. The fact that the s variables play no role on the choice of the interpolation set vindicates the construction of the subspaces based upon the x variables only. This paper is organized as follows. Section 2 introduces the proposed trust-funnel algorithm step by step. The subsection 2. explains the subspace minimization approach, while the subsections 2.2 and 2.3 give details on how the computation of the composite step components is conducted. The subsections 2.4 and 2.5 then address the whole mechanism of the algorithm itself. Section 3 concerns the application of the method to derivative-free optimization problems and Section 4 gives the description of the final method, which assembles the techniques elaborated in Sections 2 and 3. In Section 5, we discuss some initial numerical experiments and compare its performance to COBYLA s and LINCOA s in a selected set of test problems. Ultimately, we draw some final conclusions in Section 6. Notation. Unless otherwise specified, our norm is the standard Euclidean norm. Given any vector x, we denote its i-th component by [x] i. We let B(z; ) denote the closed Euclidian ball centered at z, with radius >. Given any set A, A denotes the cardinality of A. By P d n, we mean the space of all polynomials of degree at most d in IR n. Finally, given any subspace S, we denote its dimension by dim(s). 2 An active-set trust-funnel method Our method generates a sequence of points {(x, s )} such that, at each iteration, the bound constraints below are satisfied l s s u s, (2.3) l x x u x. (2.4) By using a composite-step approach, each trial step d = (d x, ds )T is decomposed as d = ( d x d s ) = ( n x n s ) + ( t x t s ) = n + t, where the normal step component n = (n x, ns )T aims to improve feasibility, and the tangent step component t = (t x, ts )T reduces the objective function model s value while preserving any gains in feasibility obtained through n. In the following subsections, we will describe how each component is computed. Before that, we briefly explain how the subspace minimization is employed in our algorithm. 2. Subspace minimization As in the method proposed by Gratton et al. [2] for bound-constrained optimization problems, our algorithm maes use of an active-set approach where the minimization is restricted to subspaces ined by the active x variables. At each iteration, we ine the subspace S as follows S = {x IR n [x] i = [l x ] i for i L and [x] i = [u x ] i for i U }, where L = {i [x ] i [l x ] i ɛ b } and U = {i [u x ] i [x ] i ɛ b } ine the index sets of (nearly) active variables at their bounds, for some small constant ɛ b > ined a priori. After that S has been

5 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 4 ined, the minimization at iteration is then restricted to the new subspace S. Once a direction d for (x, s ) has been computed, we set (x +, s + ) = (x, s ) + d if is a successful iteration; otherwise, we set (x +, s + ) = (x, s ). If a solution ( x, s ) for the subproblem ined by S satisfies the optimality conditions for the subproblem, we chec whether it is also optimal for the original problem (2.5). If it is, the solution encountered is returned to the user and the algorithm halts; otherwise, it proceeds in the full space by computing a new direction for ( x, s ) and repeats the above process at iteration + by ining S The normal step For any point (x, s), we measure the constraint violation by v(x, s) = 2 c(x, s) 2. (2.5) Analogous to the trust-funnel method for equality-constrained problems proposed in [] and in [22], we also have a funnel bound v max for v such that, for each iteration, v v max, where v = v(x, s ). As this bound is monotonically decreased, the algorithm is driven towards feasibility, guaranteeing the convergence of the algorithm. In order to ensure that the normal step is indeed normal, we add the following condition that ass that it mostly lies in the space spanned by the columns of the matrix J(x, s ) T : n κ n c(x, s ), (2.6) for some κ n. We then perform the calculation of n by solving the trust-region bound-constrained linear least-squares problem min n=(n x,n s ) c(x 2, s ) + J(x, s )n 2 s.t.: l s s + n s u s, l x x + n x u x, x + n x S, n N, where J(x, s) = (J(x) I) represents the Jacobian of c(x, s) with respect to (x, s) and for some trust-region radius c >. 2.3 The tangent step (2.7) N = {z IR n+m z min [ c, κ n c(x, s ) ] }, (2.8) After attempting to reduce infeasibility through the normal step, a tangent step is calculated with the aim of improving optimality. The computation of the latter is conducted carefully so that the gains in feasibility obtained by the former are not abandoned without good reasons. The SQP model for the function f is ined as ψ ((x, s ) + d) = f + g, d + 2 d, B d, (2.9)

6 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 5 where f = f(x, s ), g = (x,s) f(x, s ), and B is the approximate Hessian of the Lagrangian function L(x, s, µ, z s, w s, z x, w x ) = f(x) + µ, c(x, s)) + w s, s u s + z s, l s s + w x, x u x + z x, l x x with respect to (x, s), given by ( H + m B = i= [ˆµ ) ] i C i, (2.) where z s and w s are the Lagrange multipliers associated to the lower and upper bounds, respectively, on the slac variables s, and z x and w x, the Lagrange multipliers associated to the lower and upper bounds on the x variables. In (2.), H is a bounded symmetric approximation of 2 xxf(x, s ) = 2 f(x ), the matrices C i are bounded symmetric approximations of the constraints Hessians 2 xxc i (x, s ) = 2 c i (x ) and the vector ˆµ may be viewed as a local approximation of the Lagrange multipliers with respect to the equality constraints c(x, s). By using the decomposition d = n + t, we then have that where ψ ((x, s ) + n + t) = ψ ((x, s ) + n ) + g N, t + 2 t, B t, (2.) g N = g + B n. (2.2) In the interest of assuring that (2.) it is a proper local approximation for the function f((x, s ) + n + t), the complete step d = n + t must belong to T = {d IR n+m d f }, (2.3) for some radius f. The minimization of (2.) should then be restricted to the intersection of N and T, which imposes that the tangent step t results in a complete step d = n + t that satisfies the inclusion where the radius of R is thus given by d R = N T = {d IR n+m d }, (2.4) = min[ c, f ]. (2.5) Similarly to the computation of the normal step, we intend to remain in the subspace S after waling along the tangent direction. We accomplish that by imposing the following condition x + n x + tx S. Additionaly, the conditions (2.3) and (2.4) must be satisfied at the final point (x, s ) + d, i.e. we must have l s s + n s + ts u s, l x x + n x + tx u x. Due to (2.4), we as n to belong to R before attempting the computation of t by requiring that n κ R, (2.6)

7 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 6 for some κ R (, ). If (2.6) happens, which means that there is enough space left to mae another step without crossing the trust region border, the tangent step is finally computed by solving the following problem min g t=(t x,t s N, t + t, B 2 t ) s.t.: J(x, s )t =, l s s + n s + ts u s, l x x + n x + tx u x, x + n x + tx S. n + t R. (2.7) We ine our f-criticality measure as π f = g N, r, (2.8) where r is the projected Cauchy direction obtained by solving the linear optimization problem min g r=(r x,r s N, r ) s.t.: J(x, s )r =, l s s + n s + rs u s, l x x + n x + (2.9) rx u x, x + n x + rx S. r. By inition, π f measures how much decrease could be obtained locally along the projection of the negative of the approximate gradient g N onto the nullspace of J(x, s ) intersected with the region delimited by the bounds. A new local estimate of the Lagrange multipliers (µ, z s, ws, zx, wx ) are computed by solving the following bound-constrained linear least-squares problem { min M (µ,ẑ s,ŵ s,ẑ x,ŵ x 2 (µ, ẑ s, ŵ s, ẑ x, ŵ x ) 2. ) (2.2) s.t.: ẑ s, ŵ s, ẑ x, ŵ x, where M (µ, ẑ s, ŵ s, ẑ x, ŵ x ) = ( g N + ( I x w ) ( J(x ) + T I ) ŵ x + ( I x z ) ( µ + ) ẑ x, I s w ) ( ŵ s + Iz s ) ẑ s the matrix I is the m m identity matrix, the matrices Iz s and Iw s are obtained from I by removing the columns whose indices are not associated to any active (lower and upper, respectively) bound at s + n s, the matrices Ix z and Iw x are obtained from the n n identity matrix by removing the columns whose indices are not associated to any active (lower and upper, respectively) bound at x + n x, and the Lagrange multipliers (ẑ s, ŵ s, ẑ x, ŵ x ) are those in (z s, w s, z x, w x ) associated to active bounds at s + n s and x + n x. All the other Lagrange multipliers are set to zero.

8 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization Which steps to compute and retain We now explain when the normal step n and the tangent step t should be computed at iteration given the constraint violation and the measure of optimality described in the previous subsection. The normal step is computed when = or the current constraint violation is significant, which is now formally ined by the conditions c(x, s ) > ω n (π f ) or v > κ vv v max, (2.2) where ω n is some bounding function, κ vv (, ) is a constant. If (2.2) fails, the computation of the normal step is not required and so we set n =. We ine a v-criticality measure that indicates how much decrease could be obtained locally along the projection of the negative gradient of the Gauss-Newton model of v at (x, s ) onto the region delimited by the bounds as = J(x, s ) T c(x, s ), b, where the projected Cauchy direction b is given by the solution of min J(x, s ) T c(x, s ), b b=(b x,b s ) s.t.: l s s + b s u s, l x x + b x u x, x + b x S, b. π v We say that (x, s ) is an infeasible stationary point if c(x, s ) and π v =. The procedure for the calculation of normal step is given in the algorithm below. (2.22) Algorithm 2.: NormalStep(x, s, π v, πf, v, v max ) Step : If c(x, s ) and π v =, STOP (infeasible stationary point). Step 2: If = or if (2.2) holds, compute a normal step n by solving (2.7). Otherwise, set n =. If the solution of (2.9) is r =, then by (2.8) we have π f =. In this case, the computation of the tangent step is sipped, and we simply set t =. If π f is unsubstantial compared to the current infeasibility, i.e. for a given a monotonic bounding function ω t, the condition π f > ω t( c(x, s ) ) (2.23) fails, then the current iterate is still too far from feasibility to worry about optimality, and we again sip the tangent step computation by setting t =. While (2.23) and (2.2) together provide considerable flexibility in our algorithm in that a normal or tangent step is only computed when relevant, our setting also produces the possibility that both these conditions fail. In this case, we have that d = n +t is identically zero, and the sole computation in the iteration is that of the new Lagrange multipliers (2.2). Finally, we may evaluate the usefulness of the tangent step t after (or during) its computation, in the sense that we would lie a relatively large tangent step to cause a clear decrease in the model (2.) of the objective function. We therefore chec whether the conditions t > κ CS n (2.24)

9 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 8 and where and δ f,t δ f = δ f,t + δf,n κ δ δ f,t, (2.25) = ψ ((x, s ) + n ) ψ ((x, s ) + n + t ) (2.26) δ f,n = ψ (x, s ) ψ ((x, s ) + n ), (2.27) are satisfied for some κ CS > and for κ δ (, ). The inequality (2.25) indicates that the predicted improvement in the objective function obtained in the tangent step is not negligible compared to the predicted change in f resulting from the normal step. If (2.24) holds but (2.25) fails, the tangent step is not useful in the sense just discussed, and we choose to ignore it by resetting t =. The computation of the tangent step is described algorithmically as follows. Algorithm 2.2: TangentStep(x, s, n ) Step : If (2.6) holds, then Step.: select a vector ˆµ and ine B as in (2.); Step.2: compute µ by solving (2.2); Step.3: compute the modified Cauchy direction r by solving (2.9) and ine π f Step.4: if (2.23) holds, compute a tangent step t by solving (2.7). as (2.8); Step 2: If (2.6) fails, set µ = µ. In this case, or if (2.23) fails, or if (2.24) holds but (2.25) fails, set t = and d = n. Step 3: Define (x +, s+ ) = (x, s ) + d. 2.5 Iterations types Once we have computed the step d and the trial point (x +, s+ ) = (x, s ) + d, (2.28) we are left with the tas of accepting or rejecting it. If n = t =, iteration is said to be a µ-iteration because the only computation potentially performed is that of a new vector of Lagrange multiplier estimates. We will say that iteration is an f-iteration if t, (2.25) holds, and v(x +, s+ ) vmax. (2.29) Condition (2.29) ensures that the step maintains feasibility within reasonable bounds. Thus the iteration s expected major achievement is, in this case, a decrease in the value of the objective function f, hence its name. If iteration is neither a µ-iteration nor a f-iteration, then it is said to be a c- iteration. If (2.25) fails, then the expected major achievement (or failure) of iteration is, a contrario, to improve feasibility, which is also the case when the step only contains its normal component. The main idea behind the technique for accepting the trial point is to measure whether the major expected achievement of the iteration has been realized.

10 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 9 If iteration is a µ-iteration, we do not have any other choice than to restart with (x +, s + ) = (x, s ) using the new multipliers. We then ine f + = f and c + = c (2.3) and eep the current value of the maximal infeasibility v max + = vmax. If iteration is an f-iteration, we accept the trial point (i.e. (x +, s + ) = (x +, s+ )) if ρ f = f(x, s ) f(x +, s+ ) δ f η, (2.3) and reject it (i.e. (x +, s + ) = (x, s )), otherwise. The value of the maximal infeasibility measure is left unchanged, that is v+ max = vmax. Note that δ f > (because of (2.26) and (2.25)) unless (x, s ) is first-order critical, and hence that condition (2.3) is well-ined. If iteration is a c-iteration, we accept the trial point if the improvement in feasibility is comparable to its predicted value δ c = 2 c(x, s ) 2 2 c(x, s ) + J(x, s )d 2, and the latter is itself comparable to its predicted decrease along the normal step, that is n, δ c κ cnδ c,n and ρ c = v(x, s ) v(x +, s+ ) δ c η, (2.32) for some κ cn (, κ tg ] and where δ c,n = 2 c(x, s ) 2 2 c(x, s ) + J(x, s )n 2. (2.33) If (2.32) fails, the trial point is rejected. We update the value of the maximal infeasibility by v max { [ max + = κtx v max, v(x +, s+ ) + κ tx2(v(x, s ) v(x +, s+ ))] if (2.32) hold, v max otherwise, (2.34) for some κ tx (, ) and κ tx2 (, ). Note that we only chec the third condition in (2.32) after the first two conditions have been verified. Assuming that n, the Cauchy condition (2.33) and c(x, s ) ensure that δ c,n > provided π v. Thus the third condition is well ined, unless (x, s ) is an infeasible stationary point, in which case the algorithm is terminated. 3 Application to derivative-free optimization Our algorithm maes use of surrogate models m f (x) and m c (x) = (m c (x), mc 2 (x),..., mc m(x)) built from polynomial interpolation over a set of sample points Y = {y, y,..., y p } that replace the functions f(x) and c(x) = (c (x), c 2 (x),..., c m (x)). Therefore, at each iteration, the following interpolation conditions are satisfied m f (y i ) = f(y i ), m c j(y i ) = c j (y i ), (3.35)

11 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization for all y i Y. Whenever Y changes, the models m c and mf are updated to satisfy the interpolation conditions (3.35) for the new set Y +, thereby implying that new function evaluations of f and c are carried out for the additional points obtained at iteration. The algorithm developed in this wor employs the now standard idea of starting with incomplete interpolation models with linear accuracy and then enhancing them with curvature information, which provides an actual accuracy at least as good as that for linear models and, hopefully, better. We thus consider underdetermined quadratic interpolation i.e. n + Y (n + )(n + 2)/2 with initial sample sets that are poised for linear interpolation. As there are many possibilities for building underdetermined quadratic interpolation-based models, we consider three distinct ways whose complete descriptions are available in Conn, Scheinberg and Vicente [6], namely: subbasis selection approach, minimum l 2 -norm or Euclidian norm models (referred to as minimum-norm interpolating polynomials in Section 5. of [6]), whose coefficients vector associated to the natural basis is the minimum Euclidian norm solution of the linear system resulting from the interpolation conditions, and minimum Frobenius norm models where the coefficients related to the quadratic part of the natural basis are minimized (see the optimization problem (5.6) in [6]), which is equivalent to the minimization of the Frobenius norm of the Hessian of the models. In addition, we also consider least-squares regression for the construction of quadratic models, in which case one has n + Y (n + )(n + 2), giving the user a total of four possibilities for model building. For any function f : IR n IR and set Y = {y,..., y p } IR n, there exists an unique polynomial m(x) that interpolates f(x) on Y such that m(x) = p i= f(yi )l i (x), where {l i (x)} p i= is the basis of Lagrange polynomials, if Y satisfies a property of nonsingularity called poisedness (see Definition 3., page 37, in [6]). Before going further into details of the algorithm, we first introduce a condition to the interpolation set that is used to measure the error between the original functions and the interpolation models as well as their derivatives. Definition 3.. Let Y = {y, y,..., y p } be a poised interpolation set and P d n be a space of polynomials of degree less than or equal to d on IR n. Let Λ > and {l (x), l (x),..., l p (x)} be the basis of Lagrange polynomials associated with Y. Then, the set Y is said to be Λ-poised in B for P d n (in the interpolation sense) if and only if max max l i(x) Λ. i p x B As it is shown in [6], the error bounds for at most fully quadratic models depend linearly on the constant Λ; the smaller it is, the better the interpolation models approximate the original functions. We also note that the error bounds for undetermined quadratic interpolation models are linear in the diameter of the smallest ball B(Y) containing Y for the first derivatives and quadratic for the function values. 3. Recursive call in subspaces As pointed out before, by reducing the dimension of the problem we attempt to diminish the chances of a possible degeneration of the interpolation set when the sample points become too close to each other and thus affinely dependent. Such case might happen, for instance, as the optimal solution is approached. In our algorithm, we apply the recursive approach found in the derivative-free method proposed by Gratton et al. [2] for bound-constrained optimization problems. Once the subspace S has been ined at iteration, the algorithm calls itself recursively and the dimension of the problem is then reduced to ˆn = n L U, where n denotes here the dimension of IR n. A new well-poised

12 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization interpolation set Z is then constructed from a suitable choice of points in X S, where X is the set of all points obtained up to iteration. In order to save function evaluations in the building process of the new interpolation set, all the points in X that are nearly but not in S are projected onto S and used to build Z with their model values instead of their real function values. In a more formal description, we ine the set of the points that are close to (but not on) the active bounds at x as A = where, for at least one i, the strict inequality or { y X [y] i [l x ] i ɛ b for i L and [u x ] i [y] i ɛ b for i U < [y] i [l x ] i, i L, < [u x ] i [y] i, i U, must hold. We then project all the points y A onto S, obtaining new dummy points y s that are added to X with associated values m f (y s) and m c (y s) rather than the values of the original functions. As it is shown in Section 3.3, these dummy points are progressively replaced by other points with true function values with high priority during the minimization in S. We denote by Dum the set of dummy points at iteration. Convergence in a subspace is only declared if the interpolation set contains no dummy points. If a solution has been found for a subspace and there are still dummy points in the interpolation set, evaluations of the original functions f(x) and c(x) at such points are carried out and the interpolating models are recomputed from the original function values. Once convergence has been declared in a subspace S, the L U fixed components x i associated with the active bounds and the component x of the approximate solution found in S of dimension ˆn = n L U are assembled to compose a full-dimensional vector x S in IRn. The algorithm then checs whether (x S, s S ) is optimal for the full-dimensional problem or not. Firstly, a full-space interpolation set of degree n + is built in an ɛ-neighborhood around the point x S. Subsequently, the corresponding interpolating models mf and m c are recomputed and the f-criticality measure πf is calculated anew using information of the updated models. Finally, the criticality step in the full space is then entered. The following algorithm gives a full description of the ideas explained above. }, Algorithm 3.: SubspaceMinimization(X, Y, x, s, f, c, vmax ) Step : Chec for (nearly) active bounds at x and ine S. If there is no (nearly) active bound or if S has already been explored, go to Step 6. If all bounds are active, go to Step 5. Step 2: Project points in X which lie close to the (nearly) active bounds on S and associate with them suitable function values estimates. Add new dummy points to Dum, if any. Step 3: Build a new interpolation set Z in S including the projected points, if any. Step 4: Call recursively DEFT-FUNNEL(S, X, Z, x, s, f, c, vmax ) and let (x S, s S ) be the solution of the subspace problem after adding the fixed components. Step 5: If dim(s ) < n, return (x S, s S ). Otherwise, reset (x, s ) = (x S, s S ), construct new set Y around x, build m f and mc and recompute πf.

13 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 2 Step 6: If S has already been explored, set (x +, s + ) = (x, s ), reduce the trust regions radii f + = γ f and c + = γ c, set + = min[ f +, c + ] and build a new poised set Y + in B(x + ; + ). 3.2 Criticality step Convergence is declared when both feasibility and optimality have been achieved and the error between the real functions and the models is expected to be sufficiently small. As it was mentioned before, this error is directly lined to the Λ-poisedness measure given in Definition 3.. The criticality step in DEFT-FUNNEL is described in the next algorithm. Algorithm 3.2: CriticalityStep(Y, π f, ɛ i) Step : Define ˆm f i = mf, ˆmc i = mc and ˆπf i = π f. ] Step 2: If c(x, s ) ɛ i and ˆπ f i ɛ i, set ɛ i+ = max [α c(x, s ), αˆπ f i, ɛ and modify Y as needed to ensure it is Λ-poised in B(x, ɛ i+ ). If Y was modified, compute new models ˆm f i and ˆm c i, calculate ˆr i and ˆπ f i and increment i by one. If c(x, s ) ɛ and ˆπ f i ɛ, return (x, s ); otherwise, start Step 2 again; [ ] Step 3: Set m f = ˆmf i, mc = ˆmc i, πf = ˆπf i, = β max c(x, s ), π f and ine ϑ i = x if a new model has been computed. 3.3 Maintenance of the interpolation set and trust-region updating strategy Many derivative-based trust-regions methods decrease the trust region radius at unsuccessful iterations to converge. However, in derivative-free optimization, unsuccessful iterations in trust-region methods might be due to a bad quality of the interpolating model rather than a large trust region size. In order to maintain the quality of the surrogate models without resorting to costly model improvement steps at unsuccessful iterations, which might require another optimization problem to be globally solved, we apply a self-correcting geometry scheme proposed by Scheinberg and Toint [23] for the management of the geometry of the interpolation set. Such scheme is based upon the idea that unsuccessful trial points can still be useful as they may improve the geometry of the interpolation set. At such iterations, the trust region radii are reduced only if the algorithm fails to improve the geometry of the interpolation set by replacing one of its points by the trial point, thereby implying that there is no lac of poisedness in this case. Firstly, an interpolation point that is far from the current point x is sought to be replaced by the trial point x +. If there is no such a far point, one tries to replace an interpolation point y,j whose associated Lagrange polynomial value at x + in absolute value, l,j(x + ), is bigger than a preined threshold Λ. By doing that, one attempts to obtain a Λ-poised set Y +, or, since at most one interpolation point is replaced by iteration, improve its poisedness, at least. If no interpolation point is replaced, the trust regions are then shrun. A slight modification on the scheme is used on DEFT-FUNNEL. Following the idea proposed in [2], the trust region radii can also be reduced in cases where it is possible to improve poisedness. The

14 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 3 number of reductions of f and c allowed in such cases, however, is limited by constants νmax f > and νc max > preined by the user as a means to prevent the trust regions from becoming too small. The management of the interpolation set in DEFT-FUNNEL is described in the next algorithm. It depends on the criterion used to ine successful iterations, which is passed to the algorithm through the parameter criterion. Algorithm 3.3: UpdateInterpolationSet(Y, x, x +,, ɛ i, criterion) Step : Augment the interpolation set. If Y < p max, then ine Y + = Y {x + }. Step 2: Replace a dummy interpolation point. If Y = p max and the set D = {y,j Dum Y } (3.36) is non-empty, then ine Y + = Y \ {y,r } {x + } where l,r(x + ) and r is an index of any point in D such that r = arg max y,j x + j 2 l,j (x + ). (3.37) Step 3: Successful iteration. If Y = p max, D =, and criterion holds, then ine Y + = Y \ {y,r } {x + } for y,r = arg max y,j x + y,j 2 l,j (x + ). (3.38) Y Step 4: Replace a far interpolation point. If Y = p max, D =, criterion fails, either x ϑ i or ɛ i, and the set F = {y,j Y such that y,j x > ζ and l,j (x + ) } (3.39) is non-empty, then ine Y + = Y \ {y,r } {x + }, where y,r = arg max y,j x + y,j 2 l,j (x + ). (3.4) F Step 5: Replace a close interpolation point. If Y = p max, criterion fails, either x ϑ i or ɛ i, the set D F is empty, and the set C = {y,j Y such that y,j x ζ and l,j (x + ) > Λ} (3.4) is non-empty, then ine Y + = Y \ {y,r } {x + }, where y,r = arg max y,j x + y,j 2 l,j (x + ). (3.42) C Step 6: No replacements. If Y = p max, criterion fails and either [x = ϑ i and > ɛ i ] or D F C =, then ine Y + = Y.

15 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 4 We now give the details of the procedures related to f- and c-iterations, starting with the former. The operations involved in such procedures concern the acceptance of the trial point and the update of the trust regions radii. The relation between the trust region updating strategy at unsuccessful iterations and the management of the interpolation set discussed above is now formally described. Algorithm 3.4: f-iteration(x, s, x +, s+, f, c ) Step : Successful iteration. If ρ f η, set (x +, s + ) = (x +, s+ ) and ν f =. If ρ f η 2, set f + = min[max[γ 2 d, f ], max ]; otherwise, set f + = f. If v(x +, s+ ) < η 3 v max, set c + = min[max[γ 2 n, c ], max ]; otherwise, set c + = c. Step 2: Unsuccessful iteration. If ρ f < η, set (x +, s + ) = (x, s ) and c + = c. If Y + Y, set f + = γ d and ν f = ν f + if ν f ν max f If Y + = Y, set f + = γ d. or f + = f otherwise. The operations related to c-iterations follow below. Algorithm 3.5: c-iteration(x, s, x +, s+, f, c ) Step : Successful iteration. If (2.32) holds, set (x +, s + ) = (x +, s+ ), f + = f and ν c =. If ρ c η 2, set c + = min[max[γ 2 n, c ], max ]; otherwise, set c + = c. Step 2: Unsuccessful iteration. If (2.32) fails, set (x +, s + ) = (x, s ) and f + = f. If Y + Y and ν c νc max, set c + = γ n if n =, or c + = γ c otherwise ( n = ). If ν c > νc max, set c + = c. If ν c νc max, update ν c = ν c +. If Y + = Y, set c + = γ n if n, or c + = γ c otherwise ( n = ). 4 The algorithm We now provide a formal description of our complete algorithm for solving nonlinear optimization problems with general nonlinear constraints without derivatives. Algorithm 4.: DEFT-FUNNEL(S, X, Y, x, s, f, c, v max ) Step : Initialization. Choose an initial vector of Lagrange multipliers µ and parameters ɛ >, ɛ >, f >, c >, α (, ), < γ < < γ 2, ζ, < η < η 2 <, Λ >, β >, η 3 >. Define = min[ f, c ] max. Initialize Y, with x Y B(x, ) and

16 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 5 Y n +, as well as the maximum number of interpolation points p max Y. Compute the associated models m f and mc around x and Lagrange polynomials {l,j } p j=. Define vmax = max[κ ca, κ cr v(x, s )], where κ ca > and κ cr >. Compute r by solving (2.9) with normal step n = and ine π f as in (2.8). Define νmax f > and νc max > and set ν f = ν c =. Set = and i =. Step : SubspaceMinimization(X, Y, x, s, f, c, vmax ). Step 2: CriticalityStep(Y, π f, ɛ i). Step 3: NormalStep(x, s, π v, πf, v, v max ). Step 4: TangentStep(x, s, n ). Step 5: Conclude a µ-iteration. If n = t =, then Step 5.: set (x +, s + ) = (x, s ), f + = f and c + = c ; Step 5.2: set + = min[ f +, c + ], vmax + = vmax and Y + = Y. Step 6: Conclude an f-iteration. If t and (2.25) and (2.29) hold, Step 6.: UpdateInterpolationSet(Y,x, x +,, ɛ i, ρ f η ); Step 6.2: f-iteration(x, s, x +, s+, f, c ); Step 6.3: Set + = min[ f +, c + ] and vmax + = vmax. Step 7: Conclude a c-iteration. If either n and t =, or either one of (2.25) or (2.29) fails, Step 7.: UpdateInterpolationSet(Y,x, x +,, ɛ i, (2.32) ); Step 7.2: c-iteration(x, s, x +, s+, f, c ); Step 7.3: Set + = min[ f +, c + ] and update vmax using (2.34). Step 8: Update the models and the Lagrange polynomials. If Y + Y, compute the interpolation models m f + and mc + around x + using Y + and the associated Lagrange polynomials {l +,j } p j=. Increment by one and go to Step. 5 Numerical experiments We implemented DEFT-FUNNEL in Matlab and tested it first on a set of 8 small-scale constrained problems from the CUTEst [9] collection. The problems contain at least one equality or inequality constraint, many of them containing both types and some containing simple bounds as well. 63 of the problems in our experiments are in the range of the first 3 test examples from the Hoc-Schittowsi collection [3], while the remaining are nonlinearly constrained optimization problems found in [3]. For the calculation of the normal step and the approximate Lagrange multipliers, we used a Matlab code named BLLS developed in collaboration with Ane Tröltzsch for solving bound-constrained linear least-squares problems. This method is intended for small-dimensional problems and is an active-set algorithm where the unconstrained problem is solved at each iteration in the subspace ined by the currently active bounds, which are determined by a projected Cauchy step. As for the computation of

17 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 6 the tangent step, we used a nonmonotone spectral projected gradient method [2] to solve the (possibly) ininite quadratic subproblems (2.7). Finally, we ine the bounding functions ω n (t) and ω t (t) as ω n (t) =. min[, t] and ω t (t) =. min[, t 2 ]. (5.43) We compare the results of the four variants of our method with regard to the way of building the interpolating models subbasis selection, mininum Frobenius norm, minimum l 2 -norm and regression to those obtained with the popular Fortran code COBYLA [9], a trust-region method for constrained problems that models the objective and constraint functions by linear interpolation. The only criterion for comparison is the number of calls on the routine that evaluates the objective function and the constraints at the same time at the required points, which is motivated by the fact that these costs very often dominate those of all other algorithmic computations. Therefore, we do not count evaluations of the objective function and the constraints separately, but rather the number of calls to the single routine that calculates both simultaneously. Comparisons of the computational results obtained by DEFT-FUNNEL for equality-constrained problems only also have been made with another method named CDFO, a derivative-free filter-sqp algorithm proposed in [4]. The results are reported in [22] and show that DEFT-FUNNEL was superior to CDFO in our tests. The threshold ɛ for declaring convergence in the criticality step in DEFT-FUNNEL is set to ɛ = 4. The stopping criterion used in COBYLA is based on the trust region radius ρ from the interval [ρ end, ρ beg ], where ρ beg and ρ end are constants preined by the user. The parameter ρ is decreased by a constant factor during the execution of the algorithm and is never increased. The algorithm stops when ρ = ρ end. Therefore, ρ end should have the magnitude of the required accuracy in the final values of the variables. In our experiments, we set ρ end = 4. In DEFT-FUNNEL, we fixed the trust-region parameters to =, η =., η 2 =.9, η 3 =.5, γ =.5, γ 2 = 2.5 and max =. The parameter ζ used in the inition of the sets F and C of far points and close points, respectively, is set to ζ =. For the limit number of times to reduce the trust regions sizes when a far or close interpolation point is replaced at unsuccessful iterations, we choose νf max =. We set p max = (n + )(n + 2)/2 for the subbasis, minimum Frobenius norm and minimum l 2 -norm approaches, and p max = (n + )(n + 2) for the regression case. Finally, we set ɛ =. as the initial value for the loop in the criticality step, α =. and β =. The performance of the methods are first compared by means of their performance profiles [8]. = ν max c Denote the set of solvers by S, and the set of problems by P. We compare the performance on problem p P by solver s S with the best performance by any solver on this problem; that is, we use the performance ratio r p,s = t p,s min{t p,s : s S}, where t p,s is the number of function evaluations required to solve problem p by solver s. If we ine ρ s () = P {p P : r p,s }, then ρ s () is an approximation to the probability for solver s S that a performance ratio r p,s is within a factor IR of the best possible ratio. In Figure 5., we provide the performance profiles of the four variants of DEFT-FUNNEL, whereas, in Figure 5.2, we compare the performance of each variant of our method to COBYLA s individually. As it can be seen, DEFT-FUNNEL has shown superior results on the set of test problems when the interpolating models are built from subbasis selection, minimum Frobenius norm and minimum l 2 -norm approaches. For the regression variant, Figure 5.2d reveals that COBYLA was faster than DEFT-FUNNEL in most of the problems, although the latter was able to solve more problems than the former for large values of.

18 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 7 Log 2 Scaled Performance Profile on Subset of CUTEst ρs() Subbasis Min. Frobenius norm Min. l norm. 2 Regression Figure 5.: Log 2 -scaled performance profiles of DEFT-FUNNEL with different approaches to build the models on a set of 8 problems from CUTEst. We also compare the methods by using data profiles. Such profiles were proposed by Moré and Wild [8] as a means to provide the user an accurate view of the performance of derivative-free solvers when the computational budget is limited. Let t p,s be the number of function evaluations required to solve problem p by solver s as before and denote by n p the dimension of the vector x in the problem p. The data profile of solver s S is ined by d s () = P { p P : } t p,s n p +. Because of the scaling choice of dividing by n p +, we may interpret d s () as the percentage of problems that can be solved with simplex gradient estimates. In Figure 5.3, the data profiles show that the subbasis, minimum Frobenius norm and minimum l 2 -norm variants have similar performance to COBYLA when the computational budget is limited to a value between and 2. In particular, when only 2 simplex gradient estimates are allowed, they solve roughly 8% of the problems. However, as the budget limit increases, the difference in the performance between these three variants and COBYLA becomes progressively larger, revealing the superiority of DEFT-FUNNEL. For the regression variant, it is difficult to tell which method wins from Figure 5.3d. For a very small budget (up to 7 simplex gradient estimates), both solve approximately the same amount of problems. For a budget of simplex gradients where κ (7, 35], COBYLA solves more problems than DEFT-FUNNEL, while for larger values of κ DEFT-FUNNEL solves more problems. We also tested DEFT-FUNNEL on a set of 2 small-scale linearly constrained optimization problems from CUTEst and compared its results with those of the software LINCOA, a newly developed trust-region method by Powell [2] for linearly constrained optimization without derivatives. His software combines active-set methods with truncated conjugate gradients and uses quadratic interpolation models for the objective function. The user is required to provide the Jacobian matrix and the vector of constants in the right side of the constraints as inputs, as well as a feasible starting point. Since many of CUTEst problems provide an infeasible starting point, we chose a feasible one instead for the sae of comparison. The stopping criterion used in LINCOA is the same found in COBYLA, i.e. the algorithm stops when the trust region radius ρ [ρ end, ρ beg ] reaches the smallest value allowed ρ end.

19 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 8 Log 2 Scaled Performance Profile on Subset of CUTEst Log 2 Scaled Performance Profile on Subset of CUTEst ρs().5 ρs() DEFT FUNNEL/Subbasis COBYLA DEFT FUNNEL/Min. Frobenius norm COBYLA (a) Subbasis (b) Min. Frobenius norm Log 2 Scaled Performance Profile on Subset of CUTEst Log 2 Scaled Performance Profile on Subset of CUTEst ρs().5 ρs() DEFT FUNNEL/Min. l 2 norm COBYLA DEFT FUNNEL/Regression COBYLA (c) Min. l 2-norm (d) Regression Figure 5.2: Log 2 -scaled performance profiles of the methods DEFT-FUNNEL (with different approaches to build the models) and COBYLA on a set of 8 problems from CUTEst. In our experiments, we set ρ end = 4. Regarding the parameter setting in DEFT-FUNNEL, the same values were ept for comparison with LINCOA. In Figure 5.4, we provide the performance profiles of the four variants of DEFT-FUNNEL for the 2 linearly constrained problems, whereas, in Figure 5.5, we compare the performance of each variant of our method to LINCOA s individually. All the variants of DEFT-FUNNEL except the minimum l 2 -norm approach have solved all the 2 linearly constrained problems. The results reported in Figures 5.5a and 5.5b reveal a superior performance of LINCOA over DEFT-FUNNEL/Subbasis and DEFT-FUNNEL/Frobenius. On the other hand, Figure 5.5c shows that DEFT-FUNNEL/Min. l 2 -norm was faster than LINCOA and, in Figure 5.5d, it can be seen that DEFT-FUNNEL/Regression also was slightly faster, although the latter presented superior performance for large values of, which indicates more robustness. Data profiles of DEFT-FUNNEL and LINCOA are presented in Figure 5.6 and show that, for very small budgets (less than simplex gradient estimates), DEFT-FUNNEL was able to solve more

20 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 9 Data Profile on Subset of CUTEst Data Profile on Subset of CUTEst ds().5 ds() DEFT FUNNEL/Subbasis COBYLA DEFT FUNNEL/Min. Frobenius norm COBYLA (a) Subbasis (b) Min. Frobenius norm Data Profile on Subset of CUTEst Data Profile on Subset of CUTEst ds().5 ds() DEFT FUNNEL/Min. l. 2 norm COBYLA DEFT FUNNEL/Regression COBYLA (c) Min. l 2-norm (d) Regression Figure 5.3: Data profiles of the methods DEFT-FUNNEL (with different approaches to build the models) and COBYLA on a set of 8 problems from CUTEst. problems than LINCOA, while the latter performed better than the former for larger budgets. We conclude this section by noticing that the performance profiles in the Figures 5.2 and 5.5 and the data profiles in the Figures 5.3 and 5.6 give clear indication that DEFT-FUNNEL provides encouraging results for small-scale nonlinear optimization problems with general nonlinear constraints. For the case where the user nows that the constraint functions are linear and he is able to provide their gradients, it might be interesting to the handle those constraints separately rather than lumping them and other general nonlinear constraints together in c(x). The number of function evaluations required by each method to converge in our numerical experiments are reported in the appendix.

21 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 2 Log 2 Scaled Performance Profile on Subset of CUTEst ρs() Subbasis Min. Frobenius norm Min. l. 2 norm Regression Figure 5.4: Log 2 -scaled performance profiles of DEFT-FUNNEL with different approaches to build the models on a set of 2 linearly constrained problems from CUTEst.

22 Sampaio, Toint: A derivative-free trust-funnel method for constrained nonlinear optimization 2 Log 2 Scaled Performance Profile on Subset of CUTEst Log 2 Scaled Performance Profile on Subset of CUTEst ρs().5 ρs() DEFT FUNNEL/Subbasis LINCOA.5.5. DEFT FUNNEL/Min. Frobenius norm LINCOA (a) Subbasis (b) Min. Frobenius norm Log 2 Scaled Performance Profile on Subset of CUTEst Log 2 Scaled Performance Profile on Subset of CUTEst ρs().5 ρs() DEFT FUNNEL/Min. l 2 norm LINCOA DEFT FUNNEL/Regression LINCOA (c) Min. l 2-norm (d) Regression Figure 5.5: Log 2 -scaled performance profiles of the methods DEFT-FUNNEL (with different approaches to build the models) and LINCOA a set of 2 linearly constrained problems from CUTEst.

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Recent Developments in Model-based Derivative-free Optimization

Recent Developments in Model-based Derivative-free Optimization Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:

More information

A progressive barrier derivative-free trust-region algorithm for constrained optimization

A progressive barrier derivative-free trust-region algorithm for constrained optimization A progressive barrier derivative-free trust-region algorithm for constrained optimization Charles Audet Andrew R. Conn Sébastien Le Digabel Mathilde Peyrega June 28, 2016 Abstract: We study derivative-free

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

A derivative-free trust-region augmented Lagrangian algorithm

A derivative-free trust-region augmented Lagrangian algorithm A derivative-free trust-region augmented Lagrangian algorithm Charles Audet Sébastien Le Digabel Mathilde Peyrega July 5, 2016 Abstract We present a new derivative-free trust-region (DFTR) algorithm to

More information

Mathematical Programming and Research Methods (Part II)

Mathematical Programming and Research Methods (Part II) Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types

More information

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction

More information

Computational Methods. Constrained Optimization

Computational Methods. Constrained Optimization Computational Methods Constrained Optimization Manfred Huber 2010 1 Constrained Optimization Unconstrained Optimization finds a minimum of a function under the assumption that the parameters can take on

More information

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained

More information

Characterizing Improving Directions Unconstrained Optimization

Characterizing Improving Directions Unconstrained Optimization Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not

More information

MATH3016: OPTIMIZATION

MATH3016: OPTIMIZATION MATH3016: OPTIMIZATION Lecturer: Dr Huifu Xu School of Mathematics University of Southampton Highfield SO17 1BJ Southampton Email: h.xu@soton.ac.uk 1 Introduction What is optimization? Optimization is

More information

Handling of constraints

Handling of constraints Handling of constraints MTH6418 S. Le Digabel, École Polytechnique de Montréal Fall 2015 (v3) MTH6418: Constraints 1/41 Plan Taxonomy of constraints Approaches The Progressive Barrier (PB) References MTH6418:

More information

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming Gianni Di Pillo (dipillo@dis.uniroma1.it) Giampaolo Liuzzi (liuzzi@iasi.cnr.it) Stefano Lucidi (lucidi@dis.uniroma1.it)

More information

A penalty based filters method in direct search optimization

A penalty based filters method in direct search optimization A penalty based filters method in direct search optimization ALDINA CORREIA CIICESI/ESTG P.PORTO Felgueiras PORTUGAL aic@estg.ipp.pt JOÃO MATIAS CM-UTAD Vila Real PORTUGAL j matias@utad.pt PEDRO MESTRE

More information

USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS

USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS A. L. CUSTÓDIO AND L. N. VICENTE Abstract. Pattern search methods can be made more efficient if past function evaluations are appropriately

More information

Convexity Theory and Gradient Methods

Convexity Theory and Gradient Methods Convexity Theory and Gradient Methods Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Convex Functions Optimality

More information

Chapter 3 Numerical Methods

Chapter 3 Numerical Methods Chapter 3 Numerical Methods Part 1 3.1 Linearization and Optimization of Functions of Vectors 1 Problem Notation 2 Outline 3.1.1 Linearization 3.1.2 Optimization of Objective Functions 3.1.3 Constrained

More information

Tutorial on Convex Optimization for Engineers

Tutorial on Convex Optimization for Engineers Tutorial on Convex Optimization for Engineers M.Sc. Jens Steinwandt Communications Research Laboratory Ilmenau University of Technology PO Box 100565 D-98684 Ilmenau, Germany jens.steinwandt@tu-ilmenau.de

More information

A penalty based filters method in direct search optimization

A penalty based filters method in direct search optimization A penalty based filters method in direct search optimization Aldina Correia CIICESI / ESTG P.PORTO Felgueiras, Portugal aic@estg.ipp.pt João Matias CM-UTAD UTAD Vila Real, Portugal j matias@utad.pt Pedro

More information

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year CONLIN & MMA solvers Pierre DUYSINX LTAS Automotive Engineering Academic year 2018-2019 1 CONLIN METHOD 2 LAY-OUT CONLIN SUBPROBLEMS DUAL METHOD APPROACH FOR CONLIN SUBPROBLEMS SEQUENTIAL QUADRATIC PROGRAMMING

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case Ellipsoid Algorithm 15-853:Algorithms in the Real World Linear and Integer Programming II Ellipsoid algorithm Interior point methods First polynomial-time algorithm for linear programming (Khachian 79)

More information

Using Subspace Constraints to Improve Feature Tracking Presented by Bryan Poling. Based on work by Bryan Poling, Gilad Lerman, and Arthur Szlam

Using Subspace Constraints to Improve Feature Tracking Presented by Bryan Poling. Based on work by Bryan Poling, Gilad Lerman, and Arthur Szlam Presented by Based on work by, Gilad Lerman, and Arthur Szlam What is Tracking? Broad Definition Tracking, or Object tracking, is a general term for following some thing through multiple frames of a video

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

3 No-Wait Job Shops with Variable Processing Times

3 No-Wait Job Shops with Variable Processing Times 3 No-Wait Job Shops with Variable Processing Times In this chapter we assume that, on top of the classical no-wait job shop setting, we are given a set of processing times for each operation. We may select

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm

Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm Numerical Experiments with a Population Shrinking Strategy within a Electromagnetism-like Algorithm Ana Maria A. C. Rocha and Edite M. G. P. Fernandes Abstract This paper extends our previous work done

More information

5 Machine Learning Abstractions and Numerical Optimization

5 Machine Learning Abstractions and Numerical Optimization Machine Learning Abstractions and Numerical Optimization 25 5 Machine Learning Abstractions and Numerical Optimization ML ABSTRACTIONS [some meta comments on machine learning] [When you write a large computer

More information

SID-PSM: A PATTERN SEARCH METHOD GUIDED BY SIMPLEX DERIVATIVES FOR USE IN DERIVATIVE-FREE OPTIMIZATION

SID-PSM: A PATTERN SEARCH METHOD GUIDED BY SIMPLEX DERIVATIVES FOR USE IN DERIVATIVE-FREE OPTIMIZATION SID-PSM: A PATTERN SEARCH METHOD GUIDED BY SIMPLEX DERIVATIVES FOR USE IN DERIVATIVE-FREE OPTIMIZATION A. L. CUSTÓDIO AND L. N. VICENTE Abstract. SID-PSM (version 1.3) is a suite of MATLAB [1] functions

More information

NOTATION AND TERMINOLOGY

NOTATION AND TERMINOLOGY 15.053x, Optimization Methods in Business Analytics Fall, 2016 October 4, 2016 A glossary of notation and terms used in 15.053x Weeks 1, 2, 3, 4 and 5. (The most recent week's terms are in blue). NOTATION

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i

More information

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize. Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject

More information

A derivative-free trust-region algorithm for reliability-based optimization

A derivative-free trust-region algorithm for reliability-based optimization Struct Multidisc Optim DOI 10.1007/s00158-016-1587-y BRIEF NOTE A derivative-free trust-region algorithm for reliability-based optimization Tian Gao 1 Jinglai Li 2 Received: 3 June 2016 / Revised: 4 September

More information

Numerical Method in Optimization as a Multi-stage Decision Control System

Numerical Method in Optimization as a Multi-stage Decision Control System Numerical Method in Optimization as a Multi-stage Decision Control System B.S. GOH Institute of Mathematical Sciences University of Malaya 50603 Kuala Lumpur MLYSI gohoptimum@gmail.com bstract: - Numerical

More information

Introduction to Modern Control Systems

Introduction to Modern Control Systems Introduction to Modern Control Systems Convex Optimization, Duality and Linear Matrix Inequalities Kostas Margellos University of Oxford AIMS CDT 2016-17 Introduction to Modern Control Systems November

More information

Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem

Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem CENTER FOR MACHINE PERCEPTION CZECH TECHNICAL UNIVERSITY Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem Woring document of the EU project COSPAL IST-004176 Vojtěch Franc, Miro

More information

Some Advanced Topics in Linear Programming

Some Advanced Topics in Linear Programming Some Advanced Topics in Linear Programming Matthew J. Saltzman July 2, 995 Connections with Algebra and Geometry In this section, we will explore how some of the ideas in linear programming, duality theory,

More information

Constrained and Unconstrained Optimization

Constrained and Unconstrained Optimization Constrained and Unconstrained Optimization Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt2@illinois.edu Oct 10th, 2017 C. Hurtado (UIUC - Economics) Numerical

More information

CMU-Q Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization. Teacher: Gianni A. Di Caro

CMU-Q Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization. Teacher: Gianni A. Di Caro CMU-Q 15-381 Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization Teacher: Gianni A. Di Caro GLOBAL FUNCTION OPTIMIZATION Find the global maximum of the function f x (and

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,

More information

Lecture 2: August 29, 2018

Lecture 2: August 29, 2018 10-725/36-725: Convex Optimization Fall 2018 Lecturer: Ryan Tibshirani Lecture 2: August 29, 2018 Scribes: Adam Harley Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have

More information

3 Interior Point Method

3 Interior Point Method 3 Interior Point Method Linear programming (LP) is one of the most useful mathematical techniques. Recent advances in computer technology and algorithms have improved computational speed by several orders

More information

Lecture 2 Optimization with equality constraints

Lecture 2 Optimization with equality constraints Lecture 2 Optimization with equality constraints Constrained optimization The idea of constrained optimisation is that the choice of one variable often affects the amount of another variable that can be

More information

Lecture 3: Linear Classification

Lecture 3: Linear Classification Lecture 3: Linear Classification Roger Grosse 1 Introduction Last week, we saw an example of a learning task called regression. There, the goal was to predict a scalar-valued target from a set of features.

More information

Convex Optimization - Chapter 1-2. Xiangru Lian August 28, 2015

Convex Optimization - Chapter 1-2. Xiangru Lian August 28, 2015 Convex Optimization - Chapter 1-2 Xiangru Lian August 28, 2015 1 Mathematical optimization minimize f 0 (x) s.t. f j (x) 0, j=1,,m, (1) x S x. (x 1,,x n ). optimization variable. f 0. R n R. objective

More information

A trust-region method for derivative-free nonlinear constrained stochastic optimization

A trust-region method for derivative-free nonlinear constrained stochastic optimization arxiv:73.456v [math.oc] Mar 7 A trust-region method for derivative-free nonlinear constrained stochastic optimization F. Augustin and Y. M. Marzou Massachusetts Institute of Technology, 77 Massachusetts

More information

USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS (COMPLETE NUMERICAL RESULTS)

USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS (COMPLETE NUMERICAL RESULTS) USING SAMPLING AND SIMPLEX DERIVATIVES IN PATTERN SEARCH METHODS (COMPLETE NUMERICAL RESULTS) A. L. CUSTÓDIO AND L. N. VICENTE Abstract. In this paper, we introduce a number of ways of making pattern search

More information

LP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008

LP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008 LP-Modelling dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven January 30, 2008 1 Linear and Integer Programming After a brief check with the backgrounds of the participants it seems that the following

More information

Lecture 2 - Introduction to Polytopes

Lecture 2 - Introduction to Polytopes Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.

More information

Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization. Author: Martin Jaggi Presenter: Zhongxing Peng

Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization. Author: Martin Jaggi Presenter: Zhongxing Peng Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization Author: Martin Jaggi Presenter: Zhongxing Peng Outline 1. Theoretical Results 2. Applications Outline 1. Theoretical Results 2. Applications

More information

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report

More information

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection Paul Armand 1 Ngoc Nguyen Tran 2 Institut de Recherche XLIM Université de Limoges Journées annuelles

More information

Numerical Optimization: Introduction and gradient-based methods

Numerical Optimization: Introduction and gradient-based methods Numerical Optimization: Introduction and gradient-based methods Master 2 Recherche LRI Apprentissage Statistique et Optimisation Anne Auger Inria Saclay-Ile-de-France November 2011 http://tao.lri.fr/tiki-index.php?page=courses

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing

More information

Chapter 15 Introduction to Linear Programming

Chapter 15 Introduction to Linear Programming Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of

More information

USE OF QUADRATIC MODELS WITH MESH ADAPTIVE DIRECT SEARCH FOR CONSTRAINED BLACK BOX OPTIMIZATION. minf(x) (1.1)

USE OF QUADRATIC MODELS WITH MESH ADAPTIVE DIRECT SEARCH FOR CONSTRAINED BLACK BOX OPTIMIZATION. minf(x) (1.1) USE OF QUADRATIC MODELS WITH MESH ADAPTIVE DIRECT SEARCH FOR CONSTRAINED BLACK BOX OPTIMIZATION ANDREW R. CONN AND SÉBASTIEN LE DIGABEL Abstract: We consider derivative-free optimization, and in particular

More information

Lecture 4: Convexity

Lecture 4: Convexity 10-725: Convex Optimization Fall 2013 Lecture 4: Convexity Lecturer: Barnabás Póczos Scribes: Jessica Chemali, David Fouhey, Yuxiong Wang Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Mathematical and Algorithmic Foundations Linear Programming and Matchings Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 6

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 6 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 6 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 19, 2012 Andre Tkacenko

More information

Math 734 Aug 22, Differential Geometry Fall 2002, USC

Math 734 Aug 22, Differential Geometry Fall 2002, USC Math 734 Aug 22, 2002 1 Differential Geometry Fall 2002, USC Lecture Notes 1 1 Topological Manifolds The basic objects of study in this class are manifolds. Roughly speaking, these are objects which locally

More information

Parameterization of triangular meshes

Parameterization of triangular meshes Parameterization of triangular meshes Michael S. Floater November 10, 2009 Triangular meshes are often used to represent surfaces, at least initially, one reason being that meshes are relatively easy to

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

Programs. Introduction

Programs. Introduction 16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen

More information

POLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if

POLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if POLYHEDRAL GEOMETRY Mathematical Programming Niels Lauritzen 7.9.2007 Convex functions and sets Recall that a subset C R n is convex if {λx + (1 λ)y 0 λ 1} C for every x, y C and 0 λ 1. A function f :

More information

16.410/413 Principles of Autonomy and Decision Making

16.410/413 Principles of Autonomy and Decision Making 16.410/413 Principles of Autonomy and Decision Making Lecture 17: The Simplex Method Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology November 10, 2010 Frazzoli (MIT)

More information

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini DM545 Linear and Integer Programming Lecture 2 The Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 3. 4. Standard Form Basic Feasible Solutions

More information

Methodologies and Software for Derivative-free Optimization

Methodologies and Software for Derivative-free Optimization Methodologies and Software for Derivative-free Optimization A. L. Custódio 1 K. Scheinberg 2 L. N. Vicente 3 March 14, 2017 1 Department of Mathematics, FCT-UNL-CMA, Quinta da Torre, 2829-516 Caparica,

More information

LANCELOT naive, a simple interface for LANCELOT B by N. I. M. Gould 1, D. Orban 2 and Ph. L. Toint 3 Report 07/12 26th November 2007

LANCELOT naive, a simple interface for LANCELOT B by N. I. M. Gould 1, D. Orban 2 and Ph. L. Toint 3 Report 07/12 26th November 2007 LANCELOT naive, a simple interface for LANCELOT B by N. I. M. Gould 1, D. Orban 2 and Ph. L. Toint 3 Report 07/12 26th November 2007 1 Oxford University Computing Laboratory, Wolfson Building, Parks Road,

More information

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.

More information

Convex sets and convex functions

Convex sets and convex functions Convex sets and convex functions Convex optimization problems Convex sets and their examples Separating and supporting hyperplanes Projections on convex sets Convex functions, conjugate functions ECE 602,

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

CS 450 Numerical Analysis. Chapter 7: Interpolation

CS 450 Numerical Analysis. Chapter 7: Interpolation Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

CS321 Introduction To Numerical Methods

CS321 Introduction To Numerical Methods CS3 Introduction To Numerical Methods Fuhua (Frank) Cheng Department of Computer Science University of Kentucky Lexington KY 456-46 - - Table of Contents Errors and Number Representations 3 Error Types

More information

Convex sets and convex functions

Convex sets and convex functions Convex sets and convex functions Convex optimization problems Convex sets and their examples Separating and supporting hyperplanes Projections on convex sets Convex functions, conjugate functions ECE 602,

More information

A Short SVM (Support Vector Machine) Tutorial

A Short SVM (Support Vector Machine) Tutorial A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange

More information

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem Computational Learning Theory Fall Semester, 2012/13 Lecture 10: SVM Lecturer: Yishay Mansour Scribe: Gitit Kehat, Yogev Vaknin and Ezra Levin 1 10.1 Lecture Overview In this lecture we present in detail

More information

Introduction. Linear because it requires linear functions. Programming as synonymous of planning.

Introduction. Linear because it requires linear functions. Programming as synonymous of planning. LINEAR PROGRAMMING Introduction Development of linear programming was among the most important scientific advances of mid-20th cent. Most common type of applications: allocate limited resources to competing

More information

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics 1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!

More information

Lecture 15: Log Barrier Method

Lecture 15: Log Barrier Method 10-725/36-725: Convex Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 15: Log Barrier Method Scribes: Pradeep Dasigi, Mohammad Gowayyed Note: LaTeX template courtesy of UC Berkeley EECS dept.

More information

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material

More information

Machine Learning for Signal Processing Lecture 4: Optimization

Machine Learning for Signal Processing Lecture 4: Optimization Machine Learning for Signal Processing Lecture 4: Optimization 13 Sep 2015 Instructor: Bhiksha Raj (slides largely by Najim Dehak, JHU) 11-755/18-797 1 Index 1. The problem of optimization 2. Direct optimization

More information

A Brief History of Filter Methods

A Brief History of Filter Methods ARGONNE NATIONAL LABORATORY 9700 South Cass Avenue Argonne, Illinois 60439 A Brief History of Filter Methods Roger Fletcher, Sven Leyffer, and Philippe Toint Mathematics and Computer Science Division Preprint

More information

MA651 Topology. Lecture 4. Topological spaces 2

MA651 Topology. Lecture 4. Topological spaces 2 MA651 Topology. Lecture 4. Topological spaces 2 This text is based on the following books: Linear Algebra and Analysis by Marc Zamansky Topology by James Dugundgji Fundamental concepts of topology by Peter

More information

Solution Methods Numerical Algorithms

Solution Methods Numerical Algorithms Solution Methods Numerical Algorithms Evelien van der Hurk DTU Managment Engineering Class Exercises From Last Time 2 DTU Management Engineering 42111: Static and Dynamic Optimization (6) 09/10/2017 Class

More information

A FACTOR GRAPH APPROACH TO CONSTRAINED OPTIMIZATION. A Thesis Presented to The Academic Faculty. Ivan Dario Jimenez

A FACTOR GRAPH APPROACH TO CONSTRAINED OPTIMIZATION. A Thesis Presented to The Academic Faculty. Ivan Dario Jimenez A FACTOR GRAPH APPROACH TO CONSTRAINED OPTIMIZATION A Thesis Presented to The Academic Faculty By Ivan Dario Jimenez In Partial Fulfillment of the Requirements for the Degree B.S. in Computer Science with

More information

2. Convex sets. x 1. x 2. affine set: contains the line through any two distinct points in the set

2. Convex sets. x 1. x 2. affine set: contains the line through any two distinct points in the set 2. Convex sets Convex Optimization Boyd & Vandenberghe affine and convex sets some important examples operations that preserve convexity generalized inequalities separating and supporting hyperplanes dual

More information

Point-Set Topology 1. TOPOLOGICAL SPACES AND CONTINUOUS FUNCTIONS

Point-Set Topology 1. TOPOLOGICAL SPACES AND CONTINUOUS FUNCTIONS Point-Set Topology 1. TOPOLOGICAL SPACES AND CONTINUOUS FUNCTIONS Definition 1.1. Let X be a set and T a subset of the power set P(X) of X. Then T is a topology on X if and only if all of the following

More information

Algorithms for convex optimization

Algorithms for convex optimization Algorithms for convex optimization Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University kocvara@utia.cas.cz http://www.utia.cas.cz/kocvara

More information

Chapter 6. Curves and Surfaces. 6.1 Graphs as Surfaces

Chapter 6. Curves and Surfaces. 6.1 Graphs as Surfaces Chapter 6 Curves and Surfaces In Chapter 2 a plane is defined as the zero set of a linear function in R 3. It is expected a surface is the zero set of a differentiable function in R n. To motivate, graphs

More information

Convex Optimization. Convex Sets. ENSAE: Optimisation 1/24

Convex Optimization. Convex Sets. ENSAE: Optimisation 1/24 Convex Optimization Convex Sets ENSAE: Optimisation 1/24 Today affine and convex sets some important examples operations that preserve convexity generalized inequalities separating and supporting hyperplanes

More information

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy Proc. of the 9th WSEAS Int. Conf. on Mathematical and Computational Methods in Science and Engineering, Trinidad and Tobago, November -7, 7 A new Electromagnetism-like Algorithm with a Population Shrinking

More information

Lecture 2: August 31

Lecture 2: August 31 10-725/36-725: Convex Optimization Fall 2016 Lecture 2: August 31 Lecturer: Lecturer: Ryan Tibshirani Scribes: Scribes: Lidan Mu, Simon Du, Binxuan Huang 2.1 Review A convex optimization problem is of

More information

Lecture 2: August 29, 2018

Lecture 2: August 29, 2018 10-725/36-725: Convex Optimization Fall 2018 Lecturer: Ryan Tibshirani Lecture 2: August 29, 2018 Scribes: Yingjing Lu, Adam Harley, Ruosong Wang Note: LaTeX template courtesy of UC Berkeley EECS dept.

More information

Mixture Models and the EM Algorithm

Mixture Models and the EM Algorithm Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Finite Mixture Models Say we have a data set D = {x 1,..., x N } where x i is

More information

Convexity: an introduction

Convexity: an introduction Convexity: an introduction Geir Dahl CMA, Dept. of Mathematics and Dept. of Informatics University of Oslo 1 / 74 1. Introduction 1. Introduction what is convexity where does it arise main concepts and

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information