ADAPTATION OF THE UOBYQA ALGORITHM FOR NOISY FUNCTIONS

Size: px
Start display at page:

Download "ADAPTATION OF THE UOBYQA ALGORITHM FOR NOISY FUNCTIONS"

Transcription

1 Proceedings of the 2006 Winter Simulation Conference L. F. Perrone, F. P. Wieland, J. Liu, B. G. Lawson, D. M. Nicol, and R. M. Fujimoto, eds. ADAPTATION OF THE UOBYQA ALGORITHM FOR NOISY FUNCTIONS Geng Deng Department of Mathematics University of Wisconsin 480 Lincoln Dr. Madison, WI 53706, U.S.A. Michael C. Ferris Computer Sciences Department University of Wisconsin 1210 W. Dayton Street Madison, WI 53706, U.S.A. ABSTRACT In many real-world optimization problems, the objective function may come from a simulation evaluation so that it is (a) subject to various levels of noise, (b) not differentiable, and (c) computationally hard to evaluate. In this paper, we modify Powell s UOBYQA algorithm to handle those real-world simulation problems. Our modifications apply Bayesian techniques to guide appropriate sampling strategies to estimate the objective function. We aim to make the underlying UOBYQA algorithm proceed efficiently while simultaneously controlling the amount of computational effort. 1 INTRODUCTION Powell s UOBYQA (Unconstrained Optimization BY Quadratic Approximation) (Powell 2002) is a derivativefree algorithm designed for (small scale) unconstrained optimization. The general structure of UOBYQA follows a model-based approach, which constructs a series of local quadratic models that approximate the objective function. The algorithm generates its iterates using a trust region framework (Nocedal and Wright 1999). It differs from a classical trust region method in that it creates quadratic models by interpolating a set of sample points instead of using the gradient and Hessian values of the objective function (thus making it a derivative-free tool). Besides UOBYQA, other model-based software includes WEDGE (Marazzi and Nocedal 2002) and NEWUOA (Powell 2004). We develop a variant of the original UOBYQA, called noisy UOBYQA, that is adapted for noisy optimization problems. These are very common in real-world applications, in particular when the objective function is an associated measurement of an experimental simulation. In this situation, the objective value may be difficult to obtain and the inaccuracy of the objective function often complicates the optimization process. For example, the derivative value is typically unavailable, thus many standard algorithms are not applicable. We consider the stochastic problem in the following parametric form: min F(x) = E[ f(x,ω(x))]. (1) x Rn The underlying function F( ) is unknown and must be estimated. The function f(, ω(x)) is a sample response function, which is affected by a random factor ω(x) at x. We normally assume the noise term ω(x) has mean zero and is independently distributed. As a special case, we consider ω(x) as white noise in our paper and use the additive form: f(x,ω(x)) = F(x)+ω(x). When noise is present, UOBYQA may behave precariously. For example, a subproblem that minimizes the quadratic model within a trust region may generate poor solutions. The idea of our modification is to control the random error by averaging multiple evaluations per point, helping the algorithm to proceed appropriately. Therefore, we will face the primary issue of handling the tradeoff between two objectives: having an efficient algorithm versus minimizing computational effort. On the one hand, we sample more replications to increase the accuracy of estimation of the underlying function F and hence the accuracy of the algorithm; on the other hand, we want to use as little computation as possible. In our modifications, we apply Bayesian techniques to establish rules of evaluating appropriateness of performing a step in UOBYQA. Only when such rules are satisfied, do we proceed with this step of the algorithm. The noisy UOBYQA algorithm shares similarities to the Response Surface Methodology (RSM) (Box and Draper 1986). Both of the methods construct a series of models to the simulation response during the optimization process. However, many features implemented in UOBYQA, such as the quadratic model update, and trust region control, have advantages over the classical RSM approach /06/$ IEEE 312

2 The remainder of the paper is arranged as follows. In Section 2, we give a brief outline of the deterministic UOBYQA algorithm. In Section 3, we present the noisy UOBYQA algorithm, including the following modifications: stabilizing the quadratic models, comparing two candidate points and a new termination criterion. We also show how to optimally allocate computational resources in these procedures. In Section 4, we apply the new noisy UOBYQA algorithm to several numerical examples and compare it with other noisy algorithms. Finally, we design and solve a real simulation model. 2 THE UOBYQA ALGORITHM We first outline the structure of the standard UOBYQA in the noiseless case. In the following description, we denote the deterministic objective function as f( ). As we have mentioned, UOBYQA is essentially in the framework of a model-based approach. Starting the algorithm requires an initial trial point x 0 and an initial trust region radius 0. At each iteration k, the derivative estimate of the objective function f is contained in a quadratic model Q k (x k + s) = c Q + g T Qs+ 1 2 st G Q s, (2) which is constructed by interpolating a set of well-positioned points I = {y 1,y 2,...,y L }, Q k (y i ) = f(y i ), i = 1,2,...,L. The point x k acts as the center of the trust region, and is the best point in the set I. The interpolative model is expected to well approximate f around the base point x k, such that the parameters c Q,g Q and G Q approximate the Taylor expansion coefficients of f around x k. To ensure a unique quadratic interpolator, the number of interpolating points should satisfy L = 1 (n+1)(n+2). (3) 2 As in a classical trust region method, a new promising point is found from a subproblem: min Q k(x k + s), subject to s 2 k. (4) s R n The solution s of (4) is accepted (or not) by evaluating the degree of agreement between f and Q k : γ = f(x k) f(x k + s ) Q k (x k ) Q k (x k + s ). (5) If the ratio γ is large enough and the new point maintains the good geometry of the set I, then the point is accepted into I. Otherwise, we either improve the geometry of I or update the trust region radius. We always keep track of the best point in the set I. Whenever a new point x + enters, it is compared with f(x k ) to determine the best point, which becomes the next iterate x k+1 argmin{ f(x k ), f(x + )}. The following outline of the UOBYQA algorithm contains only the key points regarding the modifications; details can be found in (Powell 2002). The UOBYQA Algorithm A starting point x 0, an initial trust region radius 0, and a terminating trust region radius end are given. 1. Generate initial trial points in the interpolation set I. Let iterate x 1 be the best point in I. 2. For iterations k = 1,2,... (a) (b) (c) (d) (e) Construct a quadratic model Q k (x) of the form (2) that interpolates points in I. Solve the trust region subproblem (4). Evaluate the function at the new point x k + s and compute the agreement ratio γ in (5). Test whether the point is acceptable in the set I. If not, improve the quality I or update the trust radius k. If a point is added to the set I, another element in I should be removed. When a new point x + is added, determine the best point in I. If k end, terminate the loop. 3. Evaluate and return the final solution point. 3 MODIFICATIONS In the following subsections, we will present our modifications to the UOBYQA algorithm to deal with problematic points of the algorithm in the noisy case. 3.1 Reducing Quadratic Model Variance When there is uncertainty in the objective function, the existence of noise can cause erroneous estimations of coefficients of the quadratic model Q(x), say c Q,g Q,G Q, and as a result, generate an unstable solution x k +s. To reduce the variance of the quadratic model, we consider generating multiple function values for points y j, j = 1,2,...,L in the set I and use the averaged function value for the interpolation process. In UOBYQA, the quadratic function is constructed as a linear combination of Lagrange functions l j (x), Q k (x) = L f(y j )l j (x), x R n, (6) 313

3 where L is the cardinality of the set I, as defined in (3). Each l j (x) is a quadratic polynomial from R n to R l j (x k + s) = c j + g T j s+ 1 2 st G j s, j = 1,2,...,L, that has the property l j (y i ) = δ i j, i = 1,2,...,L, where δ i j is 1 if i = j and 0 otherwise. It follows from (2) and (6) that the parameters of Q k are derived as c Q = L f(y j )c j, g Q = L f(y j )g j, and G Q = L f(y j )G j. Note that the parameters c j, g j, and G j in each Lagrange function l j are uniquely determined when the y j are given, regardless of the objective function f( ). When we have r j function evaluations for the point y j, we can compute more accurate estimates of c Q,g Q, and G Q using mean values in (7) in place of f(y j ). We employ Bayesian tools to analytically quantify the distributions of the parameters in (7), which helps us determine the appropriate number of evaluations r j. In the Bayesian framework, the unknown mean µ(y j ) and variance σ 2 (y j ) of f(y j,ω) are considered as random variables, whose distributions are inferred by Bayes rule. By assuming a non-informative prior distribution, we can estimate the joint posterior distributions of µ(y j ) and 1/σ 2 (y j ) as 1 σ 2 (y j ) X Gamma((r j 1)/2, ˆσ 2 (y j )(r j 1)/2), µ ( y j ) σ 2 (y j ),X N( µ(y j ),σ 2 (y j )/r j ). Here we denote the sample mean and sample variance of the data by µ(y j ) and ˆσ 2 (y j ). The gamma distribution Gamma(α,β) has mean α/β and variance α/β 2. We use a non-informative prior since there is no prior information on the real distribution. The distribution of the mean value µ(y j ) is of most interest to us. When the sample size is large, we can replace the variance σ 2 (y j ) with the sample variance ˆσ 2 (y j ) in (8), and can asymptotically derive the posterior distribution of µ(y j ) X as (7) (8) µ(y j ) X N( µ(y j ), ˆσ 2 (y j )/r j ). (9) According to (7), we treat the c Q,g Q and G Q as random variables from a Bayesian perspective. They all follow normal distributions whose means are estimated using (9) as (all of these are posterior estimates) E[c Q ] [ L ] = E f(y j,ω)c j [ L ] = E µ(y j )c j = L µ(y j )c j, E[g Q ] E[G Q ] = L µ(y j )g j, = L µ(y j (10) )G j and the variances are estimated as ( L ) var(c Q ) = var µ(y j )c j = L c2 j ˆσ2 (y j )/r j, var(g Q (i )) = L g2 j (i )ˆσ 2 (y j )/r j, var(g Q (i, j )) = L G2 j (i, j )ˆσ 2 (y j )/r j, i, j = 1,...,n. (11) As r j increases to infinity, the variance decreases to zero. A more precise computation involving the use of the Student s t-distribution was found to perform similarly. To increase the stability of a quadratic model, we want to quantify how the randomness in the coefficients c Q,g Q, and G Q affects the solution of the subproblem (4). Suppose we solve N t trial subproblems whose quadratic model coefficients are extracted from their posterior distributions, we expect that solutions s (i),i = 1,2,...,N t have a small overall variance (see Figure 1). Therefore, we introduce a criterion that constrains the standard deviation of solutions in each coordinate direction, requiring them to be smaller than a threshold value β times the trust region radius: n max std([s (1) ( j),s (2) ( j),...,s (Nt) ( j)]) β k. (12) Increasing r j should help reduce the volatility of coefficients, thus reduce the standard deviations of solutions. Therefore, satisfying the above criterion necessitates a sufficiently large r j for the point y j. Figure 1: The Trial Solutions within a Trust Region are Projected to Each Coordinate Direction and the Standard Deviations of the Projected Values are Evaluated. Sequentially allocating computational resources: A new resource allocation question arises (based on r j function evaluations at site y j ) about which site to assign new function replications in order to satisfy the constraint (12) with the minimum total number of function evaluations. In solving the subproblem (4), we know that only g Q and G Q matter 314

4 for determining the solution s from the definition of Q k (x) in (2). Instead of satisfying the constraint directly, we aim to control the variance of the most volatile coefficient. We minimize the corresponding ratio of the standard deviation to the expected value ( ) std(g max Q (i )) i, j E[g Q (i )], std(g Q(i, j )), i E[G Q (i, j )], j = 1,...,n. (13) We propose a fast but sub-optimal strategy here, one that assigns additional computational resources sequentially. We assume that when an additional batch of r replications at site y j are newly produced, the sample mean µ(y j ) and variance ˆσ 2 (y j ) remain invariant. Under this assumption, the posterior variance of µ(y j ) (see (9)) changes from ˆσ 2 (y j )/r j ˆσ 2 (y j )/(r j + r). First, let φ( r) (here r = [r 1,r 2,...,r L ]) denote the largest quantity in (13): φ( r) = max = max i, j ( std(gq (i )) i, j E[g Q (i )], std(g Q(i, j )) E[G Q (i, j )] s s LP LP g 2 j (i )ˆσ 2 (y j )/r j G 2 j (i, j )ˆσ 2 (y j )/r j, LP LP µ(y j )g j (i ) µ(y j )G j (i, j ) ), i, j = 1,...,n. (14) To achieve a good allocation scheme, we invest our new resources in the point y j in order to most sharply decrease the quantity φ( r). After assigning r samples to the point y j, we obtain a new vector r+re j, where e j is the standard unit vector in R L with 1 on the jth component. We select the site y j with the index: arg max j φ( r+ re j ). (15) The best index is determined by comparing the L different possible options. Procedure to stabilize the quadratic model: Given initial sample size r 0, batch size r, and the threshold value β. 1. Generate r 0 function evaluations for each point for pre-estimations of sample mean and sample variance. Set r j r 0, j = 1,2,...,L. 2. Determine the largest quantity φ( r) which corresponds to the most volatile coefficient. 3. Select the site for further evaluations using (15). 4. Evaluate r function replications on the selected site and update the sample mean and sample variance. 5. Repeat Steps 2-4, until constraint (12) is satisfied. 3.2 Selecting the Best Point If x + is a new point entering the interpolation set I, we encounter the problem of selecting the best point in I. Since x k is known to be the best in the previous set, the question becomes how we determine the order of x k and x +. When there is noise in the function output, we need more precise estimations of the underlying function value to make the correct decision. Without loss of generality, the problem is equivalent to solving a discrete-event optimization problem: arg min j {1,2} E[ f(y j,ω)]. (16) An extension of the problem is to select the best point among K (K 2) points. This general problem is often regarded selecting the best system, which is considered one of the principal problems in discrete-event simulation studies. Kim and Nelson (2003) provide a good survey on IZ-R&S procedures. In our case, since there are only two points involved, we implement a simplified variant of a ranking and selection procedure OCBA (Chen, Chen, and Yucesan 1999). Suppose we have replicated r j function values for y j, let µ(y i ) and σ 2 (y j ) be the unknown mean and variance of the output at y j. The point evidenced by a smaller sample mean is selected: { 1, if µ(y Choose point 1 ) µ(y 2 ); 2, otherwise. We will quantify our decision-making by the so-called probability of correct selection (PCS): PCS = Pr(select y (1) µ(y (1) ) µ(y (2) )). (17) Here the notation µ(y (1) ) µ(y (2) ) reveals the underlying order of the means. In OCBA, Bayesian methods are applied to estimate the posterior PCS. The idea is to construct a posterior distribution for the underlying mean µ(y j ) using observed data and estimate the PCS via joint distributions. We increase the number of samples until the desired accuracy is achieved PCS 1 α, (18) where α represents a significance level. We perform a similar analysis using Bayesian inference as in Section 3.1. Without loss of generality, suppose we observe µ(y 1 ) µ(y 2 ) and select the first point. The PCS 315

5 corresponds to the probability of the event {µ(y 1 ) µ(y 2 )}, PCS = Pr(µ(y 1 ) µ(y 2 )) Pr(µ(y 1 ) µ(y 2 ) X) = Pr(µ(y 1 ) X µ(y 2 ) X 0). (19) The approximate PSC in (19) is a tail probability of a normal distribution: PCS (20) ( ( Pr N µ(y 1 ) µ(y 2 ), ˆσ2 (y 1 ) + ˆσ2 (y 2 ) ) ) 0. r 1 r 2 Sequentially allocating computational resources We consider a sequential allocation strategy to assign new computational resources to the points y 1 and y 2, so that we can use the least number of function evaluations to satisfy the rule (18). The change of sample mean and sample variance follows the assumptions in Section (3.1). To determine the potentially better point, we compare the derivative value: max j {1,2} r j Pr ( ( ) ) N µ(y 1 ) µ(y 2 ), ˆσ2 (y 1 ) r 1 + ˆσ2 (y 2 ) r 2 0. (21) It is not hard to find that, since the mean of the joint distribution µ(y 1 ) µ(y 2 ) is unchanged, we desire the largest decrease in the variance ψ( r) := ˆσ 2 (y 1 )/r 1 + ˆσ 2 (y 2 )/r 2. Therefore, the problem (21) becomes min ψ( r+ re j). (22) j {1,2} This determines the index of the potentially better point. Procedure to select the best point in I: Given initial sample size r 0, batch size r, and a significance parameter α. 1. Evaluate r 0 function evaluations for each point for pre-estimations of sample mean and sample variance. Set r j r 0, j = 1,2. 2. Select the point to carry out further replications using (22). 3. Evaluate r further function replications on the point selected and update the sample mean and sample variance. 4. Repeat Steps 2 and 3 until the constraint (18) is satisfied. 3.3 New Termination Criterion In UOBYQA and other model-based optimization algorithms, a test on the norm of the gradient g k or the trust region size k is typically treated as termination criteria; i.e., k end = However, these criteria are not suitable for noisy cases. When k gets smaller, the estimated value of G Q, which is a gauge for curvature of the quadratic model, will approach zero. This will inevitably require much more function evaluations in later iterations to reduce the variance of the quadratic model in order to retain accuracy. The new termination criterion is designed to relax the value of end, such that the algorithm terminates much earlier. We specify a parameter N max that controls the maximum number of replications per site in the algorithm. N max represents the amount of computing we are willing to spend at any site. In each iteration k, we will check that any point x on the edge of the subregion {x x x b k } is separable from the center x k, given the control N max on the number of replications per site. Here separable means that one point is better than the other with high accuracy (PCS 1 α). The difficulty occurs in estimations of sample mean µ(x) and sample variance ˆσ 2 (x) of an edge point x, which demand additional function evaluations. We simplify this procedure by performing the separability test using Q k ( ) instead of the original F( ), because Q k is a good surrogate model of the underlying mean function F when k is small. We can (a) approximate µ(x) with Q k (x) and (b) approximate ˆσ 2 (x) with ˆσ 2 (x k ). The second approximation is valid because the variances of function output at two points are very close within a small trust region. In fact, we enumerate 2n edge points (of {x x x k < k }) which are in standard coordinate directions from x k, and check their separability from x k. The 2n points consist of a representative set of edge points. The stopping criterion is met when a portion of the representative point set, i.e., 80% of the 2n points, are separable from x k. In practice, what we first calculate is a least separable distance d. By observing the posterior distribution with the approximations: µ(x) µ(x k ) X N(F(x) F(x k ), ˆσ 2 (x)+ ˆσ 2 (x k )/N max ) N(Q k (x) Q k (x k ),2ˆσ 2 (x k )/N max ). We compute d, satisfying Pr(N(d,2ˆσ 2 (x k )/N max ) 0) 1 α, via an inverse cdf function evaluation. Then, we say x and x k are separable if and only if the difference Q k (x) Q k (x b ) d. 316

6 4 NUMERICAL RESULTS 4.1 Numerical Functions We tested the noisy UOBYQA algorithm on several numerical examples and compared it with two other noisy optimization tools, NOMADm (Nonlinear Optimization for Mixed variables and Derivatives in Matlab) (Abramson 2006) and SPSA (Simultaneous Perturbation Stochastic Approximation) (Spall 2003). NOMADm is a pattern search algorithm that implements ranking and selection; and SPSA is a line search algorithm that applies simultaneous perturbation for gradient estimation. The test function we employed was the well-known extended Rosenbrock function: n 1 F(x) = 100(x i+1 xi 2 ) 2 +(x i 1) 2. i=1 The function has a global optima at the point (1,1,...,1) n, at which the optimal objective value attains 0. A noisy optimization problem was formulated by adding a white noise term to the objective (Section 1). We also applied our algorithm to the ARGLINC function with very similar results. We assume the random term ω was independent of x and followed a normal distribution with mean zero and variance σ 2, indicating the levels of noise. The starting point x 0 was ( 1.2,1, 1.2,1,..., 1.2,1) n and the initial trust region radius 0 was set to 2. Table 1 presents the details about a single-run of noisy UOBYQA to the two-dimensional Rosenbrock function. The variance σ 2 was set to 0.01, which was moderately noisy from our tested noise levels. We used the following default setting of parameters: initial sampling number r 0 = 3, which is small but enough to generate sample mean and sample variance; the significance level α = 0.2, the threshold value β = 0.4, the number of trail solutions N t = 20, and the maximum number of evaluations per point N max = 60. The new termination criterion we introduced terminated the algorithm earlier, at iteration 81. In fact, we did not observe noticeable improvement in terms of objective value ( to ) between the iteration 80 and 120, because the presence of noise deteriorated the accuracy in these iterations. In order to keep the algorithm running and maintaining correct decisions, a large number of function evaluations were necessary. It was therefore preferable to stop the algorithm early. In the earlier iterations, the algorithm used relatively few function replications. This implied that when noisy effects were small compared to the large function values, the basic operation of the method was unchanged. In Table 3, we further compare numerical results of UOBYQA, NOMADm, and SPSA. We tested the Rosen- Table 1: The Performance of Noisy UOBYQA for the Rosenbrock Function, with n = 2 and σ 2 = Iteration (k) FN F(x k ) k Stops with the new termination criterion Stops with the termination criterion k 10 4 brock function of dimension 2 and 10. Increasing the dimension significantly increased computational burden. To solve a 10-dimensional problem, we need function evaluations in total. The different scales of noise variance σ 2 were specified to be 0.001, 0.01, 0.1, and 1. We constrained the algorithm with various total number of function evaluations. The parameter setting was the same as before, except that we used different N max when the total number of function evaluations or the variance changed. A practically useful formula for N max was N max = Max FN Iteration # (n) δ(σ2 ). Here the iteration # represented the estimated number of iterations that the algorithm should maintain for a given dimension. This is a rough estimate and the actual number of iterations varies in different problems. δ(σ 2 ) was an adjustment value associated with the variance σ 2. The algorithm used a relatively smaller N max when the variance was small. We provide the following suggested values in Table 2. Table 2: Suggested Values to Define N max n Iteration # σ δ(σ 2 ) As shown in Table 3, for the 2 dimensional case, our algorithms performed better than the other two algorithms. The mean value was the smallest among the peer algorithms. For the 10 dimensional case, when the variance σ 2 was large, our algorithm did not perform well. We should note a fact that making the variance 10 times larger, from our previous theoretical analysis, we potentially require 10 times more function evaluations in order to achieve the same scale of accuracy. That is why the quality of the solutions 317

7 decreased sharply as σ 2 increased. Another aspect that influences the performance may be the dimension of the problem. Powell has pointed out that in higher dimensional problems, UOBYQA may not be practically useful because the number of interpolation points L is huge. We think that noisy UOBYQA inherits the limitation in the same way. Table 3: Apply Noisy UOBYQA to the Rosenbrock Function (Results are Based on 10 Replications of the Algorithms). n 2 10 Noise level σ 2 Max FN Noisy NOMADm SPSA UOBYQA Mean Error Mean Error Mean Error Simulation Problems In this subsection, we applied the noisy UOBYQA algorithm to solve a pricing problem via simulation. The parameters we considered were the prices of M similar goods in a store, say, p 1, p 2,..., p M. When a customer arrives at the store, he is sequentially exposed to the M goods by the store keeper, from the most expensive good which is of the best quality to the cheapest good. He decides to purchase the item or not after viewing it, but he will buy at most one of them. For good i, we set the probability that the customer purchases the item as Prob i = exp( p i /η i ), i = 1,2,...,M. The objective of the store keeper is to determine the best combination of prices p = [p 1, p 2,..., p M ] which yields the largest expected profit. Running the simulation for a period of time, if we denote the total number of customers as m, and the number of customers who purchase good i as m i, we formulate the stochastic problem as: max(expected Profit), where Profit := p The precise solution is obtained from: M i=1 m i m p i. max(expected Profit) (23) p M i 1 := max (1 Prob j ) Prob i p i. p i=1 We used Arena 9.0 from Rockwell software to model the pricing system. We considered cases M = 2 and M = 10, where M also indicated the dimension of the variable p. In computation of the Prob i, the parameter η i can reflect the goodness of the merchandize. We set η 1 = 50 and η 2 = 20 in the two-dimensional case; set η 1 = 50,η 2 = 48,...,η 10 = 32 in the ten-dimensional case. The noisy UOBYQA algorithm and OptQuest (the optimization add-on for Arena,< were compared on the stochastic simulation problem. (See results in Table 4.) We implemented the same parametric setting of the algorithm as in the Section 4.1, except that we assigned the initial trust region as 10, and the upper bound for replication usage as 200 and 2000 for the two cases respectively. By varying the number of customers generated in each simulation run, we can have several levels of noisy profit value. The table shows the different levels of variances of the output. We compared with the real optimal solutions solved using the same η i s via (23). As we can see in the table, noisy UOBYQA did a uniformly better job than OptQuest, with a higher quality of solutions. For all the tested cases, the gap to the optimal solution was reduced around 50%. Table 4: Optimization Results of the Pricing Model (over Average Value of 10 Runs), Where the Real Solution for M = 2 is and for M = 10 is Model Estimated variance Max FN Noisy UOBYQA OptQuest M= M=

8 5 CONCLUSIONS In this paper we modify Powell s UOBYQA algorithm to solve noisy optimization problems. UOBYQA belongs to the class of model-based algorithms that do not require evaluations of the gradient of the objective, therefore, this algorithm is particularly applicable to real-world simulation problems. In the noisy setting, performing more function evaluations can reduce the randomness in function responses. Our approach is motivated by applying analytical Bayesian inference to determine the appropriate number of replications and provide statistical accuracy of the algorithm. At the same time, we have attempted to control the total computational effort. Note that our method does not assume Common Random Numbers (CRN) (Law and Kelton 2000) are implemented in the simulations. Therefore, we cannot generate a deterministic approximation function N 1 N i=1 f(,ω i) to F( ) by fixing a sequence of samples ω i,i = 1,2,...,N. The use of CRN is a topic for future research. The modifications can be generalized to other modelbased algorithms, such as the WEDGE algorithm. Our modifications are not intended to be applied to linear model based algorithms, since linear models are more sensitive to noise than quadratic models. In a stochastic algorithm, quadratic models are robust against noise and preferable to use. REFERENCES Abramson, M NOMADm: mash adaptive direct search Matlab toolbox. <http: // MAbramson/NOMADm.html>. Box, E. P. G., and N. R. Draper Empirical modelbuilding and response surfaces. New York: John Wiley & Sons. Chen, H.-C., C.-H. Chen, and E. Yucesan An asymptotic allocation for simultaneous simulation experiments. In Proceedings of the 1999 Winter Simulation Conference, Kim, S.-H., and B. L. Nelson Selecting the best system: theory and methods. In Proceedings of the 2003 Winter Simulation Conference, Law, A., and W. Kelton Simulation modeling and analysis. Third ed. New York: McGraw-Hill. Marazzi, M., and J. Nocedal Wedge trust region methods for derivative free optimization. Mathematical Programming 91: Nocedal, J., and S. J. Wright Numerical optimization. New York: Springer. Powell, M. J. D UOBYQA: Unconstrained optimization by quadratic approximation. Mathematical Programming 92: Powell, M. J. D The NEWUOA software for unconstrained optimization with derivatives. DAMTP Report 2004/NA05, University of Cambridge. Spall, J. C Introduction to stochastic search and optimization. New York: John Wiley & Sons. ACKNOWLEDGMENTS This research was partially supported by Air Force Office of Scientific Research Grant FA , and National Science Foundation Grants DMS and IIS AUTHOR BIOGRAPHIES GENG DENG is a Ph.D. student at the University of Wisconsin Madison. He is interested in simulation-based optimization, dynamic programming, and applied statistics. He is a member of INFORMS, SIAM, and AMS. He can be reached via <geng@cs.wisc.edu>. MICHAEL C. FERRIS is Professor of Computer Sciences and Industrial Engineering at the University of Wisconsin Madison. His research interests are in computational algorithms and modeling, coupled with applications in economics, structural and mechanical engineering and medicine. His work emphasizes practical solution of large scale optimization and complementarity problems. He is a member of the Mathematical Programming Society, SIAM, and INFORMS. His and web addresses are <ferris@cs.wisc.edu> and < ferris>. 319

A derivative-free trust-region algorithm for reliability-based optimization

A derivative-free trust-region algorithm for reliability-based optimization Struct Multidisc Optim DOI 10.1007/s00158-016-1587-y BRIEF NOTE A derivative-free trust-region algorithm for reliability-based optimization Tian Gao 1 Jinglai Li 2 Received: 3 June 2016 / Revised: 4 September

More information

MATH3016: OPTIMIZATION

MATH3016: OPTIMIZATION MATH3016: OPTIMIZATION Lecturer: Dr Huifu Xu School of Mathematics University of Southampton Highfield SO17 1BJ Southampton Email: h.xu@soton.ac.uk 1 Introduction What is optimization? Optimization is

More information

A HYBRID METHOD FOR SIMULATION FACTOR SCREENING. Hua Shen Hong Wan

A HYBRID METHOD FOR SIMULATION FACTOR SCREENING. Hua Shen Hong Wan Proceedings of the 2006 Winter Simulation Conference L. F. Perrone, F. P. Wieland, J. Liu, B. G. Lawson, D. M. Nicol, and R. M. Fujimoto, eds. A HYBRID METHOD FOR SIMULATION FACTOR SCREENING Hua Shen Hong

More information

Recent Developments in Model-based Derivative-free Optimization

Recent Developments in Model-based Derivative-free Optimization Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:

More information

Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization

Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization 10 th World Congress on Structural and Multidisciplinary Optimization May 19-24, 2013, Orlando, Florida, USA Inclusion of Aleatory and Epistemic Uncertainty in Design Optimization Sirisha Rangavajhala

More information

Simultaneous Perturbation Stochastic Approximation Algorithm Combined with Neural Network and Fuzzy Simulation

Simultaneous Perturbation Stochastic Approximation Algorithm Combined with Neural Network and Fuzzy Simulation .--- Simultaneous Perturbation Stochastic Approximation Algorithm Combined with Neural Networ and Fuzzy Simulation Abstract - - - - Keywords: Many optimization problems contain fuzzy information. Possibility

More information

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

Solving singularity issues in the estimation of econometric models

Solving singularity issues in the estimation of econometric models Solving singularity issues in the estimation of econometric models Michel Bierlaire Michaël Thémans Transport and Mobility Laboratory École Polytechnique Fédérale de Lausanne Solving singularity issues

More information

SIMULATION AND MONTE CARLO

SIMULATION AND MONTE CARLO JHU course no. 550.790, Modeling, Simulation, and Monte Carlo SIMULATION AND MONTE CARLO Some General Principles James C. Spall Johns Hopkins University Applied Physics Laboratory September 2004 Overview

More information

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.

More information

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used.

4.12 Generalization. In back-propagation learning, as many training examples as possible are typically used. 1 4.12 Generalization In back-propagation learning, as many training examples as possible are typically used. It is hoped that the network so designed generalizes well. A network generalizes well when

More information

Proceedings of the 2016 Winter Simulation Conference T. M. K. Roeder, P. I. Frazier, R. Szechtman, E. Zhou, T. Huschka, and S. E. Chick, eds.

Proceedings of the 2016 Winter Simulation Conference T. M. K. Roeder, P. I. Frazier, R. Szechtman, E. Zhou, T. Huschka, and S. E. Chick, eds. Proceedings of the 2016 Winter Simulation Conference T. M. K. Roeder, P. I. Frazier, R. Szechtman, E. Zhou, T. Huschka, and S. E. Chick, eds. A QUANTILE-BASED NESTED PARTITION ALGORITHM FOR BLACK-BOX FUNCTIONS

More information

A New Statistical Procedure for Validation of Simulation and Stochastic Models

A New Statistical Procedure for Validation of Simulation and Stochastic Models Syracuse University SURFACE Electrical Engineering and Computer Science L.C. Smith College of Engineering and Computer Science 11-18-2010 A New Statistical Procedure for Validation of Simulation and Stochastic

More information

SELECTION OF THE BEST WITH STOCHASTIC CONSTRAINTS. Engineering 3211 Providence Drive Black Engineering 3004

SELECTION OF THE BEST WITH STOCHASTIC CONSTRAINTS. Engineering 3211 Providence Drive Black Engineering 3004 Proceedings of the 2009 Winter Simulation Conference M. D. Rossetti, R. R. Hill, B. Johansson, A. Dunkin and R. G. Ingalls, eds. SELECTION OF THE BEST WITH STOCHASTIC CONSTRAINTS Alireza Kabirian Sigurdur

More information

Metaheuristic Optimization with Evolver, Genocop and OptQuest

Metaheuristic Optimization with Evolver, Genocop and OptQuest Metaheuristic Optimization with Evolver, Genocop and OptQuest MANUEL LAGUNA Graduate School of Business Administration University of Colorado, Boulder, CO 80309-0419 Manuel.Laguna@Colorado.EDU Last revision:

More information

A Deterministic Dynamic Programming Approach for Optimization Problem with Quadratic Objective Function and Linear Constraints

A Deterministic Dynamic Programming Approach for Optimization Problem with Quadratic Objective Function and Linear Constraints A Deterministic Dynamic Programming Approach for Optimization Problem with Quadratic Objective Function and Linear Constraints S. Kavitha, Nirmala P. Ratchagar International Science Index, Mathematical

More information

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information

Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum

Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum Bayesian Sequential Sampling Policies and Sufficient Conditions for Convergence to a Global Optimum Peter Frazier Warren Powell 2 Operations Research & Information Engineering, Cornell University 2 Operations

More information

DETC APPROXIMATE MOTION SYNTHESIS OF SPHERICAL KINEMATIC CHAINS

DETC APPROXIMATE MOTION SYNTHESIS OF SPHERICAL KINEMATIC CHAINS Proceedings of the ASME 2007 International Design Engineering Technical Conferences & Computers and Information in Engineering Conference IDETC/CIE 2007 September 4-7, 2007, Las Vegas, Nevada, USA DETC2007-34372

More information

Monte Carlo for Spatial Models

Monte Carlo for Spatial Models Monte Carlo for Spatial Models Murali Haran Department of Statistics Penn State University Penn State Computational Science Lectures April 2007 Spatial Models Lots of scientific questions involve analyzing

More information

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini

Metaheuristic Development Methodology. Fall 2009 Instructor: Dr. Masoud Yaghini Metaheuristic Development Methodology Fall 2009 Instructor: Dr. Masoud Yaghini Phases and Steps Phases and Steps Phase 1: Understanding Problem Step 1: State the Problem Step 2: Review of Existing Solution

More information

Digital Image Processing Laboratory: MAP Image Restoration

Digital Image Processing Laboratory: MAP Image Restoration Purdue University: Digital Image Processing Laboratories 1 Digital Image Processing Laboratory: MAP Image Restoration October, 015 1 Introduction This laboratory explores the use of maximum a posteriori

More information

Truss structural configuration optimization using the linear extended interior penalty function method

Truss structural configuration optimization using the linear extended interior penalty function method ANZIAM J. 46 (E) pp.c1311 C1326, 2006 C1311 Truss structural configuration optimization using the linear extended interior penalty function method Wahyu Kuntjoro Jamaluddin Mahmud (Received 25 October

More information

ME 555: Distributed Optimization

ME 555: Distributed Optimization ME 555: Distributed Optimization Duke University Spring 2015 1 Administrative Course: ME 555: Distributed Optimization (Spring 2015) Instructor: Time: Location: Office hours: Website: Soomin Lee (email:

More information

APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING

APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING Second Edition P. Venkataraman Rochester Institute of Technology WILEY JOHN WILEY & SONS, INC. CONTENTS PREFACE xiii 1 Introduction 1 1.1. Optimization Fundamentals

More information

Introduction to Stochastic Combinatorial Optimization

Introduction to Stochastic Combinatorial Optimization Introduction to Stochastic Combinatorial Optimization Stefanie Kosuch PostDok at TCSLab www.kosuch.eu/stefanie/ Guest Lecture at the CUGS PhD course Heuristic Algorithms for Combinatorial Optimization

More information

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling

Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Locally Weighted Least Squares Regression for Image Denoising, Reconstruction and Up-sampling Moritz Baecher May 15, 29 1 Introduction Edge-preserving smoothing and super-resolution are classic and important

More information

A Brief Look at Optimization

A Brief Look at Optimization A Brief Look at Optimization CSC 412/2506 Tutorial David Madras January 18, 2018 Slides adapted from last year s version Overview Introduction Classes of optimization problems Linear programming Steepest

More information

A noninformative Bayesian approach to small area estimation

A noninformative Bayesian approach to small area estimation A noninformative Bayesian approach to small area estimation Glen Meeden School of Statistics University of Minnesota Minneapolis, MN 55455 glen@stat.umn.edu September 2001 Revised May 2002 Research supported

More information

AM205: lecture 2. 1 These have been shifted to MD 323 for the rest of the semester.

AM205: lecture 2. 1 These have been shifted to MD 323 for the rest of the semester. AM205: lecture 2 Luna and Gary will hold a Python tutorial on Wednesday in 60 Oxford Street, Room 330 Assignment 1 will be posted this week Chris will hold office hours on Thursday (1:30pm 3:30pm, Pierce

More information

A simulation-based optimization framework for urban traffic control

A simulation-based optimization framework for urban traffic control A simulation-based optimization framework for urban traffic control Carolina Osorio Michel Bierlaire December 3, 2 Report TRANSP-OR 23 Transport and Mobility Laboratory School of Architecture, Civil and

More information

Simulation Calibration with Correlated Knowledge-Gradients

Simulation Calibration with Correlated Knowledge-Gradients Simulation Calibration with Correlated Knowledge-Gradients Peter Frazier Warren Powell Hugo Simão Operations Research & Information Engineering, Cornell University Operations Research & Financial Engineering,

More information

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Parallelization in Non-hierarchical Analytical Target Cascading Yongsu Jung Department of Mechanical Engineering,

More information

You ve already read basics of simulation now I will be taking up method of simulation, that is Random Number Generation

You ve already read basics of simulation now I will be taking up method of simulation, that is Random Number Generation Unit 5 SIMULATION THEORY Lesson 39 Learning objective: To learn random number generation. Methods of simulation. Monte Carlo method of simulation You ve already read basics of simulation now I will be

More information

Approximate Linear Programming for Average-Cost Dynamic Programming

Approximate Linear Programming for Average-Cost Dynamic Programming Approximate Linear Programming for Average-Cost Dynamic Programming Daniela Pucci de Farias IBM Almaden Research Center 65 Harry Road, San Jose, CA 51 pucci@mitedu Benjamin Van Roy Department of Management

More information

Quickest Search Over Multiple Sequences with Mixed Observations

Quickest Search Over Multiple Sequences with Mixed Observations Quicest Search Over Multiple Sequences with Mixed Observations Jun Geng Worcester Polytechnic Institute Email: geng@wpi.edu Weiyu Xu Univ. of Iowa Email: weiyu-xu@uiowa.edu Lifeng Lai Worcester Polytechnic

More information

A Deterministic Global Optimization Method for Variational Inference

A Deterministic Global Optimization Method for Variational Inference A Deterministic Global Optimization Method for Variational Inference Hachem Saddiki Mathematics and Statistics University of Massachusetts, Amherst saddiki@math.umass.edu Andrew C. Trapp Operations and

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Algebraic Iterative Methods for Computed Tomography

Algebraic Iterative Methods for Computed Tomography Algebraic Iterative Methods for Computed Tomography Per Christian Hansen DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Per Christian Hansen Algebraic

More information

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation Xingguo Li Tuo Zhao Xiaoming Yuan Han Liu Abstract This paper describes an R package named flare, which implements

More information

ELEG Compressive Sensing and Sparse Signal Representations

ELEG Compressive Sensing and Sparse Signal Representations ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /

More information

Level-set MCMC Curve Sampling and Geometric Conditional Simulation

Level-set MCMC Curve Sampling and Geometric Conditional Simulation Level-set MCMC Curve Sampling and Geometric Conditional Simulation Ayres Fan John W. Fisher III Alan S. Willsky February 16, 2007 Outline 1. Overview 2. Curve evolution 3. Markov chain Monte Carlo 4. Curve

More information

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will

Recent advances in Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will Lectures Recent advances in Metamodel of Optimal Prognosis Thomas Most & Johannes Will presented at the Weimar Optimization and Stochastic Days 2010 Source: www.dynardo.de/en/library Recent advances in

More information

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu

FMA901F: Machine Learning Lecture 3: Linear Models for Regression. Cristian Sminchisescu FMA901F: Machine Learning Lecture 3: Linear Models for Regression Cristian Sminchisescu Machine Learning: Frequentist vs. Bayesian In the frequentist setting, we seek a fixed parameter (vector), with value(s)

More information

Modeling with Uncertainty Interval Computations Using Fuzzy Sets

Modeling with Uncertainty Interval Computations Using Fuzzy Sets Modeling with Uncertainty Interval Computations Using Fuzzy Sets J. Honda, R. Tankelevich Department of Mathematical and Computer Sciences, Colorado School of Mines, Golden, CO, U.S.A. Abstract A new method

More information

Numerical Method in Optimization as a Multi-stage Decision Control System

Numerical Method in Optimization as a Multi-stage Decision Control System Numerical Method in Optimization as a Multi-stage Decision Control System B.S. GOH Institute of Mathematical Sciences University of Malaya 50603 Kuala Lumpur MLYSI gohoptimum@gmail.com bstract: - Numerical

More information

Dealing with Categorical Data Types in a Designed Experiment

Dealing with Categorical Data Types in a Designed Experiment Dealing with Categorical Data Types in a Designed Experiment Part II: Sizing a Designed Experiment When Using a Binary Response Best Practice Authored by: Francisco Ortiz, PhD STAT T&E COE The goal of

More information

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy Proc. of the 9th WSEAS Int. Conf. on Mathematical and Computational Methods in Science and Engineering, Trinidad and Tobago, November -7, 7 A new Electromagnetism-like Algorithm with a Population Shrinking

More information

A PRACTICAL APPROACH TO SAMPLE-PATH SIMULATION OPTIMIZATION. Michael C. Ferris Todd S. Munson Krung Sinapiromsaran

A PRACTICAL APPROACH TO SAMPLE-PATH SIMULATION OPTIMIZATION. Michael C. Ferris Todd S. Munson Krung Sinapiromsaran A PRACTICAL APPROACH TO SAMPLE-PATH SIMULATION OPTIMIZATION Michael C. Ferris Todd S. Munson Krung Sinapiromsaran Computer Sciences Department, University of Wisconsin 1210 West Dayton Street, Madison,

More information

The Mesh Adaptive Direct Search Algorithm for Discrete Blackbox Optimization

The Mesh Adaptive Direct Search Algorithm for Discrete Blackbox Optimization The Mesh Adaptive Direct Search Algorithm for Discrete Blackbox Optimization Sébastien Le Digabel Charles Audet Christophe Tribes GERAD and École Polytechnique de Montréal ICCOPT, Tokyo 2016 08 08 BBO:

More information

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation

Bayesian Estimation for Skew Normal Distributions Using Data Augmentation The Korean Communications in Statistics Vol. 12 No. 2, 2005 pp. 323-333 Bayesian Estimation for Skew Normal Distributions Using Data Augmentation Hea-Jung Kim 1) Abstract In this paper, we develop a MCMC

More information

Algorithms for convex optimization

Algorithms for convex optimization Algorithms for convex optimization Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University kocvara@utia.cas.cz http://www.utia.cas.cz/kocvara

More information

Staffing and Scheduling in Multiskill and Blend Call Centers

Staffing and Scheduling in Multiskill and Blend Call Centers Staffing and Scheduling in Multiskill and Blend Call Centers 1 Pierre L Ecuyer GERAD and DIRO, Université de Montréal, Canada (Joint work with Tolga Cezik, Éric Buist, and Thanos Avramidis) Staffing and

More information

Lagrange Multipliers and Problem Formulation

Lagrange Multipliers and Problem Formulation Lagrange Multipliers and Problem Formulation Steven J. Miller Department of Mathematics and Statistics Williams College Williamstown, MA 01267 Abstract The method of Lagrange Multipliers (and its generalizations)

More information

IMAGE DE-NOISING IN WAVELET DOMAIN

IMAGE DE-NOISING IN WAVELET DOMAIN IMAGE DE-NOISING IN WAVELET DOMAIN Aaditya Verma a, Shrey Agarwal a a Department of Civil Engineering, Indian Institute of Technology, Kanpur, India - (aaditya, ashrey)@iitk.ac.in KEY WORDS: Wavelets,

More information

Numerical Optimization

Numerical Optimization Numerical Optimization Quantitative Macroeconomics Raül Santaeulàlia-Llopis MOVE-UAB and Barcelona GSE Fall 2018 Raül Santaeulàlia-Llopis (MOVE-UAB,BGSE) QM: Numerical Optimization Fall 2018 1 / 46 1 Introduction

More information

Module 4 : Solving Linear Algebraic Equations Section 11 Appendix C: Steepest Descent / Gradient Search Method

Module 4 : Solving Linear Algebraic Equations Section 11 Appendix C: Steepest Descent / Gradient Search Method Module 4 : Solving Linear Algebraic Equations Section 11 Appendix C: Steepest Descent / Gradient Search Method 11 Appendix C: Steepest Descent / Gradient Search Method In the module on Problem Discretization

More information

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy

A new Electromagnetism-like Algorithm with a Population Shrinking Strategy 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November, 7 7 A new Electromagnetism-like Algorithm with a Population Shrinking Strategy ANA MARIA A.

More information

A PACKAGE FOR DEVELOPMENT OF ALGORITHMS FOR GLOBAL OPTIMIZATION 1

A PACKAGE FOR DEVELOPMENT OF ALGORITHMS FOR GLOBAL OPTIMIZATION 1 Mathematical Modelling and Analysis 2005. Pages 185 190 Proceedings of the 10 th International Conference MMA2005&CMAM2, Trakai c 2005 Technika ISBN 9986-05-924-0 A PACKAGE FOR DEVELOPMENT OF ALGORITHMS

More information

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM

CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 20 CHAPTER 2 CONVENTIONAL AND NON-CONVENTIONAL TECHNIQUES TO SOLVE ORPD PROBLEM 2.1 CLASSIFICATION OF CONVENTIONAL TECHNIQUES Classical optimization methods can be classified into two distinct groups:

More information

EFFICIENT COMPUTING BUDGET ALLOCATION BY USING REGRESSION WITH SEQUENTIAL SAMPLING CONSTRAINT HU XIANG

EFFICIENT COMPUTING BUDGET ALLOCATION BY USING REGRESSION WITH SEQUENTIAL SAMPLING CONSTRAINT HU XIANG EFFICIENT COMPUTING BUDGET ALLOCATION BY USING REGRESSION WITH SEQUENTIAL SAMPLING CONSTRAINT HU XIANG NATIONAL UNIVERSITY OF SINGAPORE 2012 EFFICIENT COMPUTING BUDGET ALLOCATION BY USING REGRESSION WITH

More information

A robust optimization based approach to the general solution of mp-milp problems

A robust optimization based approach to the general solution of mp-milp problems 21 st European Symposium on Computer Aided Process Engineering ESCAPE 21 E.N. Pistikopoulos, M.C. Georgiadis and A. Kokossis (Editors) 2011 Elsevier B.V. All rights reserved. A robust optimization based

More information

Hellenic Complex Systems Laboratory

Hellenic Complex Systems Laboratory Hellenic Complex Systems Laboratory Technical Report No IV Calculation of the confidence bounds for the fraction nonconforming of normal populations of measurements in clinical laboratory medicine Aristides

More information

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms

Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 13, NO. 5, SEPTEMBER 2002 1225 Efficient Tuning of SVM Hyperparameters Using Radius/Margin Bound and Iterative Algorithms S. Sathiya Keerthi Abstract This paper

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

A penalty based filters method in direct search optimization

A penalty based filters method in direct search optimization A penalty based filters method in direct search optimization ALDINA CORREIA CIICESI/ESTG P.PORTO Felgueiras PORTUGAL aic@estg.ipp.pt JOÃO MATIAS CM-UTAD Vila Real PORTUGAL j matias@utad.pt PEDRO MESTRE

More information

INVERSE PROBLEMS IN GROUNDWATER MODELING

INVERSE PROBLEMS IN GROUNDWATER MODELING INVERSE PROBLEMS IN GROUNDWATER MODELING Theory and Applications of Transport in Porous Media Series Editor: Jacob Bear, Technion - Israel Institute of Technology, Haifa, Israel Volume 6 The titles published

More information

Ranking and selection problem with 1,016,127 systems. requiring 95% probability of correct selection. solved in less than 40 minutes

Ranking and selection problem with 1,016,127 systems. requiring 95% probability of correct selection. solved in less than 40 minutes Ranking and selection problem with 1,016,127 systems requiring 95% probability of correct selection solved in less than 40 minutes with 600 parallel cores with near-linear scaling.. Wallclock time (sec)

More information

Scale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract

Scale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses

More information

ENABLING SIMHEURISTICS THROUGH DESIGNS FOR TENS OF VARIABLES: COSTING MODELS AND ONLINE AVAILABILITY

ENABLING SIMHEURISTICS THROUGH DESIGNS FOR TENS OF VARIABLES: COSTING MODELS AND ONLINE AVAILABILITY Proceedings of the 2014 Winter Simulation Conference A. Tolk, S. Y. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds. ENABLING SIMHEURISTICS THROUGH DESIGNS FOR TENS OF VARIABLES: COSTING

More information

A Parallel Hardware Architecture for Information-Theoretic Adaptive Filtering

A Parallel Hardware Architecture for Information-Theoretic Adaptive Filtering A Parallel Hardware Architecture for Information-Theoretic Adaptive Filtering HPRCTA 2010 Stefan Craciun Dr. Alan D. George Dr. Herman Lam Dr. Jose C. Principe November 14, 2010 NSF CHREC Center ECE Department,

More information

Cover Page. The handle holds various files of this Leiden University dissertation.

Cover Page. The handle   holds various files of this Leiden University dissertation. Cover Page The handle http://hdl.handle.net/1887/22055 holds various files of this Leiden University dissertation. Author: Koch, Patrick Title: Efficient tuning in supervised machine learning Issue Date:

More information

A REVISED SIMPLEX SEARCH PROCEDURE FOR STOCHASTIC SIMULATION RESPONSE-SURFACE OPTIMIZATION

A REVISED SIMPLEX SEARCH PROCEDURE FOR STOCHASTIC SIMULATION RESPONSE-SURFACE OPTIMIZATION Proceedings of the 1998 Winter Simulation Conference D.J. Medeiros, E.F. Watson, J.S. Carson and M.S. Manivannan, eds. A REVISED SIMPLEX SEARCH PROCEDURE FOR STOCHASTIC SIMULATION RESPONSE-SURFACE OPTIMIZATION

More information

Lecture 25 Nonlinear Programming. November 9, 2009

Lecture 25 Nonlinear Programming. November 9, 2009 Nonlinear Programming November 9, 2009 Outline Nonlinear Programming Another example of NLP problem What makes these problems complex Scalar Function Unconstrained Problem Local and global optima: definition,

More information

Constrained and Unconstrained Optimization

Constrained and Unconstrained Optimization Constrained and Unconstrained Optimization Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt2@illinois.edu Oct 10th, 2017 C. Hurtado (UIUC - Economics) Numerical

More information

CONVERGENCE PROPERTIES OF DIRECT SEARCH METHODS FOR STOCHASTIC OPTIMIZATION. Sujin Kim Dali Zhang

CONVERGENCE PROPERTIES OF DIRECT SEARCH METHODS FOR STOCHASTIC OPTIMIZATION. Sujin Kim Dali Zhang Proceedings of the 2010 Winter Simulation Conference B. Johansson, S. Jain, J. Montoya-Torres, J. Hugan, and E. Yücesan, eds. CONVERGENCE PROPERTIES OF DIRECT SEARCH METHODS FOR STOCHASTIC OPTIMIZATION

More information

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies. azemi* (University of Alberta) & H.R. Siahkoohi (University of Tehran) SUMMARY Petrophysical reservoir properties,

More information

Global Minimization via Piecewise-Linear Underestimation

Global Minimization via Piecewise-Linear Underestimation Journal of Global Optimization,, 1 9 (2004) c 2004 Kluwer Academic Publishers, Boston. Manufactured in The Netherlands. Global Minimization via Piecewise-Linear Underestimation O. L. MANGASARIAN olvi@cs.wisc.edu

More information

The Cross-Entropy Method for Mathematical Programming

The Cross-Entropy Method for Mathematical Programming The Cross-Entropy Method for Mathematical Programming Dirk P. Kroese Reuven Y. Rubinstein Department of Mathematics, The University of Queensland, Australia Faculty of Industrial Engineering and Management,

More information

ONLY AVAILABLE IN ELECTRONIC FORM

ONLY AVAILABLE IN ELECTRONIC FORM MANAGEMENT SCIENCE doi 10.1287/mnsc.1070.0812ec pp. ec1 ec7 e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 2008 INFORMS Electronic Companion Customized Bundle Pricing for Information Goods: A Nonlinear

More information

Blackbox Optimization

Blackbox Optimization Blackbox Optimization Algorithm and Applications Sébastien Le Digabel Charles Audet Christophe Tribes GERAD and École Polytechnique de Montréal Workshop on Nonlinear Optimization Algorithms and Industrial

More information

CS 450 Numerical Analysis. Chapter 7: Interpolation

CS 450 Numerical Analysis. Chapter 7: Interpolation Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Combine the PA Algorithm with a Proximal Classifier

Combine the PA Algorithm with a Proximal Classifier Combine the Passive and Aggressive Algorithm with a Proximal Classifier Yuh-Jye Lee Joint work with Y.-C. Tseng Dept. of Computer Science & Information Engineering TaiwanTech. Dept. of Statistics@NCKU

More information

Auto-Extraction of Modelica Code from Finite Element Analysis or Measurement Data

Auto-Extraction of Modelica Code from Finite Element Analysis or Measurement Data Auto-Extraction of Modelica Code from Finite Element Analysis or Measurement Data The-Quan Pham, Alfred Kamusella, Holger Neubert OptiY ek, Aschaffenburg Germany, e-mail: pham@optiyeu Technische Universität

More information

Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation

Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation Fast Automated Estimation of Variance in Discrete Quantitative Stochastic Simulation November 2010 Nelson Shaw njd50@uclive.ac.nz Department of Computer Science and Software Engineering University of Canterbury,

More information

Model Parameter Estimation

Model Parameter Estimation Model Parameter Estimation Shan He School for Computational Science University of Birmingham Module 06-23836: Computational Modelling with MATLAB Outline Outline of Topics Concepts about model parameter

More information

Truncation Errors. Applied Numerical Methods with MATLAB for Engineers and Scientists, 2nd ed., Steven C. Chapra, McGraw Hill, 2008, Ch. 4.

Truncation Errors. Applied Numerical Methods with MATLAB for Engineers and Scientists, 2nd ed., Steven C. Chapra, McGraw Hill, 2008, Ch. 4. Chapter 4: Roundoff and Truncation Errors Applied Numerical Methods with MATLAB for Engineers and Scientists, 2nd ed., Steven C. Chapra, McGraw Hill, 2008, Ch. 4. 1 Outline Errors Accuracy and Precision

More information

A Geometric Analysis of Subspace Clustering with Outliers

A Geometric Analysis of Subspace Clustering with Outliers A Geometric Analysis of Subspace Clustering with Outliers Mahdi Soltanolkotabi and Emmanuel Candés Stanford University Fundamental Tool in Data Mining : PCA Fundamental Tool in Data Mining : PCA Subspace

More information

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions Chalmers/GU Mathematics EXAM TMA946/MAN280 APPLIED OPTIMIZATION Date: 03 05 28 Time: House V, morning Aids: Text memory-less calculator Number of questions: 7; passed on one question requires 2 points

More information

Function approximation using RBF network. 10 basis functions and 25 data points.

Function approximation using RBF network. 10 basis functions and 25 data points. 1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data

More information

PASS Sample Size Software. Randomization Lists

PASS Sample Size Software. Randomization Lists Chapter 880 Introduction This module is used to create a randomization list for assigning subjects to one of up to eight groups or treatments. Six randomization algorithms are available. Four of the algorithms

More information

Optimal Denoising of Natural Images and their Multiscale Geometry and Density

Optimal Denoising of Natural Images and their Multiscale Geometry and Density Optimal Denoising of Natural Images and their Multiscale Geometry and Density Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Israel. Joint work with Anat Levin (WIS),

More information

Table : IEEE Single Format ± a a 2 a 3 :::a 8 b b 2 b 3 :::b 23 If exponent bitstring a :::a 8 is Then numerical value represented is ( ) 2 = (

Table : IEEE Single Format ± a a 2 a 3 :::a 8 b b 2 b 3 :::b 23 If exponent bitstring a :::a 8 is Then numerical value represented is ( ) 2 = ( Floating Point Numbers in Java by Michael L. Overton Virtually all modern computers follow the IEEE 2 floating point standard in their representation of floating point numbers. The Java programming language

More information

Discrete Optimization. Lecture Notes 2

Discrete Optimization. Lecture Notes 2 Discrete Optimization. Lecture Notes 2 Disjunctive Constraints Defining variables and formulating linear constraints can be straightforward or more sophisticated, depending on the problem structure. The

More information

EVALUATION OF METHODS USED TO DETECT WARM-UP PERIOD IN STEADY STATE SIMULATION. Prasad S. Mahajan Ricki G. Ingalls

EVALUATION OF METHODS USED TO DETECT WARM-UP PERIOD IN STEADY STATE SIMULATION. Prasad S. Mahajan Ricki G. Ingalls Proceedings of the 2004 Winter Simulation Conference R.G. Ingalls, M. D. Rossetti, J. S. Smith, and B. A. Peters, eds. EVALUATION OF METHODS USED TO DETECT WARM-UP PERIOD IN STEADY STATE SIMULATION Prasad

More information

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes On the Complexity of the Policy Improvement Algorithm for Markov Decision Processes Mary Melekopoglou Anne Condon Computer Sciences Department University of Wisconsin - Madison 0 West Dayton Street Madison,

More information

HYBRID GENETIC ALGORITHM WITH GREAT DELUGE TO SOLVE CONSTRAINED OPTIMIZATION PROBLEMS

HYBRID GENETIC ALGORITHM WITH GREAT DELUGE TO SOLVE CONSTRAINED OPTIMIZATION PROBLEMS HYBRID GENETIC ALGORITHM WITH GREAT DELUGE TO SOLVE CONSTRAINED OPTIMIZATION PROBLEMS NABEEL AL-MILLI Financial and Business Administration and Computer Science Department Zarqa University College Al-Balqa'

More information

An Extension of the Multicut L-Shaped Method. INEN Large-Scale Stochastic Optimization Semester project. Svyatoslav Trukhanov

An Extension of the Multicut L-Shaped Method. INEN Large-Scale Stochastic Optimization Semester project. Svyatoslav Trukhanov An Extension of the Multicut L-Shaped Method INEN 698 - Large-Scale Stochastic Optimization Semester project Svyatoslav Trukhanov December 13, 2005 1 Contents 1 Introduction and Literature Review 3 2 Formal

More information