On the complexity of the steepest-descent with exact linesearches

Size: px
Start display at page:

Download "On the complexity of the steepest-descent with exact linesearches"

Transcription

1 On the complexity of the steepest-descent with exact linesearches Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint 9 September 22 Abstract The worst-case complexity of the steepest-descent algorithm with exact linesearches for unconstrained smooth optimization is analyzed, and it is shown that the number of iterations of this algorithm which may be necessary to find an iterate at which the norm of the objective function s gradient is less that a prescribed ǫ is, essentially, a multiple of /ǫ 2, as is the case for variants of the same algorithms using inexact linesearches. Introduction The worst-case analysis of optimization algorithms for finding unconstrained stationary points of nonlinear non-convex functions has recently been considered in a number of contributions (see Nesterov, 24, Nesterov and Polyak, 26, Cartis, Gould and Toint, 2a, 2b, 2c, 22a, 22b, 22c, Vicente, 2, Bian, Chen and Ye, 22, Gratton, Sartenaer and Toint, 28, or Jarre, 2, to cite a few. In particular, the study of the steepest-descent method, the most archetypal method for unconstrained nonlinear optimization, was considered by several authors, whose analysis differ primarily by the particular technique used for (possibly approximately minimizing the objective function along the steepest-descent direction. An upper bound on the number of iterations required to obtain an approximate stationary point was given by Nesterov (24 using a variant of the algorithm where the step is computed using the knowledge of a global Lipschitz constant on the gradient of the objective function. He showed that at most O(ǫ 2 iterations might be needed to find an iterate at which the Euclidean norm of the gradient is below a generic tolerance ǫ >. As it turns out, his result also applies to the pure steepest-descent algorithm, that is the variant using exact linesearches. A lower complexity bound was also obtained by Cartis, Gould and Toint (2, where it was shown that the bound of O(ǫ 2 iterations is essentially tight for a version using a Goldstein type linesearch. However, this result depends on a one-dimensional counter-example where the objective function is monotonically decreasing, in which case an exact linesearch would obviously give much better results. The purpose of this short paper is to close the remaining conceptual gap, that is to show that the lower bound of O(ǫ 2 iterations also holds for the steepest-descent algorithm with exact linesearches when applied on functions with globally Lipschitz continuous gradient. The next section recalls the algorithms and the assumptions required for our complexity analysis. Section 3 proposes an example of worst-case behaviour for the method, while School of Mathematics, University of Edinburgh, The King s Buildings, Edinburgh, EH9 3JZ, Scotland, UK. coralia.cartis@ed.ac.uk. The work of this author was supported by EPSRC Grant EP/I28854/. Computational Science and Engineering Department, Rutherford Appleton Laboratory, Chilton, Oxfordshire, OX QX, England, UK. nick.gould@stfc.ac.uk. Department of Mathematics, FUNDP - University of Namur, 6, rue de Bruxelles, B-5, Namur, Belgium. philippe.toint@fundp.ac.be.

2 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 2 Section 4 is devoted to verifying that the example does satisfy the assumptions stated. A few words of conclusion are presented in Section 5. 2 The steepest-descent method with exact linesearches We consider the unconstrained minimization problem min x IR nf(x (2. where f(x is a smooth function from IR n into IR. One of the simplest and oldest algorithm for solving this problem is the steepest-descent method by Cauchy (847, whose iterates are defined, for a given initial guess x, by the simple iteration x k+ = argmin t f(x k tg k, (k (2.2 where g k = x f(x k and where ties are broken by choosing the first minimizer of f(x k tg k if there is more than one (say. This choice is of course most often numerically unrealistic, except for special functions f(x such as quadratics, where the minimizer can be determined analytically. But it remains an ideal that numerically sounder techniques attempt to imitate, justifying our curiosity. The assumptions we make on problem (2. are as follows. AF. f(x is bounded below on IR n, that is there exists a constant κ lbf such that, for all x IR n, f(x κ lbf. AF. f(x is continuously differentiable on IR n. AF.2 g(x = x f(x is Lipschitz continuous on IR n, that is there exists a constant L g such that, for all x,y IR n, g(x g(y L g x y. Here and below, stands for the Euclidean norm. We now briefly recall the upper complexity bound for algorithm (2.2 by suitably reformulating the result of Nesterov (24. Theorem 2. Suppose that AF. AF.2 hold. Then there exists a constant κ upp depending on x and possibly on n such that, for all ǫ (, at most κupp ǫ 2 (2.3 iterations of method (2.2 are needed to obtain an iterate x k such that g k ǫ. Proof. We first note that AF., Taylor s expansion at x k and AF.2 give that, for each k, f(x k f(x k tg k f(x k f(x k +t g k 2 2t 2 L g g k 2 for any t. Maximizing the right-hand side of this inequality with respect to t, we obtain that f(x k f (x k Lg g k g k 2 ǫ2 (2.4 2L g 2L g

3 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 3 for each iteration k, as long as g k > ǫ. But (2.2 ensures that the slope of f(x k tg k must be zero at x k+ = x k t k g k, giving that, for all k, = g k,g(x k+ = g k 2 + g k,g(x k t k g k g k g k 2 ( L g t k, where we used the Cauchy-Schwarz inequality and AF.2. This implies that t k, the argument of the (first minimum in (2.2, is such that t k /L g and therefore, because of (2.4, that, for each k, f(x k f(x k t k g k f(x k f (x k Lg g k ǫ2 2L g as long as g k > ǫ. Thus a maximum number of 2Lg (f(x κ lbf ǫ 2 def = κupp ǫ 2 such iterations may take place before x k is found such that g k ǫ. The purpose of the present paper is to show that the bound (2.3 is essentially tight, which cannot be deduced from the one-dimensional example of Cartis et al. (2. The next section describes how to build a new two-dimensional example where algorithm (2.2 essentially requires O(ǫ 2 iterations to achieve g k ǫ. 3 Constructing a counter-example Because, as in Cartis et al. (2, our example is based on polynomial Hermite interpolation, we first state and prove crucial properties of this type of interpolation. Theorem 3. Assume that real values f, g, h, f T, g T, h T and T > are known. Then there exists a fifth order polynomial p(t def = c +c t+c 2 t 2 +c 3 t 3 +c 4 t 4 +c 5 t 5, t [,T], such that p( = f, p ( = g and p ( = h, p(t = f T, p (T = g T and p (T = h T. The coefficients of this polynomial are given by where c = f, c = g, c 2 = 2h, c 3 = T (r 4r + 2r 2, c 4 = T 2( 5r +7r r 2 and c 5 = T 3(6r 3r + 2r 2, r = T 2(f T f g T 2h T 2, r = T (g T g h T and r 2 = h T h. Moreover, if there are non-negative constants κ, κ and κ 2 such that (3.5 r κ, r κ and r 2 κ 2, (3.6 Then there exists κ f, κ g and κ h only depending on κ, κ and κ 2 such that, for all t [,T], p(t f + g T + 2 h T 2 +κ f T 2, (3.7 p (t g + 2 h T +κ g T, and p (t h +κ h. (3.8

4 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 4 Proof. (See Cartis et al., 2c. Using the form of p(t, we write the desired interpolation conditions as p( = c = f, p ( = c = g, p ( = 2c 2 = h (3.9 (which immeditaley gives the desired values for c, c and c 2 and p(t = c +c T +c 2 T 2 +c 3 T 3 +c 4 T 4 +c 5 T 5 = f T, p (T = c +2c 2 T +3c 3 T 2 +4c 4 T 3 +5c 5 T 4 = g T, p (T = 2c 2 +6c 3 T +2c 4 T 2 +2c 5 T 3 = h T. These conditions can the be re-expressed as a linear system with unknowns c 3, c 4 an c 5, whose solution exists and turns out to be c 3 T 4 2 [f T c 4 = T f g T 2h T 2 ] T c 2 T[g T g h T], T 3 2 h T h completeing the proof of (3.5. Taking absolute values in this relation, we obtain that c 3 T[κ +4κ + 2κ 2 ] κ c3 /T c 4 [5κ T c κ +κ 2 ] def = κ c4 /T 2. [6κ T 3 +3κ + 2κ 2 ] κ c5 /T 3 As a consequence, we have that, for all t [,T], p(t f +g T + 2h T 2 +(κ c3 +κ c4 +κ c5 T 2, (3. which gives (3.7 with κ f def = κ + κ c3 + κ c4 + κ c5. Similarly, we obtain that, for all t [,T], p (t g +h T +(3κ c3 +4κ c4 +5κ c5 T (3. yieldling the first part of (3.8 with κ g def = κ +3κ c3 +4κ c4 +5κ c5, and p (t h +(6κ c3 +2κ c4 +2κ c5, (3.2 from which the second part of (3.8 finally follows with κ h def = κ 2 +6κ c3 +2κ c4 +2κ c5. We now turn to construction our worst-case example for the steepest-descent method (2.2. The idea is to fix an arbitrary τ (, 3] and then to define f(x,y, the objective function in the example as the sum of f (x and f 2 (x,y. As in Cartis et al. (2, f (x is defined by piecewise Hermite polynomial interpolation between the sequence of iterates of the values x =, x k+ = x k +σ k (k (3.3 f (x = ζ(+2η, f (x k+ = f (x k σ 2 k, f (x k = σ k, and f (x k =, (3.4 where ζ( is the Riemann zeta function and η = η(τ def = 2 τ 2 = τ ( 4 2τ (, 2 def 2 +η and σ k =. (3.5 k + From (3.5, we then find that, for x [x k,x k+ ] and t = (x x k /σ k, f (x = f (x k σ 2 kt+σ k (σ k σ k+ [ 4t 3 +7t 4 3t 5],

5 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 5 and It is easy to verify that, for t [,] and thus, using (3.6, that In addition, taking into account that f (x = σ k +(σ k σ k+ [ 2t 2 +28t 3 5t 4], (3.6 f (x = σ k σ k+ σ k [ 24t+84t 6t 3 ]. (3.7 2t 2 +28t 3 5t 4 = t 2 [2 28t+5t 2 ] f (x σ k+ < for all x [x k,x k+ ]. (3.8 < σ k σ k+ σ k = ( k + 2 +η <, k +2 for k and that t [,] if x [x k,x k+ ], we obtain by a straightforward majoration in (3.7 that f (x < 68 (3.9 for x [x k,x k+ ], which in turn implies that f (x is uniformly bounded for all x. The behaviour of f (x and of its first and second derivatives are pictured in Figure 3. on this page. 2.5 x Figure 3.: The function f (x and its first two derivatives (from top to bottom and left to right on the first 8 intervals We now turn to the specification of the function f 2 (x,y, whose role is to limit the iterates in the y-direction to a progressively narrower corridor, thereby forcing the iteration path to oscillate between its lower and upper limits. We have already prescribed

6 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 6 that the x-components of the successive iterates are given by (3.3, and f 2 (x,y will be constructed in such a way that its gradient at the iterates is equal in norm to that of f (x but alternating in sign, thus generating the necessary orthogonality conditions and the oscillating iteration path. More specifically (and in accordance with (3.3, define, for all k, with ( x y = (, ( xk+ = y k+ ( xk y k + ( σ k ( k σ k σ k = f (x k, (3.2 defining the zig-zaging piecewise linear iteration path y(x illustrated in Figure Figure 3.2: The iteration path y(x (plain, the upper and lower boundaries y low (x and y up (x (dashed and y mid (x (dotted for k =,...,8 and η = 5. We now define the lower and upper boundaries of the corridor containing the iterates. This is achieved by defining the lower boundary y low (x as a twice continuously differentiable curve that interpolates the y coordinates of the iterates of index 2k (k and is constant on the intervals [x 2k,x 2k ], yielding y low (x 2k = y low (x 2k = y 2k. Polynomial Hermite interpolation is used to twice continuously connect the constants parts. The upper boundary y up (x is defined in the same way to interpolate the y coordinates of the iterates of index 2k + (k, being constant on [x 2k,x 2k+ ], yielding y up (x 2k = y up (x 2k+ = y 2k+. Both y low (x and y up (x are shown on Figure 3.2, as well as their average y mid (x = 2(y up (x+y low (x. If we define (the corridor width at x, we note that, by construction, δ(x def = y up (x y low (x, (3.2 δ(x k = σ k for all k. (3.22 Moreover, since the interpolation conditions defining y up (x are given (for x [x 2k,x 2k ], say by y up (x 2k = ( i σ i and y up (x 2k = y up (x 2k+ = 2k 2 i= 2k i= ( i σ i, (3.23 y up(x 2k = y up(x 2k = and y up(x 2k = y up(x 2k =, (3.24

7 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 7 acloserinspectionoftheinterpolatingpolynomial(see(3.5revealsthat,forx [x 2k,x 2k ], y up (x = y up (x 2k (σ 2k σ 2k [ t 3 5t 4 +6t 5], (3.25 where t = (x x 2k /σ 2k. Symmetrically, we have that, for x [x 2k,x 2k+ ], y low (x = y low (x 2k +(σ 2k σ 2k+ [ t 3 5t 4 +6t 5], (3.26 where t = (x x 2k /σ 2k. We thus obtain from (3.25 and (3.26, using (3.22 and defining t = (x x k /σ k, that, for x [x k,x k+ ] and δ(x = σ k (σ k σ k+ [ t 3 5t 4 +6t 5] (3.27 y mid (x = y mid (x k + 2( k (σ k σ k+ [ t 3 5t 4 +6t 5]. (3.28 These two last relations yield that and also that δ (x = 2( k+ y mid(x = 3 σ k σ k+ σ k [ t 2 2t 3 +t 4], (3.29 δ (x = 2( k+ y mid(x = 6 σ k σ k+ [ t 3t 2 σk 2 +2t 3]. (3.3 The last inequality in (3.29 results from the decreasing nature of σ k and the fact that 2t + t 2 = ( t 2 for t [,]. It immediately implies, with (3.22 and (3.29, that δ(x is non-increasing and that σ k = δ(x k δ(x δ(x k+ = σ k+ for x [x k,x k+ ]. (3.3 The next step is to define, for each x, f 2 (x,y as a twice continuously differentiable function of y whose value is small between y low (x and y up (x and first increases before levelling off when the distance of y to the corridor increases, thereby keeping the iterates within the corridor. The details of f 2 (x,y are given by 8δ(x 2 if y y low (x f 2 (x,y = (y y mid (x 2 if y [y low (x,y up (x] (3.32 8δ(x 2 if y y up (x+ where Hermite interpolation is once more used to twice continuously connect the first and second interval, as well as the second and third. In the first of these intervals, f 2 (x,y is thus defined by a fifth order polynomial translated to [, ], with boundary conditions on this latter interval given by and p( = 8δ(x 2, p ( =, p ( = p( = (y low (x y mid (x 2, p ( = 2(y low (x y mid (x = δ(x and p ( = 2. The interpolation conditions on the second interval are symmetrically defined. Figure 3.3 shows the shape of f 2 (x,y for fixed x. Note that f 2 (x,y is symmetric in y with respect to y mid (x by construction. Note also that, using (3.2, the definition of y mid (x and (3.22, f 2 y (x 2k,y 2k = f 2 y (x 2k,y low (x 2k = 2(y low (x 2k y mid (x 2k = δ(x 2k = σ 2k (3.33

8 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches Figure 3.3: The shape of f 2 (x,y for x = x 2 and η = 5, the vertical lines indicating the values of y low (x 2 and y up (x 2. and, similarly, f 2 y (x 2k+,y 2k+ = f 2 y (x 2k+,y up (x 2k+ = 2(y up (x 2k+ y mid (x 2k+ = δ(x 2k+ = σ 2k+. (3.34 Note also that, because of (3.32 and (3.29 taken at x = x k (i.e. t =, f 2 x (x k,y k = 2(y k y mid (x k y mid(x k =. (3.35 We finally define the objective function of our minimization problem (2. by f(x,y def = 2 [ f (x+f 2 (x,y ] 2 [ f (+xf (+f 2 (,y ] for x,y IR, for x <,y IR, (3.36 whose contour lines, superimposed on the path of iterates, are shown in Figure 3.4. We thus obtain, using (3.4, (3.33-(3.34 and (3.35, that g SD2 (x k,y k = ( σ k 2 ( k, (3.37 σ k and therefore that g SD2 (x k = σ k. Because of the definition of σ k in (3.5, this implies that the algorithm will require, for any ǫ (,, at least ǫ 2 τ (3.38 iterations to produce an iterate x k such that g k ǫ. This allows us to conclude, as desired, that the evaluation complexity bound of O(ǫ 2 is essentially sharp, provided we can show that f(x,y is bounded below and has a globally Lipschitz continuous gradient, and that the slope of f(x,y is always non-positive along the trajectory. This is the object of the next section. 4 Verifying the example We start by a useful auxiliary result.

9 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches Figure 3.4: The countour lines of f(x,y and the path of iterates for η = 5. Lemma 4. The values of f 2 (x,y, f 2 x (x,y, f 2 y (x,y, 2 f 2 x 2 (x,y, 2 f 2 y 2 (x,y and 2 f 2 x y (x,y are uniformly bounded (in absolute value for all x and y [y low (x,y low (x] [y up (x,y up (x+]. Proof. Because, for each x and y [y low (x,y low (x], f 2 (x,y is a polynomial in y on an interval of length one, its values and that of its first and second derivatives with respect to y are uniformly bounded (in absolute value provided its coefficients are uniformly bounded, which is the case (see (3.5 with T = in Theorem 3., page 3 if the quantities 8δ(x 2 (y up (x y mid (x 2 2 and δ(x (4.39 are themselves uniformly bounded (the third component of the right-hand side of (3.5 being identically equal to 2. But this is the case for the first term in (4.39 since 8δ(x 2 (y up (x y mid (x 2 = 8δ(x 2 4δ(x 2 < 8δ(x 2 8, and for the second because of (3.3 and the bound σ k. What about the derivatives with respect to x (for y [y low (x,y low (x]? Since f 2 (x,y is defined, in this interval, as a polynomial in y shifted to [,], the dependence in x is entirely captured by the coefficients c,...c 5 of this polynomial, themselves depending on the boundary conditions c = 8δ(x 2, c = and c 2 = (4.4 and(3.5. Theboundednessofthefirstandsecondderivativesofc,...,c 5 (asfunctions of x are then implied by (4.42 and the boundedness of the two terms in (4.39, which

10 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches we already verified. Finally, the second derivative of f 2 (x,y with respect to x and y (for y [y low (x,y low (x] is also a polynomial on an (shifted interval of length one, obtained by differentiating c,...,c 5 with respect to x in the polynomial corresponding to the derivative of f 2 (x,y with respect to y. Because we just verified that the first derivatives of c,...,c 5 with respect to x are themselves uniformly bounded in x, this must also be the case of the cross-derivatives of f 2 (x,y. By symmetry, the conclusion of the lemma also holds for all x and y [y up (x,y up (x+]. Theorem 4.2 The function f(x,y is uniformly bounded below on IR 2. Proof. Observe first that (3.8 implies that f (x is bounded below because [ f (x k f (x k+ ] = σk 2 = ζ(+2η <. i= Moreover, it also results from this last observation that f (x for all x (and thus also for all x IR. The fact that f 2 (x,y is also uniformly bounded below results from its definition in (3.32 and Lemma 4.. The desired conclusion then follows from (3.36. The verification that the gradient of f(x,y admits a uniform Lipschitz constant is a more lengthy calculation, which is the object of the next theorem. It depends on the observation that i= σ k σ k+ = ( 2 +η ( k + ( 2 +η( k η k η ( k + = ( 2 +η( 2 +η ( k +2 k + k +2 (4.4 ( 2 +ησ 2 k where we used the bound η < < 2 and the resulting concavity of t 2 +η. Theorem 4.3 The gradient of function f(x, y is uniformly Lipschitz continuous on IR 2. Proof. Let us consider the functions δ(x and y mid (x. Remembering (3.29, (3.3 and (4.4 and using the fact that t [,] when x [x k,x k+ ], we easily deduce that, for x in this interval, max [ δ (x, δ (x, y mid(x, y mid(x ] 36( 2 +η def = κ dy. (4.42 We now turn to the analysis of the second derivatives of f 2 (x,y. Consider first the case where y [y low (x,y up (x]. In this interval, we obtain, for x [x k,x k+ ], that 2 f 2 (x,y = 2. (4.43 y2

11 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches Moreover and thus f 2 x (x,y = 2(y y mid(xy mid(x ( f 2 x 2 (x,y = 2y mid(x 2 2(y y mid (xy mid(x. Taking absolute values and noting that, because of the definition of y mid (x and (4.42, y y mid (x 2δ(x 2σ k (4.45 for x [x k,x k+ ] and y [y low (x,y sup (x], we obtain, for x and y in these intervals, that 2 f 2 x 2 (x,y 2κ2 dy +σ k κ dy 2κ 2 dy +κ dy, (4.46 where we also used (4.42 and the bound σ k. Finally, for x and y in the same intervals, we have that 2 f 2 x y (x,y = 2y mid(x 2κ dy where we used (4.44 and (4.42. Considering this last relation together with (4.43 and (4.46, we thus conclude that the second derivatives of f 2 (x,y are uniformly bounded for all x and all y [y low (x,y sup (x]. The case where y [y low (x,y low (x] [y up (x,y up (x+] is covered by Lemma 4.. To conclude our analysis, we are thus left with checking the boundedness of the second derivatives of f 2 (x,y for y y up + and y y low. In these intervals, f 2 (x,y = 8δ(x 2, whose second derivatives are bounded because of (4.42. We may therefore finally assess that the second derivatives of f 2 (x,y are bounded for all x and all y. (Figure 4.5 shows the second derivative of f 2 (x,y with respect to y for x = x 2. We may now combine this last conclusion with (3.9 and (3.36 to deduce that f(x,y has uniformly bounded second derivatives for all (x,y IR 2. The desired Lipschitz continuity of its gradient then follows Figure 4.5: The second derivative of f 2 (x 2,y, for η = 5. We conclude the construction of our example by verifying that the sequence of iterates can indeed be obtained from the steepest-descent method with exact linesearches.

12 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 2 Theorem 4.4 The iterate (x k+,y k+ is the first minimizer along the steepest descent direction from (x k,y k. Proof. The theorem statement is equivalent to verifying that the slope ( ω(x = f(x,y(x, 2 ( k [ = 2 f (x+ f ] 2 f 2 x (x,y+( k y (x,y of f(x,y on [x k,x k+ ], which is given by ω(x = 2f (x+[y(x y mid (x][( k y mid(x], (4.47 is always non-positive and is zero only at the iterates (the corners of the trajectory. To prove this property, we first observe that, because of (3.29, ( k y mid(x = ( k[ y mid(x ] (4.48 Observe now that (4.4 and the decreasing nature of σ k together give that, for k >, σ 2k σ 2k σ 2k ( 2 +ησ 2k < 2 +η.6. where the last inequality follows from the bound η. Hence, recalling (3.29 and using the fact that max t [,] t 2 ( t 2 = 6, we obtain that, for x [x 2k,x 2k ], y mid(x < 5.6 max t [,] t2 ( t 2 <.57. (4.49 Similarly, (4.4 and the decreasing nature of σ k imply that, for k >, σ 2k+ σ 2k σ 2k ( 2 +ησ 2k < 2 +η.6 while, for k =, σ σ ( σ = 2 2 +η <.6. This thus gives that y mid(x < 5.6 max t [,] t2 ( t 2 <.57 (4.5 for x [x 2k,x 2k+ ]. Combining (4.49 and (4.5, we obtain that y mid(x < for all x, and therefore, using (4.48, that ( k y mid(x, for all x [x k,x k+ ], where the inequality is strict except at x k and x k+ since y mid(x k = y mid(x k+ =. Hence we obtain, using (4.45, that, for x [x k,x k+ ], [y(x ymid (x][( k y mid(x] 2δ(x. (4.5 Moreover, since, at the leftmost boundary of [x 2k,x 2k ], y(x 2k y mid (x 2k = 2δ(x 2k = 2σ 2k

13 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 3 and, at the leftmost boundary of [x 2k,x 2k+ ], y(x 2k y mid (x 2k = 2δ(x 2k = 2σ 2k (where we used (3.3, we deduce from (4.48 that the inequality in (4.5 can only hold as an equality at x k+. Our penultimate step to is note that (3.6 and (3.27 together give that, for x [x k,x k+ ] and t = (x x k /σ k, f (x+δ(x = (σ k σ k+ [ 2t 2 +8t 3 6t 5], (4.52 where, again, the inequality is strict in the interior of the interval (see Figure Figure 4.6: The polynomial 2t 2 +8t 3 6t 5 on [,]. Combining finally (4.47, (4.5 and (4.52, we obtain that, for all k, ω(x < for x [x k,x k+, (4.53 and (x k+,y k+ is indeed the first local minimizer of f(x,y along the steepest-descent direction at iterate (x k,y k. This last theorem is illustrated in Figure 4.7, and completes the construction of our example. 5 Conclusions We have constructed an example where, for an arbitrary τ >, the steepest-descent method with exact linesearches takes at least a multiple of ǫ 2+τ iterations to find an approximate stationary point at which g k ǫ, for any ǫ (,. This result closes the gap left by Cartis et al. (2 who could not accomodate this type of linesearch corresponding to the archetypal, if very often impractical, definition of the method. Given that we have shown in this last paper that it is impossible to obtain an O(ǫ 2 worst-case complexity for all ǫ, this is probably the best result that can be obtained. As was the case in this last paper, our example may furthermore be adapted to cover the case where the level sets of the objective are finite by extending f(x,y beyond the approximate minimizer. This is achieved by smoothly prolongating f (x beyond this point with a suitably increasing function and by, say, keeping the width of the corridor constant in this part of the plane. Such an example may therefore be constructed for every ǫ (,.

14 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches References Figure 4.7: The behaviour of ω(x for η = 5. W. Bian, X. Chen, and Y. Ye. Complexity analysis of interior point algorithms for non- Lipschitz and nonconvex minimization. Mathematical Programming, Series A, (submitted, 22. C. Cartis, N. I. M. Gould, and Ph. L. Toint. On the complexity of steepest descent, Newton s and regularized Newton s methods for nonconvex unconstrained optimization. SIAM Journal on Optimization, 2(6, , 2. C. Cartis, N. I. M. Gould, and Ph. L. Toint. Adaptive cubic overestimation methods for unconstrained optimization. Part I: motivation, convergence and numerical results. Mathematical Programming, Series A, 27(2, , 2a. C. Cartis, N. I. M. Gould, and Ph. L. Toint. An adaptive cubic regularisation algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity. IMA Journal of Numerical Analysis, (to appear, 2b. C. Cartis, N. I. M. Gould, and Ph. L. Toint. Optimal Newton-type methods for nonconvex optimization. Technical Report naxys-7-2, Namur Centre for Complex Systems (naxys, FUNDP-University of Namur, Namur, Belgium, 2c. C. Cartis, N. I. M. Gould, and Ph. L. Toint. Complexity bounds for second-order optimality in unconstrained optimization. Journal of Complexity, 28, 93 8, 22a. C. Cartis, N. I. M. Gould, and Ph. L. Toint. A note about the complexity of minimizing Nesterov s smooth Chebyshev-Rosenbrock function. Optimization Methods and Software, (to appear, 22b. C. Cartis, N. I. M. Gould, and Ph. L. Toint. On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization. SIAM Journal on Optimization, 22(, 66 86, 22c. A. Cauchy. Méthode générale pour la résolution des systèmes d équations simultannées. Comptes Rendus de l Académie des Sciences, pp , 847. S. Gratton, A. Sartenaer, and Ph. L. Toint. Recursive trust-region methods for multiscale nonlinear optimization. SIAM Journal on Optimization, 9(, , 28. F. Jarre. On Nesterovs smooth Chebyshev-Rosenbrock function. Technical report, University of Düsseldorf, Düsseldorf, Germany, May 2.

15 Cartis, Gould, Toint: Complexity of steepest descent with exact linesearches 5 Yu. Nesterov. Introductory Lectures on Convex Optimization. Applied Optimization. Kluwer Academic Publishers, Dordrecht, The Netherlands, 24. Yu. Nesterov and B. T. Polyak. Cubic regularization of Newton method and its global performance. Mathematical Programming, Series A, 8(, 77 25, 26. L. N. Vicente. Worst case complexity of direct search. Technical report, Department of Mathematics, University of Coimbra, Coimbra, Portugal, May 2. Preprint -7, revised 2.

Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization

Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint October 30, 200; Revised March 30, 20 Abstract

More information

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity Coralia Cartis, Nick Gould and Philippe Toint Department of Mathematics,

More information

Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity

Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity Coralia Cartis,, Nicholas I. M. Gould, and Philippe L. Toint September

More information

On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization

On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization C. Cartis, N. I. M. Gould and Ph. L. Toint 22 September 2011 Abstract The (optimal) function/gradient

More information

Adaptive cubic overestimation methods for unconstrained optimization

Adaptive cubic overestimation methods for unconstrained optimization Report no. NA-07/20 Adaptive cubic overestimation methods for unconstrained optimization Coralia Cartis School of Mathematics, University of Edinburgh, The King s Buildings, Edinburgh, EH9 3JZ, Scotland,

More information

Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models

Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models Worst-case evaluation comlexity for unconstrained nonlinear otimization using high-order regularized models E. G. Birgin, J. L. Gardenghi, J. M. Martínez, S. A. Santos and Ph. L. Toint 2 Aril 26 Abstract

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Universal regularization methods varying the power, the smoothness and the accuracy arxiv: v1 [math.oc] 16 Nov 2018

Universal regularization methods varying the power, the smoothness and the accuracy arxiv: v1 [math.oc] 16 Nov 2018 Universal regularization methods varying the power, the smoothness and the accuracy arxiv:1811.07057v1 [math.oc] 16 Nov 2018 Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint Revision completed

More information

1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016 AM 22: Advanced Optimization Spring 206 Prof. Yaron Singer Lecture 9 February 24th Overview In the previous lecture we reviewed results from multivariate calculus in preparation for our journey into convex

More information

Trust Region Methods for Unconstrained Optimisation

Trust Region Methods for Unconstrained Optimisation Trust Region Methods for Unconstrained Optimisation Lecture 9, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Trust

More information

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0.

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0. Outline Coordinate Minimization Daniel P. Robinson Department of Applied Mathematics and Statistics Johns Hopkins University November 27, 208 Introduction 2 Algorithms Cyclic order with exact minimization

More information

Convergence of trust-region methods based on probabilistic models

Convergence of trust-region methods based on probabilistic models Convergence of trust-region methods based on probabilistic models A. S. Bandeira K. Scheinberg L. N. Vicente October 24, 2013 Abstract In this paper we consider the use of probabilistic or random models

More information

A Trust Region Algorithm for Heterogeneous Multiobjective Optimization

A Trust Region Algorithm for Heterogeneous Multiobjective Optimization A Trust Region Algorithm for Heterogeneous Multiobjective Optimization Jana Thomann and Gabriele Eichfelder 8.0.018 Abstract This paper presents a new trust region method for multiobjective heterogeneous

More information

Global convergence rate analysis of unconstrained optimization methods based on probabilistic models

Global convergence rate analysis of unconstrained optimization methods based on probabilistic models Math. Program., Ser. A DOI 10.1007/s10107-017-1137-4 FULL LENGTH PAPER Global convergence rate analysis of unconstrained optimization methods based on probabilistic models C. Cartis 1 K. Scheinberg 2 Received:

More information

University of Edinburgh, Edinburgh EH9 3JZ, United Kingdom.

University of Edinburgh, Edinburgh EH9 3JZ, United Kingdom. An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity by C. Cartis 1, N. I. M. Gould 2 and Ph. L. Toint 3 February 20, 2009;

More information

Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization

Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization C. Cartis, N. I. M. Gould and Ph. L. Toint 11th November, 2014 Abstract In a recent paper (Cartis

More information

GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS

GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS ANDREW R. CONN, KATYA SCHEINBERG, AND LUíS N. VICENTE Abstract. In this paper we prove global

More information

Interpolation. 1 What is interpolation? 2 Why are we interested in this?

Interpolation. 1 What is interpolation? 2 Why are we interested in this? Interpolation 1 What is interpolation? For a certain function f (x we know only the values y 1 = f (x 1,,y n = f (x n For a point x different from x 1,,x n we would then like to approximate f ( x using

More information

Lecture Quantitative Finance Spring Term 2015

Lecture Quantitative Finance Spring Term 2015 implied Lecture Quantitative Finance Spring Term 2015 : May 7, 2015 1 / 28 implied 1 implied 2 / 28 Motivation and setup implied the goal of this chapter is to treat the implied which requires an algorithm

More information

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE GÜNTER ROTE Abstract. A salesperson wants to visit each of n objects that move on a line at given constant speeds in the shortest possible time,

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Nonlinear programming without a penalty function or a filter N I M Gould Ph L Toint October 1, 2007 RAL-TR-2007-016 c Science and Technology Facilities Council Enquires about copyright, reproduction and

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Report no. NA-07/09 Nonlinear programming without a penalty function or a filter Nicholas I. M. Gould Oxford University, Numerical Analysis Group Philippe L. Toint Department of Mathematics, FUNDP-University

More information

Steepest descent and conjugate gradient methods with variable preconditioning

Steepest descent and conjugate gradient methods with variable preconditioning Ilya Lashuk and Andrew Knyazev 1 Steepest descent and conjugate gradient methods with variable preconditioning Ilya Lashuk (the speaker) and Andrew Knyazev Department of Mathematics and Center for Computational

More information

What can we do with numerical optimization?

What can we do with numerical optimization? Optimization motivation and background Eddie Wadbro Introduction to PDE Constrained Optimization, 2016 February 15 16, 2016 Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Math. Program., Ser. A (2010) 122:155 196 DOI 10.1007/s10107-008-0244-7 FULL LENGTH PAPER Nonlinear programming without a penalty function or a filter N. I. M. Gould Ph.L.Toint Received: 11 December 2007

More information

The Correlation Smile Recovery

The Correlation Smile Recovery Fortis Bank Equity & Credit Derivatives Quantitative Research The Correlation Smile Recovery E. Vandenbrande, A. Vandendorpe, Y. Nesterov, P. Van Dooren draft version : March 2, 2009 1 Introduction Pricing

More information

Game Theory: Normal Form Games

Game Theory: Normal Form Games Game Theory: Normal Form Games Michael Levet June 23, 2016 1 Introduction Game Theory is a mathematical field that studies how rational agents make decisions in both competitive and cooperative situations.

More information

Yao s Minimax Principle

Yao s Minimax Principle Complexity of algorithms The complexity of an algorithm is usually measured with respect to the size of the input, where size may for example refer to the length of a binary word describing the input,

More information

Worst-case evaluation complexity of regularization methods for smooth unconstrained optimization using Hölder continuous gradients

Worst-case evaluation complexity of regularization methods for smooth unconstrained optimization using Hölder continuous gradients Worst-case evaluation comlexity of regularization methods for smooth unconstrained otimization using Hölder continuous gradients C Cartis N I M Gould and Ph L Toint 26 June 205 Abstract The worst-case

More information

Sensitivity Analysis with Data Tables. 10% annual interest now =$110 one year later. 10% annual interest now =$121 one year later

Sensitivity Analysis with Data Tables. 10% annual interest now =$110 one year later. 10% annual interest now =$121 one year later Sensitivity Analysis with Data Tables Time Value of Money: A Special kind of Trade-Off: $100 @ 10% annual interest now =$110 one year later $110 @ 10% annual interest now =$121 one year later $100 @ 10%

More information

Essays on Some Combinatorial Optimization Problems with Interval Data

Essays on Some Combinatorial Optimization Problems with Interval Data Essays on Some Combinatorial Optimization Problems with Interval Data a thesis submitted to the department of industrial engineering and the institute of engineering and sciences of bilkent university

More information

Finite Memory and Imperfect Monitoring

Finite Memory and Imperfect Monitoring Federal Reserve Bank of Minneapolis Research Department Finite Memory and Imperfect Monitoring Harold L. Cole and Narayana Kocherlakota Working Paper 604 September 2000 Cole: U.C.L.A. and Federal Reserve

More information

Intro to Economic analysis

Intro to Economic analysis Intro to Economic analysis Alberto Bisin - NYU 1 The Consumer Problem Consider an agent choosing her consumption of goods 1 and 2 for a given budget. This is the workhorse of microeconomic theory. (Notice

More information

I. More Fundamental Concepts and Definitions from Mathematics

I. More Fundamental Concepts and Definitions from Mathematics An Introduction to Optimization The core of modern economics is the notion that individuals optimize. That is to say, individuals use the resources available to them to advance their own personal objectives

More information

Is Greedy Coordinate Descent a Terrible Algorithm?

Is Greedy Coordinate Descent a Terrible Algorithm? Is Greedy Coordinate Descent a Terrible Algorithm? Julie Nutini, Mark Schmidt, Issam Laradji, Michael Friedlander, Hoyt Koepke University of British Columbia Optimization and Big Data, 2015 Context: Random

More information

Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem.

Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem. Exercise List: Proving convergence of the (Stochastic) Gradient Descent Method for the Least Squares Problem. Robert M. Gower. October 3, 07 Introduction This is an exercise in proving the convergence

More information

Online Shopping Intermediaries: The Strategic Design of Search Environments

Online Shopping Intermediaries: The Strategic Design of Search Environments Online Supplemental Appendix to Online Shopping Intermediaries: The Strategic Design of Search Environments Anthony Dukes University of Southern California Lin Liu University of Central Florida February

More information

Lecture 17: More on Markov Decision Processes. Reinforcement learning

Lecture 17: More on Markov Decision Processes. Reinforcement learning Lecture 17: More on Markov Decision Processes. Reinforcement learning Learning a model: maximum likelihood Learning a value function directly Monte Carlo Temporal-difference (TD) learning COMP-424, Lecture

More information

Support Vector Machines: Training with Stochastic Gradient Descent

Support Vector Machines: Training with Stochastic Gradient Descent Support Vector Machines: Training with Stochastic Gradient Descent Machine Learning Spring 2018 The slides are mainly from Vivek Srikumar 1 Support vector machines Training by maximizing margin The SVM

More information

Chapter 3: Black-Scholes Equation and Its Numerical Evaluation

Chapter 3: Black-Scholes Equation and Its Numerical Evaluation Chapter 3: Black-Scholes Equation and Its Numerical Evaluation 3.1 Itô Integral 3.1.1 Convergence in the Mean and Stieltjes Integral Definition 3.1 (Convergence in the Mean) A sequence {X n } n ln of random

More information

CS364A: Algorithmic Game Theory Lecture #3: Myerson s Lemma

CS364A: Algorithmic Game Theory Lecture #3: Myerson s Lemma CS364A: Algorithmic Game Theory Lecture #3: Myerson s Lemma Tim Roughgarden September 3, 23 The Story So Far Last time, we introduced the Vickrey auction and proved that it enjoys three desirable and different

More information

Penalty Functions. The Premise Quadratic Loss Problems and Solutions

Penalty Functions. The Premise Quadratic Loss Problems and Solutions Penalty Functions The Premise Quadratic Loss Problems and Solutions The Premise You may have noticed that the addition of constraints to an optimization problem has the effect of making it much more difficult.

More information

Equivalence between Semimartingales and Itô Processes

Equivalence between Semimartingales and Itô Processes International Journal of Mathematical Analysis Vol. 9, 215, no. 16, 787-791 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/1.12988/ijma.215.411358 Equivalence between Semimartingales and Itô Processes

More information

Smooth estimation of yield curves by Laguerre functions

Smooth estimation of yield curves by Laguerre functions Smooth estimation of yield curves by Laguerre functions A.S. Hurn 1, K.A. Lindsay 2 and V. Pavlov 1 1 School of Economics and Finance, Queensland University of Technology 2 Department of Mathematics, University

More information

Hints on Some of the Exercises

Hints on Some of the Exercises Hints on Some of the Exercises of the book R. Seydel: Tools for Computational Finance. Springer, 00/004/006/009/01. Preparatory Remarks: Some of the hints suggest ideas that may simplify solving the exercises

More information

On the Lower Arbitrage Bound of American Contingent Claims

On the Lower Arbitrage Bound of American Contingent Claims On the Lower Arbitrage Bound of American Contingent Claims Beatrice Acciaio Gregor Svindland December 2011 Abstract We prove that in a discrete-time market model the lower arbitrage bound of an American

More information

Chapter 7 One-Dimensional Search Methods

Chapter 7 One-Dimensional Search Methods Chapter 7 One-Dimensional Search Methods An Introduction to Optimization Spring, 2014 1 Wei-Ta Chu Golden Section Search! Determine the minimizer of a function over a closed interval, say. The only assumption

More information

Sublinear Time Algorithms Oct 19, Lecture 1

Sublinear Time Algorithms Oct 19, Lecture 1 0368.416701 Sublinear Time Algorithms Oct 19, 2009 Lecturer: Ronitt Rubinfeld Lecture 1 Scribe: Daniel Shahaf 1 Sublinear-time algorithms: motivation Twenty years ago, there was practically no investigation

More information

A Note on Error Estimates for some Interior Penalty Methods

A Note on Error Estimates for some Interior Penalty Methods A Note on Error Estimates for some Interior Penalty Methods A. F. Izmailov 1 and M. V. Solodov 2 1 Moscow State University, Faculty of Computational Mathematics and Cybernetics, Department of Operations

More information

Applied Mathematical Sciences, Vol. 8, 2014, no. 1, 1-12 HIKARI Ltd,

Applied Mathematical Sciences, Vol. 8, 2014, no. 1, 1-12 HIKARI Ltd, Applied Mathematical Sciences, Vol. 8, 2014, no. 1, 1-12 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.35258 Improving the Robustness of Difference of Convex Algorithm in the Research

More information

Forecast Horizons for Production Planning with Stochastic Demand

Forecast Horizons for Production Planning with Stochastic Demand Forecast Horizons for Production Planning with Stochastic Demand Alfredo Garcia and Robert L. Smith Department of Industrial and Operations Engineering Universityof Michigan, Ann Arbor MI 48109 December

More information

6.896 Topics in Algorithmic Game Theory February 10, Lecture 3

6.896 Topics in Algorithmic Game Theory February 10, Lecture 3 6.896 Topics in Algorithmic Game Theory February 0, 200 Lecture 3 Lecturer: Constantinos Daskalakis Scribe: Pablo Azar, Anthony Kim In the previous lecture we saw that there always exists a Nash equilibrium

More information

CS 3331 Numerical Methods Lecture 2: Functions of One Variable. Cherung Lee

CS 3331 Numerical Methods Lecture 2: Functions of One Variable. Cherung Lee CS 3331 Numerical Methods Lecture 2: Functions of One Variable Cherung Lee Outline Introduction Solving nonlinear equations: find x such that f(x ) = 0. Binary search methods: (Bisection, regula falsi)

More information

Stochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs

Stochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs Stochastic Programming and Financial Analysis IE447 Midterm Review Dr. Ted Ralphs IE447 Midterm Review 1 Forming a Mathematical Programming Model The general form of a mathematical programming model is:

More information

Richardson Extrapolation Techniques for the Pricing of American-style Options

Richardson Extrapolation Techniques for the Pricing of American-style Options Richardson Extrapolation Techniques for the Pricing of American-style Options June 1, 2005 Abstract Richardson Extrapolation Techniques for the Pricing of American-style Options In this paper we re-examine

More information

First-Order Methods. Stephen J. Wright 1. University of Wisconsin-Madison. IMA, August 2016

First-Order Methods. Stephen J. Wright 1. University of Wisconsin-Madison. IMA, August 2016 First-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) First-Order Methods IMA, August 2016 1 / 48 Smooth

More information

Quadrant marked mesh patterns in 123-avoiding permutations

Quadrant marked mesh patterns in 123-avoiding permutations Quadrant marked mesh patterns in 23-avoiding permutations Dun Qiu Department of Mathematics University of California, San Diego La Jolla, CA 92093-02. USA duqiu@math.ucsd.edu Jeffrey Remmel Department

More information

The Value of Information in Central-Place Foraging. Research Report

The Value of Information in Central-Place Foraging. Research Report The Value of Information in Central-Place Foraging. Research Report E. J. Collins A. I. Houston J. M. McNamara 22 February 2006 Abstract We consider a central place forager with two qualitatively different

More information

Principles of Financial Computing

Principles of Financial Computing Principles of Financial Computing Prof. Yuh-Dauh Lyuu Dept. Computer Science & Information Engineering and Department of Finance National Taiwan University c 2008 Prof. Yuh-Dauh Lyuu, National Taiwan University

More information

Sensitivity of American Option Prices with Different Strikes, Maturities and Volatilities

Sensitivity of American Option Prices with Different Strikes, Maturities and Volatilities Applied Mathematical Sciences, Vol. 6, 2012, no. 112, 5597-5602 Sensitivity of American Option Prices with Different Strikes, Maturities and Volatilities Nasir Rehman Department of Mathematics and Statistics

More information

Department of Mathematics. Mathematics of Financial Derivatives

Department of Mathematics. Mathematics of Financial Derivatives Department of Mathematics MA408 Mathematics of Financial Derivatives Thursday 15th January, 2009 2pm 4pm Duration: 2 hours Attempt THREE questions MA408 Page 1 of 5 1. (a) Suppose 0 < E 1 < E 3 and E 2

More information

4: SINGLE-PERIOD MARKET MODELS

4: SINGLE-PERIOD MARKET MODELS 4: SINGLE-PERIOD MARKET MODELS Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 4: Single-Period Market Models 1 / 87 General Single-Period

More information

Optimal Order Placement

Optimal Order Placement Optimal Order Placement Peter Bank joint work with Antje Fruth OMI Colloquium Oxford-Man-Institute, October 16, 2012 Optimal order execution Broker is asked to do a transaction of a significant fraction

More information

Part 1: q Theory and Irreversible Investment

Part 1: q Theory and Irreversible Investment Part 1: q Theory and Irreversible Investment Goal: Endogenize firm characteristics and risk. Value/growth Size Leverage New issues,... This lecture: q theory of investment Irreversible investment and real

More information

Handout 8: Introduction to Stochastic Dynamic Programming. 2 Examples of Stochastic Dynamic Programming Problems

Handout 8: Introduction to Stochastic Dynamic Programming. 2 Examples of Stochastic Dynamic Programming Problems SEEM 3470: Dynamic Optimization and Applications 2013 14 Second Term Handout 8: Introduction to Stochastic Dynamic Programming Instructor: Shiqian Ma March 10, 2014 Suggested Reading: Chapter 1 of Bertsekas,

More information

Applications of Good s Generalized Diversity Index. A. J. Baczkowski Department of Statistics, University of Leeds Leeds LS2 9JT, UK

Applications of Good s Generalized Diversity Index. A. J. Baczkowski Department of Statistics, University of Leeds Leeds LS2 9JT, UK Applications of Good s Generalized Diversity Index A. J. Baczkowski Department of Statistics, University of Leeds Leeds LS2 9JT, UK Internal Report STAT 98/11 September 1998 Applications of Good s Generalized

More information

CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems

CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems January 26, 2018 1 / 24 Basic information All information is available in the syllabus

More information

Optimization in Finance

Optimization in Finance Research Reports on Mathematical and Computing Sciences Series B : Operations Research Department of Mathematical and Computing Sciences Tokyo Institute of Technology 2-12-1 Oh-Okayama, Meguro-ku, Tokyo

More information

November 2006 LSE-CDAM

November 2006 LSE-CDAM NUMERICAL APPROACHES TO THE PRINCESS AND MONSTER GAME ON THE INTERVAL STEVE ALPERN, ROBBERT FOKKINK, ROY LINDELAUF, AND GEERT JAN OLSDER November 2006 LSE-CDAM-2006-18 London School of Economics, Houghton

More information

Optimal prepayment of Dutch mortgages*

Optimal prepayment of Dutch mortgages* 137 Statistica Neerlandica (2007) Vol. 61, nr. 1, pp. 137 155 Optimal prepayment of Dutch mortgages* Bart H. M. Kuijpers ABP Investments, P.O. Box 75753, NL-1118 ZX Schiphol, The Netherlands Peter C. Schotman

More information

IEOR E4703: Monte-Carlo Simulation

IEOR E4703: Monte-Carlo Simulation IEOR E4703: Monte-Carlo Simulation Other Miscellaneous Topics and Applications of Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Characterization of the Optimum

Characterization of the Optimum ECO 317 Economics of Uncertainty Fall Term 2009 Notes for lectures 5. Portfolio Allocation with One Riskless, One Risky Asset Characterization of the Optimum Consider a risk-averse, expected-utility-maximizing

More information

Collinear Triple Hypergraphs and the Finite Plane Kakeya Problem

Collinear Triple Hypergraphs and the Finite Plane Kakeya Problem Collinear Triple Hypergraphs and the Finite Plane Kakeya Problem Joshua Cooper August 14, 006 Abstract We show that the problem of counting collinear points in a permutation (previously considered by the

More information

On the Optimality of a Family of Binary Trees Techical Report TR

On the Optimality of a Family of Binary Trees Techical Report TR On the Optimality of a Family of Binary Trees Techical Report TR-011101-1 Dana Vrajitoru and William Knight Indiana University South Bend Department of Computer and Information Sciences Abstract In this

More information

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University Ellipsoid Method ellipsoid method convergence proof inequality constraints feasibility problems Prof. S. Boyd, EE364b, Stanford University Ellipsoid method developed by Shor, Nemirovsky, Yudin in 1970s

More information

Accelerated Stochastic Gradient Descent Praneeth Netrapalli MSR India

Accelerated Stochastic Gradient Descent Praneeth Netrapalli MSR India Accelerated Stochastic Gradient Descent Praneeth Netrapalli MSR India Presented at OSL workshop, Les Houches, France. Joint work with Prateek Jain, Sham M. Kakade, Rahul Kidambi and Aaron Sidford Linear

More information

Optimization Models one variable optimization and multivariable optimization

Optimization Models one variable optimization and multivariable optimization Georg-August-Universität Göttingen Optimization Models one variable optimization and multivariable optimization Wenzhong Li lwz@nju.edu.cn Feb 2011 Mathematical Optimization Problems in optimization are

More information

Singular Stochastic Control Models for Optimal Dynamic Withdrawal Policies in Variable Annuities

Singular Stochastic Control Models for Optimal Dynamic Withdrawal Policies in Variable Annuities 1/ 46 Singular Stochastic Control Models for Optimal Dynamic Withdrawal Policies in Variable Annuities Yue Kuen KWOK Department of Mathematics Hong Kong University of Science and Technology * Joint work

More information

Approximate Composite Minimization: Convergence Rates and Examples

Approximate Composite Minimization: Convergence Rates and Examples ISMP 2018 - Bordeaux Approximate Composite Minimization: Convergence Rates and S. Praneeth Karimireddy, Sebastian U. Stich, Martin Jaggi MLO Lab, EPFL, Switzerland sebastian.stich@epfl.ch July 4, 2018

More information

A class of coherent risk measures based on one-sided moments

A class of coherent risk measures based on one-sided moments A class of coherent risk measures based on one-sided moments T. Fischer Darmstadt University of Technology November 11, 2003 Abstract This brief paper explains how to obtain upper boundaries of shortfall

More information

Optimal online-list batch scheduling

Optimal online-list batch scheduling Optimal online-list batch scheduling Paulus, J.J.; Ye, Deshi; Zhang, G. Published: 01/01/2008 Document Version Publisher s PDF, also known as Version of Record (includes final page, issue and volume numbers)

More information

Constrained Sequential Resource Allocation and Guessing Games

Constrained Sequential Resource Allocation and Guessing Games 4946 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 11, NOVEMBER 2008 Constrained Sequential Resource Allocation and Guessing Games Nicholas B. Chang and Mingyan Liu, Member, IEEE Abstract In this

More information

Lecture 23: April 10

Lecture 23: April 10 CS271 Randomness & Computation Spring 2018 Instructor: Alistair Sinclair Lecture 23: April 10 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They

More information

Calibration Lecture 1: Background and Parametric Models

Calibration Lecture 1: Background and Parametric Models Calibration Lecture 1: Background and Parametric Models March 2016 Motivation What is calibration? Derivative pricing models depend on parameters: Black-Scholes σ, interest rate r, Heston reversion speed

More information

EC316a: Advanced Scientific Computation, Fall Discrete time, continuous state dynamic models: solution methods

EC316a: Advanced Scientific Computation, Fall Discrete time, continuous state dynamic models: solution methods EC316a: Advanced Scientific Computation, Fall 2003 Notes Section 4 Discrete time, continuous state dynamic models: solution methods We consider now solution methods for discrete time models in which decisions

More information

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process

An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process Computational Statistics 17 (March 2002), 17 28. An Improved Saddlepoint Approximation Based on the Negative Binomial Distribution for the General Birth Process Gordon K. Smyth and Heather M. Podlich Department

More information

Exact shape-reconstruction by one-step linearization in EIT

Exact shape-reconstruction by one-step linearization in EIT Exact shape-reconstruction by one-step linearization in EIT Bastian von Harrach harrach@ma.tum.de Department of Mathematics - M1, Technische Universität München, Germany Joint work with Jin Keun Seo, Yonsei

More information

Final Projects Introduction to Numerical Analysis Professor: Paul J. Atzberger

Final Projects Introduction to Numerical Analysis Professor: Paul J. Atzberger Final Projects Introduction to Numerical Analysis Professor: Paul J. Atzberger Due Date: Friday, December 12th Instructions: In the final project you are to apply the numerical methods developed in the

More information

1 Precautionary Savings: Prudence and Borrowing Constraints

1 Precautionary Savings: Prudence and Borrowing Constraints 1 Precautionary Savings: Prudence and Borrowing Constraints In this section we study conditions under which savings react to changes in income uncertainty. Recall that in the PIH, when you abstract from

More information

Maximum Contiguous Subsequences

Maximum Contiguous Subsequences Chapter 8 Maximum Contiguous Subsequences In this chapter, we consider a well-know problem and apply the algorithm-design techniques that we have learned thus far to this problem. While applying these

More information

Optimization for Chemical Engineers, 4G3. Written midterm, 23 February 2015

Optimization for Chemical Engineers, 4G3. Written midterm, 23 February 2015 Optimization for Chemical Engineers, 4G3 Written midterm, 23 February 2015 Kevin Dunn, kevin.dunn@mcmaster.ca McMaster University Note: No papers, other than this test and the answer booklet are allowed

More information

Regression estimation in continuous time with a view towards pricing Bermudan options

Regression estimation in continuous time with a view towards pricing Bermudan options with a view towards pricing Bermudan options Tagung des SFB 649 Ökonomisches Risiko in Motzen 04.-06.06.2009 Financial engineering in times of financial crisis Derivate... süßes Gift für die Spekulanten

More information

Final Projects Introduction to Numerical Analysis atzberg/fall2006/index.html Professor: Paul J.

Final Projects Introduction to Numerical Analysis  atzberg/fall2006/index.html Professor: Paul J. Final Projects Introduction to Numerical Analysis http://www.math.ucsb.edu/ atzberg/fall2006/index.html Professor: Paul J. Atzberger Instructions: In the final project you will apply the numerical methods

More information

Online Supplement: Price Commitments with Strategic Consumers: Why it can be Optimal to Discount More Frequently...Than Optimal

Online Supplement: Price Commitments with Strategic Consumers: Why it can be Optimal to Discount More Frequently...Than Optimal Online Supplement: Price Commitments with Strategic Consumers: Why it can be Optimal to Discount More Frequently...Than Optimal A Proofs Proof of Lemma 1. Under the no commitment policy, the indifferent

More information

On a Manufacturing Capacity Problem in High-Tech Industry

On a Manufacturing Capacity Problem in High-Tech Industry Applied Mathematical Sciences, Vol. 11, 217, no. 2, 975-983 HIKARI Ltd, www.m-hikari.com https://doi.org/1.12988/ams.217.7275 On a Manufacturing Capacity Problem in High-Tech Industry Luca Grosset and

More information

Advanced Probability and Applications (Part II)

Advanced Probability and Applications (Part II) Advanced Probability and Applications (Part II) Olivier Lévêque, IC LTHI, EPFL (with special thanks to Simon Guilloud for the figures) July 31, 018 Contents 1 Conditional expectation Week 9 1.1 Conditioning

More information

F A S C I C U L I M A T H E M A T I C I

F A S C I C U L I M A T H E M A T I C I F A S C I C U L I M A T H E M A T I C I Nr 38 27 Piotr P luciennik A MODIFIED CORRADO-MILLER IMPLIED VOLATILITY ESTIMATOR Abstract. The implied volatility, i.e. volatility calculated on the basis of option

More information

University of Toronto Department of Economics ECO 204 Summer 2013 Ajaz Hussain TEST 1 SOLUTIONS GOOD LUCK!

University of Toronto Department of Economics ECO 204 Summer 2013 Ajaz Hussain TEST 1 SOLUTIONS GOOD LUCK! University of Toronto Department of Economics ECO 204 Summer 2013 Ajaz Hussain TEST 1 SOLUTIONS TIME: 1 HOUR AND 50 MINUTES DO NOT HAVE A CELL PHONE ON YOUR DESK OR ON YOUR PERSON. ONLY AID ALLOWED: A

More information

FX Smile Modelling. 9 September September 9, 2008

FX Smile Modelling. 9 September September 9, 2008 FX Smile Modelling 9 September 008 September 9, 008 Contents 1 FX Implied Volatility 1 Interpolation.1 Parametrisation............................. Pure Interpolation.......................... Abstract

More information

GAME THEORY. Department of Economics, MIT, Follow Muhamet s slides. We need the following result for future reference.

GAME THEORY. Department of Economics, MIT, Follow Muhamet s slides. We need the following result for future reference. 14.126 GAME THEORY MIHAI MANEA Department of Economics, MIT, 1. Existence and Continuity of Nash Equilibria Follow Muhamet s slides. We need the following result for future reference. Theorem 1. Suppose

More information