Keywords: evaluation complexity, worst-case analysis, least-squares, constrained nonlinear optimization, cubic regularization methods.

Size: px
Start display at page:

Download "Keywords: evaluation complexity, worst-case analysis, least-squares, constrained nonlinear optimization, cubic regularization methods."

Transcription

1 On the evaluation complexity of cubic regularization methos for potentially rank-icient nonlinear least-squares problems an its relevance to constraine nonlinear optimization Coralia Cartis, Nicholas I. M. Goul an Philippe L. Toint March 21, 2012 Abstract We propose a new termination criteria suitable for potentially singular, zero or non-zero resiual, least-squares problems, with which cubic regularization variants take at most O(ǫ 3/2 ) resiual- an Jacobian-evaluations to rive either the Eucliean norm of the resiual or its graient below ǫ; this is the best-known boun for potentially rank-icient nonlinear least-squares problems. We then apply the new optimality measure an cubic regularization steps to a family of least-squares merit functions in the context of a target-following algorithm for nonlinear equality-constraine problems; this approach yiels the first evaluation complexity boun of orer ǫ 3/2 for nonconvexly constraine problems when higher accuracy is require for primal feasibility than for ual first-orer criticality. Keywors: evaluation complexity, worst-case analysis, least-squares, constraine nonlinear optimization, cubic regularization methos. 1 Introuction An ubiquitous challenge in scientific computing is the minimization of an appropriate norm of a given, sufficiently smooth, vector-value function r : IR n IR m. This problem formulation arises in numerous real-life applications requiring ata fitting, parameter estimation, image reconstruction, weather forecasting an so forth [23]. Crucially, it is often an essential builing block when solving constraine nonlinear programming problems, being use for example, to reuce the constraint violation in various sequential programming [2, 12, 24 26], filter [16], funnel [18] an re-weighte least squares approaches [23]. Nonlinear least-squares problems are also at the heart of the path-following metho for constraine problems which we propose an analyze here, as well. Here we focus on the Eucliean-norm case that gives rise to the equivalent nonlinear least-squares problem, min x IR nφ(x) = 2 r(x) 1 2, (1.1) now involving the smooth function Φ(x); other norms may be of interest an some are equally acceptable in this framework. We allow arbitrary values for m an n, an so both over- an uner-etermine resiuals r(x) are allowe in (1.1), as well as square nonlinear systems of equations; in the latter two cases, one may wish to reuce Φ(x) in (1.1) to zero so as to fin the zeros of the system r(x) = 0. Methos for solving (1.1) iffer not only in their practical performance, but also in the theoretical bouns known on their worst-case efficiency, which is the focus of this paper. Of the various methos propose, Gauss-Newton techniques are the most popular an well-researche[15,23]. Rather than tackling School of Mathematics, University of Einburgh, The King s Builings, Einburgh, EH9 3JZ, Scotlan, UK. coralia.cartis@e.ac.uk. All three authors are grateful to the Royal Society for its support through the International Joint Project Computational Science an Engineering Department, Rutherfor Appleton Laboratory, Chilton, Oxforshire, OX11 0QX, Englan, UK. nick.goul@stfc.ac.uk. Department of Mathematics, FUNDP - University of Namur, 61, rue e Bruxelles, B-5000, Namur, Belgium. philippe.toint@funp.ac.be. 1

2 2 C. Cartis, N. I. M. Goul an Ph. L. Toint the smooth formulation (1.1), recent algorithmic variants [1, 8, 21] attempt to minimize the un-square an hence nonsmooth, norm of r(x) instea, in an attempt to improve the conitioning of the system that ines the change to the iterates. Using only first-orer information namely, values of the resiual r(x) an its Jacobian J(x) at given x, obtaine from a so-calle black-box/oracle both classical an moern variants can be mae/shown to be globally convergent to stationary points of (1.1), namely to points satisfying x Φ(x) = J(x) T r(x) = 0; (1.2) furthermore, the number of resiual an Jacobian evaluations require to bring the norm of (1.2) or some (nonsmooth) first-orer optimality measure within some tolerance ǫ is O(ǫ 2 ), provie J(x) an r(x) are Lipschitz continuous [1, 8, 13, 21, 23]. Another possibility is to apply Newton-type methos to the unconstraine problem (1.1), which can ensure for example, fast local convergence for nonzero resiual problems an most importantly here, improve global efficiency for both zero- an non-zero resiual problems. In particular, cubic regularization methos [9,17,22] applie to (1.1) take O(ǫ 3/2 ) resiual evaluations to ensure (1.2) is within ǫ, provie r(x), J(x) an the Hessians xx r i (x), i = 1,...,m, are Lipschitz continuous; this boun is sharp for nonlinear least-squares [11], is optimal from a worstcase complexity point of view for a wie class of secon-orer methos an nonconvex unconstraine problems [4], an is the best-known complexity for secon-orer methos. This boun can be further improve for graient-ominate resiuals (such as when the singular values of the Jacobian are uniformly boune away from, or converge to, zero at the same rate as the resiual) [22]. The (natural) approximate satisfaction of (1.2) as termination criteria for the cubic regularization an other methos suffers from the isavantage that an approximate zero of r(x) is guarantee only when J(x) is uniformly full-rank, with a known lower boun on its smallest singular value this is a strong assumption. In this paper, we introuce a termination conition that can istinguish between the zero an non-zero resiual case automatically/implicitly. Namely, we argue for the use of a scale variant of (1.2), which is precisely the graient of r(x) whenever r(x) 0, as well as the inclusion of the size of the resiual in the termination conition. Without requiring a non-egenerate Jacobian, we then show that cubic regularization methos can generate either an approximate scale graient or resiual value within ǫ in at most O(ǫ 3/2 ) resiual-evaluations, thus preserving the (optimal) orer of the boun for cubic regularization. Consier now the evaluation complexity of minimizing a smooth but potentially nonconvex objective f(x) IR for x C. When C is escribe by finitely many smooth (but potentially nonconvex) equality an inequality constraints, we have shown that a first-orer exact penalty metho with boune penalty parameters [8], as well as a short-step target-following algorithm with steepest-escent-like steps [3], take O(ǫ 2 ) objective an constraint evaluations to generate an approximate KKT point of the problem or an infeasible point of the feasibility measure with respect to the constraints. Thus aing constraints oes not eteriorate the orer of worst-case evaluation complexity boun achieve in the unconstraine case when steepest-escent like methos are employe. A natural question arises as to whether an improve evaluation complexity boun, of the orer of cubic regularization, can be shown for constraine problems. In the case when C is given by convex constraints, projecte cubic regularization variants can be shown to satisfy the O(ǫ 3/2 ) evaluation boun [6]. In this paper, in a similar vein to [3], we propose a short-step target-following algorithm for problems with nonconvex equality constraints minimize f(x) such that c(x) = 0, that takes cubic regularization steps for a sequence of shifting least-squares merit functions. The evaluation complexity of the resulting algorithm is better than that for steepest-escent, an can even achieve O(ǫ 3/2 ), provie the (ual) KKT conitions are satisfie with lower accuracy than the (primal) feasibility with respect to the constraints. The structure of the paper is as follows. Section 2 summarizes aaptive cubic regularization methos [9] an relevant complexity results. Section 3.1 presents the new termination criteria for (1.1) base on the scale graient, while Section 3.2 gives the complexity result for cubic regularization applie to

3 Evaluation complexity of nonlinear least-squares an constraine problems 3 (1.1) with the new termination criteria. Sections 4 an 5 present the short-step target-following cubic regularization algorithm for the equality-constraine problem an its complexity analysis, respectively. Section 6 summarizes our contributions an iscusses possible extensions of this work. 2 Previous cubic regularization construction an results 2.1 Description of aaptive cubic regularization algorithm We consier applying the Aaptive Regularization with Cubics (ARC) algorithm [9, 10] to (1.1); here, we focus on the ARC variant that has the best known an optimal worst-case evaluation complexity, so-calle ARC (S). At each iterate x k, k 0, a step s k is compute that approximately minimizes the local cubic moel m k (s) = 1 2 r(x k ) 2 +s T J(x k ) T r(x k )+ 1 2s T B k s+ 1 3σ k s 3 (2.1) of Φ(x k +s) with respect to s, where B k is an approximation to the Hessian of Φ at x k an σ k > 0 is a regularization parameter. In this metho, the step s k is compute to satisfy s T kj(x k ) T r(x k )+s T kb k s k +σ k s k 3 = 0 (2.2) an s T kb k s k +σ k s k 3 0. (2.3) Conitions (2.2) an (2.3) are achieve whenever s k is a global minimizer of the moel m k along the irection s k, namely, arg min α IR m k (αs k ) = 1; in particular, they are satisfie whenever s k is a global minimizer of the moel m k over a(ny) subspace [10, Theorem 3.1, Lemma 3.2]. Note that if s k is chosen as the global minimizer of m k over the entire space, σ k is maintaine at a sufficiently large value an B k is the true Hessian, then ARC (S) is similar to the cubic regularization technique propose in [22]. To ensure ARC s fast local convergence, we nee to go beyon uniimensional minimization, an so we terminate the inner moel minimization when TC.s s m k (s k ) κ θ min{1, s k } J(x k ) T r(x k ), (2.4) where κ θ is any constant in (0,1); see [10, 3.2] for a etaile escription of this an other possible termination conitions. Note that s m k (0) = x Φ(x k ) = J(x k ) T r(x k ) so that (2.4) is a relative error conition, which is clearly satisfie at any minimizer s k of m k since then s m k (s k ) = 0. Generally, we hope that the inner minimization will be terminate before this inevitable outcome. Note that when s k is compute by minimizing m k over a subspace, we may increase the subspace of minimization until TC.s is satisfie. In particular, one may use a Lanczos-base approach where the subspace is the Krylov one generate by { x Φ(x k ), B k x Φ(x k ), B 2 k xφ(x k )... }. In this case, conitions (2.2) an (2.3) are also achieve [10, 3.2, 6, 7]. It remains to escribe the iterate upating an moel improvement technique in ARC. The step s k is accepte an the new iterate x k+1 set to x k +s k whenever (a reasonable fraction of) the preicte moel ecrease Φ(x k ) m k (s k ) is realize by the actual ecrease in the objective, Φ(x k ) Φ(x k +s k ). This is measure by computing the ratio ρ k in (2.5) an requiring ρ k to be greater than a prescribe positive constant η 1 (for example, η 1 = 0.1); it can be shown that ρ k is well-ine whenever x Φ(x k ) 0 [10, Lemma 2.1]. Since the current weight σ k has resulte in a successful step, there is no pressing reason to increase it, an inee there may be benefits in ecreasing it if the moel overestimates the function locally. By contrast, if ρ k is smaller than η 1, we juge that the improvement in objective is insufficient inee there is no improvement if ρ k 0. If this happens, the step will be rejecte an x k+1 left as x k. Uner these circumstances, the only recourse available is to increase the weight σ k prior to the next iteration with the implicit intention of reucing the size of the step. A summary of the ARC (S) algorithm applie to (1.1) follows.

4 4 C. Cartis, N. I. M. Goul an Ph. L. Toint Algorithm 2.1: Aaptive Regularization using Cubics (ARC (S) ) [9,10] applie to (1.1). A starting point x 0, an initial an a minimal regularization parameter σ 0 σ min > 0, an algorithmic parameters γ 2 γ 1 > 1 an 1 > η 2 η 1 > 0, are given. For k = 0,1,..., until termination, o: 1. Compute a step s k that satisfies (2.2) (2.4). 2. Compute r(x k +s k ) an ρ k = 1 2 r(x k ) 2 2 r(x 1 k +s k ) r(x k ) 2. (2.5) m k (s k ) 3. Set { xk +s x k+1 = k if ρ k η 1 otherwise. x k 4. Set σ k+1 [σ min,σ k ] if ρ k > η 2 [very successful iteration] [σ k,γ 1 σ k ] if η 1 ρ k η 2 [successful iteration] [γ 1 σ k,γ 2 σ k ] otherwise. [unsuccessful iteration] (2.6) Note that we have not yet ine the conition require for ARC (S) to terminate. In [9,10], we terminate ARC when x Φ(x k ) ǫ, an possibly also λ min ( xx Φ(x k )) ǫ, for a user-specifie tolerance ǫ (0,1). Here, we will require that either some scale graient or the resiual is within ǫ; this novel termination conition, specific to (1.1), is escribe in Section Assumptions an useful results The following assumptions are chosen to ensure that those in [9,10] are satisfie when ARC (S) is applie to (1.1), which allows us to employ some crucial ARC results from [9,10] to (1.1). Let X be an open convex set containing all the generate iterates {x k,x k + s k }, k 0. We assume that AR.1 r i C 2 (IR n ) an r i (x) is uniformly boune above on X, i {1,...,m}. (2.7) For each i {1,...,m}, the resiuals r i are Lipschitz continuous on X, namely, AR.2 r i (x) r i (y) κ ri x y, for all x, y X, an some κ ri 1. (2.8) This implies that r is Lipschitz continuous on X, with Lipschitz constant κ r = (κ r1,...,κ rm ) 1. We also assume that the Jacobian J of r is Lipschitz continuous on X, namely AR.3 J(x) J(y) κ J x y, for all x, y X, an some κ J > 0. (2.9) Note that AR.1 AR.3 imply that the graient x Φ given in (1.2) is Lipschitz continuous on X with Lipschitz constant L g = κ 2 r +r max κ J 1, (2.10) where r max > 0 enotes an upper boun on r(x), x X. (This is assumption AF.4 in [9,10].) For each i {1,...,m}, the Hessian xx r i is also assume to be globally Lipschitz continuous on the path of all generate iterates, namely, there exists a constant L i such that AR.4 xx r i (x) xx r i (x k ) L i x x k, for all x [x k,x k +s k ] an all k 0. (2.11)

5 Evaluation complexity of nonlinear least-squares an constraine problems 5 Note that AR.1 AR.4 imply that the Hessian of Φ xx Φ(x) = J(x) T J(x)+ n r i (x) xx r i (x) (2.12) is globally Lipschitz continuous on the path of all generate iterates, with Lipschitz constant L n n = 2κ J κ r +κ J κ ri + r(x 0 ) L i, (2.13) i=1 where we also use that ARC generates monotonically ecreasing function values so that r(x k ) r(x 0 ). (This is assumption AF.6 in [9,10].) Clearly, the values of the resiual r(x k ) an its Jacobian J(x k ) are require to form the moel (2.1) an estimate (2.5). Thus, as B k is an approximation to the Hessian of Φ in (2.12) at x k, only the Hessian of each r i nees to be approximate in B k an so it is natural to consier B k to be of the form where M k i=1 i=1 B k = J(x k ) T J(x k )+M k, (2.14) H Φ (x k ) = We require that M k an H Φ (x k ) in (2.15) agree along s k in the sense that n r i (x k ) xx r i (x k ). (2.15) i=1 AM.4 (H Φ (x k ) M k )s k C s k 2, for all k 0, an some constant C > 0. (2.16) This, (2.12) an (2.14) imply that [ xx Φ(x k ) B k ]s k C s k 2, for all k 0, (2.17) which is assumption AM.4 in [9,10]. The conition AM.4 is trivially satisfie with C = 0 when we set M k = H Φ (x k ) i.e., B k = xx Φ(x k ) for all k 0 in the ARC algorithm. The requirement (2.16) or (2.17) is a slight strengthening of the Dennis Moré conition [14]. The latter is achieve by some quasi- Newton upates provie further assumptions hol(see the iscussion following[10,(4.6)]). Quasi-Newton methos may still satisfy AM.4 in practice, though we are not aware if this can be ensure theoretically. We have shown in [7] that AM.4 can be achieve when B k is approximate by (forwar) finite ifferences of graient values, without changing the orer of the worst-case evaluation complexity boun as a function of the accuracy ǫ. The first lemma recalls some useful ARC properties, crucial to the complexity boun in Section 3.2. Lemma 2.1. Let AR.1 AR.4 an AM.4 hol, an apply Algorithm ARC (S) to (1.1). Then an so σ k 3 2(L+C) = k is very successful, (2.18) σ k max(σ 0, 3 2γ 2 (L+C)) = σ, for all k 0, (2.19) where L an C are ine in (2.13) an (2.16), respectively. Also, we have the function ecrease 1 2 r(x k ) r(x k+1 ) 2 α J(xk+1 ) T r(x k+1 ) 3/2 for all successful iterations k, (2.20) where α = η 1 σ min κ 3 g/6 an where κ g is the positive constant κ g = (1 κ θ )/( 1 2L+C +σ +κ θ L g ), (2.21) with κ θ, σ an L g ine in (2.4), (2.19) an (2.10), respectively.

6 6 C. Cartis, N. I. M. Goul an Ph. L. Toint Proof. The relation (2.18) an the boun (2.19) both follow from [10, Lemma 5.2], an (2.20) from (2.5), σ k σ min (ue to (2.6)), [10, Lemmas 3.3] an [9, Lemma 5.2]. Relating successful an total iteration counts The total number of (major) ARC iterations is the same as the number of resiual/function evaluations (as we also nee to evaluate r on unsuccessful iterations in orer to be able to compute ρ k in (2.5)), while the number of successful ARC iterations is the same as that of Jacobian/graient evaluations. Let us introuce some useful notation. Throughout, enote the inex set S = {k 0 : k successful or very successful in the sense of (2.6)}, (2.22) an, given any j 0, let S j = {k j : k S}, (2.23) with S j enoting the carinality of the latter. The lower boun on σ k an the construction of Steps 2 4 of ARC (S) allow us to quantify the total iteration count as a function of the successful ones. Theorem 2.2. For any fixe j 0, let S j be ine in (2.23). Assume that there exists σ > 0 be such that σ k σ, for all k j. (2.24) Then j 1+ 2 ( ) σ log S j. (2.25) logγ 1 σ min Proof. The upates (2.6) imply that σ k σ min for all k. Now apply [9, Theorem 2.1], namely, the boun [9, (2.14)] on the number of unsuccessful iterations up to j, an use the fact that the unsuccessful iterations up to j together with S j form a partition of {0,...,j}. Values for σ in (2.24) are provie in (2.19), uner the assumptions of Lemma 2.1. Thus, base on Theorem 2.2, it remains to boun the successful iteration count S j since the total iteration count up to j is of the same orer in ǫ as S j. 3 Evaluation complexity of cubic regularization for potentially rank-icient nonlinear least-squares problems 3.1 A suitable termination conition for ARC (S) Here, we epart from the stanar choice of termination criterion for erivative-base optimization algorithms such as ARC (S) when applie to (1.1), namely, requiring a sufficiently small graient Φ x (x k ) = J(x k ) T r(x k ) ǫ, where ǫ > 0 is the user-specifie accuracy tolerance. Such a conition is only guaranteetoprovieanapproximatezerooftheresiualr whenj(x)isuniformlyfull-rankanalowerbounon its smallest singular values is known, which are limiting assumptions. Such assumptions are not require for steepest-escent-like methos if appropriate optimality measures are employe [3, 8], but the complexity of such methos is worse than the best secon-orer methos [8, 11]. Thus, we introuce a termination conition that can istinguish between the zero an non-zero resiual case automatically/implicitly. We propose the following termination for ARC (S), termination : r(x k ) ǫ p or g r (x k ) ǫ, (3.1)

7 Evaluation complexity of nonlinear least-squares an constraine problems 7 where ǫ p > 0 an ǫ > 0 are the require accuracy tolerances an where J(x) T r(x) g r (x), whenever r(x) 0; = r(x) 0, otherwise. (3.2) Note that the scale graient g r (x) in (3.2) is precisely the graient of r(x) whenever r(x) 0. If r(x) = 0, we are at the global minimum of r an so g r (x) = 0 ( r(x) ) [19, VI.3]. In the termination conition (3.1), the scale graient g r (x k ) may be boune away from zero for instance, when the singular values of the Jacobian are uniformly boune away from zero then, as we show in the next section, the resiual values converge to zero, an so (3.1) can be achieve. When the iterates approach a nonzero resiual value, then g r converges to zero an so again, (3.1) can be satisfie. Another suitable termination conition with similar properties is given after the main result in the next section. Inthenextsection, weshowthatarc (S) cangenerateeitheranapproximatescalegraientorresiual value within ǫ in at most O(ǫ 3/2 ) resiual-evaluations, thus preserving the (optimal) orer of the boun for cubic regularization. 3.2 Evaluation complexity of ARC (S) with termination conition (3.1) The first lemma exploits (2.20) to give new lower bouns on the function ecrease that epen on the resiual an the scale graient (3.2); the bouns below will also be use for the constraine case. Lemma 3.1. Let AR.1 AR.4 an AM.4 hol, an apply the ARC (S) algorithm to (1.1). Then, for all successful iterations k for which r(x k ) 0, we have { } r(x k ) r(x k+1 ) min αβ 3/2 g r (x k+1 ) 3/2 r(x k ) 1/2, (1 β) r(x k ) (3.3) an r(x k ) 1/2 r(x k+1 ) 1/2 min { 1 2αβ 3/2 g r (x k+1 ) 3/2,(β 1/2 1) r(x k+1 ) 1/2}, (3.4) where α is ine just after (2.20) an β (0,1) is any fixe problem-inepenent constant. Proof. Suppose that r(x k ) 0, let β (0,1) an enote S β = {k S : r(x k+1 ) > β r(x k ) }, (3.5) where S is ine in (2.22). We first analyze the function ecrease for iterations k S β an then, for the ones in S \S β. Let k S β ; then r(x k+1 ) 0 since r(x k ) 0. From (2.20), (3.2) an (3.5), we euce r(x k ) 2 r(x k+1 ) 2 ( 2α J(x k+1 ) T r(x k+1 ) 3/2 J(xk+1 ) T ) 3/2 r(x k+1 ) = 2α r(x k+1) 3/2 r(x k+1 ) = 2α g r (x k+1 ) 3/2 r(x k+1 ) 3/2 2αβ 3/2 g r (x k+1 ) 3/2 r(x k ) 3/2. Conjugacy properties an the monotonicity relation r(x k ) r(x k+1 ) give r(x k ) r(x k+1 ) = r(x k) 2 r(x k+1 ) 2 r(x k ) + r(x k+1 ) r(x k) 2 r(x k+1 ) 2 2 r(x k ) (3.6) (3.7)

8 8 C. Cartis, N. I. M. Goul an Ph. L. Toint an furthermore r(xk ) r(x k+1 ) = r(x k) r(x k+1 ) r(xk ) + r(x k+1 ) r(x k) 2 r(x k+1 ) 2 4 r(x k ) 3/2. (3.8) Employing the last inequality in (3.6) into (3.7) an (3.8), respectively, we obtain an r(x k ) r(x k+1 ) αβ 3/2 g r (x k+1 ) 3/2 r(x k ) 1/2, for all k S β, (3.9) r(x k ) 1/2 r(x k+1 ) 1/2 αβ3/2 g r (x k+1 ) 3/2, for all k S β. (3.10) 2 Conversely, let k S \S β, which gives r(x k+1 ) β r(x k ), (3.11) an so the resiual values ecrease linearly on such iterations. It follows from (3.11) that on such iterations we have the following function ecrease an r(x k ) r(x k+1 ) (1 β) r(x k ) for all k S \S β. (3.12) r(x k ) 1/2 r(x k+1 ) 1/2 (1 β) r(x k ) 1/2 1 β β r(x k+1 ) 1/2 for all k S \S β. (3.13) (Note that (3.12) an (3.13) continue to hol if r(x k+1 ) = 0.) The boun (3.3) now follows from (3.9) an (3.12), an (3.4) from (3.10) an (3.13). The next theorem gives a general evaluation complexity result for ARC (S) applie to (1.1) when the termination conition (3.1) is employe. Theorem 3.2. LetAR.1 AR.4anAM.4hol, anletǫ p, ǫ (0,1). ConsierapplyingtheARC (S) algorithm with the termination conition (3.1) to minimizing (1.1). Then ARC (S) terminates after at most max{κ 1 ǫ 3/2,κ 2 ǫ 1/2 p } +1 (3.14) successful iterations or equivalently, Jacobian-evaluations an at most κ S max{κ 1 ǫ 3/2,κ 2 ǫ 1/2 p } +1 (3.15) total (successful an unsuccessful) iterations or equivalently, resiual-evaluations, where κ 1 = 2 r(x 0 ) 1/2 α 1 β 3/2, κ 2 = r(x 0 ) 1/2 (β 1/2 1) 1, (3.16) κ S = 2(1+κ u S) an κ u S = 2log(σ/σ min )/logγ 1, (3.17) with α ine just after (2.20), σ, in (2.19), an β (0,1) a fixe problem-inepenent constant. Proof. Clearly, if (3.1) is satisfie at the starting point, there is nothing left to prove. Assume now that (3.1) fails at k = 0. For any iteration (k+1) at which ARC (S) oes not terminate, it follows from (3.1) that we have r(x k+1 ) > ǫ p an g r (x k+1 ) > ǫ. (3.18)

9 Evaluation complexity of nonlinear least-squares an constraine problems 9 From (3.4) an (3.18), we euce r(x k ) 1/2 r(x k+1 ) 1/2 min{ 1 2αβ 3/2 ǫ 3/2,(β 1/2 1) p } for all k S for which (3.18) hols. (3.19) Summing up (3.19) over all iterations k S for which (3.18) hols, with say j ǫ as the largest inex, an using that the ARC (S) iterates remain unchange over unsuccessful iterations, we obtain r(x 0 ) 1/2 r(x jǫ ) 1/2 = j ǫ 1 k=0,k S S ǫ min [ r(x k ) 1/2 r(x k+1 ) 1/2] { } 1 2αβ 3/2 ǫ 3/2,(β 1/2 1) p (3.20) where S ǫ enotes the number of successful iterations up to iteration j ǫ. Using that r(x jǫ ) 1/2 0, we further obtain from (3.20) that j ǫ < an that r(x 0 ) 1/2 S ǫ { }, 1 min 2αβ 3/2 ǫ 3/2,(β 1/2 1) p which gives (3.14) since S ǫ must be an integer an since the termination conition is checke at the next iteration; see [9, (5.21), (5.22)] for full etails. To erive (3.15), apply Theorem 2.2 with j = j ǫ, σ ine in (2.19), an use also that ǫ p, ǫ (0,1). The next corollary gives the main complexity result of this section, whose proof follows immeiately from Theorem 3.2. It shows that the evaluation complexity of ARC (S) riving either r(x) or its graient below ǫ is O(ǫ 3/2 ), an improvement of existing ARC (S) results which can only ensure that the graient of r(x) 2 goes below ǫ in that same-orer number of evaluations. Corollary 3.3. Let AR.1 AR.4 an AM.4 hol, an let ǫ = min{ǫ p,ǫ } (0,1). Consier applying the ARC (S) algorithm with the termination conition (3.1) to minimizing (1.1). Then ARC (S) terminates after at most κ s Sǫ 3/2 +1 (3.21) successful iterations or equivalently, Jacobian-evaluations an at most κ S κ s Sǫ 3/2 +1 (3.22) total (successful an unsuccessful) iterations or equivalently, resiual-evaluations, where κ s S = r(x 0 ) 1/2 /min{ 1 2αβ 3/2,β 1/2 1}, (3.23) with α ine just after (2.20), κ S, in (3.17) an β (0,1) a fixe problem-inepenent constant. Some remarks on the above theorem/corollary an its proof follow: Note that in the non-zero resiual case, namely, when { r(x k ) } converges to some r > 0, the monotonicity of this sequence implies that r(x k+1 ) β r(x k ) for all k, with β = r / r(x 0 ) (0,1). Thus in this case, there is no nee to consier the iterations (3.11) of faster linear convergence. The function ecrease in (3.4) implies that instea of (3.1), we coul have use the conition termination 2 : r(x k ) 1/3 ǫ p or g r (x k ) ǫ, (3.24)

10 10 C. Cartis, N. I. M. Goul an Ph. L. Toint as termination for the ARC (S) algorithm, without changing the orer of the complexity boun as a function of (ǫ p,ǫ ) or even of ǫ = min{ǫ p,ǫ }. In fact, using the conition (3.24) improves the boun/accuracy for the resiual values reaching within ǫ p. Note that the boun (3.14) is a boun on the total number of successful iterations for which (3.18) hols. Thus espite the measure (3.1) being non-monotonic, after (3.14) iterations are taken, this measure woul remain below (ǫ p,ǫ ) for the remaining ARC (S) iterations, if any are taken. The use of conjugacy in the above proof is reminful of re-weighte least-squares techniques [23]. However, our attempts at applying (some moifie) ARC to such variants of (1.1) have not been successful. 3.3 Is the boun (3.15) sharp for the nonlinear least-squares problem (1.1)? Recall the example in [11, 5] that shows that ARC (S) takes essentially ǫ 3/2 iterations/evaluations to ensure that the norm of the graient is less than ǫ. The univariate function f : IR IR in question is positive for all x 0 an at the iterates, an it is zero at infinity, minimum to which ARC (S) converges. Thus this example can be viewe as a least-squares, zero-resiual problem, with r in (1.1) ine as r = f. It shows that ARC (S) with the termination conition that the absolute value of (1.2) which in this case, is precisely the first erivative of f is less than ǫ takes essentially ǫ 3/2 iterations/evaluations an so the ARC (S) complexity boun is sharp for nonlinear least-squares. (Note that although f(x) an its erivatives may not be globally Lipschitz continuous as x, the first an secon erivatives of r 2 = f have this property, as we have shown in [11, 5]. The latter conitions are sufficient for the O(ǫ 3/2 ) boun to hol for ARC (S).) It is unclear whether the boun (3.15) for ARC (S) with the termination conition (3.1) is also sharp. 3.4 Further improving the evaluation complexity of cubic regularization for nonlinear least-squares with special structure Suppose that r(x) in (1.1) is graient-ominate of egree 2 [22], namely, J(x) T r(x) r(x) σ min (J(x)) τ 2 > 0, x IR n, (3.25) where σ min (J(x)) enotes the smallest singular value of J(x); this implies that g r in (3.1) is boune away from zero for all r(x) 0. Then uner the conitions of Theorem 3.2, one can euce from (3.4) an (3.19) that r(x k ) must converge to zero as k, an that the asymptotic rate of this convergence is superlinear (i.e., linear with any convergence factor β (0, 1)); also, the algorithm takes a (problemepenent) constant number of steps to enter this region of superlinear convergence. We o not give the etails of this result here as a (slightly stronger) result of this form where the size of the neighbourhoo of fast local convergence oes not epen on β an r(x 0 ) enters the complexity boun in a polynomial way was given in [22, Theorem 7] for cubic regularization; the latter result continues to hol here for ARC (S) when applie to problems which we know a-priori satisfy (3.25) since then (3.1) is no longer require explicitly. An avantage of our (slightly weaker) approach here is that the termination conition (3.1) senses naturally when (3.25) hols an ensures ARC (S) behaves accoringly. Similarly, assume now that the smallest singular value of the Jacobian of r(x) converges to zero at the same rate as r(x), or that there exists τ 1 > 0 such that J(x) T r(x) / r(x) τ 1 r(x) for all x, which is the same as r(x) being graient-ominate of egree 1 [22]. Then again we can euce improve complexity bouns from (3.4) in the same vein as [22, Theorem 6], giving that ARC (S) requires at most O ( ǫ 1) evaluations to ensure r(x k ) ǫ. (Note the unerstanably weaker boun in this case since we minimize the square of the resiual, when compare to the ARC boun of orer O ( ǫ 1/2) for minimizing general unconstraine graient-ominate functions of egree 1 [5, 22].) The cases of graient-ominate resiuals of some intermeiate egree with value between 1 an 2, can be similarly analyze, yieling improvement over the boun (3.15).

11 Evaluation complexity of nonlinear least-squares an constraine problems 11 4 The ShS-ARC algorithm for equality-constraine problems Consier now the equality constraine problem where f : IR n IR an c : IR n IR m with m n. We assume that minimize f(x) such that c(x) = 0, (4.1) AC.1 The function c is twice continuously ifferentiable on IR n an f is twice continuously ifferentiable in an open neighbourhoo of C 1 = {x IR n c(x) κ c }, where κ c > 0 is (small) constant. (4.2) The algorithm we now escribe consists of two phases; see Figure 4.1 (a). In the first, ARC (S) with termination conition (3.1) is applie to (1.1) with r = c, so as to minimize 2 c(x) 1 2 (inepenently of the objective function f), resulting in a point which is either (approximately) feasible, or is an approximate infeasible stationary point of c(x). The latter outcome is not esirable if one wishes to solve (4.1), but cannot be avoie by any algorithm not relying on global minimization or if C 1 is empty. If an (approximate) feasible point has been foun, Phase 2 of the algorithm then performs short cubic regularization steps for a parametrize family of least-squares functions so long as first-orer criticality is not attaine. These steps are compute by attempting to preserve approximate feasibility of the iterates while proucing values of the objective function that are close to a sequence of ecreasing targets. To be specific, one or more ARC (S) iterations are applie to minimizing the least-squares function Φ(x,t) = 1 2 r(x,t) 2 with respect to x, where ( r(x,t) c(x) = f(x) t an where t is a target value for f(x). Clearly, the Jacobian A(x,t) of the resiual function r(x,t) in (4.3) satisfies ( ) A(x,t) J(x) = A(x) =, (4.4) g(x) where J(x) is the Jacobian of the constraint function c(x) an g(x) is the graient of f(x). Thus x Φ(x,t) = A(x,t) T r(x,t) an the scale graient (3.2) has the expression A(x,t) T r(x,t) g r (x,t) = r(x,t) 0, otherwise. ) = J(x)T c(x)+(f(x) t)g(x), whenever r(x,t) 0; r(x,t) We are now reay to summarize our Short-Step ARC (ShS-ARC) algorithm. Algorithm 4.1: The Short-Step ARC (ShS-ARC) algorithm for (4.1). A starting point x 0, initial regularization parameters σ 0 an σ 1 an a minimal one σ min such that min{σ 0,σ 1 } σ min > 0, an algorithmic parameters γ 2 γ 1 > 1 an 1 > η 2 η 1 > 0, as well as the tolerances ǫ p (0,1) an ǫ (0,1), are given. Phase 1: Starting from x 0, apply ARC (S) to minimize 2 c(x) 1 2 until a point x 1 is foun such that (3.1) is satisfie, namely, J(x 1 ) T c(x 1 ) c(x 1 ) ǫ p or ǫ. (4.6) c(x 1 ) If c(x 1 ) > ǫ p, terminate [locally infeasible]. (4.3) (4.5)

12 12 C. Cartis, N. I. M. Goul an Ph. L. Toint Phase 2: 1. Set t 1 = f(x 1 ) 2. For k = 1,2,..., o: ǫ 2 p c(x 1 ) 2 an k = 1. 2a. Startingfromx k, applyoneiterationofarc (S) toapproximatelyminimize 1 2 r(x,t k ) 2 in (4.3). 2b. If g r (x k+1,t k ) ǫ, terminate. 2c. If ρ k η 1, set t k+1 = f(x k+1 ) r(x k,t k ) 2 r(x k+1,t k ) 2 +(f(x k+1 ) t k ) 2. (4.7) Otherwise, set t k+1 = t k. Note that the monotonicity property of the ARC (S) iterates [10, (2.5), (3.19)] generate in Step 2a of Phase 2 of ShS-ARC provies r(x k,t k ) r(x k+1,t k ) for all k 1, (4.8) an so the upating proceure for t k in (4.7) is well ine. Furthermore, (4.7) gives t k t k+1 = (f(x k+1 ) t k )+ r(x k,t k ) 2 r(x k+1,t k ) 2 +(f(x k+1 ) t k ) 2, (4.9) for any successful k 1, which we use to show next that the target values t k ecrease monotonically. Lemma 4.1. In Phase 2 of the ShS-ARC algorithm, the target values satisfy t k t k+1 for all k 1. (4.10) Proof. Due to (4.9), (4.10) follows immeiately in the case when f(x k+1 ) t k. Otherwise, when f(x k+1 ) > t k, conjugacy properties an (4.9) give r(x k,t k ) 2 r(x k+1,t k ) 2 t k t k+1 = 0, f(x k+1 ) t k + r(x k,t k ) 2 r(x k+1,t k ) 2 +(f(x k+1 ) t k ) 2 where in the last inequality, we also use (4.8). Phase 2 of the ShS-ARC terminates when g r (x k+1,t k ) ǫ, (4.11) where g r is ine in (4.5) an ǫ (0,1) is fixe at the start of the algorithm. Allowing ifferent primal an ual accuracy tolerances makes sense if one consiers the possibly ifferent scalings of the (primal) resiuals an (ual) graients. The latter may occur for instance when the Jacobian A(x) in (4.4) is not full rank, which is the case at KKT points of (4.1). The next lemma connects (4.11) to relative KKT points of (4.1) an to approximate critical points of the feasibility measure c(x).

13 Evaluation complexity of nonlinear least-squares an constraine problems 13 Lemma 4.2. For some (x,t), assume that the scale graient (4.5) of r(x,t) in (4.3) satisfies g r (x,t) = J(x)T c(x)+(f(x) t)g(x) r(x,t) ǫ. (4.12) Then either or J(x) T y(x,t)+g(x) (y(x,t),1) J(x) T c(x) c(x) ǫ, (4.13) ǫ, with y(x,t) = c(x) f(x) t. (4.14) Proof. We istinguish two possible cases. Firstly, assume that f(x) = t. Then (4.3) an (4.12) straightforwarly imply (4.13). Alternatively, we must have that f(x) t. This allows us to ivie in the numerator of (4.12) by f(x) t, which then provies, also using r(x,t) = (c(x),f(x) t), J(x) T c(x) f(x) t +g(x) r(x,t) ( ) f(x) t ǫ = c(x) f(x) t,1 ǫ, which gives (4.14). The conition (4.14) is an instance of the relative ual KKT stopping criterion J(x) T y +g(x) ǫ, (4.15) (y,1) for some multiplier y IR m. The relative error conition (4.15) can be justifie by means of a perturbation argument. Namely, consiering the perturbations x = x +δx an y = y +δy to some KKT point x an to a corresponing multiplier y, a Taylor expansion an the KKT conition J(x ) T y +g(x ) = 0 give that the perturbe ual feasibility resiual J(x) T y+g(x) is to first orer [H(x )+ m i=1 y i xxc i (x )]δx+ J(x ) T δy. Thepresenceofthemultipliery inthelatterremainerillustratesthatthesizeofthemultiplier shoul not be ignore when measuring KKT equation resiuals. If we enter Phase 2 of ShS-ARC, we have c(x 1 ) ǫ p. We show in the next section that in fact, we remain sufficiently close to the constraints for all subsequent iterates so that c(x k ) ǫ p. This an Lemma 4.2 imply that when the ShS-ARC algorithm terminates with (4.11), then either we are close to a feasible critical point of the feasibility measure c(x) or we are close to a (relative) KKT point of (4.1). In the next section, we establish that ShS-ARC remains close to the constraints at each step, an that the target values t k ecrease by a fixe amount in each iteration. Thus either (4.11) hols for some k an so we are approximately critical for (4.1) or for the constraints or the targets reach f, the global minimum of f over the set of constraints, in which case again (4.11) must hol. Thus ShS-ARC will terminate; furthermore, when ǫ p = ǫ an ǫ = ǫ 2/3, its worst-case evaluation complexity is O ( ǫ 3/2), just like in the unconstraine case. 5 Complexity of the ShS-ARC algorithm for the equality constraine problem Before analyzing the complexity of Algorithm ShS-ARC, we state our assumptions formally (in aition to AC.1): AC.2 The norm of c is uniformly boune above on IR n by c up an its Jacobian J(x) is globally Lipschitz continuous in IR n with Lipschitz constant L J > 0. The components c i an xx c i (x) are globally Lipschitz continuous on IR n with Lipschitz constants L ci > 0 an L H,ci, for i {1,...,m}.

14 14 C. Cartis, N. I. M. Goul an Ph. L. Toint Φ(x,t) f ε 3/2 Φ(x +,t) t Phase 1 t + Phase 2 ε ε c Figure 4.1: (a) Illustration of ShS-ARC Phase 1 & 2. (b) A successful iteration of ShS-ARC s Phase 2 in the case where ǫ p = ǫ an ǫ = ǫ 2/3. AC.3 f(x), g(x) an xx f(x) are Lipschitz continuous in C 1 which is ine in (4.2), with Lipschitz constants L f, L g,f > 0 an L H,f, respectively. AC.4 The objective f(x) is boune above an below in C 1, that is there exist constants f low an f low f up +1 such that f low f(x) f up for all x C 1. The assumptions AC.1 AC.4, the construction of ShS-ARC an (4.10) imply that AR.1 AR.4 hol for 1 each of the least-squares functions that we employ in ShS-ARC, namely, 2 c(x) 2 an 1 2 r(x,t k ) 2 for k 1; furthermore, the resulting constants are inepenent of k. In particular, the corresponing values of L g in (2.10) for 2 c(x) 1 2 an 2 r(x,t 1 k ) 2 are, respectively, L g,1 = L 2 c +c up L J an L g,2 = (L c,l f ) 2 +max{c up, f up + t 1 } (L J,L g,f ), (5.1) where L c = (L c1,...,l cm ) is the Lipschitz constant of c, while the corresponing values of L in (2.13) for 1 2 c(x) 2 an 1 2 r(x,t k ) 2 are, respectively, m m = 2L J L c +L J L ci + c(x 0 ) L H,ci, an (5.2) L 1 i=1 i=1 ( ( ) m m L 2 = (L J,L g,f ) 2 (L c,l f ) +L f + L ci )+max{c up, f up + t 1 } L H,f + L H,ci. (5.3) i=1 ThenextlemmashowsthatPhase2ofShS-ARCconsistsof(atmost)aconstantnumberofunsuccessful ARC (S) stepsfollowebyasuccessfuloneforminimizing 1 2 r(x,t k ) 2 forfixet k,afterwhicht k isecrease accoring to (4.7). i=1

15 Evaluation complexity of nonlinear least-squares an constraine problems 15 Lemma 5.1. Let AC.1 AC.4 hol, as well as AM.4 for the Hessian of 1 2 r(x,t k ) 2 an its approximation. Then the Phase 2 iterations of the ShS-ARC algorithm satisfy σ k max(σ 1, 3 2γ 2 (L 2 +C)) = σ sh, for all k 1, (5.4) where L 2 is ine in (5.3). Also, at most L sh = 1+ 2 logγ 1 log ( σsh σ min ) (5.5) ShS-ARC/ARC (S) iterations are performe for each istinct target value t k. Proof. The implication (2.18) in Lemma 2.1 irectly applies to the Phase 2 iterations of ShS-ARC, with constants L = L 2 ine in (5.3) an C given in AM.4, inepenent of k. The construction of a Phase 2 iteration of ShS-ARC an (2.6) imply that whenever σ k is large in the sense of (2.18), we have σ k+1 σ k. Thus (5.4) follows, noting that the factor γ 2 in σ sh is allowe for the case when σ k is only slightly less than 3(L 2 +C)/2 an k is not very successful, while the term σ 1 in (5.4) accounts for choices at the start of Phase 2. Note that Theorem 2.2 irectly applies to the Phase 2 iterations of ShS-ARC that employ the same target value t k. Thus the boun (5.5) follows irectly from (2.25), (5.4), the use of parameters γ 1 an σ min in Phase 2 of ShS-ARC, as well as the fact that we only take one successful ShS-ARC/ARC (S) iteration for each fixe t k (an so, here, S j = 1 in (2.25)). The next lemma gives an auxiliary result to be use in Lemma 5.3. Lemma 5.2. Consier the following optimization problem in two variables min (f,c) IR 2F(f,c) = f + ǫ 2 c 2 subject to f 2 +c 2 α 2, (5.6) where 0 < α < ǫ. The global minimum of (5.6) is attaine at (f,c ) = (α,0) an it is given by F(f,c ) = α+ǫ. Proof. Note that for any feasible (f,c), ǫ 2 c 2 > 0 since α < ǫ. We have ( ) c F(f,c) = 1, 0, for all (f,c). ǫ2 c 2 Thus the solution of (5.6) is attaine on the bounary of the feasible region, namely f 2 +c 2 = α 2. Also, (f,c ) satisfies the KKT conitions for (5.6), namely, 1+2λ f = 0 c +2λ c = 0 ǫ2 c 2 λ 0 an f 2 +c 2 = α 2, where λ is the multiplier at the minimizer (f,c ). The first conition above implies that f 0 an λ 0. Thus λ = 1/(2f ), which substitute into the secon KKT conition gives ( ) 1 1 c = 0. f ǫ2 c 2

16 16 C. Cartis, N. I. M. Goul an Ph. L. Toint Note that f ǫ 2 c 2 since f 2 + c 2 = α 2 an α < ǫ. Thus c = 0, an f = ±α. Since we are minimizing, the smallest value of F(f,c ) is at f = α. The next lemma proves the crucial observation that all Phase 2 iterates remain (approximately) feasible, an that the targets t k ecrease by a quantity boune below by a multiple of ǫ 3/2 p at every successful iteration k until termination. Lemma 5.3. Suppose that AC.1 AC.4 hol, as well as AM.4 for the Hessian of 1 2 r(x,t k ) 2 an its approximation. For every Phase 2 iteration k 1 of the ShS-ARC algorithm for which (4.11) fails, we have that f(x k+1 ) t k+1 0, (5.7) Moreover, if iteration k is successful an ǫ ǫ 1/3 p, then r(x k,t k ) = ǫ p, (5.8) c(x k ) ǫ p an f(x k ) t k ǫ p. (5.9) t k t k+1 κ t ǫ 3/2 p (5.10) for some problem-epenent constant κ t = min{αβ 3/2,1 β}, where β (0,1) is any fixe probleminepenent constant, α = η 1 σ min κ 3 g,r an κ g,r = (1 κ θ )/( 1 2L 2 +C +σ sh +κ θ L g,2 ), (5.11) with κ θ, L 2, C, σ sh an L g,2 ine in (2.4), (5.3), (2.16), (5.4) an (5.1), respectively. Proof. We start by observing that (5.7) immeiately follows from (4.7) an (4.8). Also, (5.9) follows straightforwarly from (5.8). Next, we prove (5.8), by inuction on k. Firstly, note that this inequality hols by construction for k = 1. Assume now that iteration k > 1 is successful an that Then r(x k,t k ) = ǫ p. (5.12) (f(x k+1 ) t k+1 ) 2 = r(x k,t k ) 2 r(x k+1,t k ) 2 +(f(x k+1 ) t k ) 2 = r(x k,t k ) 2 c(x k+1 ) 2, where (5.7) an (4.7) give the first ientity, while the secon equality follows from (4.3). Thus we euce, also using (4.3), that which conclues our inuction step ue to (5.12). r(x k+1,t k+1 ) 2 = r(x k,t k ) 2, It remains to establish (5.10). Lemma 3.1 applies to minimizing 1 2 r(x,t k ) 2, an so (3.3) implies that for any successful k 1, we have { } r(x k,t k ) r(x k+1,t k ) κ t min g r (x k+1,t k ) 3/2 r(x k,t k ) 1/2, r(x k,t k ), (5.13) where κ t is ine below (5.10). Thus for any successful k 1 for which (4.11) fails, (5.13) becomes r(x k,t k ) r(x k+1,t k ) κ t min { ǫ 3/2 p,ǫ p } = κ t ǫ 3/2 p, (5.14)

17 Evaluation complexity of nonlinear least-squares an constraine problems 17 where we also use (5.8) in the first inequality an ǫ ǫ 1/3 p, in the secon ientity. Using (4.3) an the properties of the l 2 -norm, (4.9) becomes t k t k+1 = (f(x k+1 ) t k )+ r(x k,t k ) 2 c(x k+1 ) 2 = (f(x k+1 ) t k )+ ǫ 2 p c(x k+1 ) 2, where we use (5.8) in the secon equality. It follows from (4.3) that (f(x k+1 ) t k ) 2 + c(x k+1 ) 2 = r(x k+1,t k ) 2 (5.15) ( ) 2 ( ) 2 r(x k,t k ) κ t ǫ 3/2 p = ǫ p κ t ǫ 3/2 p (5.16) where in the first inequality we use (5.14) an in the secon, (5.8). We now apply Lemma 5.2 to the thir right-han sie of (5.15), letting f = f(x k+1 ) t k, c = c(x k+1 ), ǫ = ǫ p an α = ǫ p κ t ǫ 3/2 p. We euce from this Lemma, (5.15) an (5.16) that t k t k+1 α+ǫ p = (ǫ p κ t ǫ 3/2 p )+ǫ p = κ t ǫ 3/2 p which proves (5.10). Figure 4.1 (b) illustrates the workings of one successful Phase 2 iteration for ǫ = ǫ p an ǫ = ǫ 2/3, the case of most interest to us as it coincies with the evaluation complexity of ARC for the unconstraine case. The figure exemplifies that the amount of ecrease in the target values is inherite from the merit function ecrease (5.14). Note that the ShS-ARC algorithm requires one evaluation of the objective function, its graient (an possibly Hessian) an one evaluation of the vector of constraint functions, its Jacobian (an possibly Hessians) per iteration. We are now reay to give the main complexity result for ShS-ARC applie to (4.1). Theorem 5.4. Suppose that AC.1 AC.4 hol, an that ShS-ARC is applie to minimizing (4.1) with ǫ ǫ 1/3 p. Assume also that AM.4 hols for the Hessians of 1 2 c(x) 2 an 2 r(x,t 1 k ) 2 an its approximations. Then the ShS-ARC algorithm generates an iterate x k satisfying either a relative KKT conition for (4.1), namely, c(x k ) ǫ p an J(x k) T y k +g(x k ) (y k,1) ǫ (5.17) for some y k IR m, or an approximate first-orer criticality conition for the feasibility measure c(x), namely, J(x k ) T c(x k ) ǫ (5.18) c(x k ) in at most κ f,c ǫ 3/2 ǫ 1/2 p (5.19) evaluations of c an f (an their erivatives), where κ f,c > 0 is a problem-epenent constant, inepenent of ǫ p, an x 0. Proof. The evaluation complexity of Phase 1 follows irectly from Theorem 3.2 with Φ(x) = 1 2 c(x) 2. In particular, the evaluation complexity of obtaining x 1 is boune above by κ S max{κ 1,κ 2 }max{ǫ 3/2,ǫ 1/2 p } (5.20) where κ 1,2 an κ S are ine in (3.16) an (3.17) with r(x 0 ) = c(x 0 ), L = L 1 given in (5.2) an L g = L g,1 in (5.1). If the ShS-ARC algorithm terminates at this stage, then (4.6) implies that (5.18) hols with

18 18 C. Cartis, N. I. M. Goul an Ph. L. Toint k = 1 an c(x 1 ) > ǫ p. Assume now that Phase 2 of the ShS-ARC algorithm is entere. From AC.4 an (5.9), we have f low f(x k ) t k +ǫ p t 1 i k κ t ǫ 3/2 p +ǫ p f(x 1 ) i k κ t ǫ 3/2 p +ǫ p where i k is the number of successful ShS-ARC iterations from 1 to k for which (4.11) fails, an where we have also use (5.10) an the inition of t 1 in the ShS-ARC algorithm. Hence, we obtain from the inequality f(x 1 ) f up (itself implie by AC.4 again) an ǫ p (0,1) that f up f low +1 i k κ t ǫ 3/2 p = L s sh. (5.21) Since for each istinct value of t k we have one successful iteration, (5.5) in Lemma 5.1 implies that the total number of Phase 2 iterations for which (4.11) fails is boune above by L s sh L sh, where L sh is ine in (5.5) an L s sh, in (5.21). Thus the ShS-ARC algorithm must terminate after this many iterations at most, yieling, because of Lemma 4.2, an iterate satisfying c(x k ) ǫ p an either (4.14) or (4.13); thus either (5.17) or (5.18) will hol in this case. Recalling that only one evaluation of c an f (an their erivatives, if successful) occurs per iteration, the boun (5.19) now follows by summing up the Phase 1 an Phase 2 iteration bouns, an using that ǫ p (0,1) which gives that the Phase 2 boun ominates in the orer of (ǫ p,ǫ ). If ǫ = ǫ 2/3 p, then Theorem 5.4 implies that the evaluation complexity of ShS-ARC is at most O(ǫ 3/2 p ), the same as for the unconstraine case. However, if ǫ = ǫ p, then this complexity boun worsens to O(ǫ 2 p ), the same in orer as for steepest-escent-type methos for both constraine an unconstraine problems [3, 11]. 6 Conclusions We have shown that with an appropriate an practical termination conition, the (optimal) cubic regularization variant ARC (S) takes at most O(ǫ 3/2 ) evaluations to rive the resiual or the scale graient of the potentially singular least-squares problem (1.1) below ǫ. Our analysis has focuse on the Eucliean norm case, but it can be easily extene to general inner proucts an inuce norms, an to smooth l p -norms for p > 2. Though the orer ǫ 3/2 of the ARC boun is optimal for unconstraine optimization when secon-orer methos are employe [4], an it is sharp for nonlinear least-squares when ensuring (1.2) is sufficiently small, it is unclear whether it is optimal or even sharp for ARC (S) with the novel termination conition (3.1). For the equality-constraine potentially nonconvex programming problem (4.1), we presente a targetfollowing technique ShS-ARC that applies ARC (S) to target-epenent least-squares merit functions tracking a path of approximately feasible points (if an initial such point can be foun). Furthermore, in orer to ensure approximate first-orer conitions for (4.1) or for a feasibility measure within ǫ p for the constraint feasibility an within ǫ for ual (first-orer) feasibility ShS-ARC requires at most O(ǫ 3/2 ǫ 1/2 p ) problem-evaluations, which epening on the choice of tolerances ǫ p an ǫ can take any value between the complexity O(ǫ 3/2 p ) of ARC (namely, when ǫ = ǫ 2/3 p ) an O(ǫ 2 p ) of steepest-escent (when ǫ = ǫ p ). Though it is natural for the primal an ual feasibility resiuals to vary at ifferent rates, an hence require ifferent optimality tolerances (with higher accuracy for primal feasibility than for ual being common), it is an open question at the moment whether an algorithm for nonconvexly constraine problems can be evise that has worst-case evaluation complexity of orer ǫ 3/2 where ǫ = ǫ p = ǫ. Also, extening ShS- ARC or other cubic regularization approaches to problems with nonconvex inequality constraints remains to be consiere.

19 Evaluation complexity of nonlinear least-squares an constraine problems 19 References [1] S. Bellavia, C. Cartis, N. I. M. Goul, B. Morini an Ph. L. Toint. Convergence of a Regularize Eucliean Resiual algorithm for nonlinear least-squares. SIAM Journal on Numerical Analysis, 48(1): 1 29, [2] R. H. Byr, R. B. Schnabel an G. A. Shultz. A trust region algorithm for nonlinearly constraine optimization. SIAM Journal on Numerical Analysis, 24: , [3] C. Cartis, N. I. M. Goul an Ph. L. Toint. On the complexity of fining first-orer critical points in constraine nonlinear programming. ERGO Technical Report , School of Mathematics, University of Einburgh, [4] C. Cartis, N. I. M. Goul an Ph. L. Toint. Optimal Newton-type methos for nonconvex smooth optimization problems. ERGO Technical Report , School of Mathematics, University of Einburgh, [5] C. Cartis, N. I. M. Goul an Ph. L. Toint. Evaluation complexity of aaptive cubic regularization methos for convex unconstraine optimization. Optimization Methos an Software, DOI: / , [6] C. Cartis, N. I. M. Goul an Ph. L. Toint. An aaptive cubic regularization algorithm for nonconvex optimization with convex constraints an its function-evaluation complexity. IMA Journal of Numerical Analysis, oi: /imanum/rr035, [7] C. Cartis, N. I. M. Goul an Ph. L. Toint. On the oracle complexity of first-orer an erivative-free algorithms for smooth nonconvex minimization. SIAM Journal on Optimization, 22(1):66 86, [8] C. Cartis, N. I. M. Goul an Ph. L. Toint. On the evaluation complexity of composite function minimization with applications to nonconvex nonlinear programming. SIAM Journal on Optimization, 21(4): , [9] C. Cartis, N. I. M. Goul an Ph. L. Toint. Aaptive cubic regularisation methos for unconstraine optimization. Part II: worst-case function- an erivative-evaluation complexity. Mathematical Programming, 130(2): , [10] C. Cartis, N. I. M. Goul an Ph. L. Toint. Aaptive cubic regularisation methos for unconstraine optimization. Part I: motivation, convergence an numerical results. Mathematical Programming, 127(2): , [11] C. Cartis, N. I. M. Goul an Ph. L. Toint. On the complexity of steepest escent, Newton s an regularize Newton s methos for nonconvex unconstraine optimization. SIAM Journal on Optimization, 20(6): , [12] M. R. Celis. A trust region strategy for nonlinear equality constraine optimization. Technical Report Department of Computational an Applie Mathematics, Rice University, Houston, Texas, USA, [13] A. R. Conn, N. I. M. Goul an Ph. L. Toint. Trust-Region Methos. SIAM, Philaelphia, USA, [14] J. E. Dennis an J. J. Moré. A characterization of superlinear convergence an its application to quasi-newton methos. Mathematics of Computation, 28(126): , [15] J. E. Dennis an R. B. Schnabel. Numerical Methos for Unconstraine Optimization an Nonlinear Equations. Prentice-Hall, Englewoo Cliffs, NJ, USA, Reprinte as Classics in Applie Mathematics 16, SIAM, Philaelphia, USA, 1996.

Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity

Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function- and derivative-evaluation complexity Coralia Cartis,, Nicholas I. M. Gould, and Philippe L. Toint September

More information

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity Coralia Cartis, Nick Gould and Philippe Toint Department of Mathematics,

More information

Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization

Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization Evaluation complexity of adaptive cubic regularization methods for convex unconstrained optimization Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint October 30, 200; Revised March 30, 20 Abstract

More information

Adaptive cubic overestimation methods for unconstrained optimization

Adaptive cubic overestimation methods for unconstrained optimization Report no. NA-07/20 Adaptive cubic overestimation methods for unconstrained optimization Coralia Cartis School of Mathematics, University of Edinburgh, The King s Buildings, Edinburgh, EH9 3JZ, Scotland,

More information

Numerical solution of conservation laws applied to the Shallow Water Wave Equations

Numerical solution of conservation laws applied to the Shallow Water Wave Equations Numerical solution of conservation laws applie to the Shallow Water Wave Equations Stephen G Roberts Mathematical Sciences Institute, Australian National University Upate January 17, 2013 (base on notes

More information

Modes of Convergence

Modes of Convergence Moes of Convergence Electrical Engineering 126 (UC Berkeley Spring 2018 There is only one sense in which a sequence of real numbers (a n n N is sai to converge to a limit. Namely, a n a if for every ε

More information

Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization

Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization Corrigendum: On the complexity of finding first-order critical points in constrained nonlinear optimization C. Cartis, N. I. M. Gould and Ph. L. Toint 11th November, 2014 Abstract In a recent paper (Cartis

More information

An investment strategy with optimal sharpe ratio

An investment strategy with optimal sharpe ratio The 22 n Annual Meeting in Mathematics (AMM 2017) Department of Mathematics, Faculty of Science Chiang Mai University, Chiang Mai, Thailan An investment strategy with optimal sharpe ratio S. Jansai a,

More information

SPLITTING FIELDS KEITH CONRAD

SPLITTING FIELDS KEITH CONRAD SPLITTING FIELDS EITH CONRAD 1. Introuction When is a fiel an f(t ) [T ] is nonconstant, there is a fiel extension / in which f(t ) picks up a root, say α. Then f(t ) = (T α)g(t ) where g(t ) [T ] an eg

More information

On the complexity of the steepest-descent with exact linesearches

On the complexity of the steepest-descent with exact linesearches On the complexity of the steepest-descent with exact linesearches Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint 9 September 22 Abstract The worst-case complexity of the steepest-descent algorithm

More information

On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization

On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization On the oracle complexity of first-order and derivative-free algorithms for smooth nonconvex minimization C. Cartis, N. I. M. Gould and Ph. L. Toint 22 September 2011 Abstract The (optimal) function/gradient

More information

A GENERALIZED COUPON COLLECTOR PROBLEM

A GENERALIZED COUPON COLLECTOR PROBLEM J. Appl. Prob. 48, 08 094 (20) Printe in Englan Applie Probability Trust 20 A GENERALIZED COUPON COLLECTOR PROBLEM WEIYU XU an A. KEVIN TANG, Cornell University Abstract This paper presents an analysis

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Universal regularization methods varying the power, the smoothness and the accuracy arxiv: v1 [math.oc] 16 Nov 2018

Universal regularization methods varying the power, the smoothness and the accuracy arxiv: v1 [math.oc] 16 Nov 2018 Universal regularization methods varying the power, the smoothness and the accuracy arxiv:1811.07057v1 [math.oc] 16 Nov 2018 Coralia Cartis, Nicholas I. M. Gould and Philippe L. Toint Revision completed

More information

GAINS FROM TRADE UNDER MONOPOLISTIC COMPETITION

GAINS FROM TRADE UNDER MONOPOLISTIC COMPETITION bs_bs_banner Pacific Economic Review, 2: (206) pp. 35 44 oi: 0./468-006.250 GAINS FROM TRADE UNDER MONOPOLISTIC COMPETITION ROBERT C. FEENSTRA* University of California, Davis an National Bureau of Economic

More information

A NOTE ON THE DYNAMIC ROLE OF MONOPOLISTIC COMPETITION IN THE MONETARY ECONOMY. abstract

A NOTE ON THE DYNAMIC ROLE OF MONOPOLISTIC COMPETITION IN THE MONETARY ECONOMY. abstract A NOTE ON THE DYNAMIC ROLE OF MONOPOLISTIC COMPETITION IN THE MONETARY ECONOMY abstract In the new Keynesian economics, monopolistic competition plays an important role. Much static research is base on

More information

Abstract Stanar Risk Aversion an the Deman for Risky Assets in the Presence of Backgroun Risk We consier the eman for state contingent claims in the p

Abstract Stanar Risk Aversion an the Deman for Risky Assets in the Presence of Backgroun Risk We consier the eman for state contingent claims in the p Stanar Risk Aversion an the Deman for Risky Assets in the Presence of Backgroun Risk Günter Franke 1, Richar C. Stapleton 2, an Marti G. Subrahmanyam. 3 November 2000 1 Fakultät für Wirtschaftswissenschaften

More information

Trust Region Methods for Unconstrained Optimisation

Trust Region Methods for Unconstrained Optimisation Trust Region Methods for Unconstrained Optimisation Lecture 9, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Trust

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Report no. NA-07/09 Nonlinear programming without a penalty function or a filter Nicholas I. M. Gould Oxford University, Numerical Analysis Group Philippe L. Toint Department of Mathematics, FUNDP-University

More information

Capacity Constraint OPRE 6377 Lecture Notes by Metin Çakanyıldırım Compiled at 15:30 on Tuesday 22 nd August, 2017

Capacity Constraint OPRE 6377 Lecture Notes by Metin Çakanyıldırım Compiled at 15:30 on Tuesday 22 nd August, 2017 apacity onstraint OPRE 6377 Lecture Notes by Metin Çakanyılırım ompile at 5:30 on Tuesay 22 n August, 207 Solve Exercises. [Marginal Opportunity ost of apacity for Deman with onstant Elasticity] We suppose

More information

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0.

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0. Outline Coordinate Minimization Daniel P. Robinson Department of Applied Mathematics and Statistics Johns Hopkins University November 27, 208 Introduction 2 Algorithms Cyclic order with exact minimization

More information

Global convergence rate analysis of unconstrained optimization methods based on probabilistic models

Global convergence rate analysis of unconstrained optimization methods based on probabilistic models Math. Program., Ser. A DOI 10.1007/s10107-017-1137-4 FULL LENGTH PAPER Global convergence rate analysis of unconstrained optimization methods based on probabilistic models C. Cartis 1 K. Scheinberg 2 Received:

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Math. Program., Ser. A (2010) 122:155 196 DOI 10.1007/s10107-008-0244-7 FULL LENGTH PAPER Nonlinear programming without a penalty function or a filter N. I. M. Gould Ph.L.Toint Received: 11 December 2007

More information

Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models

Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models Worst-case evaluation comlexity for unconstrained nonlinear otimization using high-order regularized models E. G. Birgin, J. L. Gardenghi, J. M. Martínez, S. A. Santos and Ph. L. Toint 2 Aril 26 Abstract

More information

CDO TRANCHE PRICING BASED ON THE STABLE LAW VOLUME II: R ELAXING THE LHP. Abstract

CDO TRANCHE PRICING BASED ON THE STABLE LAW VOLUME II: R ELAXING THE LHP. Abstract CDO TRANCHE PRICING BASED ON THE STABLE LAW VOLUME II: R ELAXING THE ASSUMPTION German Bernhart XAIA Investment GmbH Sonnenstraße 9, 833 München, Germany german.bernhart@xaia.com First Version: July 26,

More information

Nonlinear programming without a penalty function or a filter

Nonlinear programming without a penalty function or a filter Nonlinear programming without a penalty function or a filter N I M Gould Ph L Toint October 1, 2007 RAL-TR-2007-016 c Science and Technology Facilities Council Enquires about copyright, reproduction and

More information

PERFORMANCE OF THE CROATIAN INSURANCE COMPANIES - MULTICRITERIAL APPROACH

PERFORMANCE OF THE CROATIAN INSURANCE COMPANIES - MULTICRITERIAL APPROACH PERFORMANCE OF THE CROATIAN INSURANCE COMPANIES - MULTICRITERIAL APPROACH Davorka Davosir Pongrac Zagreb school of economics an management Joranovac 110, 10000 Zagreb E-mail: avorka.avosir@zsem.hr Višna

More information

Repos, Fire Sales, and Bankruptcy Policy

Repos, Fire Sales, and Bankruptcy Policy Repos, Fire Sales, an Bankruptcy Policy Gaetano Antinolfi Francesca Carapella Charles Kahn Antoine Martin Davi Mills E Nosal Preliminary an Incomplete May 25, 2012 Abstract The events from the 2007-2009

More information

1. An insurance company models claim sizes as having the following survival function. 25(x + 1) (x 2 + 2x + 5) 2 x 0. S(x) =

1. An insurance company models claim sizes as having the following survival function. 25(x + 1) (x 2 + 2x + 5) 2 x 0. S(x) = ACSC/STAT 373, Actuarial Moels I Further Probability with Applications to Actuarial Science WINTER 5 Toby Kenney Sample Final Eamination Moel Solutions This Sample eamination has more questions than the

More information

1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016 AM 22: Advanced Optimization Spring 206 Prof. Yaron Singer Lecture 9 February 24th Overview In the previous lecture we reviewed results from multivariate calculus in preparation for our journey into convex

More information

Partial State-Owned Bank Interest Margin, Default Risk, and Structural Breaks: A Model of Financial Engineering

Partial State-Owned Bank Interest Margin, Default Risk, and Structural Breaks: A Model of Financial Engineering Partial State-Owne Bank Interest Margin, Default Risk, an Structural Breaks: A Moel of Financial Engineering JYH-HORNG IN,CHING-HUI CHANG * AND ROSEMARY JOU Grauate Institute of International Business

More information

What can we do with numerical optimization?

What can we do with numerical optimization? Optimization motivation and background Eddie Wadbro Introduction to PDE Constrained Optimization, 2016 February 15 16, 2016 Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016

More information

REAL OPTION MODELING FOR VALUING WORKER FLEXIBILITY

REAL OPTION MODELING FOR VALUING WORKER FLEXIBILITY REAL OPTION MODELING FOR VALUING WORKER FLEXIBILITY Harriet Black Nembhar Davi A. Nembhar Ayse P. Gurses Department of Inustrial Engineering University of Wisconsin-Maison 53 University Avenue Maison,

More information

Data Center Demand Response in Deregulated Electricity Markets

Data Center Demand Response in Deregulated Electricity Markets Data Center Deman Response in Deregulate Electricity Markets Shahab Bahrami, Stuent Member, IEEE, Vincent W.S. Wong, Fellow, IEEE, an Jianwei Huang, Fellow, IEEE Abstract With the evelopment of eregulate

More information

Efficient Numerical Methods for Pricing American Options Under Stochastic Volatility

Efficient Numerical Methods for Pricing American Options Under Stochastic Volatility Efficient Numerical Methos for Pricing American Options Uner Stochastic Volatility Samuli Ikonen, 1 Jari Toivanen 2 1 Norea Markets, Norea FI-00020, Finlan 2 Department of Mathematical Information Technology,

More information

An efficient method for computing the Expected Value of Sample Information. A non-parametric regression approach

An efficient method for computing the Expected Value of Sample Information. A non-parametric regression approach ScHARR Working Paper An efficient metho for computing the Expecte Value of Sample Information. A non-parametric regression approach Mark Strong,, eremy E. Oakley 2, Alan Brennan. School of Health an Relate

More information

Appendix. Confidence Banking and Strategic Default. Guillermo Ordoñez. University of Pennsylvania and NBER

Appendix. Confidence Banking and Strategic Default. Guillermo Ordoñez. University of Pennsylvania and NBER Appenix Confience Banking an Strategic Default Guillermo Oroñez University of Pennsylvania an NBER 1 Proofs 1.1 Proof of Proposition 1 Since s ( ) is the signal that makes a goo firm with a given reputation

More information

New Trade Models, New Welfare Implications

New Trade Models, New Welfare Implications New rae Moels, New Welfare Implications he Harvar community has mae this article openly available. Please share how this access benefits you. Your story matters Citation Melitz, Marc J., an Stephen J.

More information

A Costless Way to Increase Equity

A Costless Way to Increase Equity A Costless Way to Increase Equity Raphael Flore October 27, 2016 Abstract This paper complements stanar theories of optimal capital structure by allowing firms to invest in the financial markets in which

More information

A Game Theoretic Model of Deposit Contracts between the Bank and the Depositor - Extend Study on the Economic Analysis of Bank Run

A Game Theoretic Model of Deposit Contracts between the Bank and the Depositor - Extend Study on the Economic Analysis of Bank Run wwwscieuca/ijfr International Journal of Financial Research Vol 5, No 3; 04 A Game Theoretic Moel of Deposit Contracts between the Bank an the Depositor - Exten Stuy on the Economic Analysis of Bank Run

More information

GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS

GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS ANDREW R. CONN, KATYA SCHEINBERG, AND LUíS N. VICENTE Abstract. In this paper we prove global

More information

transfers in orer to keep income of the hospital sector unchange, then a larger welfare gain woul be obtaine, even if the government implements a bala

transfers in orer to keep income of the hospital sector unchange, then a larger welfare gain woul be obtaine, even if the government implements a bala The Impact of Marginal Tax Reforms on the Supply of Health Relate Services in Japan * Ryuta Ray Kato 1. Introuction This paper presents a computable general equilibrium (CGE) framework to numerically examine

More information

Data Center Demand Response in Deregulated Electricity Markets

Data Center Demand Response in Deregulated Electricity Markets This article has been accepte for publication in a future issue of this journal, but has not been fully eite. Content may change prior to final publication. Citation information: DOI 0.09/TSG.208.280830,

More information

Dynamic Pricing through Customer Discounts for Optimizing Multi-Class Customers Demand Fulfillment

Dynamic Pricing through Customer Discounts for Optimizing Multi-Class Customers Demand Fulfillment Dynamic Pricing through Customer Discounts for Optimizing ulti-class Customers Deman Fulfillment Qing Ding Panos Kouvelis an Joseph ilner# John. Olin School of Business Washington University St. Louis,

More information

The Comprehensive Business Income Tax System: A Proposal for Ultimate Neutrality between Debt and New Equity Issues?

The Comprehensive Business Income Tax System: A Proposal for Ultimate Neutrality between Debt and New Equity Issues? International Journal of Sciences: Basic an Applie Research (IJSBAR) ISSN 2307-4531 (Print & Online) http://gssrr.org/inex.php?journaljournalofbasicanapplie ---------------------------------------------------------------------------------------------------------------------------

More information

A Trust Region Algorithm for Heterogeneous Multiobjective Optimization

A Trust Region Algorithm for Heterogeneous Multiobjective Optimization A Trust Region Algorithm for Heterogeneous Multiobjective Optimization Jana Thomann and Gabriele Eichfelder 8.0.018 Abstract This paper presents a new trust region method for multiobjective heterogeneous

More information

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models IEOR E4707: Foundations of Financial Engineering c 206 by Martin Haugh Martingale Pricing Theory in Discrete-Time and Discrete-Space Models These notes develop the theory of martingale pricing in a discrete-time,

More information

University of Edinburgh, Edinburgh EH9 3JZ, United Kingdom.

University of Edinburgh, Edinburgh EH9 3JZ, United Kingdom. An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity by C. Cartis 1, N. I. M. Gould 2 and Ph. L. Toint 3 February 20, 2009;

More information

An Evaluation of Shareholder Activism

An Evaluation of Shareholder Activism An Evaluation of Shareholer Activism Barbara G. Katz Stern School of Business, New York University 44 W. 4th St., New York, NY 10012 bkatz@stern.nyu.eu; tel: 212 998 0865; fax: 212 995 4218 corresponing

More information

2. Lattice Methods. Outline. A Simple Binomial Model. 1. No-Arbitrage Evaluation 2. Its relationship to risk-neutral valuation.

2. Lattice Methods. Outline. A Simple Binomial Model. 1. No-Arbitrage Evaluation 2. Its relationship to risk-neutral valuation. . Lattice Methos. One-step binomial tree moel (Hull, Chap., page 4) Math69 S8, HM Zhu Outline. No-Arbitrage Evaluation. Its relationship to risk-neutral valuation. A Simple Binomial Moel A stock price

More information

Yao s Minimax Principle

Yao s Minimax Principle Complexity of algorithms The complexity of an algorithm is usually measured with respect to the size of the input, where size may for example refer to the length of a binary word describing the input,

More information

arxiv: v2 [quant-ph] 29 Mar 2018

arxiv: v2 [quant-ph] 29 Mar 2018 The Commutation Relation for Cavity Moe Operators arxiv:1611.01003v2 [quant-ph] 29 Mar 2018 Fesseha Kassahun Department of Physics, Ais Ababa University P. O. Box 33761, Ais Ababa, Ethiopia October 21,

More information

Lecture 5: Iterative Combinatorial Auctions

Lecture 5: Iterative Combinatorial Auctions COMS 6998-3: Algorithmic Game Theory October 6, 2008 Lecture 5: Iterative Combinatorial Auctions Lecturer: Sébastien Lahaie Scribe: Sébastien Lahaie In this lecture we examine a procedure that generalizes

More information

Fuzzy EOQ Model for Time-Deteriorating Items Using Penalty Cost

Fuzzy EOQ Model for Time-Deteriorating Items Using Penalty Cost merican Journal of Operational Research 6 6(: -8 OI:.59/j.ajor.66. Fuzzy EOQ Moel for ime-eteriorating Items Using Penalty ost Nalini Prava Behera Praip Kumar ripathy epartment of Statistics Utkal University

More information

Introduction to Financial Derivatives

Introduction to Financial Derivatives 55.444 Introuction to Financial Derivatives Week of December n, 3 he Greeks an Wrap-Up Where we are Previously Moeling the Stochastic Process for Derivative Analysis (Chapter 3, OFOD) Black-Scholes-Merton

More information

NBER WORKING PAPER SERIES PROFIT SHIFTING AND TRADE AGREEMENTS IN IMPERFECTLY COMPETITIVE MARKETS. Kyle Bagwell Robert W. Staiger

NBER WORKING PAPER SERIES PROFIT SHIFTING AND TRADE AGREEMENTS IN IMPERFECTLY COMPETITIVE MARKETS. Kyle Bagwell Robert W. Staiger NBER WORKING PAPER SERIES PROFIT SHIFTING AND TRADE AGREEMENTS IN IMPERFECTLY COMPETITIVE MARKETS Kyle Bagwell Robert W. Staiger Working Paper 14803 http://www.nber.org/papers/w14803 NATIONAL BUREAU OF

More information

CS 3331 Numerical Methods Lecture 2: Functions of One Variable. Cherung Lee

CS 3331 Numerical Methods Lecture 2: Functions of One Variable. Cherung Lee CS 3331 Numerical Methods Lecture 2: Functions of One Variable Cherung Lee Outline Introduction Solving nonlinear equations: find x such that f(x ) = 0. Binary search methods: (Bisection, regula falsi)

More information

P. Manju Priya 1, M.Phil Scholar. G. Michael Rosario 2, Associate Professor , Tamil Nadu, INDIA)

P. Manju Priya 1, M.Phil Scholar. G. Michael Rosario 2, Associate Professor , Tamil Nadu, INDIA) International Journal of Computational an Applie Mathematics. ISSN 89-4966 Volume, Number (07 Research Inia Publications http://www.ripublication.com AN ORDERING POLICY UNDER WO-LEVEL RADE CREDI POLICY

More information

Monopolistic Competition

Monopolistic Competition Welfare Ranking of A-valorem an Specific Tariffs in a Moel of Monopolistic Competition Esra Durceylan Bilkent University May 3, 2010 Abstract This paper compares the welfare implications of a-valorem an

More information

Macro Dynamics and Labor-Saving Innovation: US vs. Japan

Macro Dynamics and Labor-Saving Innovation: US vs. Japan CIRJE-F-528 Macro Dynamics an Labor-Saving Innovation: US vs. Japan Ryuzo Sato New York University an University of Tokyo Tamaki Morita National Grauate Institute for Policy Stuies (GRIPS) November 2007

More information

Full file at

Full file at Chapter 2 Supply an eman Analysis Solutions to Review uestions 1. Excess eman occurs when price falls below the equilibrium price. In this situation, consumers are emaning a higher quantity than is being

More information

Simple Robust Hedging with Nearby Contracts

Simple Robust Hedging with Nearby Contracts Simple Robust Heging with Nearby Contracts Liuren Wu Zicklin School of Business, Baruch College Jingyi Zhu University of Utah Abstract Most existing heging approaches are base on neutralizing risk exposures

More information

Development Economics and Public Policy WORKING PAPER SERIES

Development Economics and Public Policy WORKING PAPER SERIES Development Economics an Public Policy WORKING PAPER SERIES Paper No. 5 DO TECHNOLOGY SHOCKS SHIFT OUTPUT? AN EMPIRICAL ANALYSIS OF A TWO FACTOR MODEL Hulya Ulku University of Manchester May 005 ISBN:

More information

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming

Dynamic Programming: An overview. 1 Preliminaries: The basic principle underlying dynamic programming Dynamic Programming: An overview These notes summarize some key properties of the Dynamic Programming principle to optimize a function or cost that depends on an interval or stages. This plays a key role

More information

4: SINGLE-PERIOD MARKET MODELS

4: SINGLE-PERIOD MARKET MODELS 4: SINGLE-PERIOD MARKET MODELS Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 4: Single-Period Market Models 1 / 87 General Single-Period

More information

Preferences, Prices, and Performance in Monopoly and Duopoly

Preferences, Prices, and Performance in Monopoly and Duopoly Preferences, Prices, an Performance in Monopoly an Duopoly Yongmin Chen y an Michael H. Rioran z November 20, 2009 Abstract. This paper takes the new approach of using a copula to characterize consumer

More information

Forecast Horizons for Production Planning with Stochastic Demand

Forecast Horizons for Production Planning with Stochastic Demand Forecast Horizons for Production Planning with Stochastic Demand Alfredo Garcia and Robert L. Smith Department of Industrial and Operations Engineering Universityof Michigan, Ann Arbor MI 48109 December

More information

A Moment Matching Approach to the Valuation of a Volume Weighted Average Price Option

A Moment Matching Approach to the Valuation of a Volume Weighted Average Price Option A Moment Matching Approach to the Valuation of a Volume Weighte Average Price Option Antony William Stace Department of Mathematics, University of Queenslan, Brisbane, Queenslan 472, Australia aws@maths.uq.eu.au

More information

If you have ever spoken with your grandparents about what their lives were like

If you have ever spoken with your grandparents about what their lives were like CHAPTER 7 Economic Growth I: Capital Accumulation an Population Growth The question of growth is nothing new but a new isguise for an age-ol issue, one which has always intrigue an preoccupie economics:

More information

Ch 10. Arithmetic Average Options and Asian Opitons

Ch 10. Arithmetic Average Options and Asian Opitons Ch 10. Arithmetic Average Options an Asian Opitons I. Asian Options an Their Analytic Pricing Formulas II. Binomial Tree Moel to Price Average Options III. Combination of Arithmetic Average an Reset Options

More information

The Intriguing Nexus Between Corruption and Capital Account Restrictions

The Intriguing Nexus Between Corruption and Capital Account Restrictions The Intriguing Nexus Between Corruption an Capital Account Restrictions Axel Dreher Lars-H.R. Siemers June 2003 Abstract In a simple theoretical moel we ientify a mutual relationship between corruption

More information

BOUNDS FOR THE LEAST SQUARES RESIDUAL USING SCALED TOTAL LEAST SQUARES

BOUNDS FOR THE LEAST SQUARES RESIDUAL USING SCALED TOTAL LEAST SQUARES BOUNDS FOR THE LEAST SQUARES RESIDUAL USING SCALED TOTAL LEAST SQUARES Christopher C. Paige School of Computer Science, McGill University Montreal, Quebec, Canada, H3A 2A7 paige@cs.mcgill.ca Zdeněk Strakoš

More information

On the Superlinear Local Convergence of a Filter-SQP Method. Stefan Ulbrich Zentrum Mathematik Technische Universität München München, Germany

On the Superlinear Local Convergence of a Filter-SQP Method. Stefan Ulbrich Zentrum Mathematik Technische Universität München München, Germany On the Superlinear Local Convergence of a Filter-SQP Method Stefan Ulbrich Zentrum Mathemati Technische Universität München München, Germany Technical Report, October 2002. Mathematical Programming manuscript

More information

The Joint Dynamics of Electricity Spot and Forward Markets: Implications on Formulating Dynamic Hedging Strategies

The Joint Dynamics of Electricity Spot and Forward Markets: Implications on Formulating Dynamic Hedging Strategies Energy Laboratory MI EL 00-005 Massachusetts Institute of echnology he Joint Dynamics of Electricity Spot an Forwar Markets: Implications on Formulating Dynamic Heging Strategies ovember 2000 he Joint

More information

Lecture Quantitative Finance Spring Term 2015

Lecture Quantitative Finance Spring Term 2015 implied Lecture Quantitative Finance Spring Term 2015 : May 7, 2015 1 / 28 implied 1 implied 2 / 28 Motivation and setup implied the goal of this chapter is to treat the implied which requires an algorithm

More information

Key words: financial intermediation, entrepreneurship, economic growth

Key words: financial intermediation, entrepreneurship, economic growth DEPARTMENT OF ECONOMICS ISSN 1441-5429 DISCUSSION PAPER 18/07 FINANCIA INTERMEDIATION, ENTREPRENEURSHIP AND ECONOMIC GROWTH Wenli Cheng * Abstract: This paper presents a simple general equilibrium moel

More information

Lecture 7: Bayesian approach to MAB - Gittins index

Lecture 7: Bayesian approach to MAB - Gittins index Advanced Topics in Machine Learning and Algorithmic Game Theory Lecture 7: Bayesian approach to MAB - Gittins index Lecturer: Yishay Mansour Scribe: Mariano Schain 7.1 Introduction In the Bayesian approach

More information

Pricing Multi-Dimensional Options by Grid Stretching and High Order Finite Differences

Pricing Multi-Dimensional Options by Grid Stretching and High Order Finite Differences Pricing Multi-Dimensional Options by Gri Stretching an High Orer Finite Differences Kees Oosterlee Numerical Analysis Group, Delft University of Technology Joint work with Coen Leentvaar Southern Ontario

More information

Deterrence and Risk Preferences in Sequential Attacker Defender Games with Continuous Efforts

Deterrence and Risk Preferences in Sequential Attacker Defender Games with Continuous Efforts Risk Analysis DOI:./risa.768 Deterrence an Risk Preferences in Sequential Attacker Defener Games with Continuous Efforts Vineet M. Payappalli, Jun Zhuang,, an Victor Richmon R. Jose Most attacker efener

More information

Is Greedy Coordinate Descent a Terrible Algorithm?

Is Greedy Coordinate Descent a Terrible Algorithm? Is Greedy Coordinate Descent a Terrible Algorithm? Julie Nutini, Mark Schmidt, Issam Laradji, Michael Friedlander, Hoyt Koepke University of British Columbia Optimization and Big Data, 2015 Context: Random

More information

Introduction to Financial Derivatives

Introduction to Financial Derivatives 55.444 Introuction to Financial Derivatives Week of December 3 r, he Greeks an Wrap-Up Where we are Previously Moeling the Stochastic Process for Derivative Analysis (Chapter 3, OFOD) Black-Scholes-Merton

More information

LGD Risk Resolved. Abstract

LGD Risk Resolved. Abstract LGD Risk Resolve Jon Frye (corresponing author) Senior Economist Feeral Reserve Bank of Chicago 230 South LaSalle Street Chicago, IL 60604 Jon.Frye@chi.frb.org 32-322-5035 Michael Jacobs Jr. Senior Financial

More information

The use of Expected Utility Theory (EUT) in Taxpayers Behaviour Modelling

The use of Expected Utility Theory (EUT) in Taxpayers Behaviour Modelling American Journal of Applie Sciences Original Research Paper The use of Expecte Utility Theory (EUT) in Taxpayers Behaviour Moelling Fari Ameur an Mohame Tkiouat Stuies an Research Laboratory in Applie

More information

Convergence of trust-region methods based on probabilistic models

Convergence of trust-region methods based on probabilistic models Convergence of trust-region methods based on probabilistic models A. S. Bandeira K. Scheinberg L. N. Vicente October 24, 2013 Abstract In this paper we consider the use of probabilistic or random models

More information

Essays on Some Combinatorial Optimization Problems with Interval Data

Essays on Some Combinatorial Optimization Problems with Interval Data Essays on Some Combinatorial Optimization Problems with Interval Data a thesis submitted to the department of industrial engineering and the institute of engineering and sciences of bilkent university

More information

MULTISTAGE PORTFOLIO OPTIMIZATION AS A STOCHASTIC OPTIMAL CONTROL PROBLEM

MULTISTAGE PORTFOLIO OPTIMIZATION AS A STOCHASTIC OPTIMAL CONTROL PROBLEM K Y B E R N E T I K A M A N U S C R I P T P R E V I E W MULTISTAGE PORTFOLIO OPTIMIZATION AS A STOCHASTIC OPTIMAL CONTROL PROBLEM Martin Lauko Each portfolio optimization problem is a trade off between

More information

Assessment of Acceptance Sampling Plans Using Posterior Distribution for a Dependent Process

Assessment of Acceptance Sampling Plans Using Posterior Distribution for a Dependent Process Rochester Institute of Technology RIT Scholar Works Articles 1-21-2010 Assessment of Acceptance Sampling Plans Using Posterior Distribution for a Depenent Process A. Erhan Mergen Rochester Institute of

More information

Why Has Swedish Stock Market Volatility Increased?

Why Has Swedish Stock Market Volatility Increased? Why Has Seish Stock Market Volatility Increase? by John Hassler Institute for International Economic Stuies This revision: May 29, 1995 Preliminary Abstract Is the increase volatility on the Seish stock

More information

Laurence Boxer and Ismet KARACA

Laurence Boxer and Ismet KARACA SOME PROPERTIES OF DIGITAL COVERING SPACES Laurence Boxer and Ismet KARACA Abstract. In this paper we study digital versions of some properties of covering spaces from algebraic topology. We correct and

More information

First-Order Methods. Stephen J. Wright 1. University of Wisconsin-Madison. IMA, August 2016

First-Order Methods. Stephen J. Wright 1. University of Wisconsin-Madison. IMA, August 2016 First-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) First-Order Methods IMA, August 2016 1 / 48 Smooth

More information

Methods and Models of Loss Reserving Based on Run Off Triangles: A Unifying Survey

Methods and Models of Loss Reserving Based on Run Off Triangles: A Unifying Survey Methods and Models of Loss Reserving Based on Run Off Triangles: A Unifying Survey By Klaus D Schmidt Lehrstuhl für Versicherungsmathematik Technische Universität Dresden Abstract The present paper provides

More information

Appendix B: Yields and Yield Curves

Appendix B: Yields and Yield Curves Pension Finance By Davi Blake Copyright 006 Davi Blake Appenix B: Yiels an Yiel Curves Bons, with their regular an generally reliable stream of payments, are often consiere to be natural assets for pension

More information

Protection and International Sourcing

Protection and International Sourcing Protection an International Sourcing Emanuel Ornelas Lonon School of Economics John L. Turner University of Georgia November 2008 Abstract We stuy the impact of import protection on relationship-specific

More information

Financial Integration, Growth, and Volatility

Financial Integration, Growth, and Volatility W/05/67 Financial Integration, Growth, an Volatility Anne paular an Aue ommeret 005 International Monetary Fun W/05/67 IMF Working aper IMF Institute Financial Integration, Growth, an Volatility repare

More information

Mitigating Double Taxation in an Open Economy

Mitigating Double Taxation in an Open Economy Mitigating ouble Taxation in an Open Economy Tobias Linhe This version: February 2001 Abstract The interaction o various methos o mitigating economic an international ouble taxation o corporate source

More information

1. Basic currency board equations.

1. Basic currency board equations. A Supply-Sie oel Uner Currency Boar Rules George Chobanov Faculty of Economics an Business St. Kliment Ohriski University Tsarigrasko chaussee 125, Block 3 1113 Sofia, Bulgaria Abstract This paper provies

More information

Dynamic Demand for New and Used Durable Goods without Physical Depreciation: The Case of Japanese Video Games

Dynamic Demand for New and Used Durable Goods without Physical Depreciation: The Case of Japanese Video Games Dynamic Deman for New an Use Durable Goos without Physical Depreciation: The Case of Japanese Vieo Games Masakazu Ishihara Stern School of Business New York University Anrew Ching Rotman School of Management

More information

arxiv: v1 [math.ds] 30 Mar 2017

arxiv: v1 [math.ds] 30 Mar 2017 MEAN CURVATURE FLOW WITH DRIVING FORCE ON FIXED EXTREME POINTS LONGJIE ZHANG arxiv:173.179v1 [math.ds] 3 Mar 217 Abstract: In this paper, we consier the mean curvature flow with riving force on fixe extreme

More information

Forthcoming in The Journal of Banking and Finance

Forthcoming in The Journal of Banking and Finance Forthcoming in The Journal of Banking an Finance June, 000 Strategic Choices of Quality, Differentiation an Pricing in Financial Services *, ** Saneep Mahajan The Worl Bank (O) 0-458-087 Fax 0-5-530 email:

More information

Econ 582 Nonlinear Regression

Econ 582 Nonlinear Regression Econ 582 Nonlinear Regression Eric Zivot June 3, 2013 Nonlinear Regression In linear regression models = x 0 β (1 )( 1) + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β it is assumed that the regression

More information