What can we do with numerical optimization?

Similar documents
1 Overview. 2 The Gradient Descent Algorithm. AM 221: Advanced Optimization Spring 2016

Trust Region Methods for Unconstrained Optimisation

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Stochastic Programming and Financial Analysis IE447. Midterm Review. Dr. Ted Ralphs

Lecture Quantitative Finance Spring Term 2015

Decomposition Methods

Optimization in Finance

An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity

CS 3331 Numerical Methods Lecture 2: Functions of One Variable. Cherung Lee

First-Order Methods. Stephen J. Wright 1. University of Wisconsin-Madison. IMA, August 2016

Optimization Methods in Finance

Statistics and Machine Learning Homework1

Outline. 1 Introduction. 2 Algorithms. 3 Examples. Algorithm 1 General coordinate minimization framework. 1: Choose x 0 R n and set k 0.

GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE TRUST-REGION ALGORITHMS TO FIRST AND SECOND ORDER CRITICAL POINTS

Chapter 7 One-Dimensional Search Methods

Calibration Lecture 1: Background and Parametric Models

I. More Fundamental Concepts and Definitions from Mathematics

Lecture 17: More on Markov Decision Processes. Reinforcement learning

Convergence Analysis of Monte Carlo Calibration of Financial Market Models

CS227-Scientific Computing. Lecture 6: Nonlinear Equations

Market Risk Analysis Volume I

1 Explicit Euler Scheme (or Euler Forward Scheme )

A Trust Region Algorithm for Heterogeneous Multiobjective Optimization

EC316a: Advanced Scientific Computation, Fall Discrete time, continuous state dynamic models: solution methods

Chapter 5 Portfolio. O. Afonso, P. B. Vasconcelos. Computational Economics: a concise introduction

CSCI 1951-G Optimization Methods in Finance Part 00: Course Logistics Introduction to Finance Optimization Problems

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Budget Management In GSP (2018)

56:171 Operations Research Midterm Exam Solutions October 22, 1993

Dynamic Portfolio Execution Detailed Proofs

SYLLABUS AND SAMPLE QUESTIONS FOR MSQE (Program Code: MQEK and MQED) Syllabus for PEA (Mathematics), 2013

Interior-Point Algorithm for CLP II. yyye

1 Explicit Euler Scheme (or Euler Forward Scheme )

Advanced Operations Research Prof. G. Srinivasan Dept of Management Studies Indian Institute of Technology, Madras

Final exam solutions

Taylor Series & Binomial Series

Financial Optimization ISE 347/447. Lecture 15. Dr. Ted Ralphs

Convergence of trust-region methods based on probabilistic models

Sensitivity Analysis with Data Tables. 10% annual interest now =$110 one year later. 10% annual interest now =$121 one year later

Stochastic Optimal Control

Course notes for EE394V Restructured Electricity Markets: Locational Marginal Pricing

Economics 2010c: Lecture 4 Precautionary Savings and Liquidity Constraints

Adaptive cubic overestimation methods for unconstrained optimization

CSCI 1951-G Optimization Methods in Finance Part 07: Portfolio Optimization

FINANCIAL OPTIMIZATION

Answer Key for M. A. Economics Entrance Examination 2017 (Main version)

CONSUMER OPTIMISATION

The Correlation Smile Recovery

IE 495 Lecture 11. The LShaped Method. Prof. Jeff Linderoth. February 19, February 19, 2003 Stochastic Programming Lecture 11 Slide 1

Is Greedy Coordinate Descent a Terrible Algorithm?

Martingale Pricing Theory in Discrete-Time and Discrete-Space Models

Contents Critique 26. portfolio optimization 32

Ellipsoid Method. ellipsoid method. convergence proof. inequality constraints. feasibility problems. Prof. S. Boyd, EE364b, Stanford University

Technical Report Doc ID: TR April-2009 (Last revised: 02-June-2009)

Log-Robust Portfolio Management

Online Shopping Intermediaries: The Strategic Design of Search Environments

The Optimization Process: An example of portfolio optimization

A model reduction approach to numerical inversion for parabolic partial differential equations

MACROECONOMICS. Prelim Exam

Yao s Minimax Principle

Introduction to Operations Research

Penalty Functions. The Premise Quadratic Loss Problems and Solutions

An advanced method for preserving skewness in single-variate, multivariate, and disaggregation models in stochastic hydrology

Multi-period Portfolio Choice and Bayesian Dynamic Models

Heinz W. Engl. Industrial Mathematics Institute Johannes Kepler Universität Linz, Austria

Optimization for Chemical Engineers, 4G3. Written midterm, 23 February 2015

LP OPTIMUM FOUND AT STEP 2 OBJECTIVE FUNCTION VALUE

Numerical schemes for SDEs

Characterization of the Optimum

Systems of Ordinary Differential Equations. Lectures INF2320 p. 1/48

1 Dynamic programming

A model reduction approach to numerical inversion for parabolic partial differential equations

Modelling long term interest rates for pension funds

Lecture 7: Bayesian approach to MAB - Gittins index

Consistency of option prices under bid-ask spreads

Global convergence rate analysis of unconstrained optimization methods based on probabilistic models

Estimating Macroeconomic Models of Financial Crises: An Endogenous Regime-Switching Approach

Microeconomics II. CIDE, MsC Economics. List of Problems

6.896 Topics in Algorithmic Game Theory February 10, Lecture 3

Gradient Descent and the Structure of Neural Network Cost Functions. presentation by Ian Goodfellow

Multi-armed bandits in dynamic pricing

A Stochastic Levenberg-Marquardt Method Using Random Models with Application to Data Assimilation

Accelerated Stochastic Gradient Descent Praneeth Netrapalli MSR India

6.231 DYNAMIC PROGRAMMING LECTURE 10 LECTURE OUTLINE

1 Precautionary Savings: Prudence and Borrowing Constraints

Online Appendix Optimal Time-Consistent Government Debt Maturity D. Debortoli, R. Nunes, P. Yared. A. Proofs

An Efficient Monte Carlo Method for Optimal Control Problems with Uncertainty

Agricultural and Applied Economics 637 Applied Econometrics II

Short-time-to-expiry expansion for a digital European put option under the CEV model. November 1, 2017

Applied Mathematical Sciences, Vol. 8, 2014, no. 1, 1-12 HIKARI Ltd,

Machine Learning (CSE 446): Pratical issues: optimization and learning

Lecture 4. Finite difference and finite element methods

Optimal Security Liquidation Algorithms

THE TRAVELING SALESMAN PROBLEM FOR MOVING POINTS ON A LINE

Reduced Basis Methods for MREIT

ECS171: Machine Learning

Interpolation. 1 What is interpolation? 2 Why are we interested in this?

The mean-variance portfolio choice framework and its generalizations

Non-Deterministic Search

Statistical and Computational Inverse Problems with Applications Part 5B: Electrical impedance tomography

Transcription:

Optimization motivation and background Eddie Wadbro Introduction to PDE Constrained Optimization, 2016 February 15 16, 2016 Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (1 : 21) Acknowledgement This mini-course is supported by an initiation grant from STINT, the Swedish Foundation for International Cooperation in Research and Higher Education Handout version of the slides presented during the lectures can be downloaded from http://www8.cs.umu.se/~eddiew/optpde2016/ Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (2 : 21) What can we do with numerical optimization? Operations research Allocation of resources for industrial production Finding logistics, scheduling, or transportation solutions Crew scheduling for airline cabin personel Managing investment portfolios Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (3 : 21)

What can we do with numerical optimization? Inverse problems Estimate material parameters from measurements Oil exploration; exploit data to obtain subsurface images Medical tomography Non-destructive testing Estimate initial conditions for numerical weather models from weather data Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (4 : 21) What can we do with numerical optimization? Engineering Design Which shape is the best? Which material composition is the right one?... Example: Optimization of a cantilever beam. Use 50 % material while minimizing the compliance of the beam Ω D f Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (5 : 21) Course objectives for todays lecture Become familiar with the language of optimization Seeing examples of practical problems formulated as optimization problems Introduce important classes of optimization problems Terminology Characterization of solutions Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (6 : 21)

Course objectives Introduce important numerical methods to solve optimization problems Newton-type methods for unconstrained optimization Feasible-point and barrier/penalty methods for nonlinear programming Obtaining hands-on experience with software Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (7 : 21) A general optimization problem Minimize f (x) among all x such that g i (x) = 0 i = 1, 2,..., n eq h i (x) 0 i = 1, 2,..., n ineq l i x i u i for some i s where f, g i, h i are functions from R n to R Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (8 : 21) Terminology f x g i (x) = 0 h i (x) 0 l i x i u i objective function descision variables equality constraints inequality constraints box constraints (special case of inequality constraints) An x is feasible if the constraints are satisfied at x. Otherwise x is infeasible. The feasible set S consist of all x that satisfies all constraints. Using S the general problem can be written as min x S f (x). Assume that x is feasible for an inequality constraint h i (x) 0 The constraint is active (or binding) if h i (x) = 0 Otherwise the constraint is inactive (nonbinding, or slack). A problem is unconstrained if there are no constrains. Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (9 : 21)

Global and local minimizers S R n f (x) Global minimizer: x S such that f (x ) f (x) Global x S S minimizer: x S such that, for some ɛ > 0, f (x ) f (x) x S B(x ; ɛ), where B(x ; ɛ) is a ball of radius ɛ surrounding x, that is, (B(x ; ɛ) = {x R n x x ɛ}) Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (10 : 21) Convexity and optimality A convex function f on a convex set S: f is below or on the linear interpolant between any two points in the set αf (x 1) + (1 α)f (x 2) f (x) x 1 x 2 x 1, x 2 S f ( ) αx 1 + (1 α)x 2 αf (x1 ) + (1 α)f (x 2 ) α [0, 1] Examples of convex functions on R n (a) f (x) = c T x (b) f (x) = γ + c T x + 1 2 x T Qx (also concave!) (Q positive semidefinite matrix) Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (11 : 21) S R n Global minimizer: x S such that f (x ) f (x) x S f (x) minimizer: x S such that, for some ɛ > 0, f (x ) f (x) x S B(x ; ɛ) where B(x; ɛ): ball of radius ɛ centered at x Global S Theorem. For convex functions on convex sets holds that each local minimizer is a global minimizer Example: min f (x) such that x Ax b, where f (x) = c T x (linear program) or f (x) = γ+c T x+ 1 2 x T Qx, Q positive semidefinite (quadratic program) Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (12 : 21)

Unconstrained optimization Will start to consider unconstrained optimization min f (x) or, equivalently, Find x R n such that x R n f (x ) f (x) x R n Function f is nonlinear in x. Unconstrained optimization meaningless for linear f, since linear f on R n are unbounded or constant Most common application for unconstrained optimization: inverse problems, parameter estimation Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (13 : 21) Unconstrained optimization: example application At time intervals t 1, t 2,..., t m a physical process generates a time sequence of m observations (measurements) b 1, b 2,..., b m A model of the process says that b k b(t k ) where b(t) = x 1 + x 2 e x 3t + x 4 e x 5t Model not exact: measuring errors (noise), modeling errors We want to find the coefficients x 1,..., x 5 that best matches the observations Define the residual r k = b k (x 1 + x 2 e x 3t k + x 4 e x 5t k ) and solve the unconstrained optimization problem m min r x 1,...,x k 2 5 k=1 Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (14 : 21) A general iterative algorithm for unconstrained optimization Problem: min f (x) x R n Many optimization algorithms are of the type 1. Specify an initial guess x 0 2. For k = 0, 1,... 2.1 If x k optimal stop 2.2 Determina a search direction p k and a step length α k 2.3 Set x k+1 = x k + p k α k For most problems, optimum cannot be reached within a finite number of steps Important: convergence rate, the behavior of x k x as k Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (15 : 21)

Convergence rate Definition The sequence {x k } k=1 converges to x with rate p and rate constant C < if x k x and x k+1 x lim k x k x p = C. Linear: p = 1 and 0 < C < 1. The error is essentially multiplied by C each iteration. Quadratic: p = 2. Roughly a doubling of the correct digits each iteration. Superlinear: p = 1 and C = 0. Faster than linear". Includes quadratic convergence but also "intermediaterates. Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (16 : 21) Positive definite matrices An n-by-n real matrix A is positive semidefinite if v T Av 0, v R n It is positive definite if v T Av > 0, v {x R n x 0} A positive definite matrix is nonsingular Matrix A is positive definite if and only if matrix A 1 is positive definite A symmetric matrix A is positive definite if and only if All eigenvalues are strictly positive A = LL T with L lower triangular and L ii > 0 (Cholesky factorization). Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (17 : 21) Quadratic functions Let φ : R n R be definied by φ(s) = αg T s + 1 2 st Hs, where α R, g and s are n-vectors, and H an n-by-n matrix. Theorem. If H is symmetric and positive definite, then the solution to is the unique minimizer of φ Hs = g Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (18 : 21)

Finding minima Finding global minimum: very hard! Reason: in general refers to all points minimum is also very hard to find if nothing more about f is known mimumum easier if f is differentiable, since gradients provide information about local behaviour of functions. Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (19 : 21) Taylor series with remainder term Let f : R n R be of class C 2 (twice continuously differentiable). Then x, y R n, f (y) = f (x) + (y x) T f (x) + 1 2 (y x)t 2 f (ξ)(y x), where ξ = αx + (1 α)y for some α [0, 1]. The gradient of f : The Hessian of f : f (x) = 2 f (x) = ( f,..., f ) T x 1 x n 2 f... 2 f x 1 2..... 2 f x n x 1... 2 f xn 2 x 1 x n Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (20 : 21) Optimality conditions for unconstrained optimization First order necessary condition: Assume f : R n R has a local minimum at x = x and that f is differentiable at x = x. Then f (x ) = 0. Second order necessary condition: Assume f : R n R has a local minimum at x = x and that f is of class C 2. Then 2 f (x ) is positive semidefinite. Second order suffcient condition: Assume f : R n R is of class C 2. If f (x ) = 0 and 2 f (x ) is positive definite then x is a local minimizer. Eddie Wadbro, Introduction to PDE Constrained Optimization, February 15 16, 2016 (21 : 21)