introduction to unconstrained optimization: part 1

48
Introduction Optimality Conditions Functions Models and Algorithms Quadratic Forms and Scaling Introduction to Unconstrained Optimization: Part 1 James Allison ME 555 January 29, 2007 James Allison Introduction to Unconstrained Optimization: Part 1

Upload: others

Post on 14-Apr-2022

17 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Introduction to Unconstrained Optimization:Part 1

James Allison

ME 555January 29, 2007

James Allison Introduction to Unconstrained Optimization: Part 1

Page 2: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Monotonicity Analysis

What can MA be used for?

Problem size reduction (elimination of variables andconstraints)

Identification of problems such as unboundedness

Solution of optimization problem in some cases

What can be done if MA does not lead to a solution?

Application of optimality conditions

Use optimality conditions to derive analytical solution

Use numerical algorithms based on optimality conditions

James Allison Introduction to Unconstrained Optimization: Part 1

Page 3: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Monotonicity Analysis

What can MA be used for?

Problem size reduction (elimination of variables andconstraints)

Identification of problems such as unboundedness

Solution of optimization problem in some cases

What can be done if MA does not lead to a solution?

Application of optimality conditions

Use optimality conditions to derive analytical solution

Use numerical algorithms based on optimality conditions

James Allison Introduction to Unconstrained Optimization: Part 1

Page 4: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Monotonicity Analysis

What can MA be used for?

Problem size reduction (elimination of variables andconstraints)

Identification of problems such as unboundedness

Solution of optimization problem in some cases

What can be done if MA does not lead to a solution?

Application of optimality conditions

Use optimality conditions to derive analytical solution

Use numerical algorithms based on optimality conditions

James Allison Introduction to Unconstrained Optimization: Part 1

Page 5: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Monotonicity Analysis

What can MA be used for?

Problem size reduction (elimination of variables andconstraints)

Identification of problems such as unboundedness

Solution of optimization problem in some cases

What can be done if MA does not lead to a solution?

Application of optimality conditions

Use optimality conditions to derive analytical solution

Use numerical algorithms based on optimality conditions

James Allison Introduction to Unconstrained Optimization: Part 1

Page 6: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Scope

Ch. 4: Unconstrained Optimization

Concerned only with objective function

Constrained optimization covered in Ch. 5

minx

f (x)

Assumptions

Functions and variables are continuous

Functions are C 2 smooth

James Allison Introduction to Unconstrained Optimization: Part 1

Page 7: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Scope

Ch. 4: Unconstrained Optimization

Concerned only with objective function

Constrained optimization covered in Ch. 5

minx

f (x)

Assumptions

Functions and variables are continuous

Functions are C 2 smooth

James Allison Introduction to Unconstrained Optimization: Part 1

Page 8: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

MotivationOutline

Lecture Outline

Derivation of optimality conditions

Analytical solutions

Function approximations

Numerical methods

First order methods (gradient descent)Second order methods (Newton’s Method)

Problem scaling

James Allison Introduction to Unconstrained Optimization: Part 1

Page 9: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Optimality Conditions

Necessary Conditions

B ⇒ AB is true only if A is true (A is necessary for B)

Sufficient Conditions

A⇒ BB is true if A is true (A is sufficient for B)

Necessary and Sufficient Conditions

B ⇐⇒ AB is true if and only if A is true (A is necessary and sufficient for B)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 10: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review

Gradient: ∇f (x)

Multidimensional derivative

Vector valued (column in my documents, row in POD2)

∇f (x) = [∂f /∂x1, ∂f /∂x2, . . . , ∂f /∂xn]T

Points in direction of steepest ascent

Example:

f (x , y) = x2 +2y2−xy

x

y

x2 + 2 y2 − x y

x=[1 1] T

∇ f(x) = [1 3] T

0 1 2 3 4 50

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

James Allison Introduction to Unconstrained Optimization: Part 1

Page 11: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review

Gradient: ∇f (x)

Multidimensional derivative

Vector valued (column in my documents, row in POD2)

∇f (x) = [∂f /∂x1, ∂f /∂x2, . . . , ∂f /∂xn]T

Points in direction of steepest ascent

Example:

f (x , y) = x2 +2y2−xy

x

y

x2 + 2 y2 − x y

x=[1 1] T

∇ f(x) = [1 3] T

0 1 2 3 4 50

0.5

1

1.5

2

2.5

3

3.5

4

4.5

5

James Allison Introduction to Unconstrained Optimization: Part 1

Page 12: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review

Hessian: H, sometimes written ∇2f (x)

Multidimensional second derivative

Matrix valued (symmetric)

Provides function shape information

H =

∂2f (x)∂x2

1

∂2f (x)∂x1x2

∂2f (x)∂x2x1

∂2f (x)∂x2

2

Example:

f (x , y) = x2 + 2y2 − xy

James Allison Introduction to Unconstrained Optimization: Part 1

Page 13: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review

Hessian: H, sometimes written ∇2f (x)

Multidimensional second derivative

Matrix valued (symmetric)

Provides function shape information

H =

∂2f (x)∂x2

1

∂2f (x)∂x1x2

∂2f (x)∂x2x1

∂2f (x)∂x2

2

Example:

f (x , y) = x2 + 2y2 − xy

James Allison Introduction to Unconstrained Optimization: Part 1

Page 14: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review: Taylor’s series expansion

Function of one variable:

f (x) =∞∑

n=0

f (n)(x0)

n(x − x0)

n

Function of multiple variables:

f (x) = f (x0) +n∑

i=1

∂f (x0)

∂xi(xi − xi0)

+1

2

n∑i=1

n∑j=1

∂2f (x0)

∂xi∂xj(xi − xi0)(xj − xj0) + o

(‖x− x0‖2

)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 15: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Math Review: Taylor’s series expansion

Function of one variable:

f (x) =∞∑

n=0

f (n)(x0)

n(x − x0)

n

Function of multiple variables: (matrix form)

∂f , f (x)− f (x0) = ∇f (x0)∂x + ∂xTH∂x + o(‖x− x0‖2

)where ∂x , x− x0 is the perturbation vector

James Allison Introduction to Unconstrained Optimization: Part 1

Page 16: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Derivation of Optimality Conditions

First order necessity

Suppose that x∗ is a local minimum of f (x).

no perturbations about x∗ will result in a function decrease

first order function approximation can be used to derivenecessary conditions (i.e., if x∗ is a minimum, then what mustbe true?)

If x∗ minimizes f (x), then ∇f (x∗) = 0If ∇f (x†) = 0, then x† is a stationary point, but not necessarily aminimizer

James Allison Introduction to Unconstrained Optimization: Part 1

Page 17: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Derivation of Optimality Conditions

First order necessity

Suppose that x∗ is a local minimum of f (x).

no perturbations about x∗ will result in a function decrease

first order function approximation can be used to derivenecessary conditions (i.e., if x∗ is a minimum, then what mustbe true?)

If x∗ minimizes f (x), then ∇f (x∗) = 0

If ∇f (x†) = 0, then x† is a stationary point, but not necessarily aminimizer

James Allison Introduction to Unconstrained Optimization: Part 1

Page 18: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Derivation of Optimality Conditions

First order necessity

Suppose that x∗ is a local minimum of f (x).

no perturbations about x∗ will result in a function decrease

first order function approximation can be used to derivenecessary conditions (i.e., if x∗ is a minimum, then what mustbe true?)

If x∗ minimizes f (x), then ∇f (x∗) = 0If ∇f (x†) = 0, then x† is a stationary point, but not necessarily aminimizer

James Allison Introduction to Unconstrained Optimization: Part 1

Page 19: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Derivation of Optimality Conditions

Second order sufficiency

Suppose that x† is a stationary point of f (x).

no perturbations about x∗ will result in a function decrease

second order function approximation can be used to derivenecessary conditions (i.e., what must be true for x† to be aminimum?)

If H � 0 at x†, then x† minimizes f (x)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 20: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Derivation of Optimality Conditions

Second order sufficiency

Suppose that x† is a stationary point of f (x).

no perturbations about x∗ will result in a function decrease

second order function approximation can be used to derivenecessary conditions (i.e., what must be true for x† to be aminimum?)

If H � 0 at x†, then x† minimizes f (x)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 21: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Analytical Solution Example

minx

f (x) = x21 + 2x2

2 − x1x2 + x1

1) Identify stationary point

x† =

[−4

7−1

7

]

2) Test stationary point

H =

[2 −1−1 4

]� 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 22: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Analytical Solution Example

minx

f (x) = x21 + 2x2

2 − x1x2 + x1

1) Identify stationary point

x† =

[−4

7−1

7

]2) Test stationary point

H =

[2 −1−1 4

]� 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 23: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

LogicMath ReviewDerivation

Analytical Solution Example

minx

f (x) = x21 + 2x2

2 − x1x2 + x1

1) Identify stationary point

x† =

[−4

7−1

7

]2) Test stationary point

H =

[2 −1−1 4

]� 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 24: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Functions Models and Algorithms

Analytical solution to optimization problem frequentlyimpossible (why?)

Numerical algorithms based on FONC and function modelscan be used

James Allison Introduction to Unconstrained Optimization: Part 1

Page 25: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

First Order Model: Gradient Descent

f (x) ≈ f (xk) +∇f (xk)∂x

1 Build a linear model about the current point xk

2 Move in the direction of steepest descent (−∇f (xk)) untilf (x) stops improving

3 Update the linear model and repeat until no descent directionexists (∇f (xk) = 0)

Iterative formula:xk+1 = xk − α∇f (xk)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 26: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Line Search

f (x) = 7x21 + 2.4x1x2 + x2

2 , xk = [10 − 5]T

minα

f (α) = x0 − α∇f (x0)

x1

x 2

Example 1: Convex Quadratic Function

xs

x0

−15 −10 −5 0 5 10 15−15

−10

−5

0

5

10

15

0 0.05 0.1 0.15 0.20

500

1000

1500

2000

2500

α

f(x 0−α

∇ f(

x 0))

Line Search View

James Allison Introduction to Unconstrained Optimization: Part 1

Page 27: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Discussion on Gradient Descent Method

Stability/optimality

Descent

Speed of convergence

James Allison Introduction to Unconstrained Optimization: Part 1

Page 28: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Second Order Model: Newton’s Method

f (x) ≈ f (xk) +∇f (xk)∂x + ∂xTH∂x

1 Build a quadratic model about the current point xk

2 Go to the quadratic approximation for the stationary point

3 Update the quadratic model and repeat until the current pointis a stationary point (∇f (xk) = 0)

Iterative formula:

xk+1 = xk −H−1∇f (xk)

James Allison Introduction to Unconstrained Optimization: Part 1

Page 29: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Second Perspective on Newton’s Method

Newton’s method for root finding (solve f (x) = 0):

xk+1 = xk − f (xk)

f ′(xk)

Multidimensional system of equations (solve f(x) = 0):

xk+1 = xk − J−1f(xk)

What system of equations do we need to solve in unconstrainedoptimization?

∇f (x) = 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 30: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Second Perspective on Newton’s Method

Newton’s method for root finding (solve f (x) = 0):

xk+1 = xk − f (xk)

f ′(xk)

Multidimensional system of equations (solve f(x) = 0):

xk+1 = xk − J−1f(xk)

What system of equations do we need to solve in unconstrainedoptimization?

∇f (x) = 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 31: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Second Perspective on Newton’s Method

Newton’s method for root finding (solve f (x) = 0):

xk+1 = xk − f (xk)

f ′(xk)

Multidimensional system of equations (solve f(x) = 0):

xk+1 = xk − J−1f(xk)

What system of equations do we need to solve in unconstrainedoptimization?

∇f (x) = 0

James Allison Introduction to Unconstrained Optimization: Part 1

Page 32: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Optimization AlgorithmsFirst Order Models and AlgorithmsSecond Order Models and Algorithms

Discussion on Newton’s Method

Stability/optimality

Descent

Speed of convergence

James Allison Introduction to Unconstrained Optimization: Part 1

Page 33: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Quadratic Forms

Quadratic Form: function is a linear combination of xixj terms

Matrix representation:

f (x) = x′Ax

Example: f (x) = 2x21 + x1x2 + x2

2 + x2x3 + x23

f (x) = [x1 x2 x3]

2 0.5 00.5 1 0.50 0.5 1

x1

x2

x3

= x′Ax

James Allison Introduction to Unconstrained Optimization: Part 1

Page 34: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Quadratic Forms

Quadratic Form: function is a linear combination of xixj terms

Matrix representation:

f (x) = x′Ax

Example: f (x) = 2x21 + x1x2 + x2

2 + x2x3 + x23

f (x) = [x1 x2 x3]

2 0.5 00.5 1 0.50 0.5 1

x1

x2

x3

= x′Ax

James Allison Introduction to Unconstrained Optimization: Part 1

Page 35: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Types of Quadratic Functions

if x′Ax > 0 ∀x, A is positive definite⇒ convex quadraticfunction

if x′Ax < 0 ∀x, A is negative definite⇒ concave quadraticfunction

if x′Ax ≶ 0, A is indefinite⇒ hyperbolic quadratic function

Physical interpretation?

James Allison Introduction to Unconstrained Optimization: Part 1

Page 36: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

� ����

���

� �������

� ���

�������

� � ���

� �����

��

��� ���� �����������������! #"$� %��� ��

�&

'( )*

�+�

),

��� ��-� �����������������! #"$� .�/� ��

� � � � ��� � � � � ��� � �� � �

� ���

� �

���

� �

� ����

���

� �������

� � �����

� � � ���

� �������

� � ���

��

��0 �� �-� ���������������% 1"2� .�/� �0

� &

'( )*

��

),

��0 �� �-� ���������������% 1"2� .�/� �0

� � � � ��� � � � � ��� � �� � �

� ���

� �

���

� �

� ����

���

� �������� � ������ � � ���� �������� � ���

� ���

�������

� �

3�4$5 � �76 ��8 �% 9�������������/�% #"$� %��� ��

� &

'( )*

��

),

3�4$5 � �76 ��8 �% 9�������������/�% #"$� %��� ��

� � � � ��� � � � � ��� � �� � �

� ���

� �

���

� �

James Allison Introduction to Unconstrained Optimization: Part 1

Page 37: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Quadratic Function Definitions

f1(x) = x′A1x

f2(x) = x′A2x

f3(x) = x′A3x

Where:

A1 =

[7 1.2

1.2 1

], A2 =

[−7 1.21.2 −1

], and A3 =

[−5 2.62.6 2

]

James Allison Introduction to Unconstrained Optimization: Part 1

Page 38: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Eigenvalues, Eigenvectors, and Function Geometry

Av = λvv eigenvector

λ eigenvalue

Provide insight into function shape

Facilitate useful coordinate system rotations

Helpful for understanding problem condition (ellipticity) andscaling

James Allison Introduction to Unconstrained Optimization: Part 1

Page 39: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Eigenvalues, Eigenvectors, and Function Geometry

Av = λvv eigenvector

λ eigenvalue

Provide insight into function shape

Facilitate useful coordinate system rotations

Helpful for understanding problem condition (ellipticity) andscaling

James Allison Introduction to Unconstrained Optimization: Part 1

Page 40: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Interpretation of Eigenvalues

f (x) = f0 + x′b + x′Ax

∇f (x) = b + 2Ax

x† = −1

2A−1b

Shift coordinate system:

f (z) = f† + z′Az

Rotate coordinate system:

f (p) = f† +n∑

i=1

λip2i

James Allison Introduction to Unconstrained Optimization: Part 1

Page 41: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Interpretation of Eigenvalues

f (x) = f0 + x′b + x′Ax

∇f (x) = b + 2Ax

x† = −1

2A−1b

Shift coordinate system:

f (z) = f† + z′Az

Rotate coordinate system:

f (p) = f† +n∑

i=1

λip2i

James Allison Introduction to Unconstrained Optimization: Part 1

Page 42: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Interpretation of Eigenvalues

f (x) = f0 + x′b + x′Ax

∇f (x) = b + 2Ax

x† = −1

2A−1b

Shift coordinate system:

f (z) = f† + z′Az

Rotate coordinate system:

f (p) = f† +n∑

i=1

λip2i

James Allison Introduction to Unconstrained Optimization: Part 1

Page 43: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Numerical Examples

Function 1:

v1 =

[.189−.982

], λ1 = .769, v2 =

[−.982−.189

], λ2 = 7.23

Function 2:

v1 =

[−.982.189

], λ1 = −.723, v2 =

[−.189−.982

], λ2 = −.769

Function 3:

v1 =

[−.949.314

], λ1 = −5.86, v2 =

[−.314−.949

], λ2 = 2.86

James Allison Introduction to Unconstrained Optimization: Part 1

Page 44: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Numerical Examples

���

���

����� � ��������������������� � � �!� �"

#�$�% % $�%#�$�&

#�$�%

#�&

%

&

$�%

$�&

���

���

���� �'� � �������������!���(�"� �'��� �"

#�$�% % $�%#�$�&

#�$�%

#�&

%

&

$�%

$�&

�"�

���

)+*", � ��- �". ����������������������� � �'��� �"

#�$�% % $�%#�$�&

#�$�%

#�&

%

&

$�%

$�&

λ1 = .769λ2 = 7.23

λ1 = −.723λ2 = −.769

λ1 = −5.86λ2 = 2.86

James Allison Introduction to Unconstrained Optimization: Part 1

Page 45: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Connection with Quadratic Form

x′Ax > 0 ∀x ⇐⇒ λi > 0 ∀i⇒ A � 0 ∧ convex quadratic function

x′Ax < 0 ∀x ⇐⇒ λi < 0 ∀i⇒ A ≺ 0 ∧ concave quadratic function

if x′Ax ≶ 0 ⇐⇒ λi ≶ 0⇒ A is indefinite ∧ hyperbolic quadratic function

James Allison Introduction to Unconstrained Optimization: Part 1

Page 46: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Problem Scaling

James Allison Introduction to Unconstrained Optimization: Part 1

Page 47: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Condition Number

C =λmax

λmin

C � 1⇒ numerical difficulties, slow convergence

C ≈ 1⇒ faster convergence

James Allison Introduction to Unconstrained Optimization: Part 1

Page 48: Introduction to Unconstrained Optimization: Part 1

IntroductionOptimality Conditions

Functions Models and AlgorithmsQuadratic Forms and Scaling

Quadratic FormsEigenvaluesProblem Scaling

Scaling Approaches

Scale design variables to be the same magnitude: y = s′x

Account for v not aligned with coordinate axes: y = S−1x

Implement scaling within algorithm:

James Allison Introduction to Unconstrained Optimization: Part 1