quadratic programming
TRANSCRIPT
Quadratic programming
Under guidance of:Dr.G.Agarwal
(Prof. Deptt.of Mech. Engg. , MNIT) Presented by : Anil Kumar Sharma (2011PMM5023) Ramniwas Saran (2010PMM133)
Quadratic programming
A linearly constrained optimization problem with a quadratic objective function is calleda quadratic program (QP). Because of its many applications, quadratic programming is often viewed as a discipline in and of itself .We begin this section by examining the Karush-Kuhn-Tucker conditions for the QP and see that they turn out to be a set of linear equalities and complementarity constraints.
The quadratic programming problem can be formulated as
Minimize with respect to xF(X)=CTX+1/2XTQXSUBJECT TO:
AX<=B Inequality constraint
EX=D equality constraint
X>=0
Equality Constraints
Maximize f(X)s.t. gi(X) = bi
Set up the Lagrangian function:
L(X,) = f(X) - ii(gi(X)-bi)
Constrained Optimization
• Equality constraints – often solvable by calculus
• Inequality constraints – sometimes solvable by numerical methods
Optimizing the Lagrangian
Differentiate the Lagrangian function withrespect to X and . Set the partial derivatives equal to zero and solve thesimultaneous equation system. Examinethe bordered Hessian for concavity conditions. The "border" of this Hessianis comprised of the first partial derivativesof the constraint function, with respect to , X1, and X2.
7
The Principles and Geometries of KKT of Optimization
X)(
)(
21
21
s.t.min
x ,...,x ,x
x ,...,x ,xf
n
n
8
Geometries of KKT: Unconstrained
• Problem:• Minimize f(x), where x is a vector that could have any values, positive or negative
• First Order Necessary Condition (min or max): f(x) = 0 (∂f/∂xi = 0 for all i) is the first order necessary condition for optimization
• Second Order Necessary Condition: 2f(x) is positive semidefinite (PSD)
• [ x • 2f(x) • x ≥ 0 for all x ]• Second Order Sufficient Condition
(Given FONC satisfied) 2f(x) is positive definite (PD)
• [ x • 2f(x) • x > 0 for all x ] ∂f/∂xi = 0
xi
f
∂2f/∂xi2 > 0
9
Geometries of KKT: Equality Constrained (one constraint)
• Problem:• Minimize f(x), where x is a vector• Subject to: h(x) = b
• First Order Necessary Condition for minimum (or for maximum): f(x) = h(x) for some free ( is a scalar)
Two surfaces must be tangent
h(x) = b and -h(x) = -b are the same;there is no sign restriction on
h(x) = b
10
Geometries of KKT: Equality Constrained (one constraint)
• First Order Necessary Condition: f(x) = h(x) for some
• Lagrangian:• L(x, ) = f(x) - [h(x)-b] , • Minimize L(x, ) over x and Maximize L(x, ) over . Use principles of
unconstrained optimizationL(x, ) = 0:xL(x, ) = f(x) - h(x) = 0L(x, ) = h(x)-b = 0
11
Geometries of KKT: Equality Constrained (multiple constraints)
• Problem:• Minimize f(x), where x is a vector• Such that: hi(x) = bi for i = 1,2,…,m
• KKT Conditions (Necessary Conditions): Exist i ,i = 1,2,…,m, such that f(x) = i=1
n ihi(x) • hi(x) = bi for i = 1,2,…,m
• Such a point (x, ) is called a KKT point, and is called the Dual Vector or the Lagrange Multipliers. Furthermore, these conditions are sufficient if f(x) is convex and hi(x), i = 1,2,…,m, are linear.
12
Geometries of KKT: Inequality Constrained (one constraint)
• Problem:• Mimimize f(x), where x is a vector• Subject to: g(x) ≥ b
• Equivalently:• f(x) = g(x)• g(x) ≥ b > 0 • [g(x) – b] = 0
13
Geometries of KKT: Inequality Constrained (two constraints)
• Problem:• Minimize f(x), where x is a vector• Subject to: g1(x) ≥ b1 and g2(x) ≥ b2
• First Order Necessary Conditions: f(x) = 1 g1(x) + 2 g2(x), 1 ≥ 0, 2 ≥ 0
f(x) lies in the cone between g1(x) and g2(x)• g1(x) > b1 1 = 0 • g2(x) > b2 2 = 0 1 [g1(x) - b1] = 0 2 [g2(x) - b2] = 0
• Shaded area is feasible set with two constraints
x1
x2 -g1(x)
-g2(x)-f(x)
Both constraints are binding
14
Geometries of KKT: Inequality Constrained (two constraints)
• Problem:• Minimize f(x), where x is a vector• Subject to: g1(x) ≥ b1 and g2(x) ≥ b2
• First Order Necessary Conditions: f(x) = 1 g1(x), 1 ≥ 0• g2(x) > b2 2 = 0 • g1(x) - b1 = 0
• Shaded area is feasible set with two constraints
x1
x2 -g1(x)
-f(x)
First constraint is binding
15
Geometries of KKT: Inequality Constrained (two constraints)
• Problem:• Minimize f(x), where x is a vector• Subject to: g1(x) ≥ b1 and g2(x) ≥ b2
• First Order Necessary Conditions: f(x) = 0• g1(x) > b1 1 = 0 • g2(x) > b2 2 = 0
• Shaded area is feasible set with two constraints
x1
x2
f(x)=0
None constraint is binding
16
Geometries of KKT: Inequality Constrained (two constraints)
• Lagrangian:• L(x, 1, 2) = f(x) - 1 [g1(x)-b1] - 2 [g2(x)-b2]
• Minimize L(x, 1, 2) over x. • Use principles of unconstrained maximization L(x, 1, 2) = 0 (gradient with respect to x only) L(x, 1, 2) = f(x) - 1 g1(x) - 2 g2(x) = 0• Thus f(x) = 1 g1(x) + 2 g2(x)
• Maximize L(x, 1, 2) over 1≥ 0, 2 ≥ 0.• g1(x)-b1 > 0, then 1=0• g2(x)-b2 > 0, then 2=0
Bordered Hessian0 g'(x1) g'(x2)
g'(x1) f11 f12g'(x2) f21 f22
For a max, the determinant of this matrix wouldbe positive. For a min, it would be negative. Forproblems with 3 or more variables, the “even”determinants are positive for max, and “odd”ones are negative. For a min, all are negative.
Note: thedeterminantis designated|H2|
Aside on Bordered Hessians
You can also set these up so that the border carriesnegative signs. And you can set these up sothat the border runs along the bottom and theright edge, with either positive or negative signs. Be sure that the concavity condition testsmatch the way you set up the bordered Hessian.
Multiple Constraints SOCM is the number of constraints in a given problem. N the number of variables in a given problem.
The bordered Principle Minor that contains f22 asthe last element is denoted |H2| as before. Iff33 is the last element, we denote |H3|, and so on.
Evaluate |Hm+1| through |Hn|. For a maximum,they alternate in sign. For a min, they all take the sign (-1)M
Interpreting the Lagrangian Multipliers
The values of the Lagrangian multipliers (i) aresimilar to the shadow prices from LP, exceptthey are true derivatives (i = L/bi) and are not usually constant over a range.
Inequality Constraints
Maximize f(X)
s.t. g(X) b X 0
Example
Minimize C = (X1 – 4)2 + (X2 –4)2
s.t. 2x1 + 3x2 ge 6 -3x1 – 2x2 ge –12
x1, x2 ge 0
Graph
76543210 1 2 3 4 5 6 7
Optimum: 2 2/13, 2 10/13
A Nonlinear Restriction
Maximize Profit = 2x1 + x2s.t. -x12 + 4x1 - x2 le 0 2x1 + 3x2 le 12
x1, x2 ge 0
Graph – Profit Max Problem
4
3
2
1
0 1 2 3 4 5 6 7
F1
F2
There is a local optimum at edge of F1, but itisn't global.
KKT Conditions: Final Notes
• KKT conditions may not lead directly to a very efficient algorithm for solving NLPs. However, they do have a number of benefits:• They give insight into what optimal solutions to NLPs look
like• They provide a way to set up and solve small problems• They provide a method to check solutions to large
problems• The Lagrange multipliers can be seen as shadow prices of
the constraints
29
The Kuhn-Tucker Conditions
• xf(X*) - * xg(X*) 0
• [ xF(X*) - * xg(X*)]X* = 0
• X* 0• g(X*) b• *(g(X*)-b) =0• * 0
represents the gradient vector (1st derivatives)
Wolfe’s method
Max. Z = 4X1 + 6X2 -2X12 – 2X1X2 – 2X2
2 = f(x) say
Subjected to constraints-X1 + 2X2 ≤ 2 =g(x),say
X1 , X2 ≥ 0
Non-negativity conditions X1 , X2 ≥ 0 as inequality constraints
Wolfe’s methodStandard form of QP problem-
X1 + 2X2 + S12 = 2
-X1 + r12 = 0
-X2 + r22 = 0
And X1, X2 , S1, r1, r2 ≥ 0
Lagrange function for necessary condition- L(X1, X2 , S1, λ1, µ1, µ2, r1, r2 )
= (4X1 + 6X2 -2X12 – 2X1X2 – 2X2
2) -λ1(X1+2X2+ S12 -2 ) -
µ1(-X1+ r12) - µ2(-X2+ r2
2)
Wolfe’s method
Necessary & sufficient condition for maximization of Lagrange function, hence objective function-
∂L/∂X1 = 4 - 4X1 - 2X2 - 1λ1 + µ1 = 0;
∂L/∂X2 = 6 - 2X1 - 4X2 - 2λ1 + µ2 = 0;
∂L/∂λ1 = -(X1+2X2+S12-2)=0 => X1+2X2+ S1
2-2 = 0
∂L/∂S1 = -2λ1S1 = 0 => 2λ1S1 = 0
∂L/∂µ1 = -(-X1+ r12) = 0; => -X1+ r1
2 = 0
∂L/∂µ2 = - (-X2+ r22) =0; => -X2+ r2
2 =0
∂L/∂r1 = -2µ1r1 = 0; => 2µ1r1 = 0
∂L/∂r2 = -2µ2r2 = 0; => 2µ2r2 = 0
Wolfe’s method
On re-arranging the conditions-4X1 + 2X2 + 1λ1 - µ1 = 4; --------12X1 + 4X2 + 2λ1 - µ2 = 6; --------2X1+ 2X2+ S1
2 = 2;
λ1S1 = 0µ1x1 = 0µ2x2 = 0
X1, X2 , λ1, S1 , µ1, µ2 ≥ 0
Complementary condition
Wolfe’s method
On introducing artificial variable A1 and A2 in constraints 1 & 2 respectively, the modified LP-
Min. Z*= A1 +A2
Subject to- 4X1 + 2X2 + 1λ1 - µ1 + A1 = 4;
2X1 + 4X2 + 2λ1 - µ2 + A2 = 6;
X1+ 2X2+ S12 = 2;
X1, X2 , λ1 , µ1, µ2 , A1, A2 ≥ 0
Initial solution
Cj 0 0 0 0 0 0 1 1
CB
Variable
in basis B
Solution value
Xb
X1 X2 λ1 µ1 µ2 S1 A1 A2
Min. Exchange
ratio
1 A1 4 ④ 2 1 -1 0 0 1 0 1→
1 A2 6 2 4 2 0 -1 0 0 1 3
0 s1 2 1 2 0 0 0 1 0 0 2
Z*=10 Zj 6 6 3 -1 -1 0 1 1
Cj - Zj -6 -6 -3 1 1 0 0 0
↑
Cont…..
Cj 0 0 0 0 0 0 1
CB
Variable in basis
B
Solution value
Xb
X1 X2 λ1 µ1 µ2 S1 A2
Min. Exchange
ratio
0 X1 1 1 1/2 1/4 -1/4 0 0 0 2
1 A2 4 0 3 3/2 1/2 -1 0 1 4/3
0 s1 1 0 3/2 -1/4 ¼ 0 1 0 2/3 →
Z*=4 Zj 0 3 3/2 ½ -1 0 1
Cj - Zj 0 -3 -3/2 -1/2 1 0 0
↑
Cont…..
Cj 0 0 0 0 0 0 1
CB
Variable in basis
B
Solution value
Xb
X1 X2 λ1 µ1 µ2 S1 A2
Min. Exchange
ratio
0 X1 2/3 1 0 1/3 -1/3 0 -1/3 0 2
1 A2 2 0 0 2 0 -1 -2 1 1 →
0 X2 2/3 0 1 -1/6 1/6 0 2/3 0 (-)ve
Z*=2 Zj 0 0 2 0 -1 -2 1
Cj - Zj 0 0 -2 0 1 2 0
↑
Cj 0 0 0 0 0 0
CB
Variable in basis
B
Solution value
Xb
X1 X2 λ1 µ1 µ2 S1
Min. Exchange
ratio
0 X1 1/3 1 0 0 -1/3 1/6 0
0 λ1 1 0 0 1 0 -1/2 -1
0 X2 5/6 0 1 0 1/6 -1/12 ½
Z*=0 Zj 0 0 0 0 0 0
Cj - Zj 0 0 0 0 0 0
Since all Cj – Zj = 0, optimul solution for phase I is reached, hence
X1 = 1/3; X2 =5/6; λ1 =1; Z*= 0, Max.Z =25/6 µ1=0; µ2 =0; S1 =0;
.
Note:-
CONCLUSIONThis presentation has presented for study about a new feature of operations research to solve the problem which has square in powers of variables, these problems are solved by the help of langarnge multiplier, wolfe’s method and with the help of kkt condition .
References
•Hamdy A.Taha,Operations Reasearch an inroduction,Pearson,eighth edition,2011•JK.Sharma,Opertions Research,macmillan,third edition•QUADRATIC PROGRAMMING BASICS,H. E. Krogstad,Spring 2005/Rev. 2008
•Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard
THANKS