avenues of geometric programming* t.r. j c.h,...a central idea here is the arithmetic-geometric mean...
TRANSCRIPT
109
NZOR volume 6 number 2 July 1978
This is the first of two articles - one on theory, the other on applications.The Editor
AVENUES OF GEOMETRIC PROGRAMMING*
T.R. Je f f e r s o n a n d C.H, Sc o t t
T he Un i v e r s i t y o f Ne w So u t h Wa l e s , Au s t r a l i a
Summary
Since its initial appearance more than a decade ago, geometric programming has proved itself to be a powerful means of solving many optimization problems, particularly those pertaining to engineering design. Recently the theory has been generalised to provide a structure for the analysis of any convex programming problem. Here we trace the development of geometric programming from its initial use as a trick for solving certain cost optimization problems to its current status as a fully fledged branch of optimization theory. Examples draum from inventory theory reinforce and motivate this development.
1. Introduction
In the early sixties, Clarence Zener was a Director of
Science at Westinghouse Research Laboratories at Pittsburgh.
In such a position, a physicist becomes involved with cost
modelling. The difference, in Zener's case, was that he o b
served a curious mathematical phenomena with regard to cost
models made up as a sum of component costs [47,48]. He was
able to optimize certain such models by mere inspection of
the exponents of the design variables (Section 2). Thus was
born a new and exciting approach to mathematical programming
which has henceforth been termed geometric programming.
Initial research concentrated on posynomial (polynomial with
positive coefficients) programs where both the objective
function and constraints are in posynomial form (Section 3).
This research culminated in the seminal book entitled " G eo
metric Programming" by Duffin, Peterson and Zener [13].
Subsequent work developed in two significantly different
directions: Signomial Programming (Section 4) and General
ised Geometric Programming (Section 5) [33].
‘M a n u s c r i p t s u b m i t t e d O c t o b e r 28, 1977
110
Signomial Programming, first developed by Passy and
Wilde [32], sought to relax the restriction of the functions
treated by Geometric Programming to the form of polynomials.
With the relaxation to polynomials one loses convexity and
hence one can only hope for a local minimum, not a global
minimum. Generalised Geometric Programming proceeds to
generalise the type of function considered from posynomial
to convex. This generalisation does maintain the convexity
of the problem and hence one is able to find the global sol
ution to the mathematical program.
The relationship among the various developments may
outwardly seem tenuous. This is because Geometric Program
ming should not be thought of as a class of mathematical
programs but as a method of analysis of mathematical programs.
In this respect, geometric programming has much in common
with dynamic programming. Dynamic Programming makes use of
recursive functions. Geometric Programming makes use of
linearity, separability, convexity, and duality.
(i) Linearity
Most mathematical programming is carried out over
linear vector spaces. The usual choice is Euclidean n-
dimensional space (En ) . Here a vector is represented by an
n-tuple of real components. Taking any two vectors x^ and
X 2 belonging to the space, and real scalars a and 3, a
linear vector space has the property that the linear combin
ation ax^+gx2 belongs to the space. In the context of
functions, a function £(x) is said to be linear if for any
scalars a and 6 , £(ax+By) = ail (x) + 0J,(y) . Linear functions
have many important properties. They are both convex and
concave. Further, a first order Taylor series expansion
approximates the linear function exactly. The value of
linearity to mathematical programming is best exemplified by
the fruitful area of linear programming.
Ill
(ii) Separability
A function f(x^,x2 »•••*xn ) is separable if it can be
written as f (x1 ,x2 ,. . . ,xn ) = f x ( x ^ + f 2 (x2)+ . . . +fn (xn ) .
This property is important as it allows one to treat an n-
dimensional function as n functions of 1-dimension. This is
much simpler from a computational point of view.
(iii) Convexity
A set A is convex if for any x 2 e A and 0 4: \ 4 1,
Xx^ + (1-X)x2 e A. To extend the concept of a function f(x)
defined for x e C, a convex set, we require the following
definition. The epigraph of a function f defined on C is
the set A defined by: A A {(a,x) | a > f(x) , x £ C } . f(x)
is a convex function iff A is a convex set. Convex func
tions have a number of powerful properties. The most impor
tant for mathematical programs with a convex objective
function and a convex constraint set is that any local
minimum is also a global minimum. Further properties of
convexity will be given in Section 5.
(iv) Duality
For a linear space X = En , linear functions may be
written in the form <a,x> = E? ,a.x., where a is a vector~ ’ ~ i = l 1 1 * -
parameter. The set of all linear functions (parameterised
by the vector a) is a space and is called the dual space
(X*) . The duality arises from the fact that <a,x> is
symmetric in a and x. For finite dimensional spaces, the
set of linear functions on X* turns out to be the original
space X. Hence there is a symmetric relationship between X
and X * .
Of importance to mathematical programming is the p a i r
ing of sub-spaces in primal and dual spaces with respect toj.
the inner product. Let X be a subspace of X. Then X =
{y | <y,x> = 0, V x e X} is called the orthogonal complemen
tary subspace of X. Any point in X is orthogonal to anyj.
point in X by construction. If £ x anc* a »6 are
112
scalars, it is easy to see that ay^+B}^ e * •
The importance of the concepts of linearity, separab
ility, convexity and duality in the development of geometric
programming will remain obscure in Sections 2, 3 and 4.
However, their paramount role will become clear in Section 5
when we introduce Generalised Geometric Programming. Section
6 uses this generalised theory to re-derive the theory of
posynomial programs, and the central role played by these
concepts will be further clarified. Throughout this survey,
theory will be reinforced by examples drawn, in the main,
from the area of inventory theory. The examples presented
are relatively simple in order to highlight the use of
geometric programming theory.
2. Unconstrained Posynomial Programming
Here we consider the following nonlinear problem:
nMinimize g(t) = Z u- (1)
i = l 1
where , i=l,...,n, are posynomials defined by
m a . .u, = c. n t. (2)
1 j = i J
c. > 0, Vi, t. > 0, Vj , and a-, are arbitrary real con- l J t
stants; t = (t.,...,tm ) and T denotes a transpose. In the
usual situation, g(t) represents a composite cost made up of
component costs u . , i = l,...,n. We term this the primal
p r o b le m.
In principle, the primal problem could be solved by the
methods of differential calculus. This will give rise to a
set of nonlinear equations which, in general, are difficult
to solve. Hence, in practice, one would invariably resort
to a numerical solution by an iterative descent-type method.
Here, we propose to construct a geometric programming dual
program to the primal. In certain cases, this admits a
trivial solution to the problem. Further, this dual
113
problem has an interesting and informative practical inter
pretation.
A central idea here is the arithmetic-geometric mean
inequality, henceforth termed the geometric inequality. In
fact, the name "Geometric Programming" comes from the impor
tance of this inequality to the original theory. It states
that
n n 6.£ fi.v. * n (v.) (3)
i=l 1 1 i“ln
where I 6. B 1, 6- » 0, Vi, v. > 0, Vi. Equality is attain-i=l 1 1 1
ed when v, = v- = ... * v . It is convenient to set 1 L n
V i = ^i in equation (3) to obtain
n n 6.Z u. > n (u./6.) (4)
i=l 1 i=l
In this form, we can use the geometric inequality to
obtain a lower bound on our primal optimization problem.
Hence, substituting equation (2) into (1) and using (4), we
have that
6i m(5)g (t) * n (C./5.) n t.
i=i 1 1 j-i J
If we choose the weights, 5., Vi, such that
nZ a. .6. = 0, j = l ,...,m (6)
i=l 1J 1
then the variables t ., j=l,...,m on the right hand side of
inequality (5) may be eliminated. In this case, we have that
n 6-g(t) > n (c • / 6 •) = v(6) (7)
i = l 1
where v(6) is the dual function which gives a lower bound
on the minimum of g(t). Equation (7) implies that
min g (t) ^ max v(<5) (8 )
114
under conditions t > 0, 6 % 0, = = 1 anc ̂ equation (6).
It may be shown that there exists an optimal t* (in the
sense of minimizing g(t)) and an optimal 6* (in the sense of
maximizing v(6)) to satisfy inequality (8) at equality. In
this case the relation between t* and 6* is given by
<5 i* = u.(t*)/g(t*) (9)
Hence the following dual program to the original primal
program may be constructed:
n 6iMaximize v(6) = II (c./6-) (10)
i = l 1 1
subject to the normalisation condition
nn . » i, 6- => o (ii)
i = l 1 1
and the orthogonality condition
nI a 6 = 0, j = l ,...,m (12)
i = 1 J
At optimality g(t*) = v(<5*).
From equation (9), we see that we can interpret the
dual variables at optimality, 6^ , Vi, as the relative c o n
tribution of each component cost u^ to the composite cost
g(t). Further, we obtain the following relationship between
the primal and dual variables
mIn (6.*v(5*)/c.) = Z a., log t.* (13)
j=l 1J J
This is a system of linear equations in log tj ,
j=l,...,m which are readily solvable once the dual program
has been solved.
We note that the dual program maximizes a nonlinear
function subject to linear equality constraints. A dual
variable 6^ is associated with each term i=l,...,n in the
primal formulation. In the case that the number of terms in
115
the primal n is equal to the number of variables in the
primal m plus one, i.e., n=m+l, the linear constraints admit
a unique solution 6* and the optimization problem in the
dual is trivial. The quantity n-(m+l) is termed the degree
of difficulty of a geometric program and is, in some sense,
a measure of the computational complexity of the dual
program.
Example
We consider the well known "Economic Lot Size" problem
from inventory theory. Items are withdrawn continuously
from inventory at a known constant rate a. Items are or de r
ed in equal numbers, Q, at a time and production is instan
taneous. The problem is to determine how often to make a
production run and how much to order, Q, each time to
minimize the cost, C, per unit time, where C * aK/Q + hQ/2
+ ac, and K is the fixed set-up cost, h is the inventory
holding cost per item per unit time, and c is the variable
production cost per item. Neglecting the constant part, ac,
we require to
minimize * aK/Q ♦ hQ/2
This is an unconstrained posynomial program. The
corresponding dual program, from equations (10), (11) and
(12), is to maximize
6 6 (aK/6^ 1 (h / 262) 2
subject to 5^ + 62 = 1 (normality)
-6i 62 = 0 (orthogonality)
61 ^ 0, 62 > 0
Here the constraint equations can be uniquely solved to
yield 6-̂ * = 62* = h and there is no maximization problem. We
have a program with zero degree of difficulty. The inter
pretation of this result is that at optimality, the set-up
116
cost per unit time and the holding cost per unit time contri
bute equal amounts to the optimal cost, i.e., the optimal
distribution of cost is an invariant with respect to the
cost coefficients. Hence, from equations (10) and (13), the
optimal cost is /2aKh , and the optimal order quantity is
JTaTTK.
Suppose now that the production process is a lengthy
one so that the assumption of instantaneous availability is
no longer acceptable. Furthermore, the size of the lot d e
termines the length of the production process, and this in
turn determines the in-process inventory holding cost. In
this case a modified economic lot size model could be to
minimize = aK/Q + hQ/2 + IQT/2, where I is in-process in
ventory holding cost per unit time, and T is the length of
the production process, given by T(Q) = nQ + m, where n and
m are empirically determined constants. In this case, we
have the unconstrained posynomial program:
Minimize = aK/Q + Q(h/2+Im/2) + In Q^/2
The corresponding dual program is given by
6 6 6Maximize (aK/6^) *((h+Im)/2 62) 2 (m/263) 3
subject to
6i + 62 + 63 = i
-61 + 62 + 26j = 0
6j_ > 0 , 62 > 0 , 63 > 0
Here we have two equations and three unknowns; in
effect, a degree of difficulty of one. One could proceed to
solve this problem as an unconstrained maximization problem
in one variable or as a constrained maximization. However,
insight may be gained without going into this detail. M a n i
pulation of the constraint equations gives the following in
equalities: 0 . 5 $ <5̂ s 1. 0 , and 0 . 3 3 ^ 6 ^ ^ 0 . 6 6 , which imply
that at optimality, the set-up cost makes a contribution
117
between 50% and 66% to the optimal composite cost. Similar
insights may be obtained from the other dual variables.
3. Constrained Posynomial Programming
Here we consider the minimization of a posynomial form
subject to constraints which are also posynomials. This is
the primal formulation.
Minimize gg(t) subject to gk (t) $ 1, k - 1 ,..
1,... ,mm a. . J
gk (j)
t. > 0, m a.. J
where g,.(t) ■ I c. n t . J , k-Q,. ie [k] j-1 J
k - 0 f. .
and m Q-l, m 1-n0+ l i...tmp -np _1+l, nn -n
,P (14,15)
(16)
(17)
(18)
p p-1 ’ p
{ lk]}, k*0,l,...,p is a sequential partition of the integers
1 to n. As before a^j are arbitrary real exponents and the
coefficients c^ are positive.
In order to handle constraints, we need a generalisat
ion of the geometric inequality in which the <5^'s are not
normalised. We let
X* = -Z rv,4i ie [k](19)
for a particular constraint k. Hence the geometric inequal
ity, equation (4) may be written in a convenient form as
( I u.) k > n (u./6.) 1 X, ie[k] ie[k]
(2 0 )
Combining equations (15), (17) and (20) we have that
A-* in a. . 6. ti » gk (t) k * n (c. n t. 1J/6±) \
ietk] j=l J (21)
for any constraint k=l,...,p. For the objective function,
we normalise the 5^, i e [0]. Hence
118
n a . . 6. g n (t) > n (c II t. XJ/6 -) 1 (22)
ietO] 1 j=l J 1
and z 6 - : = l (23)ie [0]
Combining results (21) and (22), we obtain the
inequality
n 6. p X.g0 Ct) n (C./5.) n X. (24)
i=l k-1 1
Hence, using the same lines of reasoning as for the u n
constrained problem, the dual problem is given by:
n 6. p X,Maximize v(6) = n (c./6.) II X. (25)
i=l 1 1 k=l
subject to the linear constraints
Z 6- = 1 (normalisation) (26)ie [0] nZ a . .6- = 0, j=l,...,m (orthogonality)(27)
i=l 1
5i * 0, i-l,...,n (28)
We note that the primal problem is a highly nonlinear
mathematical program, whereas the dual is the maximization
of a concave function subject to linear constraints. Once
again, if n=m+l, the program has zero degree of difficulty,
and the dual program is trivial. It may be shown [13] that
at optimality the primal variables t* and the dual variables
6* are related by
m a . .c, n t.* 1;) =
6^*v(6*), i e [0]
1 j = r j l<Si*/Xk (6*) , i e [k], Xk (6*)>0,k-l,...,p
Hence 6^ , i e [0] gives the relative contribution of
each component cost to the composite cost. Also at
optimality, the primal and dual objective functions are
(29)
119
equal in value, i.e., gQ (t*) * v(6*).
Example
We consider an Economic Lot Size model with m u l t i
products and a resource constraint. In this case we have
the following constrained posynomial program:
NMinimize Z (a.K./Q. + h.Q./2)
i=l 1 1 1 1 1
subject to a resource constraint
NZ b.Q. * W
i = l 1 1
where b^ and W are given constants. From equations (25) to
(28), we find the corresponding dual program:
N 6. 2N 6. 3N 6. .Maximize n (a-K./6-) n (h-/26.) n (b-/W6.) XA
i=l 1 1 i-N+1 1 1 i=2N+l 1 1
subject to Z 5- = 1
i*1 3NX - Z <5-
i-2N+l
’6i + 6N+ i + 62N+i " °* »• • • »N
6^ £ 0, i=l,...,3N
4 . Signomial Programming
While posynomial programming has found application in a
variety of areas, many problems of interest have fallen o u t
side the posynomial form. Often it was the positivity c o n
dition on the coefficients which was violated. However, the
benefits of posynomial programming were too tempting to be
passed up. Passy and Wilde [32] introduced a generalisation
of posynomial programming, where the coefficients of the
terms were not required to be positive. This class of p r o
grams is called signomial. In signomial programming, a
global minimum cannot be guaranteed. However, the p r o pe r
ties of the dual subspace are maintained.
120
Duffin and Peterson [11,12] have developed methods for
the analysis of signomials and an algorithm for solving them
(at least for a local solution, if not for a global solution).
To use the analysis, signomial programs must first be c o n
verted into posynomial programs. Any signomial f(t) can be
written as a difference of two posynomials, say r(t) and
s(t) and hence
f(t) j? 1 -<-+r(t)-s(t) «: 1 +-*■ r (t) s s (t)+l
f(t) >1 +-*-r(t)-s(t) > 1 +-*• r (t) » s(t)+1
f(t) $ 0 +-vr(t)-s(t) 0 +-*■ r (t) ^ s(t)
Since r(t), s(t) and s(t)+l are all non-negative for
all values of t, in each case a new variable x can be intro
duced so that
r(t) H v < s (t)+l «-+ r(t)/x < 1 and l * s ( t ) / x + l/x
r (t) > t » s (t) +1 -*->• r (t) / x >,1 and 1 ^s(t)/x + l/x
r (t) ^ x ^ s (t) *-*■ r(t)/x ^ 1 and 1 s(t)/x
These transformations on a signomial produce a posynom
ial program with reversed constraints:
where the functions g^(t) are posynomials defined by
equation (17). The difference in this formulation is the
reversed constraints (32). While the constraints (31) d e
fine a convex region, the reversed constraints make the
region non-convex. The dual to the primal program defined
by (30) to (33) is given by:
Minimize gQ(t)
subject to S 1 k=l,...,p
gk (t) > 1 k=p + l , . . . ,r
t > Q
(30)
(31)
(32)
121
6. b 6, r -6. Maximize v(S) = II (c./6.) n n (c.X,/6.) K IT II (c.X,/<5.) 1
ie[0] k=l ie[k] 1 K 1 k*p+l ie[k](34)
subject to same constraints as for posynomials, namely
E 6. = 1 (35)ie [0]
Ak " E 6i» k-l,...,r (36)n iElk>E a . .6• - o , j = l ,...,m (37)
i = l 13
6̂ i 0, i « l , . . . ,n (38)
We note that log v(6) is concave in the variables assoc
iated with the regular constraints k*l,...,p and the o bject
ive function. However, it is convex in the variables assoc
iated with the reversed constraints k ap + l ,... ,r. In the
posynomial case, the dual program required finding the m a x
imum of a concave function subject to linear constraints.
Here we are looking for the maximum of a number of saddle
points of a concave-convex function subject to linear c on
straints. If the program has degree of difficulty of zero,
it is possible that a solution can readily be found but in
the general case, the program is very difficult to solve. In
order to cope with this problem, Duffin and Peterson [11]
apply the arithmetic-harmonic mean inequality to convert the
reversed constraints into regular constraints. This a pprox
imation has the advantage that the linear constraints in the
dual are the same for all such approximations. As well, the
sequence of all regular posynomial programs generated is
monotone decreasing in the value of the objective function.
Thus one can only improve any feasible solution.
The arithmetic-harmonic inequality, given parameters
a •, i e [k], positive and E a- = 1, is
i£[k]a( E u . ) 1 n (a./u,) 1 « E ai2/u-j (39)ie[k] i e[k] 1 ie[k]
122
with equality holding when
d; = U./( Z U.) (40)ietk] 1
The means in equation (30) are the harmonic, geometric and
arithmetic respectively. Consider any reversed constraint
gk (t) > 1 or (gk (t))_1 ^ 1
2We set gv (t,ct) Z a- / u - . This is a posynomial for
k ~ ~ ie[k] 1 1 fixed a* We note that gk (t,a) ^ 1 implies that gk (t) > 1.
Using equation (40) for any t such that gk (t) 1, one can
generate an a such that gk (t,a) $ 1. Hence, by use of the
arithmetic - harmonic inequality, we may reduce a posynomial
program with reversed constraints to a regular posynomial
program parameterised with respect to a, viz.
Minimize g0 C't)
subject to Sk*--) < 1 » k = l ,. . . ,p (42)
1, k - p + 1 ,...,r (43)
t > 0 (44)
To obtain a monotone decreasing sequence of programs,
we solve the program (41) to (44) for a particular a and
then reset the a's using equation (40).
Example
We return to the multiproduct economic lot size problem
given in Section 3. Here we add consideration of the d i s
tribution effort. Sales are now a function of the allocat
ion of effort to distribution, and we seek to maximize
profit. The problem becomes:
N d. d.Maximize Z C(Pi -ci)aiD i 1 -D i -K ia iD i 1/Qi 'hiQ i/3)
Nsubject to Z b.Q. s W
i = l 1 1
123
Q. > 0, D i * 0, V i
Heje Pj*c^ is the sales price minus the cost for product
i, 1 are the sales generated by a distribution effort
of and a^ and d^ are parameters. All other symbols have
been defined previously. This is a signomial program and is
equivalent to the following program:
Minimize 1/P
Nsubject to Z b.Q./W < 1
i = 1 1 1N d. d..^1 (Pi -ci)aiD i 1 - D. - Kia iD i 1/Q i - h.Qi/2 >. P (45)
Q i » 0, > 0, V i
P is a new variable indicating profit. Taking the
negative terms in equation (45) to the right hand side and
inserting a new variable R between the sides of the inequal
ity as in the theory, we obtain the program:
Minimize 1/P
Nsubject to E b-Q./W < 1
i*l 1 1d.
P/R ♦ I(Di/R + K ia iD i 1/(Q.R) ♦ h.Q./(2R)) «: 1
d.E(pi -ci)aiD i X/R ^ 1 (46)i
Q i ^ 0, D i * 0 V i
Our original signomial program is now in the form of a
posynomial program with one reversed constraint (46) . This
may be harmonised to become
? d.Z aiZR/((pi -ci)aiD i x) « 1 (47)
where the ct̂ > 0 (and Za^ = 1) are parameters to be varied.
To find a local solution, we take a feasible set of
124
distribution efforts to evaluate ou which is given by
(see equation (40))
d. d.= (pi -ci)a.D. 1/(i:(Pi -ci)aiD i x) , V i (48)
i
We now solve the posynomial program with equation (46)
replaced by (47). With the resulting value of obtained,
we reset from equation (48) and resolve the program with
this new value of a^. We repeat this procedure until there
is no significant change in the solution.
5. Generalised Geometric Programming
Generalised geometric programming deals with the analy
sis of convex mathematical programs. Commonly these appear
in the form:
Minimize
subject to g^(z) ^ 0 , i e I (50)
z e C (51)
where C is a convex set and g^, i e {0} U I are convex
functions. In the generalised geometric programming
approach, we first separate the arguments of the constraints
and objective functions. We introduce the following
no ta ti on :
z = x° = x*, V i £ I
X = X^ X X* £ X. i£l
X A {x|x° = x * , V i e I}, a subspace of X
C. = C, for i e {0} U I
The program defined by equations (49) , (50) and (51)
may now be rewritten as:
125
Minimize go(*°)
subject to g^Cx1) $ 0 i e I (53)
x 1 e C i i e {0} U I (54)
x e X (55)
This formulation separates the constraints and the o b
jective function excepting for the subspace condition,
equation (55) which ties the problem together. The s ub
space captures the linearity of the problem.
At this stage we require further ideas relating to c o n
vexity . Recall from Section l(iii), the definition of an
epigraph to a pair [g,C] of a convex function g defined on a
convex set C. Each nonvertical hyperplane that supports the
epigraph of a convex function g at a boundary point
(x',g(x')) produces a subgradient of g at x', i.e., a vector
£ e Rn with
g(x') ♦ < £ , x- x’> « g(x) , V x e C (56)
The subgradient set 3g(x') that consists of all such
vectors ^ is generally a closed convex subset of En which
contains only a single vector iff g is differentiable at x'.
In this case the single vector is the usual gradient vector
Vg(x') .
A convex function [g,C] is said to be closed if its
epigraph is a closed set. We shall assume that all functions
are closed.
The conjugate transform [h,D] of an arbitrary convex
function [g, C] is defined by:
h(y) = sup (<y,x> - g (x )) (57)xeC
and D =* (y|sup (<y,x> - g(x)) < +<=°} (58)xeC
We note that h(y) = <y,x'> - g(x') for each y e 8g(x') .
By construction [h,D] is a closed convex function d e
fined on a convex set. A further consequence of the
126
conjugate transform is the conjugate inequality which states
that
g(x) + h(y) > < x , p (59)
for x e C and y e D, with equality iff y e 8g(x) or equival
ently x e 3h(y).
To handle constraints we require the conjugate trans
form of a particular function [0,g(x) < 0, x e C]. This
transform is related to the transform of [g,C] and is called
the positive homogeneous extension of [h,D]. This is
denoted by [h+ ,D+ ] where
[Ah(y/A) , X > 0 h (y,X) = ^ " (60)
[sup <y,x>, X = 0
xeC
and D+ = { (y ,X) |y/XeD ,X>0} U { (y ,0) | sup<y ,x> < <=°} (61)xeC
The conjugate inequality in this case is given by
0 + h + (y,X) $ <x,y> (62)
for (y,X) e D+ and x e C with equality iff y e X 3g(x) or
x e 9h+ (y,X) .
We now apply these ideas of conjugate transform theory
to the program defined by equations (52) to (55) . This
program is the primal mathematical program. For the obje c
tive function, we have from equation (59) that
r 0. V r O', ̂ . 0 0^ ,,,,g0 (x ) + h 0 (y ) <x ,y > (63)
and for each constraint, from equation (62), we have that
0 + h i+ (y1 ,Xi) » <x 1 ,y1 > , i £ I (64)
Adding the inequalities (63) and (64), we have that
g 0 (x°) + h 0 (y°) + Z h i+ (y1 ,^i) » <x,y> (65)i£l
127
where x = x° X x 1 and y = y° X y 1 * By restricting x e X
and y e X , e d i t i o n (65) becomii1
g0 (x°) ♦ h 0 (y°) + Z h. + (y1 , X .) j» 0 (66)u ' u * ie I 1
with equality when
y° e 3g0 (x°) or x° e 8h0 (y°) (67)
and y 1 e A - ^ g ^ x 1) or x 1 e a h / (y1 ,Aj) , i e I (68)
Hence the geometric programming dual to the primal
program (equations (52) to (55)) is given by:
Minimize hg(y°) + E h . + (y*,A.) (69)ie I
subject to y° e D g , (y1 , ^ ) e D i+ » * e 1 (^0) x
y e x
The dual program can have certain advantages over the
original primal program:
(i) The nonlinear constraints are incorporated into the
objective function.
(ii) If X is of dimension n and X is of dimension m, then i
X , the orthogonal complement of X, is of dimension
n-m. With the possibility of n-m being much smaller
than m and the elimination of explicit non-linear
constraints, the dual may turn out to be a much
simpler problem to solve.
(iii) The fact that the primal and dual objectives sum to, ( k . , ( k
zero at optimality, i.e., gQ (x ) + hp(y )
+ (y1 » = 0, provides a good stopping
criterion for an algorithm. The conditions for o p
timality are
128
giUi) * 0, x1 e Ci e D i+ * i e I
X E X y e X
or y e 8g0 (x )
Or y 1 e X i3gi (x1) , i e I
x° £ 3h0 (y°)
x 1 e 9hi (y1 ,Xi)
From these conditions one may calculate the optimal
solution for one problem from the optimal solution of
the other.
It is important to note the roles played by the four
concepts of linearity, separability, convexity and duality
in generalised geometric programming theory. Linearity may
occur naturally in the problem, e.g., as linear constraints,
or it may be induced by the need to separate the variables
as in the program at the beginning of this section. Any
linearity is usually conveniently captured in the subspace
condition, equation (55). Separability is necessary to fac
ilitate a simple computation of conjugate transforms. C on
vexity (and closure) guarantees that there is no duality gap,
i.e., the primal and dual objectives sum to zero at optimal
ity. Duality is the goal of generalised geometric program
ming theory.
A well known problem in inventory control is to select
a set {xt ; t=l,...,T} of production levels to minimize, over
a planning horizon of length T, the sum of production and
holding costs, whilst meeting demand. Formally the problem
may be posed a s :
Example
TMinimize Z c(x ) + h y
t = l 1(72)
subject to the inventory balance dynamics
129
y1 - *1 - dj
y t ■ y t .1 = x t - t = 2 , . . . ,T-1 (73)
-yT . j = x-p - d̂ ,
and the non-negativity constraints
y » 0, t * l ,... ,T(74)
x t 5 0, t * l , . . . ,T
Here yt denotes the inventory level in period t, d^ is
the demand in period t, c(xt) is the production cost (assum
ed convex and strictly monotonically increasing) and h^ is
the holding cost per unit in period t.
To invoke the theory of generalised geometric pr ogram
ming, we need to put the constraint equations (73) into a
subspace. Hence we introduce a new variable a^, t=l,...,T,
and restrict it to a one point domain {dt >, t=l,...,T. This
variable is then associated with an additive component of
the objective function which is identically zero. Hence we
obtain a subspace condition
y l ' X 1 + al “ 0
y t ' y t-l " x t + a t “ °» t = 2 »---»T -1 (7S)
" yT-l " XT + aT " ®
It is convenient to treat the non-negativity constraints
x t £ 0, y > 0, t = l .....T, in an implicit way, i.e.,
C q = (xt |xt » 0, y t > 0, d t = a t , t=l,...,T}. Our problem
is now in a form which is directly suitable for application
of the theory. We note that in this problem there are no
explicit constraints as in equation (53) . The dual ob ject
ive is given by
130
sup E (xtu t + y tv t + a t et - c(xt) - h ty t)x t ,yt »at t
x t > 0
y t > o
= sup E (x u - c(x )) + sup E a 6 + sup E(y v -h y ) x t ^ 0 t tt: 1 a t t y t^0 t 1 1 1 r
■ I c«(ut) * d t8t
where c*(u..) = sup x u - c(xt ) x t * 0 z r r
u t e 3c(xt)
and v t h
Usually c(xt) is a quadratic function and c*(ut) is
readily calculated. Further we require the orthogonal c o m
plement to the subspace defined by equations (75), i.e.
{(ut ,vt ,Bt)|E(utx t + v ty t + 0 ^ ) = 0, V x t ,yt ,at
satisfying equations (75)}. A straightforward calculation
shows that
V t = Pt ‘ p t + l’
u t - ~ P ■£ > t -1, ... ,T
Bt = P t , t = l , . . . ,T
Hence the dual problem is
TMinimize E c*(-pt) + d p
t-1
subject to p^ - P t+-̂ - h t ^ 0, t = l,...,T-l
For constant h^, we have a minimization over a monoton-
ically increasing set of decision variables.
131
6. The Analysis of Posynomial Programming by Generalised
Geometric Programming
Recall that in the preceding section we were able to
use the convexity of the functions of a convex mathematical
program by applying separability and then using linearity
and duality to derive a potentially simpler problem. To
bring this structure out in posynomial programming, we c o n
sider the following convex function:
E 6. log 6./c- (76)i e [k ] 1 1
defined on 5- >, 0, i e [k] and E &■ = 1 (77)1 ie [k]
Here the parameters c^ > 0, i e [k]. The conjugate
transform of the function defined by (76) taken over the set
defined by equations (77) may be shown to be
log E c. ex p (:■) (78)ie [k]
Hence the conjugate inequality from equation (59) is:
£ <5. log 6. /c • ♦ log( E c. exp (z.)) >, E 5.z. ie[k] 1 1 1 ie[k] 1 1 ie[k] 1 1
(79)
with equality when
6i ■ c- e x p (zi)/( E c i exp(zi)) (80)ie [k]
In order to handle constraints of the form
log E c- exp(z-) ^ 0 (81)ie[k] 1
we require the positive homogeneous extension (see equation
(60)) of function (76) defined over the set (77). This is
given by
E 6i log (5i/(Xkc i)) (82)ietk]
132
defined for A^ = £ 6^, 6^ £ 0, i e [k] (83)ietk] x
To relate functions of the form (78) to posynomial g e o
metric programs discussed in Section 3 we set
in equations (14) and (15) . Equation (85) is the subspace
condition in the generalised theory (see equation (55)).
Making the above substitution and taking logs, our posynomial
program is as follows:
In the above form, the theory of generalised geometric
programming may be invoked since we have already calculated
the relevant conjugate transforms (see equations (76), (77),
(82) and (83)). Hence, using equations (69), (70) and (71),
the dual to a posynomial program is given by
The above objective function is equivalent to maximizing
Xj = log t . , V j (84)
mand (85)
Minimize log Z c^ exp(z^)ie [0]
subject to log Z c- exp(z.) 0, i = l,...,p ietk] 1
m
n pMinimize Z 6- log 6 - / c - - Z A, log A,
• 1 1 1 1 1 1 K Ki=l k=l
subject to Z 6- = l is [0]
6. » 0, V i
n
i = l
133
i = l
n
which is the original form of the dual objective function
for posynomial programming (see equations (25) to (28)).
Further it is straightforward to obtain the optimality c o n
ditions between the primal and dual variables, equation (29)
from the general optimality conditions in Section 5.
7. Applications
In the previous sections we have shown the development
of the theory of geometric programming. Included in the
references are various books and papers describing applic
ations of geometric programming. In the sequel to this
paper (appearing in the next issue of NZOR) we present in
some detail the major areas of application of geometric
programming.
[1] Abrams, R.A. and M.L. Bunting, "Reduced reversed p o s y nomial programs", SIAM J. Appl. Math. Vol. 27, 1974, pp.629-640.
[2] Avriel, M . , R. Dembo and U. Passy, "Solution of generalised geometric programs", Int. J. of Numerical Methods in Eng. Vol. 9, 1975, p p. 149-169.
[3] Avriel, M . , M.S. Rijckaert, and D.J. Wilde, e d s . , Optimization and Design, Prentice-Hall, Englewood Cliffs,N.J. , 1973.
[4] Avriel, M. and A.C. Williams, "Complementary Geometric Programming", SIAM J. of Appl. Math. Vol. 19, 1970, pp. 125-141.
[5] Beightler, C.S. and D.T. Phillips, Applied Geometric Programming} John Wiley and Sons, Inc., New York, 1976.
[6] Ben-Tal, A. and A. Ben-Israel, "Primal geometric p r o gramming treated by linear programming", Center for cybernetic studies report, University of Texas, 1974.
[7] Charnes, A., K.E. Haynes, F. Phillips, and G.M. White, "Some new equivalences and dualities in the solution of the gravity model of spatial interaction using the u n constrained dual of an extended geometric programming
REFERENCES
134
[8] Dembo, R.S., "The current state-of-the-art of algorithms and computer software for geometric programming", School of Organization and Management, Yale University,1976 .
[9] Dinkel, J.J. and G.A. Kochenberger, "On sensitivity analysis in geometric programming", O p n s . Res. Vol. 25, 1977, p p.155-162.
[10] Dinkel, J.J., G.A. Kochenberger and S.N. Wong,"Entropy maximization and geometric programming", Environment and Planning A, Vol. 9, 1977, pp.419-427.
[11] Duffin, R.J. and E.L. Peterson, "Reversed geometric programs treated by harmonic means", Indiana Univ.Math. J., 1972 , p p. 531-550.
[12] Duffin, R.J. and E.L. Peterson, "Geometric programming with signomials", J. of Op. Thy. and A p p . Vol. 11,1973, p p. 3-35.
[13] Duffin, R.J., E.L. Peterson and C. Zener, Geometric Programming, John Wiley and Sons, Inc., 1967.
[14] Ecker, J.G., "A geometric programming model for optimal allocation of stream dissolved oxygen", Management Science, Vol. 21 , 1975, pp.658-668.
[15] Ecker, J.G. and W. Gochet, "A reduced gradient method for quadratic programs with quadratic constraints and Jl -constrained £ -approximation problems", Katholieke Universiteit Leuven, 1977.
[16] Ecker, J., W. Gochet and Y. Smeers, "Reversed geometric programming: a branch and bound method involving linear subproblems". Katholieke Universiteit Leuven,1977 .
[17] Ecker, J.G. and R.D. Neimi, "A dual method for q ua d ratic programs with quadratic constraints", SIAM J.A p p l . Math. Vol. 28, 1975, p p .568-576.
[18] Ecker, J.G., and R.D. Wiebking, "Optimal selection of steam turbine exhaust annulus and condenser sizes by geometric programming", Engineering Optimization, Vol.2, 1976, p p . 173-181.
[19] Ecker, J.G. and R.D. Wiebking, "Optimal design of a dry type natural-draft cooling tower by geometric programming", Core Discussion Paper 7610, 1976.
[20] Ecker, J.G. and M.J. Zoracki, "An easy primal method for geometric programming", Management Science, Vol.23 , 1976 , p p . 71-77.
[21] Ekeland, I., and R. Teman, Convex analysis and variational problems, North Holland, 1976.
[22] Falk, J., "Global solutions of signomial programs", George Washington University, 1973.
135
[23] Gochet, W. and Y. Smeers, "A branch and bound method for reversed geometric programming", Katholieke Univer- siteit Leuven, 1975.
[24] Hall, M.A., "Hydraulic network analysis using (generalised) geometric programming", Networks, Vol. 6, 1976, pp.105-130.
[25] Hall, M.A. , and E.L. Peterson, "Traffic equilibria analysed via geometric programming”, Proa. Inter. Sym. on Traffic Equilibrium Methods, M. Florian, ed., Springer-Verlag, 1975.
[26] Jefferson, T.R., "Manual for the geometric programming computer code", School of Mechanical and Industrial Engineering Report, University of New South Wales,1974.
[27] Jefferson, T.R., and C.H. Scott, "Entropy models of spatial interaction and geometric programming", Proceedings of 3rd National Conference of the Auet.Soc. for O.R., R.G. Mills, ed., 1977 , p p . 175-184.
128] Jefferson, T.R., and C.H. Scott, "Generalised geometric programming applied to problems of optimal control I. Theory", to appear J. Optim. Th. and App.
[29] Jefferson, T.R., C.H. Scott and J. Cozzolino, "The analysis of risky portfolios by geometric programming", School of Mechanical and Industrial Engineering, University of New South Wales, 1977.
[30] Nijkamp, G.L., and J.H.P. Paelinck, "A dual interpretation and generalisation of entropy maximizing models in the regional sciences", Papers of the Regional Science Association, Vol. 33, 1974, p p .13-31.
[311 Passy, U. , "Nonlinear assignment problems treated by geometric programming", Op. Res., Vol. 19, 1971, pp. 1675-1690.
[321 Passy, U. and D.J. Wilde, "Generalized polynomial optimization", SIAM Appl. Math. Vol. 15, 1967, pp.1344 ,1356 .
[33] Peterson, E.L., "Geometric programming", SIAM Review, Vol. 18, 1976, p p. 1-52.
[343 Peterson, E.L. and J.G. Ecker, "Geometric programming: duality in quadratic programming and I -approximation I", Proc. Int. Symposium on Math. Prog., H.W. Kuhn and A.W. Tucker, e d s . , Princeton University Press, Princeton, 1970, pp.445-480.
[35] Peterson, E.L. and J.G. Ecker, "Geometric programming: duality in quadratic programming and 2. -approximation II: canonical programs”, SIAM J. Appl.^Math. Vol. 17, 1969, p p. 317-340.
136
[36] Peterson, E.L. and J.G. Ecker, "Geometric programming: duality in quadratic programming and £ -approximation III: degenerate programs", J. Math. A n & l . Appl. Vol.29 , 1970 , p p . 365-383.
[37] Peterson, E.L., and R.E. Wendell, "Optimal location by geometric programming", 1975.
[381 Rijckaert, M.J. and X.M. Martens, "A comparison of generalised geometric programming algorithms", Katholieke Universiteit Leuven, 1975.
[39] Rijckaert, M.J. and X.M. Martens, "A condensationmethod for generalised geometric programming”, Math. Prog. Vol. 11 , 1976 , pp.89-93.
[401 Rockafellar, R.T., "Conjugate duality and optimization", SIAM Regional Conference Series in Applied Mathematics, No. 16, 1974.
[41] Scott, C.H. and T.R. Jefferson, "Trace optimization problems and generalised geometric programming", J. Math. Anal, and A p p ., Vol. 58 , 1977 , pp. 373- 377 .
[421 Scott, C.H. and T.R. Jefferson, "Entropy maximisingmodels of residential location via geometric programming", Geographical Analysis, April, 1977 , p p.182- 187 .
[43] Scott, C.H. and T.R. Jefferson, "Generalised geometric programming applied to various problems of operations research", Proceedings of 3rd National Conference of the Aust. Soc. for O.R., R.G. Mills, ed., 1977, pp. 185-198 .
[441 Scott, C.H. and T.R. Jefferson, "A generalisation ofgeometric programming with an application to information theory", Information Sciences, Vol. 12, 1977, pp. 263-269.
[45] Scott, C.H. and T.R. Jefferson, "Duality for convex functions: specific cases", Int. J. Systems Sc., Vol.8, 1977.
[46] Scott, C.H. and T.R. Jefferson, "Duality in infinite dimensional mathematical programming - convex integral functions", to appear J . Math. Anal, and App.
[47] Zener, C., "A mathematical aid in optimizing engineering designs”, Proc. Nat. Acad. Sci., Vol. 47, 1961, p. 537.
[481 Zener, C., "A further mathematical aid in optimizing engineering designs", Proc. Nat. Acad. Sci., Vol. 48, 1962 , p . 518.
[49] Zener, C., Engineering Design by Geometric Programming, John Wiley and Sons, Inc., 1971.