a Mathematical Programming Package for Multicriteria Dynamic Linear Problems
3.1. General remarks
The most popular methods for solving linear programming problems are based on the simplex algorithm. However, a number of other iterative non-simplex approaches have recently been developed 15-71. HYBRID belongs t o this group of non-simplex methods. T h e solution technique is based on the minimization of an augmented Lagran- gian ~ e n a l t y function using a modification of the conjugate gradient method. The Lagrange multipliers are updated using a modified version of the multiplier method [8]
(see Sections 3.2 and 3.4).
This method is useful not only for linear programming problems but also for other purposes, as described in Section 1.2. In addition, the method may be used t o solve prob- lems with non-unique solutions (as a result of regularization - see Section 3.7).
The following notation will be used:
a, denotes the i-th row of matrix A x . denotes the j-th component of vector x
3
llxll denotes the Euclidian norm of vector x
( u ) + denotes the vector composed of the non-negative elements of vector u (where nega- tive elements are replaced by zeros)
A T denotes transposition of matrix A 3.2. The multiplier method
We shall first explain how the multiplier method may be applied directly t o LP problems.
Consider the problem ( P O ) , which is equivalent t o the problem ( P ) :
min
cxwhere d E RP,
B
is a p x n matrix, and m<
p<_
2 ( m + n ) . T o apply the multiplier method t o this problem we proceed a s follows:Select initial multipliers y 0 (e.g., O = 0) and p E
R
,p>
0 . Then for k = 0,1,...,
" + '
ySlfl
wheredetermine successive values of x
,
and
where
L b , s k
)
=+ (
ll(yk+
p ( ~ x - d ) ) + l I ~ - llyk l I 2 ) l ( 2 p ) until a stopping criterion is satisfied.The method has the following basic properties:
M . Makowski, J . Sosnowski - 9 0 - HYBRID 9.01 1. A piecewise quadratic differentiable convex function is minimized a t each iteration.
2 . The algorithm terminates in a finite number of iterations for any positive p.
3. There exists a constant jT such t h a t for any p
>
jT the algorithm terminates in the second iteration.Note t h a t it is assumed above t h a t the function L ( - , ~ ~ ) is minimized exactly and t h a t the value of the penalty parameter p is fixed. Less accurate minimization may be performed provided t h a t certain conditions are fulfilled (see, e.g., [7,8]). For numerical reasons, a non-decreasing sequence of penalty parameters { p k
)
is generally used instead of a fixed p . 3.3. The conjugate gradient method for the minimization of an augmented Lagrangian penalty functionThe augmented Lagrangian function for a given vector of multipliers y will be called the augmented Lagrangian penalty function (221. For minimization of t h a t function the conjugate gradient method has been modified t o t a k e advantage of the formulation of the problem. The method may be understood as an modification of the techniques developed by Polyak [ l o ] , O'Leary [ l l ] and Hestenes [12] for minimization of a quadratic function on an interval using the conjugate gradient method.
The problem ( P ) may be reformulated as follows:
min cx
where z E
R m
are slack variables.Formulation (PS) has a number of advantages over the initial formulation (PO):
1 . The dimension of matrix A in (PS) is usually much smaller than t h a t of matrix
B
in ( P O ) .2 . The problem is one of minimization of a quadratic function in (PS), and of minimi- zation of a piecewise quadratic in ( P O ) .
3. Some computations only have to be performed for subsets of variables. Note that slack variables are introduced only for ease of interpretation and do not have t o be computed.
In (PS) the augmented Lagrangian is defined by
We shall first discuss the problem of minimizing L ( Z , Z , ~ ) for given y,p> 0, subject to lower and upper bounds for z and z . Let us consider the following augmented Lagran- gian penalty function
F ( x , z ) = ( c / p ) z
+ (
I l y l p+
‘42 -b + 112
-IIY/P
112)/2. (3.4)The gradient of F is defined by
M . Makowski, J . Sosnowski H Y B R I D 9.01
where
From the Kuhn-Tucker optimality condition, the following relations hold for the minimum point ( x i , % * ) :
and
For any given point such t h a t 1
5
z<_
u it is possible t o determine slack variables 05
z5
r in such a way t h a t the optimal it^ conditions with respect t o z are obeyed.Variables z are defined by
if
g. I<
- 0 ( a F / a z ,>
0 )if g,
>_
r, ( a F / a z ,<
0 ) (3.5) g, if r,>
g,>
0 ( a F / a z , = 0 ).
We shall use the following notation and definitions. The vector of variables x with indices t h a t belong t o a set J will be denoted by zJ, and analogous notation will be used for variables g. We shall let q denote minus the gradient of the Lagrangian penalty func- tion reduced t o z-space (q = - ( a F / a z ) ) . The following sets of indices are defined for a given point z :
The set of indices I of violated constraints, i.e., I =
{ i :
g,2
r,)U { i :
g,5
0 ).
Tis the complement of I , i.e.,
T =
{1,2,....,
m)\I.
The set of indices I can be also interpreted as a set of active simple constraints for z. The set of indices J of variables t h a t should be equal t o either the upper or the lower bound, l.e.,
J =
{ j :
z, = 1, and q - 3<
- 0 )u { j :
zj = ujand
q,2
0 ).
is the complement of J , i.e.,
. i
= { 1 , 2,...,
n)\ J.
M . Makowski, J . Sosnowski - 9 2
-
HYBRID 9.01 For the sake of illustration t h e matrixA
may be schematically split u p in the fol- lowing three ways (see the Figure below): first according t o active rows, second according t o basic columns a n d third with illustrate the p a r t of the matrixA
for which augmented Lagrangian penalty function is computed. T h e contents of the matrix A.i I (for which the augmented Lagrangian penalty function is computed) changes along with computations.In essence, t h e augmented Lagrangian penalty function is minimized using the conju- gate gradient method with the following modifications:
1 . During the minimization process z a n d z satisfy simple constraints and z enters t h e augmented Lagrangian in the form defined by (3.5).
2. T h e conjugate gradient routine is run until no new constraint becomes active, i.e., neither set
I
nor setJ
increases in size. If this occurs, the computed step length is shortened t o reach the next constraint, t h e corresponding set ( I orJ)
is enlarged and the conjugate gradient routine is re-entered with the direction set equal t o minus t h e gradient.3. Sets
J
a n dI
are defined before entering the procedure discussed in point 2 and may be only enlarged before the minimum is found. When t h e minimum with respect t o t h e variables with indices in setsJ
andI
has been found, setsJ
andI
are redefined.4 . Minimization is performed subject only t o those components of variables z whose indices belong t o set
J ,
i.e., variables t h a t are not currently equal t o a bound value.5. Minimization is performed subject only t o those components of variables z whose indices d o not belong t o s e t I , i.e., slack variables t h a t correspond t o non-active sim- ple constraints for z . Note t h a t , formally, this requires only the use of different for- mulae for z . In actual fact i t is sufficient t o know only the set
I,
which defines the minimized quadratic function.M . Makowski, J . Sosnowski - 93 - HYBRID 9.01