• Keine Ergebnisse gefunden

Mathematical Methods for the Analysis of Hierarchical Systems. 1. Problem Formulation, and Stochastic Algorithms for Solving Minimax and Multiobjective Problems

N/A
N/A
Protected

Academic year: 2022

Aktie "Mathematical Methods for the Analysis of Hierarchical Systems. 1. Problem Formulation, and Stochastic Algorithms for Solving Minimax and Multiobjective Problems"

Copied!
22
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

NOT FOR QUOTATION WITHOUT PERMISSION OF THE AUTHOR

IhWHEMATlCAL MFTHODS FOR THE ANALYSIS O F x i l E R A R C H I C A L ~ S

I. PROBLEY P O ~ T I O N . AND SrOCHASI'IC AU;OFtmlMs FOR SILYING MINIMAX

AND

YULTlOl3JECPZYE PROBLE3dS

F.I. Ereshko V.V. Fedorov

S.K.

Zavriev

May 1984 CP-84-19

Collaborative Rzpers report work which has not been performed solely a t the International Institute for Applied Systems Analysis and which has received only limited review. Views or opinions expressed herein do not necessarily represent those of the Institute, its National Member Organizations, or other organizations supporting the work

INTERNATIONAL INSX'ITUTE FOR APPLIED SYSTEMS ANALYSIS 2361 Laxenburg, Austria

(2)
(3)

PREFACE

This is t h e first of two papers dealing with mathematical methods t h a t can be used to analyze hierarchical systems.

In this paper, t h e authors look a t t h e situation t h a t arises when cer- tain decision-making powers are delegated t o various elements within a hierarchical structure. I t is found t h a t these elements inevitably begin to operate in accordance with their own interests, which are not neces- sarily those of t h e system as a whole. Thus we have the problem of how t o distribute t h e decision-making functions between t h e central body and t h e other parts of t h e system in s u c h a way that t h e efficiency of t h e control system is maximized with respect t o t h e global criterion.

The authors take a game-theoretical approach to this problem. look- ing first a t two-level hierarchical systems and using Germeyer's games as a model. They derive a number of methods for solving t h e problem t h u s formulated. and give some numerical results obtained using two of the resulting algorithms.

ANDRZEJ

WIERZBICKI Chairman

System and Decision Sciences

(4)
(5)

MATHEMATICAL METHODS FOR THE ANALYSIS OF HIEXARCHICAL SYSI'EMS I. Problem Formulation, and

Stochastic Algorithms for Solving Minimax and Multiobjective Problems

F.I. Ereshko*.

K

V. Fedorov** a n d SIC Zavriev**

*Computing Center of t h e USSR Academy of Sciences, Moscow, USSR

**Moscow University, Moscow, USSR

1. INTRODUCTZON

Hierarchical control systems form one of the most interesting classes of large systems with r e g a r d t o theoretical a n d practical applications. Hierarchi- cal control problems were first formulated in connection with t h e need t o dis- t r i b u t e t h e right t o process information and t h e responsibility lor making deci- sions among t h e various elements of t h e control system. Problems arise due t o t h e fact t h a t , when different elements of the system have these rights a n d responsibilities and c a n exercise t h e m independently, these elements inevit- ably begin t o operate according t o their own interests. which generally differ from t h e global objectives of t h e system. Thus i t is necessary to distribute t h e decision-making Functions between t h e central body and the separate p a r t s of t h e system in such a way t h a t the efficiency of t h e control system is maximized with respect t o the global criterion (we shall assume t h a t this criterion coin- cides with t h a t of t h e central body). This problem m a y be divided into two p a r t s [I-31: t h e problem of analysis, i.e.. t h e choice of a reasonable control for a Axed hierarchical system, a n d t h e problem of synthesis, i.e., the choice of t h e best s t r u c t u r e for t h e control system.

Game theory s e e m s to provide t h e best approach t o such problems. How- ever, traditional game theory does not consider a number of questions which arise in this particular case, e.g., how t o deal with problems caused by t h e s h a r - ing of information between different elements in the hierarchy, priorities in decision making, and lack of knowledge of the objective.function by s o m e ele- ments. We shall t h e r e f o r e begin by introducing a class of games in which moves a r e taken in a Axed order and t h e process of information t r a n s f e r is quite similar t o t h a t found in some hierarchical systems.

(6)

Hierarchical two-person games describe the simplest two-level hierarchical system. This is t h e most thoroughly investigated hierarchical s t r u c t u r e , a n d is of considerable importance. Let t h e objective of player 1 (representing t h e upper level of the hierarchy) be t o increase the value of the criterion ~ ( 2 , ~ ) using decision variable z E X , a n d the objective of player 2 be t o increase t h e value of t h e criterion G ( z , y ) using decision variable y E Y. The principle behind the second player's move is to maximize his gain, given t h a t t h e out- come depends on his action only.

I t is assumed t h a t player 1 h a s t h e first move and knows t h e principle on which the second player will act, a s well as being acquainted with F, G.

X,

Y.

There a r e various formulations of the games now known as Germeyer's games [I] which depend on t h e information available to player 1 about the deci- sion of player 2.

Game

GI.

Player 1 will not have a n y information on the choice made by player 2: his strategy is t o choose a certain z1 E X and report i t t o player 2.

Then t h e best g u a r a n t e e d r e s u l t of player

1

is v

=

sup inf F ( z l , y l ) ,

x i @ y i e B 1 ( z 1 )

where

B'(z')

=

ly' E

Y'(

~ ( z l , y ' )

=

rnax ~ ( z ' , z ) j ,

X ' =

X , Y'

=

Y z c r

b e

G2.

Player 1 will know t h e choice y 2 E Y made by player 2: his strategy is t o choose t h e mapping X2 N

= Is2:

Y -, Xj.

The best guaranteed r e s u l t of player 1 is

inf ~ ( z " ~ , ~ ~ ) V z

=

$Ul& u+p(i.)

b e Gg. Player 2 formulates his action as a function y ( z ) , i.e.. h e chooses a mapping

g3

E

= tq3:

X' -, Yj. Player 1 has the first move and since h e will know

g3

h e reports t o player 2 t h e mapping

z3

which is an e l e m e n t of t h e s e t

529 = 153:

P 3 -,

x'j.

The best guaranteed r e s u l t of player 1 in such games is -3 -3

inf F ( z ,y )

=

3gs:!; f 3 @ 3 ( 5 3 )

(7)

In games G2 and G3 the sets of rnultivalued mappings ~ ~ (and 83(g3) 2 ~ ) are defined (like g l ( z l ) ) as the sets of possible answers of the second player, given t h a t t h e strategy of the first player is fixed.

Increasing t h e number of iterations we can formulate games GZn, GZn+l, n 2 2 .

The s e t s of players' strategies in game G2n a r e

and the best guaranteed result of player 1 is

-2n -2n v2n

= 2z:$,

J " ~ 8 2 " ( 2 ~ ) inf ~ ( z ,y ) . In game G2n+1 we have

-*+I. y & + i

p + l

= 12

- . P - l j , p C 1 = j Y -2n+l.

.

p n - 1

,

j%?n-l]

-

inf -2n+l - h + l )

F'(. 9

Y

"&+l

-

-a:?&+i p a + ~ E ~ ~ + ~ ( ; a + i ) 2

where

& ( g k )

= igk

E

? I

~ ( ? & , y " ~ )

=

max G ( Z ~ , Z ) ]

.

ZEP

The following relationships hold for n 1 2 [4]:

Thus from the point of view of player 1 there is no point in having a stra- tegy more complicated t h a n in games G I , G2, Gg. In other words, the &st t h r e e games can be regarded as basic and we shall confine ourselves to a considera- tion of these games only.

Games G I , G2, G3 have a natural economic interpretation in t h e framework of the "Center-Producer" system [5].

1. The setting of prices z1 for the output y of t h e producer. The natural approach here is game GI, as in this case prices a r e chosen without any information about y

.

(8)

2. Decisions on fixed payments z 2 (subsidies, premiums, assignments and so on). As accounts with the producer a r e settled on receiving the final pro- duct, h e may be informed beforehand of the chosen system of fixed pay- ment (i.e., how the amount paid depends on the results of his work). Here we have game G2 on t h e s e t of strategies

p,

3. Allocation of resources z3 (raw material, equipment, labor a n d so on). It is obvious t h a t resources m u s t be allocated before the production process begins, and formally the producer has the right to dictate his terms:

g3 =

y(z3). However, since t h e center has the first move he may report his strategy as the mapping z 3 : + X. This is a typical G3 formulation, although game G1 is also possible here. The guaranteed result of player 1 in games GI, G2. Gg satisfies t h e relationship v

<

v g C v 2 , and thus the allocation of resources to t h e producer in a game G3 formulation is more profitable to the center than in GI.

3. ANALYSIS OF TWO-LEYEL HIERARCHICAL S R X l W S

Since Germeyer's games may be taken as models of two-level hierarchical systems, the analysis is reduced to t h e question of finding the solutions of the games formulated in Section 2.

G a n e GI. The problem of solving game G1 is reduced to that of solving a maxi- min problem with linked variables (see (2.1)).

Assume t h a t the criteria F and G a r e continuous on compact sets X,Y.

Then the inner infimum in (2.1) can be replaced by a minimum. However, in the general case the function

is discontinuous. Consider the simple example F

=

y

-

z2, G

=

ZIJ , X

=

Y

= [-I.

11. Here f (z) has a discontinuity a t point z

=

0 and the first player has no optimal strategy. This means that &-optimal strategies z, should be found which satisfy the inequality f (z,) 2 v l

-

e for given E

>

0. With these assumptions f (z) i s lower semicontinuous; in general i t is multiex- tremal.

In theory the problem ,may be solved using the penalty function method, which reduces it to an unconstrained optimization problem [1,6,7]. Consider

(9)

the penalty function

where d

>

0. The reduction of problem (2.1) to a maximin problem with separ- able variables is based on the following theorem:

Theorem 1 [6.7]. /f z, yields a solution of max m i n ( F

+

cJ) at fized c , then f o r any sequence ck -,

-

the points zCk form an &-optimal sequence of strategies f o r the first player.

A number of methods c a n be used to solve problems of the form rnax m i n { F ( ~ , ~ )

+

cJ]. including stochastic programming methods [7.8] and non-smooth optimization methods [9-111. In addition t o t h e nondifferentiability of the objective function there may be some difficulties connected with t h e multiple extrema of the problem, which make it necessary to develop appropriate optimization algorithms [12-141.

The use of numerical methods to search for v l and the E-optimal strategy of the first player is complicated by the fact that problem (2.1) is not neces- sarily stated correctly with respect t o the functional, in t h a t any small varia- tions in t h e second player's strategy G(z,y) (due to errors in computations, for example) can cause variations in the first player's guaranteed result.

In t h e same way, for F

=

y , G

=

f ( 2 ) , X

=

Y

=

[0.1] t h e optimal result of the first player in game

G1

is zero. If the second player's criterion is G c = G

+

~ ( y - 1 ) . where e may take any small positive value, then the guaranteed result will be equal to 1, since

for any

z

E

X

To obtain a numerically stable procedure for computing t h e best guaranteed results, it is necessary to regularize problem (2.1) using the method described in [15].

Came

G2.

We shall make use of t h e following values, sets and functions:

L2

=

max G(zP(y),y)

=

m a x min G(z , y )

Y E Y y e Y 2 E X

(10)

Here zP(y) is a penalizing strategy and z a ( y ) is the absolutely optimal strategy of the first player.

Theorem 2 [I]. Let v 2

=

max(K2, Mz). m e n the strategy if y

=

y,. K 2 > M2 i f y E E ~ , Kz<Mz z P ( y ) otherwise

isthe &-optimal strategy of the first player in game G2.

The case K2

>

M, is particularly interesting: it corresponds to the situa- tion in which the objectives of both (the levels of the hierarchical sys- tem) are in some sense similar.

The theorem formulated above shows t h a t the problem of constructing the optimal strategy in game G2 is reduced to that of solving a nonlinear program- ming problem and a maximin problem with separable variables.

h e

Gg.

Let us define

D3

=

{ ( z , y )

I

G ( z , y )

>

L3

=

min max G(z,y)l t c X ~ E Y

.

K3

=

sup ~ ( z , y ) S ~ ( z , , y , )

+

E . B

=

{z E

XI

rnax G ( z , y )

=

L2j

( . , Y ) E D s Y E Y

B ( z )

=

{y E

YI

G ( z , y )

=

rnax G ( z , z ) j t € Y

M3

=

su min ~ ( z , y ) S min ~ ( z f , y )

+

E

~ E B Y E B ( ~ ) V E B ( Z J

Theorem3 [1,4]. Let v 3

=

max(K3,M3). m e n the strategy

is the &-optimal strategy of the first player in game

G3.

(11)

Here y", is t h e strategy of t h e second player, which consists in choosing point y , E Y, and z : E B plays the role of a penalizing strategy. Thus t h e prob- lem of finding t h e optimal strategy in game G3 is reduced t o t h a t of solving a mathematical programming problem and a maximin problem with linked vari- ables (value

M g

and s t r a t e g y z: E 8).

4.

A

COhtBINF,D PENALTY

AND

Sl'OCHASLIC GRADIENT

METHOD

(CPSGM)

In the previous sect-ion we showed t h a t a necessary s t e p in t h e analysis of games ' 1 , 2 , 3 is t h e solution of t h e following minimax problem: Find

z

E XO a n d uo, where

xo = l z

E A

I

min F ( z , y )

=

u0j

Y E Y

uo

=

max m i n ~ ( z , y )

rEA Y E Y

Let us consider certain stochastic algorithms for solving problem ( 4 . 1 ) . We may assume without loss of generality t h a t

a n d also t h a t functions F ( z , y ) , q i ( z ) . i

=

1 ,

...,

m . a r e continuous together with t h e i r derivatives with respect t o z on s e t X ' x

Y

, X '

=

O J X ) . In addition, we assume t h a t

Y

is a compact s e t from E l , A # $, E~

>

0.

I t is clear t h a t

where j ~ , + ~ ( z )

=

R

-

11211. Now introduce where

(12)

Here M represents the mathematical expectation, i is a random number whose values are taken from set 11

....,

m j with probabilities p l , ...,p, ; y is a ran- dom number distributed on Y according to measure p in such a way t h a t any non-empty intersection of y with any open s e t has positive measure.

I t is shown in [I] t h a t problem (4.1) can be reduced to a sequence of prob- lems in which it is required to maximize function (4.2) with c n

=

(cy , c z ) T m

(this is the penalty function method).

The stochastic gradient method

[a]

can be used to search for the max- imum of function

Lq

a t fixed c . If t h e algorithm allows for penalty parameters c , c 2 to increase. then we obtain the following iterative procedure:

where

vector z0 E X and value E . 0

<

E

<

E ~ , a r e both chosen arbitrarily; ( y n . i n ) a r e the values of t h e random numbers ( y , i ) during t h e n - t h independent test;

r1

=

( z l , u l ) is the initial approximation; and

t % j ,

tb,j, { c n j are control sequences.

Theorem4 [16]. Let functions pi(z) , i

=

1,

...,

m+1 satisfy the condition

where

~ +( z ) l

=

ti

I

p i ( z ) ( 01

f o ~ a n y point z E X, and the control sequences satisfy the following conditions

(13)

T h e n f o r any initid a p p r o z i m a t i n n (T l,z l ) , s e q u e n c e s jrn j , lzn

1

o f s o l u t i o n s o f a l g o r i t h m (4.3) e x k t s u c h that, with p r o b a b i l i t y o n e :

( 1) A s u b s e q u e n c e o f t h e n a t u r a l s e r i e s of n u m b e r s

{? 1

e x i s t s s u c h t h a t

( 2 ) A f o l l o w s f r o m lim ,ch

=

0 that t h e l i m i t p o i n t s o f s e q u e n c e Irns]

s +-

b e l o n g t o t h e s e t o f s t a t i n n a r y p o i n t s [ l o ] o f p r o b l e m (4.1).

Remarks

1. Condition (4.4) is satisfied if y ( z )

.

i

=

1,

...

m , are concave and Slater's con- dition is satisfied.

2 . The following a r e examples of sequences which satisfy conditions (4.5):

3. The parameter ,cn is introduced into (4.3) t o follow t h e value of

ah/

ar and t o provide a m e a n s of finding t h e elements of the sequence trn'] which converges to t h e s e t of stationary points. (If

F ,

pi a r e concave with r e s p e c t t o z. t h e n sequence trnj will converge to t h e set of solutions of problem (4.1) a n d t h e r e is no n e e d to follow parameter ,cn.)

4. Theorems s i m i l a r . t o Theorem 4 but with different restrictions on sequences (4.5) and r a t h e r m o r e rigorous restrictions on functions F , pi have been proved in [?,I?, 181.

5.

A

STOcHASl3C "ERRORS' METHOD FOR FINDING

A

MAXIMIN

Let us consider problem ( 4 . 1 , assuming t h a t functions

F .

p i ( z ) , i

=

1,

...,

rn, a r e concave with respect t o z on convex compact set X

c Ek

for any y E Y (where Y E El is a compact s e t ) and t h a t both functions F, p i ( z ) and their partial derivatives with respect t o z a r e continuous on X x Y , A

$ 4 .

This problem can be reduced t o t h e following mathematical programming problem [I]: Find T

=

( z , u ) which solves

(14)

max

U

2 ,u

subject to

m

@ , ( r ) = - j l

m i n ( 0 . ~ ( 2 . y ) - u ) l q p ( d y )

- C 1

min ( 0 . p i ( z ) ) q > o ,

Y i = l

where q r 1 ,

U

is a line segment which includes

[ rnin F ( z , y ) ; max F ( z , y ) ] ,

( Z ,Y ) a x Y ( z , u ) ~ X x Y

and measure p satisfies the conditions given on p. 8 in Section 4.

Problem (5.1) is equivalent to the following problem: From the points r

=

(2.u) for which

max tPq(r)

=

0 ,

2 EX

find the point with the largest value of u .

Function Gq(r) can be treated as an "error" which characterizes t h e dis- tance of the point r from t h e feasible set of problem (5.1). This approach to solving problem ( 4 . 1 ) was suggested for the first time in [ 1 9 ] .

Note that, as in ( 4 . 2 ) , we have Gq(r)

=

M p q ( r , y . i ) , where

( p p ( r . y , i )

= -1

min (0, F ( z . y ) - u ) l q

-

( l / p i ) I min (0. p i ( z ) ) l q , and random numbers y, i a r e as defined in Section 4 .

We can now formulate the following iterative algorithm:

T n + l

=

+

% t n )

on'' =

ln [p"

+

bn ( p q ( r n , y n , i n )

-

p n ) ]

where n~ is the projection operator on R

=

X x V and vector

tn

is deflned by the formula

(15)

Here

I:(rn , y n , i n ) is a conditional &,-subgradient of function (p, (-. y n , i n ) a t point r n from set R , 8, > 0 ; y n , in a r e the values of random numbers y and i dur- ing the n - t h independent test; a n d ( T ' , ~ ' ) is the initial approximation.

Parameter (pn in algorithm (5.3) follows the value of the error

@ , ( T ~ ) , lim ( (pn

-

p q ( r n )

I =

0 P-a.s. At the n - t h step, if the value of the error n +-

is n e a r zero ((pn 2 +) the value of u increases in accordance with (5.3), but if

pn

<

-d, then the value of u changes in accordance with the stochastic quasi-

gradient of t h e error Function.

Theorem 5 [20]. Assume t h a t a c o n s t a n t k

>

0 e z i s t s s u c h t h a t

11 (11 <

k , n

=

1,

...

for a n y y ~ ! ? = f ( ~ l , i ' , ..., y n , i " . . . ) j , that s e q u e n c e s j g , ] , l b n j a n d t 4 , j , t L n j , ren{ e h t m c h t h a t

and t h a t one of the f o l l o w i n g c o n d i t i o n s iss a t i s f i e d :

%n f o r a n y initid a p p r o z i m a t i o n ( r l , pl), the s e q u e n c e rn defined b y a l g o r i t h m (5.3) c o n v e r g e s t o t h e s e t of s o l u t i o n s of p ~ o b l e m (4.1) w i t h probabil- ity o n e .

Remark The following a r e examples of sequences which satisfy the conditions of the theorem:

(16)

6. E-SUBGRADIENT DESCENT ALGOlUTHM MIR APPROXIMATION

OF THE

PARETO SET

Consider the following parametric programming problem: Find

z

( a ) E xO(a), where

Xo(a)

=

tz E XI

I

( z , a ) = max ( z ' , a ) j

Z'EX (6.1)

for all a E A .

Function ~ ( z , a ) is assumed to be continuous on convex compact s e t X E

El,

for any a E A , where A E

E'

is a bounded set. We say t h a t point z* is a solution of problem (6.1) a t a

=

a + with accuracy (&A) if p2(z*. ~ ~ ( a * ) )

<

A, where p is a metric and

Assume that values do , A. , a.

>

0 are given. Let us construct a n algorithm for finding (do. Ao). the approximate solutions of problem (6.1) at all d-nets A d

=

f d l,....dNj on A such t h a t d S d o and

Here IIrzlI

=

m v ( 4 )

.

a E ES.

t

We shall assume t h a t I ( z , a ) is concave with respect to z on X for any a E A , d i a m X s D a n d

where L

=

const

>

0. Let the solution of problem (6.1) be known with accuracy ( b o , A,,) a t values of parameter a

=

a l from d-net A d .

We shall determine t h e solution of problem (6.1) a t the nodes of net Ad using the formula

where

nx

is the projection operator on X;

(F

is t h e conditional e-subgradient of concave Function f ( - , a n ) a t point z n on set X , E

>

0; and a is a step-size mul- tiplier.

(17)

Theorem 6 [20].

Lf

p a r a m e t e r s a . d , E o f a l g o r i i h m (6.2)-(6.3) s a t i s f y t h e f o l - l o w i n g c o n d i t i o n s :

w h e r e

K

is a c o n s t a n t , s o that

11 (211 <

K , n

=

1

... N

in (6.3), and

2 1

p

(I

, Xs, (a1))

<-

Ao. t h e n all s u b s e q u e n t in

.

n

=

2.

... .

w i l l s a t i s f y c o n d i t i o n p2fzn 9 X6.(an ))

< &.

Thus. using algorithm (6.2)-(6.3) we can obtain the solution of problem (8.1) with precision ( d o ,

4)

on d - n e t Ad, 0

<

d 4 d o . For any fixed 6 0 , A. i t is always possible to find values of a , a and E which a r e sufficiently small t h a t ine- qualities (6.4) are satisfied.

We shall now show how algorithm (6.2)-(6.3) may be applied t o vector optimization problems.

Let vector criterion

be defined and positive on X c

Ek,

and

Let

n ( w )

be t h e s e t of efficient vectors from W (Pareto-optimal vectors), where

We shall use the following notation:

where

(18)

m

P ( X )

=

min \ w i ( z )

+

7

C

w i ( z )

I s i s m i =l

It is shown in [21] t h a t f o r V e

>

0 37,d0, t h e s e t

where w(X) is an a r b i t r a r y point from w7(h) and

krn

is an arbitrary d - n e t on Am

,

0

<

d 5 do, is an &-net on II(w). Thus t o And an &-net on II(w) it is sumcient t o solve t h e following parametric programming problem: Find

for a l l X ~ q . 0

<

d 5 do.

If functions w i ( z ) , i

=

1,

...,

m a r e concave and continuous on convex com- pact s e t X, then an approximate solution of problem (6.5) can be found.using algorithm (6.3).

7. NUMERICAL, RESULTS

The proposed algorithms were implemented a n d then t e s t e d on some sim- ple problems in order t o investigate their practical efficiency.

7.1. CPSGM

Algorithm (4.3) (with certain modifications) has been used t o solve (4.1) with functions

F l ( z , y )

=

cos (0.25(z1

+

z 2

+

z3)

+

y l

-

0.5)

+

+

cos (0.25(z1

+

2 z 2

+

z3)

+

y 2

-

0.5)

+

cos (0.5(z1

+

z 2 )

+

Y J

-

0.5)

deflned on the product of u n i t cubes. The following control sequences were used: [ ~ , , ] = T L - ~ ' ~ , f a n c n j = n - 1 1 / 2 0 , q = I , E = O for F1, and q = 2 ,

-

71+'85 , b,

=

-0.72

'=n

-

, Cn

- -

,0.2 for F2.

(19)

-

15-

The results are presented in Table 1.

Table 1. The results obtained with the

CPSGM

algorithm.

It can be observed t h a t a good approximation to solution z and the first approximation to

u

a r e obtained reasonably quickly. However, further refinement of t h e solution takes place very slowly.

When t h e gradient of the efficiency function is computed using t h e difference scheme, the rate of convergence of the algorithm is the same as when the precise gradient is used.

7.2.

Errors

Method

The e r r o r s method (with parameters

en =

0,

a, =

n4.', b,,

=

n-0.7g

4 =

0.01n4.') was used t o find t h e maximin of functions Func-

tion

F,

F2

i +I

~ ( z , Y )

= C

cos (z,

+

yi

-

0.5) ; i

=

1.2 ) = I

Initial approximation

zO=(O.OOO, O . O O O , ~ . O O )

uO=

1 .OOO

z0=(0.6,0.6, 0.6.0.6) u 0 = 2

Number of iterations

400 800 2400 3200 500 1500 9000 17000

defined on unit cubes.

The results of t h e computations are presented in Table 2.

Precise solution

z *=(0.000, O.000,O.OOO) ue=2.634

z *=(0.5000, 0.5000, 0.5000, O.$OOO) u =1

Approximate solution With precise

gradient z=(0.009,0.017, 0 . 0 0 9 ) ; ~ =2.634 z=(0.012,0.018, 0.0 12);u =2.626 z=(0.000,0.000,

0 . 0 0 0 ) ; ~ =2.6 19 z=(0.000,0.000, 0 . 0 0 0 ) ; ~ =2:634

With approx.

gradient z=(0.000,0.000 0 . 0 1 8 ) ; ~ =2.606 z=(0.000.0.000, 0 . 0 9 7 ) ; ~ = 1.549 z=(0.000,0.000, 0 . 0 1 6 ) ; ~ =2.6 13 z=(O.000.0.000, 0.000);~ =2.624 z =(0.4704.0.4668,

u = 1.069

z=(0.5118,0.4974,0.5014.0.5003);

~ = 1 . 0 5 9

z =(0.5005,0.4988,0.5006,0.4975);

u=1.041

z =(0.5002,0.50 12,0.4989,0.5030);

u

=

1.036

(20)

-

16-

Table 2. The results obtained using the errors method.

Func- tion

F

1

F2

No. of itera-

ti ons 200 600 5400 10600 200 600 42000 84000 Initial

approx- imation

zO=(O.OOO.

0.000);

uO=O.OOO

zO=(O.OOO, 0.000,1.000);

u 0

=

0.000

Precise solution

z *=(o.ooo.

0.000);

u *= 1.756

z *=(o.ooo.

0.000,0.000);

u *=2.634

Approximate solution z =(0.0277,0.0289);

u = 1.809

z =(0.000,0.0028);

u

=

1.900

z =(0.0133,0.0089);

u = 1.860

z =(0.0037,0.0066);

u=1.815

z =(0.0458,0.0000,0.0133);

u =2.721

~=(0.0196,0.0066.0.0172);

u =2.831

z =(0.0046.0.0076,0.0119);

u =2.791

(21)

Y.B. Germeyer. Ghmes with Non-Antagonistic Enterests. Nauka, Moscow, 1976.

Y.B. Germeyer a n d N.N. Moiseev. "On some problems in t h e theory of hierarchical control systems", in Problems in Applied Mathematics and Mechanics. Nauka. Moscow. 1971.

V.N. Burkov. Models a n d methods of functioning of hierarchical systems.

Automatika i Telemechanica. 11, pp. 106-131, 1977.

N.N. Kukushkin. The role of m u t u a l information in two-person g a m e s with non-antagonistic interests. B u r n a l Vychislitelnoj Matematiki i Matema- ticheskoj fiiki. 12(4), pp. 1029- 1034, 1972.

1 . k Vortel a n d F.I. Ereshko. Mathematics of Conflict a n d Cooperation, Znanie. Moscow. 1973.

V . A Gorelik Approximate determination of maximins with constraints linking t h e variables. Bumal Vychislitelnoj Matematiki i Matematicheskoj f i i k i . 12(2), pp. 510-517, 1972.

V.V. Fedorov. Computational M a z i m i n Methods. Nauka, Moscow, 1979.

Y.M. Ermoliev. Methods of S o c h a s t i c Programming. Nauka, Moscow, 1976.

N.Z. Shor. New directions in t h e development of methods for nonsmooth optimization. Kibemtetika, 6, pp. 87-91, 1977.

V.F. Demyanov a n d V.H. Malozemov. Entroduction to M n i m a z . Nauka, Mos- cow, 1972.

E . A Nurminski. IWmerical Methods for S l u i n g Deterministic a n d S o c h a s - tic Minimrrz Problems. Naukova Dumka. Kiev. 1979.

R.G.

Strongin. Numerical Methods in ~ t i e z t r e m a l Problems. Nauka, Mos- cow. 1978.

F.I. Ereshko a n d AS. Zlobin. An algorithm €or centralized allocation of resources among active subsystems. Economics and Mathematical Methods, 4. pp. 703-713. 1977.

kG. Sukharev. @tima1 S a r c h for @tima Moscow University Press, 1975.

D . A Molodtzov a n d V.V. Fedorov. Approximation of two-person g a m e s with information transfer. Bumal Vychislitelnoj B a t e m a t i k i i Alatematicheskoj fiiki, 13(8), pp. 1469-1464, 1973.

S.K. zldvriev. On searching for stationary points in a maximin problem with constraints. Vestnik

MGU,

S r i e s Computational Mathematics a n d Q b e r - n e t i c s , 2, pp. 48-57, 1980.

S.K Zavriev. A combined penalty and stochastic gradient m e t h o d for t h e determination of maximin. Bumal &chislitelnoj Matematiki i Matema- ticheskoj fiiki, 19(2). pp. 329-342, 1979.

N.M. Novikova. A stochastic quasigradient method for t h e determination of maximin. Buntal Vychislitelnoj Matematiki i Matematicheskoj Fizzki.

17(1), pp. 91-99, 1977.

Y.B. Germeyer a n d I.A. Krylov. Determination of maximin by a "discrepan- cie s" method. USSR Computational Mathematics and Mathematical Physics, 12(4), pp. 871-881, 1972.

(22)

20.

S.K.

Zavriev. S t o c h a s t i c G r a d i e n t A l g o r i t h m s f o r ,%Solving Minimux Prob- l e m s . Ph.D. Thesis, Moscow University, 1981.

21. P.S. Krasnoschekov e t al. Decomposition i n problems of t e c h n i c a l design.

I z u e s t i a of the

USSR

A c a d e m y of a i e n c e s , S r i e s T e c h n i c a l C y b e r n e t i c s . 2, pp. 7-17, 1979.

Referenzen

ÄHNLICHE DOKUMENTE

If the current tower base may only be placed partly on the real container floor and if there exist feasible placements of tower bases not yet used, the alternative with the largest

The parallel searches are carried out by differently configured instances of a tabu search algorithm, which cooperate by the exchange of (best) solutions at the end of defined

In particular, we prove existence and uniqueness of global regular solutions in the two-dimensional case for initial data of arbitrary “size”, and for “small” data in

(See [Grinbe15, proof of Theorem 2.74] for the details of this derivation.) It is clear how to perform induction using Theorem 2.2.3: It differs from standard induction only in that

The most widely proposed MP techniques are the hlSD formulation (Freed and Glover 198lb), which minimizes the sum of absolute exterior deviations from the classification

AB far as the parameter ko limits the value of objective function, we will name this method the Method of Constraints. The solution procedure component is a

We shall consider a hierarchical model of a production planning problem.. An algorithm for solving this game is given below.. Let us now illustrate the problem by

(Improved version of the Representation theorem for semilattices) Every semilattice (S, ∨) satisfying the ascending chain condition can be embedded into (2 M , ∪) where M denotes