• Keine Ergebnisse gefunden

Global convergence of multidirectional algorithms for unconstrained optimization in normed spaces

N/A
N/A
Protected

Academic year: 2022

Aktie "Global convergence of multidirectional algorithms for unconstrained optimization in normed spaces"

Copied!
32
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Global convergence of multidirectional algorithms for unconstrained optimization in

normed spaces

J. A. Gomez and M. Romero

Center of Mathematics and Theoretical Physics.

ICIMAF, Havana, Cuba.

Abstract

Global convergence theorems for a class of descent methods for unconstrained optimization problems in normed spaces, using multi- directional search, are proved. Exact and inexact search are considered and the results allow to dene a globally convergent algorithm for an unconstrained optimal control problem which operates, at each step, on discrete approximations of the original continuous problem.

AMS Subject Classication: 49M10, 49M37, 65K10.

1 INTRODUCTION

Global convergence theorems of multidirectional search algorithms for nite dimensional optimization problems were derived in 11]. The general global convergence theorem of Zangwill (see 13]) was systematically used and ap- plications to numerical algorithms for unconstrained nite dimensional prob- lems were given. In nite dimensional optimization problems in Hilbert space were also considered in 11], but the heavy assumption about the closedness of the point-to-set map, which is essential in the Zangwill theorem, limited

1

(2)

the applications, in optimal control problems, only to algorithms de ned by point-to-set maps with range in a nite dimensional subspace.

In this paper we consider general in nite dimensional optimization prob- lem in a real normed space. Global convergence theorems are proved for a class of descent multidirectional algorithms, using the global convergence theorem of Polak (see 17]) without any nite dimensional assumption. Exact and inexact search are considered and in this last case we use the well known Wolfe conditions for global convergence, which is a common instrument in almost all the Quasi-Newton methods for unconstrained optimization (see for example 5] or 10]).

The results are used to design an algorithm for unconstrained optimal control problems which is globally convergent to local minima of the contin- uous problem but which operates, at each iteration, with nite dimensional approximation problems which are discretizations of the continuous one. In some sense the convergence result of this algorithm justi es the usual pro- cedure for solving unconstrained optimal control problems, which takes the optimal solution of a convenient nite dimensional discrete problem as a good approximation of the original one.

At each iteration, the algorithm nd a new point (i.e. a control func- tion) which satis es Wolfe's conditions for the continuous problem at the current (function) point, but this is performed nding a new point (i.e. a control sequence) which satis es Wolfe's conditions for the discrete prob- lem at the current (sequence) point. This point of view allows to use, in the implementation, all the common Quasi-Newton optimization methods and subroutines for the discrete problems. In addition, it also diers from the usual convergence results in common publications, in which it is always proved the convergence of theoptimal solution of the discrete problem to the optimal solution of the continuous problem (see for example 6]). Finally, the convergence results of the algorithm, given in the paper, are related with recent developments which study the inuence of the discretization step size in the algorithmic convergence (see 2]), since each iteration can be seen as an attempt to select not only a direction, but also a convenient step size to improve the current solution.

2

(3)

We consider the following problem:

min f(x)

x2X (1)

whereX is a real normed space, and we will study the global convergence of multidirectional descent algorithms, with exact and inexact search. We sup- pose f 2 C1 a continuously dierentiable function, and denote by rf(x)2 X the Frechet derivative of f at x, where X is the topological dual space of X. In all the paper, the symbol hy xi denotes the evaluation of the functionaly 2X at point x:

De nition 1

An algorithmic map in X is a point-to-set map A dened in X with values in the class P(X) of all subsets of X

.

A function c : X !<

is a stopping rule for the set ; X respect to the map A in X if for all x =2; we have:

c(x0)< c(x) 8x0 2A(x):

De nition 2

An algorithm A on X for nding points of a set ; X (shortly an algorithm for ;) is given by a sequence of pairs(Ai ci)i2@ where Ai are algorithmic maps satisfying:

Ai(z)6= 8z =2; 8i2@

and ci(:) are stopping rule functions for the set ;, and the following steps:

A1) Choose x0 2X and set i = 0 A2) Compute Ai(xi)

A3) Choose yi 2Ai(xi)

A4) If ci(yi)ci(xi) stop and take xi as the last point of the sequence, If ci(yi)< ci(xi) go to step A5),

A5) Take xi+1 =yi set i + 1 !i and go to step A2).

When both sequences are constant Ai = A ci = c 8i 2 @ we say that we have a uniform algorithm A = (A c):

De nition 3

A (nite or innite) sequence generated by an algorithm A= (An cn)n2@, from the starting pointx2X is any sequencefxn n = 1 2 :::g

3

(4)

X which is obtained following the steps A1) - A5) in the last denition, with x0 = x:

An algorithm A = (An cn)n2@ for ; X is called globally convergent to

; if for all x02X and any sequence fxng generated by A from the starting point x0 we have that the last element offxngbelongs to ;, if fxng is nite, or that every accumulation point of fxng belongs to ; if fxng is innite.

Theorem 1

(Polak): Let X be a metric space and A = (A c) a uniform algorithm in X for ; X and suppose the maps A(:) and c(:) satisfy the following conditions:

i) c(:) is continuous or bounded below in X ii) 8z =2; 9" = "(z) > 0 9 = (z) < 0 :

c(z00);c(z0)< 8z002A(z0) 8z02B (z ") (2) Then the algorithm A is globally convergent to ;:

For a proof see (17]).

De nition 4

Let be 2 (0 1): We say the direction d 2 X satises Wolfe's conditions respect to f 2 C1 at x 2 X if the following inequalities hold:

f(x + d)f (x) + h5f (x) di (3)

h5f(x + d) di h 5f (x) di:

We call ;condition (respectively ;condition) the inequality corresponding to (respectively to ) in (3). We will also say that the point y = x + d satises the Wolfe conditions.

Lemma 1

Let X be a normed space and f a Frechet dierentiable function.

Let's consider 2 (0 1), > . Let be x 2X and d a descent direction in x, i.e.h 5f (x) di < 0 . Suppose the set ff (x + d) 0g is bounded below.

Then there existsy2X , with the formy = x+ d which strictly satises the Wolfe conditions:

f(y) < f (x) + h5f (x) di (4)

h5f(y) di> h 5f (x) di

and therefore, there exists an entire interval (;" + ") where the Wolfe conditions hold.

4

(5)

Proof.

We have:

f (x + dl) = f(x) +h 5f (x) dli+o(jlj)< f(x) + h 5f (x) dli (5) for small l > 0.

Since ff (x + dl) l2<g is bounded below and l ! f (x + dl) is contin- uous, there exists the least l0 > 0 such that:

f (x + dl0) =f(x) + h5f (x) dl0i:

In fact, for l near 0 we have (5) and for l!+1, the functionf (x + dl) is bounded and f(x) + h5f (x) dli!;1.

On the other hand, we have:

f (x + dl0);f(x) =D5f(x + del) dl0E for some el2(0 l0) therefore:

D

5f(x + del) dl0E=h5f(x) dl0i> h 5f (x) dl0i since > and h5f (x) di< 0, then:

D

5f(x + del) dE> h5f (x) di:

Taking y = x + d~l and recalling that el < l0 and the de nition ofl0, we have:

f(x + del) = f(y) < f (x) + elh 5f (x) di

D

rf(x + del) dE = h5f(y) di> h 5f (x) di:

The existence of an interval is a consequence of the continuity of the functions l!f (x + dl) and l!5f(x + dl): 2

Lemma 2

Let X be a normed space, f a continuously Frechet dierentiable function and 2(0 1). Let be fxngn2@X with xnn!+1! x and 5f (x)6= 0andfyngn2@X withyn=xn+ndn, n 0 kdnk= 1, such that, for all n 2@ the inequalities:

h 5f (xn) dni;k5f (xn)k

h5f(yn) dnih 5f (xn) dni

hold, and suppose ynn!+1! y: Then x6=y. 5

(6)

Proof.

We have:

h5f(xn);5f(yn)] dnih5f (xn) dni;h5f (xn) dni=

= (1;)h5f (xn) dni;(1;)k5f (xn)k< 0 (6) and

jh5f(xn);5f(yn)] dni jk5f(xn);5f(yn)kkdnk=k5f(xn);5f(yn)k: If we suppose x = y, then there exists the limit of the sequence:

jh5f(xn);5f(yn)] dnij

which, by continuity of rf satis es:

n!+1lim jh5f(xn);5f(yn)] dni j= 0 (7) and therefore, the sequence of real numbers h5f(xn);5f(yn)] dni con- verges to 0 but on the other hand:

n!+1lim h5f(xn);5f(yn)] dni

nlim

!+1

;(1;)k5f(xn)k=;(1;)k5f (x)k< 0:

We have a contradiction.2

Usually, a descent uniform algorithm A for the problem (1) is de ned through an algorithmic point-to-set mapA which is the composition of two maps A = SG, representing a "selector of directions" map and a "selector of new points" map respectively. We recall that the composition map of the point-to-set maps S and G is de ned by:

(SG)(x) =

y2G(x)S(y):

The stopping rule function c(:) of a descent uniform algorithm A for the problem (1), is almost always chosen as the objective function f(:): In de ning our rst algorithm, we will use this common point of view.

De nition 5

Let be p 2 @ and 2 (0 1): The point-to-set map of -non orthogonal descent directions Gp(z) : X !X P(Xp) is dened by:

Gp(z) = f(z D)2fzgXp j9 2<p :kD k6= 0

hrf (z) D i;krf (z)kkD k g (8) 6

(7)

where D is a p;vector D = (D1 ::: Dp), with Di 2 X i = 1 ::: p and the productD is dened by the linear combination :D =Ppi=1 iDi:The norm in Xp is the usual product norm:

kDk1= maxi

=1:::pkDik:

De nition 6

Let be p2@. The point-to-set map of exact search Sp(z D) : X Xp !X is dened by:

Sp(z D) =y 2X j y = z + D f(y) = min

2<pf(z + D ): (9)

De nition 7

The descent uniform algorithm Ap = (Ap c) for the prob- lem (1), with multidirectional and exact search, is dened by the algorithmic point-to-set map Ap =SpGp and the stopping rule function c(x) = f(x):

In words, the algorithmApselects, at stepk, a p;vector of directionsD = (d1 ::: dp) belonging toXp in such a way that the linear variety generated by those vectors contains a descent direction of the objective function f at the current pointxk: In the next step, a new point xk+1is chosen as the minimum of the objective function f(z) in the linear variety xk +Sd1 ::: dp]:

Theorem 2

Let X be a normed space, f a continuously Frechet dieren- tiable and bounded below function, and 2(0 1). For anyp2@ the descent uniform algorithm with multidirectional and exact search, Ap = (Ap f) is globally convergent to the set:

; =fx2X j5f (x) = 0g:

Proof.

If x =2; ,rf(x)6= 0, there exists a "0 > 0 such that:

rf(x0)6= 0 8x0 2B(x ") 8"2(0 "0)

then a ;descent direction always exist and therefore, Lemma (1) ensures that:

Ap(x0)6= 8x0 2B(x ") 8"2(0 "0):

: We will verify the conditions (2) of Polak's global convergence theorem 1 for the set ;.

7

(8)

i) c(:) = f(:) is continuous.

ii) By contradiction, suppose it is not true. Then9x2X with5f (x)6= 0 such that8"x> 0 and 8x< 0, 9x0 2B(x "x) and9x00 2Ap(x0) such that x f(x00);f(x0) < 0. Taking (x)n = ;n1 and ("x)n = n1 8n 2 @, we have 9nx0n

o

n2@ with x0n n!+1! x and 9nx00n

o

n2@ with x00n 2 Ap(x0n) such that f(x00n);f(x0n)n!+1! 0. But, by de nition x00n=x0n+Dn0 0n , and denoting 0n the vector of <p which satis es the ;condition in (8), we will have:

f(x00n);f(x0n) =f(x0n+D0n 0n);f(x0n)f(x0n+D0n(0nl));f(x0n) 8l2<

by optimality of x00n 2 Sp(x0n D0n). But for all n 2 @ D0n 0n

6= 0 and the vectors d0n= kDD0n0n0n0nk , d0n= 1 satisfy the inequalities:

D

5f(x0n) d0n

E

; 5f(x0n):

By Lema 1, for any 2 (0 1), > and for all n 2 @ there exists zn0 2X , zn0 =x0n+d0nln0 such that:

f(z0n) f(x0n) +D5f(x0n) d0nln0E

D

5f(zn0) d0n

E

D5f(x0n) d0n

E:

Hence, there is not a subsequence ln0k converging to 0 since we would have: x0nk k !

!+1

x 5f (x)6= 0 z0nk 2X zn0k =x0nk +d0nkl0nk

D

5f(x0nk) d0nk

E

;5f(x0nk)

D

5f(z0nk) d0nk

E

D5f(x0nk) d0nk

E

z0nk k!+1! x

which is a contradiction with Lema 2. In addition, by the de nition of Ap

we have the inequalities:

f(x00n);f(x0n)f(zn0);f(x0n)D5f(x0n) d0nl0nE< 0 8n2@ 8

(9)

and then:

f(z0n);f(x0n)n!+1! 0) D5f(x0n) d0nl0n

E

n!+1! 0: (10) On the other hand,

n!+1lim

D

5f(x0n) d0nl0nElim supn

!+1

(;)rf(x0n)ln0 =;krf(x)klim infn!+1ln0 < 0 which is a contradiction with (10). 2

The other descent uniform algorithm requires a non composite de nition:

De nition 8

Let be p2@ and 2(0 1) > : The point-to-set map Ap :X !X given by:

Ap(z) =fy2X j 9 2<p D 2Xp :D 6= 0 y = z + D

D

rf (z) D E;k rf (z)kD f(z + D )D f (z) + D5f (z) D E

5f(z + D ) D ED5f (z) D E o

(11) and the stopping rule objective function c(:) = f(:)dene the descent uniform algorithm Apw with multidirectional inexact search.:

In words,Apw nd, in the subspace generated by the directions belonging to D a linear combination D which satis es the ; ;and ;conditions at the same time. We can consider also an algorithm with variable number of directions, i.e. we use multidirectional search but in subspaces with dierent dimensions at each step:

De nition 9

Let fpng@ be a sequence of natural numbers. For 2 (0 1) > let's consider the sequences of point-to-set maps Apn dened, for p = pn as in (11). The descent algorithm with variable multidirectional inexact search is dened by Afpngw = ( Apn c)n2@ where the stopping rule sequence is constant, c(:) = f(:):

Theorem 3

Let X be a normed space and f a continuously Frechet dif- ferentiable and bounded below function. Let be 2 (0 1), > . For any sequence fpng@ the descent uniform algorithm Afpngw with variable multidirectional and inexact search, is globally convergent to the set:

; =fx2X j5f (x) = 0g: 9

(10)

Proof.

Lemma (1) ensures again that:

8x =2; 9"x > 0 :8n 2@ Apn(x0)6= 8x0 2B(x ") 8"2(0 "x):

We will use similar arguments as in the preceding proof to verify the condi- tions (2) of the Polak's global convergence theorem.

i) c(:) = f(:) is continuous.

ii) By contradiction, suppose it is not true. Then9x2X with5f (x)6= 0 such that 8"x > 0 and 8x < 0, 9x0 2 B(x "x) and 9x00 2 Apn(x0) such that x f(x00);f(x0)< 0. Taking (x)n =;n1 and ("x)n = n1 8n2@, we have9nx0non2@ with x0nn!+1! x and 9nx00non2@ with x00n2 Apn(x0n) such that f(x00n);f(x0n)n!+1! 0.

Butx00n=x0n+l0nd0n, withd0n= kDDnnnnk Dn2Xpn l0n=kDn nk

D

5f(x0n) d0n

E

;5f(x0n),D5f(x00n) d0nE D5f(x0n) d0nE,f(x00n)f(x0n)+D5f(x0n) d0nl0nE,

d0n

= 1, l0n> 0 8n2@ then, there is not a subsequencenl0nk

o

k2@such that ln0k k!+1! 0, since we would havex0nk k!+1! x and x00nk =x0nk+ln0kd0nk k!+1! x and this is a contradiction with Lema 2. Furthermore:

f(x00n);f(x0n)D5f(x0n) d0nln0

E< 0 and if we have f(x00n);f(x0n)n!

!+1

0 we obtain:

D5f(x0n) d0nl0n

E

n!+1! 0:

On the other hand

D5f(x0n) d0nln0E;5f(x0n)ln0 and then:

0 = limn

!+1D5f(x0n) d0nln0

E

limsupn

!+1

;5f(x0n)ln0 =;k5f(x)kliminfn

!+1 ln0 < 0 since there is no subsequence nl0nk

o

k2@ converging to 0. This contradiction proves the theorem.2

.

10

(11)

PROBLEMS

3.1 Discretization of Unconstrained Optimal Control Problems

We consider the following optimal control problem:

minJ(u(:)) = 'x(t1)] (12)

s:t: _x(t) = f(x(t) u(t)) a:a: t2t0:t1] x(t0) = ^x0

u(t) 2 U a:a: t2t0:t1]

whereu(:)2Lm2 =L2(t0:t1] <m) the set of square integrable functions with image in <m andx(:)2C =Ca(t0:t1] <n) the set of absolutely continuous functions with image in <n.

The maps' :<n !<andf :<n<m !<n are supposed continuously dierentiable respect to its arguments, the functionalJ(u) is assumed to be Frechet dierentiable respect to the Lm2 ;norm and the set U <m for our purposes, will be considered the whole space <m.

If we apply Euler's integration scheme we obtain a discrete approxima- tion of the problem (12) by the following nite dimensional optimal control problem:

minJN(uN) = 'xN] (13)

s:t: xi+1 = xi+hf(xi ui) i = 0 1 ::: N;1 x0 = ^x0

uN = fu0 u1 ::: uN;1g

ui 2 U i = 0 1 ::: N;1:

where h = (t1N;t0) is the integration step, which de nes a partition fi i = 0 Ng of t0 t1] inN subintervals:

i =t0+ih i = 0 ::: N

which in turn allows us to de ne, from a given feasible control sequence uN = fu0 u1 ::: uN;1g 2 LmN of the discrete problem (13), a feasible con- trol function uN(:) 2 Lm2 of the continuous problem (12), with the classical

11

(12)

piecewise constant form:

uN(t) = uj fort 2j j+1) j = 0 1 ::: N ;1: (14) We can makethis de nition from any given vector sequence de ned on any partition fig. The piecewise constant function uN(:) de ned in (14), from a vector sequence uN =fu0 u1 ::: uN;1ggiven on a partitionfi i = 0 Ng, will be called "the constant canonical extension" of the sequence to a function in Lm2 :

>From the corresponding discrete trajectory, xN = fx0 x1 ::: xNg 2

LnN+1 we can de ne the continuous function:

xNt] = xj+ (t;j)f(xj uj) t2 j j+1] (15) which is the classical polygonal Euler approximation of the solution xN(:) of the dierential equation system of (12), corresponding to the constant canonical extension uN(:), and satisfying:

x_N(t) = f(xN(t) uN(t)) t2t0:t1] xN(t0) = ^x0:

The continuous function de ned in (15), from a vector sequence xN given in a partition fig will be called "the polygonal canonical extension" of the sequence to a function in C:

We can de ne, reciprocally, from any piecewise continuous function z(t) on t0 t1] with image in <k a vector sequence zN = fz0 z1 ::: zN;1g 2LkN

de ned on a partition fi i = 0 Ng of t0 t1] in the trivial way:

zi =z(i) i = 0 ::: N;1 (16) and we will call (16) "the sequential canonical reduction" ofz(:) to a sequence in LkN:

As nal remarks about notation, we are denoting by LkN the set of nite sequence ofN vectors in<k which is isomorphic to<kNthe vector sequence associated with a partitionfi i = 1 ::: Ng with elementsfzi i = 1 ::: Ng is denoted by zN for the constant canonical extension to Lm2 we add paren- thesis zN(:) and for the polygonal canonical extension to C we add square brackets zN:] nally, for the sequential canonical reduction of z(:) to LkN, we suppress the parenthesis, add an over bar to z and a subindex with the number of vectors de ned in the partition: zN:

12

(13)

We note also that a vector sequence uN 2 <mN given on the partition

fig can be considered de ned in any partition fi0g of t0 t1] which con- tains fig (with N0 intervals, N0 > N) de ning rst the constant canonical extension uN(:) of uN to Lm2 and then taking the sequential canonical reduc- tion uN0 of uN(:) to fi0g. It's easy to see that we also have equality for the constant canonical extension of uN and uN0:

uN(t) = uN0(t) 8t 2t0:t1] 8N0 N

therefore, from now on we will identify uN(:) with uN0(:) for any N0 > N which corresponds to a partition fi0g containingfig:

There are many publications about the convergence of the optimal solu- tion of the discrete problem (13) to an optimal solution of the continuous problem (12) when N ! +1: There are even quantitative results in the speed of convergence of error estimates of optimal trajectories, controls and adjoint variables, and also many generalized results in several directions. As examples, it can be mentioned the works of Alt 1], Daniel 4], Dontchev 6], Evtuschenko 8], Hager 12], Malanowski 14], Mordukhovich 16], Teo 20]

and many others.

In our opinion, all these results are more in the "stability of optimal so- lution" framework than in the "convergence of an algorithm" context. We haven't seen that the concepts of "descent and feasible directions" or "in- exact line search", appearing naturally in the context of nite dimensional optimization algorithm (see for example 5]), have been su ciently exploited in the design of " nite dimensional approximation" algorithms for optimal control problems and in the proof of they global convergence. In this paper we will present an example of this other point of view.

To any control sequence uN = fu0 u1 ::: uN;1g 2 LmN which is feasible to the problem (13), corresponds a piecewise constant function uN(:)2Lm2 through the constant canonical extension, which is feasible to the problem (12) and reciprocally.

>From a current point ukN 2 <mN the k + 1;step in a classical descent algorithm for the discrete optimization problem (13) consists of nding a di- rection of decrease4ukN in ukN and then performing an inexact line search to nd a step length k and a new point ukN+1 = ukN+ k4ukN which satis es the ; and ;global convergence conditions of Wolfe. This k +1;step can be viewed as one step of a single-direction optimization algorithm for the contin- uous problem (12), identifying uN and 4ukN with their canonical extensions

13

(14)

to Lm2 : If any of the Wolfe conditions fails to hold or even when they both hold, we can perform several iteration now in a non classical algorithm for the discrete problem and this can be viewed as one step of a multidirectional optimization algorithm for the continuous problem. Increasing of N implies an increment of the number of variables in the discrete problem and a reduc- tion of the integration step in Euler's formula for the continuous problem.

It also can be considered as the start of searching in a new direction of a multidirectional optimization algorithm for (12).

The number of iteration made by the optimization algorithm in each approximating discrete problem until a satisfactory point was found would be the number of directions that we take at that step of the algorithm for the continuous problem. This number can be the same or can be varied in dierent iteration during computation, but in practice it is always bounded.

Therefore, since we have global convergence theorems to local minima for unconstrained multidirectional descent methods with inexact line search (using Wolfe conditions) and with possible variable number of direction at each iteration, we have the conditions to model an algorithm for the contin- uous problem which is based on iterations in the discrete one. Hence, we will examine the following questions:

a) When ;non orthogonal conditions in the discrete problems implies the same condition in the continuous problem?

b) When the global convergence ; and ;Wolfe conditions in the dis- crete problem implies the same conditions in the continuous problem?

c) Is it possible to design a global convergence algorithm for the contin- uous problem, only ensuring the (possible varying) Wolfe conditions in the discrete problems?

In the next section we will answer these questions positively, and the idea is quite simple:

1) The descent ;condition for the direction 4ukN(:) depends on the gradient of J(:) at the current point ukN(:) :

D

rJ(ukN(:)) 4ukN(:)ELm

2

;rJ(ukN(:))Lm

2

4ukN(:)Lm

2

therefore we should have the discrete gradientrJN(ukN) close, in some sense, to the continuous gradientrJ(ukN(:)) at the current point,

2) The ;condition for the direction4ukN(:) depends on the gradient of 14

(15)

J(:) at the current point ukN(:) and at the new point vkN(:) = ukN(:)+4ukN(:) :

D

rJ(vkN(:)) 4ukN(:)ELm

2

DrJ(ukN(:)) 4ukN(:)ELm

2

therefore we should also have the discrete gradient rJN(vkN) close, in the same sense as before, to the continuous gradientrJ(vkN(:)) at the new point, 3) The ;condition for the direction 4ukN(:) depends on the increment of J(:) at the new point respect to the current one, and on the gradient of J(:) at the current point :

J(vkN(:));J(ukN(:))D5J(ukN(:)) 4ukN(:)ELm

2

therefore we should have also the discrete incrementJN(vkN);JN(ukN) close to the continuous incrementJ(vkN(:));J(ukN(:)):

We can't expect that the direction 4ukN verify both the discrete and the continuous global convergence conditions for the same parameters and . Then, the main di culties are:

- rst, to nd conditions on the parameters and on the closeness of the required quantities in such a way that the corresponding ; ; or ;condition is satis ed at each problem,

- second, to prove that it is possible to choose the parameter values in such a way that they satis es all the conditions together, and

- third, to design a globally convergent algorithm for the continuous prob- lem, using the above results.

3.2 Relations between Wolfe's Conditions

We need rst to point out some relations between the scalar products and the norm of the sequence of controls and their canonical extensions to Lm2 :

In any feasible control sequence uN = fu0 u1 ::: uN;1g 2 LmN of (13), each uj is a vector in<m with euclidean norm:

kujkm =qhuj uji<m =qkujk2m=

v

u

u

t

m

X

i=1u2ji

and hence, we can de ne the norm of the sequence uN as the `2;norm:

kuNk`2 = qhuN uNiLmN =qhuN uNi<mN =

v

u

u

u

t

NX;1 j=0

huj uji<m = 15

(16)

=

v

u

u

u

t

NX;1

j=0 kujk2m =

v

u

u

u

t

NX;1 j=0

m

X

i=1u2ji =kuNkmN

i.e. the euclidean norm of the vector (u01 ::: u0m ::: uN;11 ::: uN;1:m) 2

<mN:

For any uN 2LmN the Lm2 ;norm of the function uN(:) can be calculated:

kuN(:)kLm2 =

s

Z t1

t0 kuN(t)k2mdt =

v

u

u

u

t

NX;1 j=0

Z j+1

j kujk2mdt =phkuNkmN

and for any v(:)2Lm2 we have:

hv(:) uN(:)iLm2 =Ztt1

0

vT(t)uN(t)dt = NX;1

j=0

Z j+1

j vT(t)dt

!

uj:

If, for example, v(:) is the function of Lm2 representing the gradient

rJ(uN(:)) of the objective function of the continuous problem (12) at the point uN(:), and since Lm1 is dense in Lm2 it should have the following well known formula: (see 17]):

rJ(uN(:))(t) = ;NT(t)fu(xN(t) uN(t)) a:a: t2t0:t1] where N(:) is the solution of the adjoint dierential system:

_N(t) = ;fxT(xN(t) uN(t))N(t) a:a: t2t0:t1] N(t1) = ;'Tx xN(t1)]

and xN(:) is the continuous trajectory of the problem (12) corresponding to uN(:) then we have:

hrJ(uN(:))(:) uN(:)iLm2 =;NX;1

j=0

Z j+1

j TN(t)fu(xN(t) uN(t))dt

!

uj: Another important example is when v(:) is the constant canonical ex- tension to Lm2 of the vector sequence rJN(uN), i.e. the vector sequence of the gradient of the discrete problem objective function (13) at the point

16

Referenzen

ÄHNLICHE DOKUMENTE

In Chapter 5, we introduce the Stieltjes-transform of finite measures on ( R , B). Just as the method of moments, the Stieltjes transform method is a very popular and established

Damped Newton's method, global convergence, first-order approximation, line search, path search, nonsmooth equations, normal mappings, variational inequalilies, generalized

To justify the use of probability distributions based on a (necessarily) limited num- ber of samples to calculate estimates for the optimal solutions of

Abstract: Ln optimization in lUn with m nonlinear equality constraints, we study the local convergence of reduced quasi-Newton methods, in which the updated matrix is

In Section 4 we translate the interiority information obtained in Section 3 into a pair of simple conditions on the optimization problem (1.1): these amount to a Slater condition

The method of stochastic quasi-gradients [8,9], a generalization of the method of stochastic approximations, relies on (ii) combined with an appropriate choice of step-size

Dixon's 1972 theorem states that all methods in the Broydon @ - class develop identical iterates when line searches are exact, conflicts in choice of minimum along

This paper describes the role of nondifferentiable optimization from the point of view of systems analysis, briefly describes the state of the art, and gives a