Nonhnear
Analyws,
Theory,
Methods
Pergamon
OPTIMAL
& Applrcatms, Vol. 23, NO. 1, pp. 15-U. ,994 Copyright S 1994 Elsevier Science Ltd Printed in Great Britain. All nghts reserved 0362-546X/94 $7.00+ .OO
CONTROL OF SEMILINEAR STOCHASTIC EVOLUTION EQUATIONS V. S. BORKAR~~ and T. E. GOVINDAN§~(
t Department
of Electrical Engineering and Nonlinear Studies Group, Indian Institute of Science, Bangalore 560012; and 8 University Department of Chemical Technology, University of Bombay, Matunga, Bombay 400019, India (Received 6 February
1993; received for publication
Key words and phrases: Stochastic evolution equation, selection, verification theorem, Nisio semigroup.
optimal
16 June 1993)
control,
Markov
control,
Markov
I. INTRODUCTION
THE AIM of this paper is to initiate a semigroup theory-based approach to characterization of optimal Markov controls for controlled semilinear stochastic evolution equations. (It may be recalled that Markov controls are those that depend only on the current state at each time.) For finite dimensional controlled stochastic differential equations with a nondegenerate diffusion matrix, this task is traditionally achieved through the Hamilton-Jacobi-Bellman equation of dynamic programming associated with the problem and an accompanying “verification theorem”. The latter states that an optimal Markov control can be explicitly obtained by the pointwise minimization of a “Hamiltonian” derivable from the solution of the HJB equation. Moreover, any optimal Markov control is obtainable in this manner. Thus, one simultaneously obtains the existence of, and the necessary and sufficient conditions for, optimal Markov controls [l]. (An alternative approach to necessary conditions proceeds through the stochastic maximum principle, see [2].) In finite dimensional problems with degeneracy or infinite dimensional problems, the HJB equation presents some problems. One, namely possible nonuniqueness of solutions, has been successfully tackled in recent years through the concept of viscosity solutions [3]. A good verification theorem, however, is still missing due to the absence of sufficient regularity of the viscosity solutions (see, however, [4]) and other even more formidable problems, such as the ill-posedness of a general Markov-controlled stochastic differential equation in the absence of nondegeneracy. (The only exception is the case when the dependence of the Markov control on the current state is sufficiently smooth, which is an unreasonable assumption to make.) In the case of the former difficulty, nonsmooth analysis [5] gives some hope, but it still leaves the latter issues wide open. In the case of finite dimensional controlled diffusions with degeneracy, one way these issues were handled, at least insofar as the existence of optimal Markov controls is concerned, was by adopting Krylov’s Markov selection procedure [6], originally developed for obtaining a Markov solution of ill-posed Martingale problems, to select a specific extremal solution measure of the control system which is both optimal and Markov [7, 81. Associated with this Markov process, of course, is its transition semigroup. On the other hand, the control problem leads to the construction of another,
$ Research 11Research
supported supported
by Grant No. 2611192-G from the Department by Chimanlal Choksi Award from the UDCT, 15
of Atomic Energy, Government University of Bombay.
of India.
16
V. S. BORKAR and T. E. GOVINDAN
nonlinear semigroup of operators, called the Nisio semigroup. This semigroup may be viewed as an abstract statement of the dynamic programming principle and is to the HJB equation what the transition semigroup is to the backward Kolmogorov equation [9]. Our central idea is to compare the Nisio semigroup with a modified transition semigroup corresponding to the optimal Markov solution in order to extract necessary and sufficient conditions for optimality of a Markov control. The paper is organized as follows: the next section establishes the existence of a solution to the controlled evolution equation for a prescribed admissible control process and proves the compactness of solution measures over all admissible controls under a fixed initial law, proving thereby the existence of an optimal admissible control. These are proved essentially along standard lines. Section 3 constructs the associated Nisio semigroup. Section 4 uses the Markov selection procedure to extract an optimal Markov solution. Section 5 compares the transition semigroup (to be precise, a modification thereof) of this Markov process with the Nisio semigroup to derive the necessary and sufficient conditions for optimality of a Markov solution. The rest of this section sets up the notation and describes the problem. Let H, H’ be real separable Hilbert spaces and U c H’ the closed unit ball. The control system under consideration will be described by the evolution equation dX(t)
= [AX(t)
+ F(X(t))
+ Bu(t)] dt + dw(t), (1.1)
X(O) = &, for t 2 0. Here: (i) A is the infinitesimal generator of a differentiable semigroup S(t), I > 0, on H (see [lo, p. 731 for definition and properties) with domain a)(A); (ii) F: H --* H, a bounded Lipschitz map with Lipschitz constant K; (iii) B: H’ -+ H, a continuous linear operator; (iv) w( 0) is an H-valued Wiener process with incremental covariance given by a trace class operator Q; (v) X0 an H-valued random variable independent of w( *) and with a prescribed law rrOwhich has bounded moments; and (vi) u( *) is a U-valued control process with measurable paths satisfying the following “nonanticipativity” condition: for t 5 s, w(t) - w(s) is independent of (w(y), u(y),y 5 sJ. The process u(*) as above will be called an admissible control. Call it a Markov control if if it is a timeu( -) = v(X( a)) for a measurable u: H -+ U. Call X( *) a Markov solution homogeneous Markov process. If X( .) is a Markov solution, it follows as in [ll, corollary 1.1.1, p. 131, that u(a) is a Markov control. The converse is not true in general. What is more, for a given measurable u : H -+ U, (1.1) with u( *) set equal to u(X( *)) need not in general have a solution or have a unique one if it does. The cost we seek to minimize over all admissible controls U( .) is m
J(no, 4.)) = E
ewcrrk(X(t), 0
u(t)) dt 1
where cx > 0 and k E C,(H x H’) (the space of bounded continuous maps from H x H’ to R with the topology of uniform convergence on compacts). Let U be endowed with the weak topology of H’ relativized to U. We further assume that k( *, u) is Lipschitz uniformly in u
17
Semilinear stochastic evolution equations
and k(x, a) is convex and, therefore, weakly lower semicontinuous. compact metrizable with the metric d(x,,x,)
= E 2-‘lC~,,e,)
Topologized as above, U is
- (x,,e,)l
n=l
where (e,) is a complete orthonormal basis of H’ and (., * > is the inner particular, U is Polish and k(x, e): U + R lower semicontinuous. It should be added that we consider the weak (in the probabilistic of the above control problem. That is, we seek to minimize the cost (X(a), u( *), w( *)) as above, possibly defined on different probability spaces. for a discussion of these issues. 2. EXISTENCE
OF OPTIMAL
product in H’. In sense) formulation over all processes See [ 11, pp. 17-191
CONTROLS
This section establishes the existence of an optimal admissible control. First, we recall the relevant solution concepts for (1.1). Definition 2.1. An H-valued process X( *) with continuous paths is a strong solution of (1.1) for a prescribed pair of processes (u(v), w(e)) on a probability space (Q, 5, P) with u(e) admissible, if: (i) it is adapted to the filtration (Z,], where 5, = the P-completion of ,? @NY), &% Y 5 4; (ii) X(t) E a)(A) for a.e. t, a.s.; T
(iii) I’0
IIAx(t)II dt < * t
f
(iv)
X(t) = X0
AX(s)ds
+
a.s. for all T 2 0; and
+
F(x(s))ds
+
~~Bu(s)ds
+
i:dw(.s),
t z- 0a.s.
i0
i0
This concept is in general too strong even in the linear case and the equations of delay or hyperbolic type will not usually have strong solutions [12]. This motivates the following definition. Definition (u( -), w( *))
2.2. An H-valued process X( *) is a mild solution of (1 .l) for a prescribed pair as above if (i) above holds, and
(ii)’ t + X(t) is measurable with ji IIx(s)~~~ds < 00 for all t 2 0 a.s.; (iii)’
X(t)
= s(t)&
+
t
t + c qt Jo
t - s)F(X(s)) ds +
S(t - s) dw(s),
Jo
S(t - s)Bu(s) ds
t 1 0; a.s.
0
We shall call (iii)’ the mild version of (1.1). A strong solution is also a mild solution. The conditions for the converse to hold are given by the following lemma which follows as in [13, proposition 2.31.
18
V. S.BORKARandT.E. GOVINDAN
LEMMA 2.1. uEu,t>S; (a) S(t (b) IIAS(t (c) llAS(t Then a mild
Suppose
the following
conditions
hold:
X0 E D(A)
a.s.
and
for each x E H,
s)F(x), S(t - s)Bu E D(A); - s)F(x)II 5 g,(t - s)llxll for some g,: R, -+ R which is locally integrable; - s)BuII I g,(t - s)llxll for some g,: R, 4 R which is locally integrable. solution X(e) of (1.1) is also a strong solution.
The next result proves the existence (u(.), IV(*)) as above.
of a unique
mild solution
to (1.1) for a prescribed
THEOREM 2.1.Given a Wiener process w(a), a random variable X,, and an admissible above on some probability space (Q, 5, P), (1.1) has an a.s. unique mild solution. Proof. Note that IlS(t)II I M for t E [0, T], T L 0, and M < 05 which may depend (see [lo, theorem 3.2.31). Let x(t), t 2 0, i = 1,2 be H-valued processes with continuous on (Q, 5, P) and define ‘r Zi(t)
= S(t)X”
+
pair
u( .) as
on T paths
f [S(t - s)F(yi(s))
S(t - s)dw(s),
+ S(t - s)Bu(s)] ds +
! ,0
i .O
i = 1, 2, t r 0. Fix T > 0 and let II * II denote the supremum norm on C([O, T]; H). Consider the Banach space e of H-valued processes X(t), t E [0, T] on (Q, 5, P) with continuous sample paths under the norm II . II-= [EII. //2]1’2. Then ’ [S(t - s)F(Y,(s)) II.0
- S(t - s)F(y,(s))]
ds
5 TMKIIY,(.)
- yZ (.)11-.
II-
Thus, /lZ,(*) For sufficiently small a unique fixed point repeated on [T, 2T], map and, therefore,
Z2(911-5
TMKIIy,(.)
-
W)11-.
on C and, hence, has T > 0, the map yi( *) + Zi( *) is thus a contraction X( .) which will then be a mild solution of (1.1). The argument may be [2T, 3T] and so on. Conversely, any mild solution is a fixed point of this the unique one.
as in [lo, 141 to cover locally Lipschitz F(.). Remark 2.1. This result may be “localized” See [15] for extensions of the Lipschitz condition. From now on, a solution of (1.1) is always meant in the mild sense unless otherwise indicated. We shall now establish the compactness of the laws of (X(e), u( .)) as u(a) varies over all admissible controls, the law of X0 being kept fixed at no. We shall use Oe(...) to denote “the law of . ..“. Also, for a function f(e)on [0, a), let f([O, T]), T > 0, denote its restriction to [0, T]. We topologize the path space of u(e) as follows: for T > 0, denote by B, the closed unit ball of L,[O, T] with the weak topology of &[O, T] relativized to it. B, is compact metrizable and, hence, Polish. Let B denote the closed unit ball in L,[O, KJ) with the coarsest topology required to render continuous maps B + B, that map f E B to f([O, T]) E B,. Recall the orthonormal basis (ei) of H’. Let ai = (ei, u(e)>, i 2 1. Then ai E B for each i and a(*) = [o(t(*), (Yz(.), ... ] E B”. B, B” are compact Polish spaces as well. The map lo: u E CJ -+ between I/ and u?(U) by virtue of being 1% a 20, (e,, u>, . . .] E [-1, 11” is a homeomorphism
19
Semilinear stochastic evolution equations
a continuous injection with compact domain. We identify u(s), (Y(*) via this homeomorphism and use the notation u(e) to denote either, depending on the context. Similarly, u([O, T]) may be viewed as a B,-valued random variable for T > 0. For any Polish space X, P(X) will denote the Polish space of probability measures on X with the Prohorov topology. Let JGO= ($(X(a), 4.)) I N-1 a d missible, .$(X0) = no] c P(C([O, 00); H) x B”). rTO is tight.
LEMMA 2.2.
Proof. Since B” is compact, it suffices to prove that (&Z(X(*)) /u(e) is admissible, 2(X,) = n,) is tight. Furthermore, it suffices to fix an arbitrary finite T > 0 and show that td:(X(]O, TI)) Id*> a dmissible, d=(X,,) = rrO) is tight in P(C([O, T); H)). Fix t2 > t, in [0, T]. Then
X(t*) - X(1,) = Lw,>& - w,KI1
+
‘fl
[W, - s) - S(f, - ~)lKm))
\ ,O
fl
‘fl
+
.0 I
[S(t, - s) - S(t, - s)]Bu(s) d.s +
’ f2 +
We now obtain II~(~)II, K2 =
-
s)F(X(s))
ds
+ i
i fl U
fl
estimates for each term on the right-hand side. Let K, be a bound on in the proof of lemma 2.1. Then for a suitable constant C > 0,
II1 2n
[II!
[s(t, - s) - S(t, - s)lF(X(s))
ds
0
f1
- t2”-lE 1 <
< -
S(t, - s) dw(s).
s)Bu(s) ds +
-
IIBIIand~4as
IfI
E
[S(t, - s) - S(t, - s)] dw(s) f2
S(t2
t1
moment
i0
n f2 qt,
1
ds
K2”t2”1
ll
II[S(t, - 4 - s(t, -
W'W(s))l12"
1
ds
.0
‘E I
IIS(t,
-
s)
-
S(t,
(by Jensen’s
inequality)
s)112”ds
-
1
I (K,)2”t;“-1E
C2”(t2 - tJ2” ds
(by [lo, theorem
3.2.31)
1 5 (CK,)2”T3n(t2
- tJ,
(2.1) 2n
E
[s(t, - s) - S(t, - s)] dw(s) II 1 I [n(2n - l)]“T”-’
(trace Q)” E
‘I [IS@, - s) - S(t, - s)/[~~ ds 0
1
(by [13, proposition I
[n(2n - l)]“(CT)2”
(trace Q)” (t2 - tl)n,
1.91) (2.2)
V. S. BORKAR and T. E. GOVINDAN
1
II1 2n
fl
E
[S(t, - s) - S(t, - s)]Bu(s) ds 0
dl12n ds
I K;“T2”-1E I (CK2)2”T3n(t2 -
E[llWMo
-
(2.3)
W,)Xo112”l 5 C2"T"E[II~0112"l(t,
td”.
(2.4)
- tl)“,
(2.5)
-
Similarly 2n
t2
E
illi
set, - w(m))
E and E (2.1)-(2.7),
- s)Bu(s) ds
t2
(MKJ2”T”(t2
II1
I (MK2)2”Tn(t2
- tJ,
(2.6)
2n
S(t, - s) dw(s)
I
[n(2n - 1)M2 trace Q]” (t2 - tf.
(2.7)
we obtain E]]tXt,)
for a constant claim.
I
2n
qt, [IIS 11
t1 Combining
II1 II1
ds
- -W,)ii2”l 5 alt, - t,I”
a = a(T, M, K, , K2, C, n). For n = 2, a standard
tightness
criterion
implies
the
LetX”(*), w”(a), u”(e), n = 1,2, . . . . be, respectively, H, Hand U-valued processes on some probability spaces (fi2,, 5,) P,,), n = 1,2, . . ., such that w”(m) is a Wiener process with incremental covariance Q and independent of X”(O), u”(*) is admissible with respect to w”(e), g(X”(0)) = ;rcoand (X”(v), w”(e), u”(e)) satisfy the mild version of (1.1). By the above lemma, d=(X”(*), w”(o), c?(e)), n 2 1 is a tight sequence in P(C([O, 03); H) x C([O, 00); H) x B”), theorem, it where c?(e) = [01;(e), a!;(*), . . . ] with ar = (ei, u”( *)) for i L 1. By Prohorov’s converges along a subsequence. By dropping to this subsequence, we suppose that .JZ(X”( *), w”( *), ofl( *)) + 2(X-(
*), w”( *), a-( *)).
Clearly, .$(X”(O)) = rco and w”(o) is a Wiener process with incremental covariance Q and independent of X”(O). By Skorokhod’s theorem [16, p. 91, we may suppose that space (!A, ‘5, P) and X”(e), W”(‘),CYn(*), n = 1,2, . . . . 00 are defined on a common probability (X”(e), w”(o), (u”(a)) -+ (X”(.), w”(e), cum(*)) in P(C([O, 00); H) x C([O, w); H) x B”) outside a set N E 5 of zero probability. LEMMA
2.3. There exists a U-valued
process
u”(o) such that
ocj”(t) = (ej, u”(t)), Proof. Fix a T > 0 and a sample to satisfy
Ii
point
T (aj@‘(t)
0
jl
1.
o $ N. Let n(1) = 1 and define
- c$“(t))(a,+‘(t)
- a;(t))
dt
(n(k)) inductively
< l/k.
Semilinear stochastic evolution equations
This is possible because $( *) -+ cUj(-) in B. Argue conclude that for each j 2 1,
21
as in the proof
strongly in &[O, T] and, hence, a.e. along a subsequence. extract a subsequence (m(k)) of ]m) such that
of [ll,
By a diagonal
lemma
argument,
II. 1.21 to
we may
a.e. t, j 2 1. Define
u,Jt) E U by Q(t)
p-1
=
(_!_ y
a”“)(t)
m(k) I=1
1
=
y
UNo(t),
t 2
m(k) I=1
>
0.
Then any limit point u(t) of (u,Jt)) in U must satisfy u(t) = olj”(t), j 1 1 for a.e. t. Thus, a”(t) E p(U) for a.e. t where the qualification “a.e.” may be dropped by modifying CY”(0) on a set of zero Lebesgue measure. Set u”(e) = ~-‘(cY~(*)). LEMMA2.4. The process
u”(e) above is an admissible
control
Proof. Fix t 1 s. Then w”(t) - W’(S) is independent n = 1,2,.... Since independence is preserved under convergence n = 03. The claim follows.
with respect
to w”(*).
of (w”([O, s]), c?([O, s])) for in law, this also holds true for
LEMMA2.5. For t 2 0, t
t S(t - s)Bu”(s) ds --f
i
0
S(t - s)Bu”(.s) ds
a.s.
0
Proof. Fix o $ N as above. Recall t -+ u,Jt) defined as in the proof of lemma 2.3. Then uk(t) + u”(t) in U, that is, weakly in H’, for a.e. t. Thus, Buk(t) + Bum(t) weakly in H for a.e. t and in turn, S(t - s)Buf(s) + ,S(t - s)Bu”(s) weakly for a.e. s in [0, t], t L 0. Thus, t
t S(t - s)Bv,(s)
ds -+
weakly.
At the same time, we have x”(t)
tqtI 0
and
S(t - s)Bu”(s) ds s0
i0
s(t)x”(o) s)F(X”(s))
+ X”(t),
t
+ s(t)x”(o), ds -+
i
qt
0
w”(t) -+ w”(t).
- s)F(X”(.s)) ds,
V. S. BORKAR and T. E. GOVINDAN
22
Thus,
equation
(1.1) in the mild form implies
that
‘I
!
S(t - s)Bu”(s) ds + h(t)
,0 for an h E C([O, co); H).
But then
“f
I
S(t - s)Bu”“‘(s)
S(t - s)Bu,(s) ds =
ds + h(t),
,I0 implying -1
h(t)
!
qt
=
-
s)Bu”(s) ds.
.O
This immediately
leads to the following
theorem.
2.2. I_ is compact.
THEOREM
Proof. It suffices to prove that (X”( *), u”(e), w”(a)) satisfies But this is immediate from the proof of the above lemma. Let /3 = inf J(n,, u(a)) where the infimum is over all admissible THEOREM
J(n,,
u”(a))
Proof.
2.3. If (X”(o), u”(e)), n = 1, 2, . . . above = p = lim J(7to, u”(e)). n-m
From
the uniform
Lipschitz
continuity
are
such
the mild
and,
e-“‘[k(X”(t),
that
J(7co, u”(-)) 1 /I,
of k( *, u),
u”(t)) - k(X”(t),
u”(t))] dt + 0
hence, 1, lim E rl
Thus,
for (n(j)),
(m(k)),
u”(t)) dt 1 = fi.
u,(e) as before, &
By convexity
em”’ &Y”(t),
m!‘~[ /
I
1-e -“‘k(X”(t),
u”“‘(t))
e-“‘k(X”(t),
uk(t)) dt
k-m
1
that is, BsE[ proving
dt
.O
of k(x, +), lim sup E
the claim.
<[:e -“‘k@?‘(t),
u”(t)) dt
1
of (1.1).
u(e).
‘Im I,I 0
version
I /3,
a.s.
then
Semilinear stochastic evolution equations COROLLARY 2.1. The set (d=(X( e), u( 0)) E lYr,,1J(n,,
u( *)) = /?I is compact
Proof. That this set is nonempty is already proved above. arguments by taking (X”(.), u”(*)) with J(n,, u”(e)) = p.
nonempty.
Compactness
Theorem 2.3 guarantees the existence of an optimal admissible the existence of an optimal Markov control in Section 4, following semigroup in the next section. 3. THE NISI0
23
follows by the above
control. We improve this to the construction of the Nisio
SEMIGROUP
In this section we shall construct the Nisio semigroup associated with the control system (1.1). This is a semigroup of nonlinear operators which captures the essence of the dynamic programming principle. Let C,(H) denote the Banach space of bounded uniformly continuous functions mapping H + R with the supremum norm, denoted by I(- 11.For f E C,(H), t 2 0, define Qtf(x)
=
u(.) admissible
LEMMA 3.1. Q,f(*)
bf emcus /@X(s), u(s)) ds + e-“‘f(X(t))/X(O)
E
inf
x E H.
=
U “0
E C,(H).
To prove this, we need another technical lemma. Consider prescribed probability space and for x E H, let X(x, t), t z 0, denote the mild solution
u( *), IV(*) on some of (1.1) for X0 = x.
for x,y E H, 3.2. E[llX(x, t) - X(y,s)/12] I c,(llx - y112+ It - .s12+ Jt - sl) t, s E [0, T] for some prescribed T > 0 and some constant C, > 0 which may depend on T.
LEMMA
Proof. Write out explicitly X(x, t) - X(y, s) using the mild versions them and then use the estimates as in lemma 2.2 to deduce E[llX(x,
for some inequality.
constants
0 - X0,
C’,
C”
of (1 .l) satisfied
~)~~21 5 C’(llx - _dz + it - ~1’ + It - ~1) ”T + C" EtllX(x, z) - KY, z)tf21dz /0
depending
on
T. The claim
now
follows
from
the Gronwall
Proofoflemma3.1.Let&>O.Forx,yEH,t>O, IQlf(x)
- Qrf(y)I ‘I I sup E emuS(k(X(x, s), W) UC.)I il
5 tcilx
- ;;
+ wE[lf(X(x,
- k(X(y, 9, 4s))) ds + e-“’ (f(X(x,
t)) - f(X(y,
t)) iIIIIX(x,
t) - XY,
UC.) +
;ly,yElif(X(x,
by
t)) - f(X(y,
t)) 1IiIIX@, 0 - WY,
t)ll<
all
1)) - f(X(y,
011 2 all
0))
V. S. BORKAR and T.E. GOVINDAN
24
for any 6 > 0, C being a positive constant. (The first term bounds the integral involving k( *, 0) by virtue of the uniform Lipschitz continuity of k(. , u) and lemma 3.2.) Pick 6 > 0 above such that If(x) -f(y)\ < E whenever 1(x - ~11 < 6. Then
IQrfC4 - Qrf(y)I 5 tCIlx -YII + E + ~E[IlXx, tc(Jx - yll +
I
where C’ is a bound on llfll and x + Q,f(x) is uniformly continuous. This completes the proof.
& +
$g IIX - Yl12,
C” = C,C’. Since E > 0 was arbitrary, it follows that It is clearly bounded by llfll + tC2 for any C2 L Ik(*, *)I.
Thus, Qr: C,(H) + C,(H), t 2 0, is a family of nonlinear in particular, that it forms a semigroup.
THEOREM
(i) (ii) (iii) (iv)
0 - NY, WI
operators.
The next result proves,
3.1. Q,, t 2 0, satisfying the following: Q,, = Identity; if f, g E C,(H) satisfy f L g, then Q, f L Q,g, t L 0; forf,g E H, t 2 0, IIQtf - Qrgll 5 ewa’llf - gll; IIQ,f - Q,f II--f 0 as t + s;
(v) QtQ, = Q,Qr = Qr+s; (vi) IIQtf II 5 t(l - e--01’)“,“,p(k(x, 2.4)(/a+ e-*‘ll f II.
Proof.
(i), (ii), (vi) are easy. To prove (iii), note that IQtf(4
- Q,g(x)l 5 T;
iE[ ”
-
[~e-“‘k(X(x,
t))]
r
E
[.i 0 5 eC”’
4, 4s)) do + e-“‘f(X(x,
emasW%G 4, u(s)) ds +
e-“’ g(X,(x, 0)
II
I(f - g(l.
To prove (iv), note that for t > s, “f
IQrf(4
- Q,f(x)l
I [! H 5aif -~1 +b 5 y,y
0
s
-E
,O
emayk(X(x, Y), U(Y)) dy + e-“sf(Xx,
wE[If(X(x, t)) - f(X(x, $)I1 UC.1
1 II t))
e?’ &X(x, Y), U(Y)) dy + e-aff(X(x,
E
9)
25
Semilinear stochastic evolution equations
for suitable constants a, b > 0 independent of x. Argue as in the proof of lemma 3.1 to conclude that for any E > 0 and a suitable 6 > 0, JQrf(x) - Q,f(x)\
I aJt - sJ + &b + ?(\I
- sj2 + It - sl),
where C” is as in lemma 3.1. Since E > 0 was arbitrary, (iv) follows. (v) follows from standard dynamic programming arguments as in [ 11, Sections III. 1, 111.31. Properties (i), (v) identify the semigroup property of lQ,, t r 0). This is the Nisio semigroup associated with system (1 .l). We now characterize its infinitesimal generator. Define the operator G on C,(H) by
for h E C,(H) whenever the limit exists in C,(H) and set a>(G) = (h E C,(H) ) the above limit exists]. We call G the infinitesimal generator of (Qt, t 2 0) and d)(G) its domain. Let 6 = (h E C,(H) 1h is twice continuously Frechet differentiable with bounded support and its gradient h’ and Hessian h” are uniformly continuous and uniformly bounded in norm and satisfy h’(x) E 9(A*) for all x E H, where A* is the adjoint of A]. THEOREM
3.2. 6 C a)(G) and for h E 6, Gh = inf [(Lh)(*, U) - olh(*) + k(*, u)] U
where (Lh)(x, u) = (A*h’(x),x)
+ (h’(x),F(x)
+ Bu) + t trace (Qh”(x)).
Proof. The proof depends on an approximation argument adapted from [13]. For A E p(A), the resolvent set of A, and R(I) = AR@, A), where R(A, A) = (AI - A)-‘, let X,(a) satisfy the evolution equation dXh(t) = [AX,(t)
+ R(A)F(X,(t))
+ R(A)Bu(t)] dt + R(A) dw(t), (3.1)
X,(O) = R(A)X,. By lemma 2.1, this has a unique strong solution. We now prove that E[l/X(t) - X,(t)/12] -+ 0 as A -+ 00, t 1 0. Consider t X(t)
- Xi(l)
= s(t)[x,
- R(A)&]
+
.i0
s(t - wvw)
-
t +
I,
+
ds
t qt
s 0 =
WF(X,(~))l
-
s)[Bu(s)
-
R(l)Bu(s)]
ds
+
S(t .i 0
I,,
-
s)[l
-
R(l)]
dw(.s)
26
V. S. BORKAR and T.E. GOVINDAN
where
,f I
‘f
S(t - s)[l - R(A)]F(X(s)) ds +
+
0
1
S(t - s)[Z - R(A)] dw(s).
.O
x,(ol121 5 ww,21 + ml) with E[Z:] and for a suitable E[Z;]
I
constant
M2E[II(Z
c
I 2(MK)‘t
- X,(s)I12] ds
E > 0, - R(A))Xo(12]+ M2t
(
‘E[I(I- R(A))F(X(S))~~~] ds
.i 0 t
+
M=t
E[II(Z - R(A))Bu(s)~~~] ds + M2 trace Q i .O
for M, K as in theorem dominated convergence
2.1, the last term being a consequence theorem,
E[llX(t)
- X,(t)([=] I 2(MK)2r
The claim now follows
from Gronwall’s
‘E[llX(s) i0
inequality.
E]I]Z -
W1121 ds >
of [13, proposition
1.41. By the
- X,(s)ll=] ds + o(h).
Let
I
Q:f(z)
= inf E UC.)
e --OLs k(X,(s),
u(s)) ds + e-“‘f(X,(t))/X,(O)
0
= z
, I
For h E 2, define (L,h)(x,
U) = (A*h’(x),
x> + (h’(x), R@)(F(x)
Applying Ito’s formula of [13, corollary solution of (3.1) for X0 = x, we have
+ Bu)) + * trace(R(A)QR@)h”(x)).
1.21 to h(X,(x,
t)), where h E $j and X,(x,
*) is the
f = infE 4.)
e-as ((k,h)(X,(x, 0
s), u(s)) - oh(X,(x,
s)) + &X,(x,
s), u(s))) ds
. I
Semilinear
Letting A + 00, the foregoing above) to
stochastic
evolution
and our hypothesis
27
equations
on k, h lead (after a small computation
t Q,h(x)
- h(x) = inf E UC.)
emols((Lh)(X(x, [.i
Let X0(x, t), t 2 0, denote
s), 4s))
- cuh(X(x, s)) + k(X(x,
s), u(s))) ds
0
the mild solution
- h(x) 5 inf E aeu
emols((Lh)(X”(x, ij
QtW
- W) 5
for ,any arbitrarily prescribed in x. On the other hand, Q,h(x)
-
t
h(x) z E
s), a) - cuh(X”(x, s)) + k(X’(x,
s), a))) ds
0
Using lemma 3.2 and the uniform 3.1 to conclude
1 .
to (1.1) when X0 = x and u( *) = a E U. Then
‘t Qth(x)
as
continuity
of k, h, h’, h”, argue as in the proof
1 .
of theorem
inf [(Lh)(x, a) - &z(x) + k(x, a)] + E t + o(t) >
( aeu
E > 0 independent
of x and with o(t)/t
e --Ois,‘rf, ((Lh)(X(x,
+ 0 as t 10 uniformly
s), a) - a/2(X(x, s))
0
1
inf [(,55)(x, a) - c&(x) + k(x, a)] - E (1 - e-“‘)/a new >
by analogous arguments, E > 0 again being o(t)/t + 0 as t 10 uniformly in x. Combining ‘I’&?f (QtW
arbitrarily the two,
prescribed
- W)) = SW)
and independent
‘ewus k(X”(s),
= E H
where X”( *) satisfies [ll, theorem 111.3.81.
a) ds + e-“‘f(X”(t))/X”(O)
= x
0
the mild
version
of (1.1)
4. MARKOV
for
of x and
in C,(H).
Remark 3.1. One can also show that Qt, t 2 0, is the “lower envelope” [ll, p. 811 of the affine semigroups Qf, t 2 0, a E U, defined by QPf(x)
- o(t)
u( .) = a. This
can
in the sense of
1 ,
be proved
as in
SELECTION
We shall now set up an analog of Krylov’s Markov selection procedure to extract an optimal Markov control. Let (Si] C (0, 00) be the rationals and (fi) a countable subset of C,(H) which is a separating class for P(H). For n = 0, 1,2, . . . , define F,, = C([O, co); H) x B” + R by m &(x(e),
~(a)) =
e-“’ &x(t), u(t)) dt i
0
V. S. BORKAR and T. E. GOVINDAN
28
and let [Fr , F2, . . .) be a relabelled
enumeration
of the maps
m e+‘fj(x(t))
(x(*)9 Y(.)) + i For rc E P(H),
i,j=
dt,
1,2 ,....
0
let
Jo(n, M.1) = EFo(X-), (that is, Jo(rr, u( *)) = J(n, u(e)) defined
with 2(X(O))
M*)>l
= rr,
earlier),
&I(@ = i$ Jo(n, U(‘)), A,(n) andfori=
Jitrc,
U(‘)) E r, I Job, a*))
x E H,
=
6(n)
= Amf) Ji(n, d’)), 1 IT
let S, denote
E[F,(X(‘)v
u(o))l
U(‘))
Ai For
= lUX(.)Y
= I/o(n)19
1,2,..., with $(X(O))
= loe(x(o>> u(‘)) E Ai-l(n) the
= n,
I Jj(r9 u(*)) = Y(n))*
Dirac
measure at x. Let Al(x) = Ai(6,), v’(x) = I$(6,), 2.1, it follows that Ai( i 2 0, ... . As in corollary is a family of compact nonempty subsets of P(C([O, 00); H) x B”) for each rc. It is clearly nested A,(n) = fliAi(7r) (or A;(X) = A/(X)) (that is, Ai C A,_l(lr) for all i). Thus, are compact nonempty. Let B(n) = (S(X(*)) 1Oe(X(*), u(a)) E A,(n)} (respectively, B’(x) = (JZ(X(*)) 12(X(*), u(e)) E AL(x)]). We shall prove that these are in fact singletons containing a single Markov solution and furthermore, if B’(x) = (P,], then (P,, x E Hj is a Markov family (i.e. satisfies the Chapman-Kolmogorov equation). We start with some technical lemmas. Let (X(e), w(e), u(e)) satisfy the mild version of (1.1) with 2(X(O)) = rr. Let ,u~ E P(C([O, m>; H) x C([O, co); H) x B”) denote the regular conditional law of (X(e), W(D), u(*)) given X(0) = x, defined n- a.s. uniquely. Consider the probability space (d, 5, p) defined as follows: d = C([O, 03); H) x C([O, co); H) x B”, 5 its Bore1 a-field and p = px. Define on this probability space the “canonical” processes X,(e), w,(e), u,(o) as follows: for w = (x(m), u(a), z( *)) E fi, Xx(w, t) = x(t), w,(w, t) = u(t), u,(w, t) = z(t), t > 0. The following is easily proved. Ji'(Xl U(')) = Ji(S,,U(')),i = 0, 1,2,
ni
LEMMA 4.1. For z- a.s. x, w,(e) is a Wiener process with incremental covariance admissible control w.r.t. w,( *) and X,( 0) the corresponding (unique) mild solution X,(O) = x. 4.2. If
LEMMA
da))
E
Proof.
$(X(e),
u(e)) EAT
for
some
i z 0,
then
for
II-
a.s.
A;(x). Let i = 0. Then vo(n) =
n(dx) I/d(x) = E K,‘W(O))l .
Q, u,(e) an to (1.1) with
x,
.33(X,(*),
Semilinear
stochastic
evolution
equations
29
Let E > 0 and E, = 2-“~, n L 1. Let u”( .), n h 1, be admissible controls such that for n 2 1, u”( *) is &,-optimal for initial law rr. Let X”(a) be the corresponding mild solution of (1.1). Let 2, = lx E H ( u,“(a) is c-optimal for the initial condition xl = lx E HI J,(x, uJ(.)) I V;(x) + E),
n 2 1.
Then ~(2,‘) 5 2-“, otherwise the c,-optimality of u”(e) for initial law rr is contradicted. Let B, = Zi, B, = .Z,\(Uk:‘, Z,), n 2 2. Then n(U, B,) = 1. Consider the probability space (a’, 3’, P’) defined as follows: Q’ = H x C([O, co); H) x B”, 5’ = its Bore1 a-field and P’ the probability measure defined as follows: letting (x, y, z), where y = y( *), z = z( *), denote the typical elements of H, C([O, 00); H) and B”, respectively, let P’(dx, dy, dz) = Ir(dz) v(x, dy, dz), where v(x, dy, dz) E P(C([O, co); H) x B”), x E H, is simply 2(X,“(.), u,“(e)) for x E B,, n 2 1 and arbitrary for x $ lJ, B,. Define the “canonical processes” (X(e), ~(a)) on (Sz’, F’, P’) as follows: if UJ = (a, x(.),y(*)) denotes a typical element of Q’, then _%(a, t) = x(t), U(o, t) = y(t), t > 0. It is routine to check that this is an admissible pair satisfying the mild version of (1 .l) with C(X(0)) = rc. Also,
Since E > 0 was arbitrary, it follows that v,(n) = E[&‘(X(O))] = j n(dx)E[F,(X,(*), u,(e))]. The claim now follows for i = 0. Suppose the claim holds for i = 0, 1, 2, . . . ,j. Let $(X(e), u(e)) E Aj+l(n). By hypothesis, S(X,(*), u,(e)) E A;(x) for 7~ a.s. x. Now mimick the above argument to conclude that I$+i(rr) = E[v+i(X(O))] = 1 rc(d_x)E[Fj+,(X,(*), u,(e))] and, therefore, $3(X,(*), u,(+)) E AJ+i(x) for rr- a.s. x. The claim follows by induction. Let S(X,(*), u,(a)) E Ai for some i 2 0, rc E P(H). Let 5, = the completion of n S>f 0(X,(y), ui(y), y 5 s) with respect to the underlying probability measure for t 2 0. Let q be an a.s. finite (S,)-stopping time. Let .53(X,(~)) = P and S(X,(*), u,(e)) E A,(p) for i as above. Construct (2(e), 1?(e)) satisfying the mild version of (1 .l) as follows: let f: C([O, a); H) x B” -+ [0, co) denote the measurable function such that q = j&Y,(*), ui(*)). Let Fj = f@(e), z&o)). The process (x(*), t?(e)) is constructed (say, on the canonical path space C([O, a~); H) x B” endowed with its Bore1 a-field) such that (i) .=W(Q A e), 6(1 A 0)) = J&W7 A *), U(VA e)), (ii) the regular conditional law of (_%(Q+ *), tl(J + *)) given (??(e A *), ti(ij A -)) = its regular con_ditional law given z(e) = the regular conditional law of (X,(m), u,( *)) given X,(O). Thus, (X(e), 17(*)) is a replica in law of (Xi(*), ui(.)) up to the stopping time 3 - q and (X(4 + *), $0 + 0)) is a replica in law of (X,( *), u2( m)). Let 5, = the completion of nS, f a(z(y), u”(y), y I S) with respect to the underlying probability measure for t 2 0. Let 6, (. . .) denote the “regular conditional law of . . .” given gi. LEMMA 4.3.-Ci(_@fj + *),ii(ij + *))E Al(x)(,=gcGj Proof.
a.s.
This is immediate from the preceeding lemma and our construction
of (z( *), c(e)).
V. S. BORKAR and T. E. GOVINDAN
30
LEMMA 4.4. .$(X(*),
Proof. Suppose
G(e)) E Ai(
i = 0. Then
E[F,(z( -),
U( -))I = E
+
eecrt&z(t), eealiE
?
=E
0
C(t)) dt
e-““-“’ k@?(t), C(t)) dt/$,
11
eear k(Xl(t), u,(t)) dt + emaq4(x, u~~(~))I~=~~~~.
(4.1)
1
If the claim were false, then (4.1) is strictly
greater
than
“‘I = E
b(n)
e-“‘k(X,(t), II II0
ul(t)) dt
m +
e-““E
eC”(‘-‘) k(X,(t),
u,(t)) dt/F,
li ‘I
11
‘1
=E
em”’k(X,(t), [S,O
u,(t)) dt + e-“” J&G ti,(-)) Ix=x,cllj I
(4.2)
where fi(*) = u,(q + a). Letting 4 = $(X,(q)), we have J;(x, G,(a)) 2 Vi(x) = J$x, u,,(a)) for $-- a.s. x by lemma 4.2. Thus, (4.2) is greater than or equal to (4.1). Since the opposite inequality is true in any case by the very definition of Vo(n), equality must hold. The claim J,‘(x, ti,( *)) = follows for i = 0. Suppose it holds for i = 0, 1,2, . . . . j, and furthermore, Vi(x) = J/(x, u,,(e)) for 4- a.s. x and 0 5 I 5 j. Now mimick the above argument to show that the claim is true for i = j + 1 and Jjl+,(x, c,(e)) = vj’+,(x) = J’+,(x, u,,(.)) for $- a.s. x. The claim follows by induction. COROLLARY4.1. .G,,(X,(q
+ a), U,(V + .)) E Ai(X)Ix=X,(II) a.s.
Proof. This is equivalent contained
in the proof
to: J/(x, ci,(*)) = c’(x) for dof the above lemma.
LEMMA 4.5. Let AZ(X,(*), u,(e)) E A,(n). process and ur(*) a Markov control.
Proof. The second claim follows that from corollary 4.1, &&X,(v where II is as above.
Then
X,(e)
a.s. x, the proof of which is already
is a time-homogeneous
Markov
from the first as in [II, pp. 12-131. To prove the first, note
+ .A u,(rl + -)) E ~i(~)l,=X,~s~
a.s.,
i 1 0,
Thus,
E]Fi(X,(rl
strong
+ e), u,(ul + -)W,l
= vi’(X,h))
a.s.,
i 2 0.
Semilinear
Taking
conditional
expectations
E]Fi(X,(r
stochastic
with respect
+ *), u,(rl +
-))/m)l
evolution
to X,(q)
31
equations
on both sides
= vi’(X,(rl)) = E[Fi(X,(q + -), u,(q + -))/F,] as.
Thus,
for (si], {f,j as in the beginning
(4.3)
of this section,
1, e --5f1E[f,(X,(q
+ t))/s,]
dr
\ ,0 ‘m =
e -““E[fj(X,(q
+ t))/X,(q)]
dt a.s.,
il
l,jz
1.
I0 Since (Si) are dense in (0, co), it follows E]fj(Xi(11
+
W5,l
from [ll,
= E[fj(X,(r
lemma
+ Wx,(rl)l
IV.l.51,
that
a.s.,
j 2 1, for a.e. t.
of X,(o) Use Fubini’s theorem to interchange “a.s.” and “a.e. t” and invoke path continuity to drop the “a.e”. Since (A] is a separating class for P(H), the strong Markov property follows. Time-homogeneity follows from the observation that the middle term in (4.3) does not have an explicit q-dependence. THEOREM 4.1. Also, if for x ] rc(dx)P,(da) corresponding law II.
For each x E H, B’(x) is a singleton containing the law of a Markov solution. E H, B’(x) = {Px], then (P,, x E HI is a Markov family. Finally, if P,(dw) = for some 71 E P(H) and $(X( *), u( *)) is such that .$(X(e)) = P, and u( *) the Markov control implied by lemma 4.5, then (X(e), u( *)) is optimal for initial
Proof. Suppose JZ(X(.), u(*)), Z(X’( *), u’(.)) E A L(x), for some x E H. Then the foregoing with q = 0 shows that E[Fi(X(*), ~(a))] = E[&(X’(.), u’(e))], i 2 1, leading toE[fi(X(t))] = E’[fi(X’(t))], t > 0, i 2 1, by arguments similar to those of the preceeding lemma. Since cfi) is a separating class for P(H), the laws of X(t), X’(t) agree for each t. Denote their common law of ~(x, t, dy) E P(H). By corollary 4.1, &(X(t + -)) E B(X(t)) a.s. Thus, for i 2 1, t, s 1 0,
i
fi(Z)p(X,
t + S,
dz) = E[fi(X(t
+ s))/X(O) = X]
= E[E[fi (X(t + s))/X(t)ll =E
fi(Z)P(X(t),
S, dZ)
l.i >I
= I// I
fi (z)P(Y, ~3 dz)P(x,
1 t, dY)*
By our choice of (fi], the Chapman-Kolmogorov equation follows. Thus, (p(t, x, dy), t 2 0, x E HJ are indeed the transition probabilities of a Markov process. Since transition probabilities and the initial law completely specify a Markov process in law, each B’(x) must be a singleton. From lemma 4.2, it follows that each B(n) is a singleton. The last claim is obvious.
‘TJ 3 J .103 : dq 2 + 2 : ‘J aur3ap ‘0 z 1 god ‘LLLIOUunu_ua.Idns ayl IJ$!M g + ff suo!wIn3 a~qt?.InsEauI [aloH papunoq 30 aat?ds yXue$~ ayl alouap TJ Ia7 *ssaDo.Id ~oy.~em Zkupuodsa.LIo3 ayi Zu!aq (.)x ‘(( .&)n = (.)n ~O.IWO~~oyni?n B 01 %npuodsal.IoD (1.1) 30 sanweaw uoynlos 30 (H 3 x ‘“d) 1cpu1~?3AO~IEIAJ 1? .Iap!suoD MON '0 < 1 ‘!a 30 uyod paxg E si A ‘snyL ‘0 7 I .103
~yc2wdp p.wpuws
‘LIIEU!~ *(~)“3
01 peal [I’III uorvag ‘1 I] y se ssuawnXn? 8uyuuue.Go.Id 3 h ‘snyL ~~UBISUO~aIqv!ns v 8uIaq 0 < 3 ‘Z’E r?wuraI dq
(‘)n il I (.)n i’lI 114- xl12 +
3 5
0’
IP ((W
‘(I ‘OxPi
- (W
‘(I ‘x)x)Y) ,,_a
2
dns + ? 5
_LL,
I[
0
2~ Wn
‘0 ‘43~
- HIP ‘0 ‘xlxh)
IL
2
,,_a
dns 5 [((c)h - (x)h I
a%IEl 0 < _z YXd ‘0 < 3 pur! H 3 ic ‘x la-I *papunoq ICl.IEap s! hi .anbyun aq .r!ayl ‘alo3alayl ‘pue suoyXwu0~ a.w ‘0 < .t “6 wqi ~~01103 II ‘(~!)I’E uraloayi ‘0 < _I “0
30 lu!od paxy anbrun 0
.
x=
(O)X/IP
(‘)?I
(W ‘(~lxh ,,_a
B
11~~
sw!od pax!3 tu0.1~ *Joor&
ays s! h ‘I’S vruns?
qq!ss!mpr!
= (x)/h
34
,=<11
[
Icq pauyap
s! s!y~ wa[qold
.iuapyAa aq 111~ se ‘uogsas ixau ayi u! waloayi uogez!3Ila.h ayi 30 uogejnw~o3 Ino spage s!y~ *s1oguo3 AO~.IE~J .Iaylo 103 uoyyos au0 uaAa 30 asualsrxa ayl alnsua li saop .ION ~suop!puos pi?!l~u! 1112 10 XLIOS.103 suoyn[os ‘Ipwydouou dIq!ssod ‘laylo ar\ey llew ui pa%nId ((.)x), n = (.)n qil~ (1-1) ieq Aiyq~ssod ayi aicu!uuIa $0~ = (.)*n aq ~0.1iuo~ saop s!yi w?yl pu!tu u! lday aq pinoys 11 ~a~qwns~atu n + jg f *n ‘((.)x)*0 9wpuodsaJlo3 ayl lay ‘Ieurgdo ale (1’ 1) 30 (.)_y Suo~lnIos ~0y.q~~ %upuodsal.IoD ysns {H 3 x cxd) dpwe3 ~oy.qy e 30 aauais!xa ayi away aki ‘az!.n?wuns 01
~oyqq
ayi
wyl
NV(INIAOD
‘3
‘J
pLI?2 XVXXOH
‘S
‘A
ZE
33
Semilinear stochastic evolution equations
semigroup of operators where k,(x) = k(x, v(x)), x E H, and (T,‘, t 1 0) is the transition E -+ E associated with the Markov process X( *). For A > 0, define the resolvent operators R,(A): E + E by m f E E. e-” T,‘f(x) dx, R,(L)f(x) = i0 These satisfy the resolvent
identity
%(A)&(P)
= R&)R,@)
= -(A - P)-~(R,@)
- R,(P)),
for A, fi E (0, a~), implying in particular that range (R,(A)) is independent of A. Let E, denote its closure in E. Then E. is a closed subspace of E given by [17, p. 1101 Eo=(f~E~AR,(A)f+finEas,I-+co).
(5.1)
From (5.1), it follows that T,’ maps E, + E. for t 2 0 and thus may be viewed as a semigroup of operators on E,. Define an operator G: with domain a)(G:) c E, as follows: ZIJ(G:) = R,(A)E, (which is again independent of A), which is then dense in E, by (5.1), and G:: D(G:) + E. is unambiguously defined by (A - G:)R,(A)f = f for all 1 > 0, f E E, [17, p. 1111. Then the following hold [17, pp. llO-1131: (i) R,(A) = (A - G/,-l on E,; (ii) for f E E,, f E a>(G:) if and only if the limit
exists in E, and then G;f LEMMA 5.2.
infinitesimal
= g.
t 1 0) restricts to a semigroup of bounded affine operators generator G, satisfying D(G,) = a)(G:) and for f E D(G:),
(T,,
G,f
E, + E, with
= G:f - af + (a! - G:)R,(a)k,.
(5.2)
Remark 5.1. If k, E E,, the last term on the right in (5.2) is simply k,. Proof.
For f E E, , t > 0, 7;f(x)
- f(x)
= eC”’ T,‘f(x)
- f(x)
+
m emcuS T,’ k,(x) ds 0
=
ema’
q’f(x)
- f(x)
me-“$ T; k,(x) ds i t
+ R,(cr)k,(x)
- em*’ T,’ R,(ol)k,(x).
From (5.3), it follows that 7; maps E, to E,. Semigroup property follows property. (q) are clearly affine. Finally, divide (5.3) on both sides by t and as t 10. Note that R,(oc)f E ‘B(G:) and thus the limit on the right-hand side the left-hand side) exists if and only if f E a)(G:). If it does, the left-hand tends to G,f and the right-hand side is seen to tend to the right-hand side LEMMA 5.3. For t > 0, T,’ : E. + Ho is a contraction
V,(X) = E
co emartk,(X(t)) 0
and has a unique
1
dt 1X(0) = x .
(5.3)
from the Markov take the limit in E, (and, therefore, on side by definition of (5.2).
fixed point
wv given by
34
V. S. BORKAR and T. E. GOVINDAN
Proof.
For
f, g E E,, t > 0, one easily checks that 1T’f(x)
- T,’ &)I
= ~E[e-“‘(fW(O) 5
The first claim follows.
By the Markov
eCatllf - gll.
property, m
*t w,(x) = E
eFas k,(X(s))
ds + e-“‘E
[! 0
e-“‘“-” WWI
ds + emcuty(X(t))/X(O)
[!.O
proving
the second
Our verification
ds/X(0]/WX
= x]
i.i t
'I emolsk,(X(s))
= E
- sW(t)))li
= x
, 1
claim. theorem
is the following
theorem.
5.1. Let {Px, x E H) be a Markov family corresponding to a Markov control u( *) = v(X( e)), X( *) being the corresponding Markov process. Define (T,‘, t L 01, E, as above. (a) The above Markov family is optimal for any initial law if and only if there exists a I?/ E E, fl C,(H) which is the common fixed point of (Qt, t > 0) and (c’, t > 0). (b) Alternatively, the above Markov family is optimal if and only if the value function (I/ defined at the beginning of this section satisfies w E D(G) n a)(G,) and THEOREM
G,i,u=t?=Gv where 19denotes the zero vector in H. (c) At least one optimal Markov family
(5.4)
exists.
Proof. The first part is immediate from the observation that I,Y= wV implies optimality of the Markov family for all initial conditions. The second part follows immediately from the first. The third is simply a restatement of the results of Section 3 and ensures that the first two statements are not vacuous. Note the clear similarity between (5.4) and the “pointwise minimization of the Hamiltonian” that occurs in the finite dimensional verification theorem. In conclusion, it should be noted that this work leaves a few loose ends and should, therefore, be considered only as a first step towards obtaining a good verification theorem for controlled stochastic evolution equations. For one thing, we have made several restrictive assumptions. Another problem is the fact that we do not have a good characterization of the space E,. Also, the uniqueness issue for the evolution of the Nisio semigroup needs to be studied (see, for example, [18]). We hope, however, to have successfully underscored the following point: any statement of verification theorem for controlled stochastic evolution equations in infinite dimensions should involve a Markov family of solution measures rather than an individual Markov control because of the possible ill-posedness of a general Markov-controlled stochastic evolution equation.
Semilinear
stochastic
evolution
equations
35
REFERENCES 1. BENSOUSSANA., Stochastic Control by Functional Analysis Methods. North-Holland, Amsterdam (1982). 2. HAUSSMANNU. G., A stochastic maximum principle for optimal control of diffusions, Pifman Research Notes in Mathematics, Vol. 151. Longman Scientific and Technical, Harlow (1986). 3. LIONS P. L., Viscosity solutions of fully nonlinear second-order equations and optimal stochastic control in infinite dimensions, Part 1: The case of bounded stochastic evolutions, Acta Math. 161, 243-278 (1988). 4. CANNARSAP. & DA PRATOG., Second-order Hamilton-Jacobi equations in infinite dimensions, SIAM J. Control
Optim. 29, 414-492 (1991). 5. CLARKE F. H., Optimization and Nonsmooth Analysis. Wiley, New York (1983). 6. KRYLOV N. V., Selection
of a Markov
process
from a Markov
system of processes,
Izv. Akad. Nauka USSR Ser.
Math. 37, 691-708 (1973). (In Russian.) 7. EL KAROUI N., NGUYEN D. H. & JEANBLANC-PIQUE M., Compactification methods in the control of degenerate diffusions: existence of an optimal control, Sfochastics 20, 169-219 (1987). 8. HAUSSMANN, U. G., Existence of optimal Markovian controls for degenerate diffusions, in Stochastic Differential Systems (Edited by N. CHRISTOPEIT, K. HELMES and M. KOHLMANN), Lecture Notes in Control and Information Sciences, Vol. 78, pp. 171-186. Springer, Berlin (1986). 9. NISIO M., Lectures on Stochastic Control Theory, ISI Lecture Notes, Vol. 9, Macmillan, India (1983). 10. AHMED N. U., Semigroup theory with applications to systems and control, Pitman Research Notes in Mathematics, Vol. 246. Longman Scientific and Technical, Harlow (1991). 11. BORKAR V. S., Optimal control of diffusion processes, Pitman Research Notes in Mathematics, Vol. 203. Longman Scientific and Technical, Harlow (1989). 12. ARNOLD L., CURTAIN R. F. & KOTELENEZ P., Nonlinear stochastic evolution equations in Hilbert space, Report No. 17, University of Bremen (1980). 13. ICHIKAWA A., Stability of semilinear stochastic evolution equations, J. math. Analysis Applic. 90, 12-44 (1982). 14. DA PRATO G. & ZABCZYK J., Nonexplosion, boundedness and ergodicity for stochastic semilinear equations, J. diff Eqns 98, 181-195 (1992). 15. GOVINDAN T. E. & JOSHI M. C., Stability and optimal control of stochastic functional differential equations with memory, Num. funct. Anal. Optim. 13, 249-265 (1992). 16. IKEDA N. & WATANABE S., Stochastic Differential Equations and Diffusion Processes. North-Holland, Tokyo (1981). 17. WILLIAMS D., Diffusions, Markov Processes and Martingales, Vol. 1: Foundations. Wiley, Chichester (1979). 18. NISIO M., Note on uniqueness of semigroup associated with Bellman operator, in Advances in Filtering and Optimal Stochastic Control (Edited by W. FLEMING and L. E. GOROSTIZA), Lecture Notes in Control and Information Sciences, Vol. 42, pp. 267-275. Springer, New York (1982).