• Aucun résultat trouvé

MULTI-PERIOD MEAN VARIANCE OPTIMAL CONTROL OF MARKOV JUMP WITH MULTIPLICATIVE NOISE SYSTEMS

N/A
N/A
Protected

Academic year: 2022

Partager "MULTI-PERIOD MEAN VARIANCE OPTIMAL CONTROL OF MARKOV JUMP WITH MULTIPLICATIVE NOISE SYSTEMS"

Copied!
14
0
0

Texte intégral

(1)

on the occasion of his 70th birthday

MULTI-PERIOD MEAN VARIANCE OPTIMAL CONTROL OF MARKOV JUMP WITH MULTIPLICATIVE NOISE SYSTEMS

OSWALDO L.V. COSTA and RODRIGO T. OKIMURA

We consider the multi-period mean variance stochastic optimal control problem of discrete-time Markov jump with multiplicative noise linear systems. First, we consider the performance criterion to be a linear combination of the final variance and expected value of the output of the system. We analytically derive an optimal control policy for this problem. By using this solution, we consider next the cases in which the performance criterion is to minimize the final variance subject to a restriction on the final expected value of the output, and to maximize the final expected value subject to a restriction on the final variance of the output of the system. The optimal control strategies are obtained from a set of interconnected Riccati difference equations.

AMS 2000 Subject Classification: 49N10, 60J10, 91B28, 93E20.

Key words: mean variance control, Markov jump system, multiplicative noise, sto- chastic optimal control.

1. INTRODUCTION

The uni-period mean-variance optimization is a classical financial prob- lem introduced by [10] which paved the foundation for the modern portfo- lio theory. Using a stochastic linear quadratic theory developed in [1], the continuous-time version of Markowitz’s problem was studied in [15], with closed-form efficient policies derived, along with an explicit expression of the efficient frontier. In [5] the authors extended the mean-variance allocation problem to the discrete-time multi-period case while in [16] they considered a multi-period generalized mean-variance formulation for the risk control over bankruptcy. A geometric approach to these problems was presented in [4], considering assets as well as liabilities in the portfolios. In [13] the authors considered the discrete-time multi-period mean-variance allocation problem in the case where the parameters are subject to Markovian jumps, following an approach closely related to that in [5], while in [14] they studied a simi- lar problem under a different point of view. As pointed out in [16], one of

MATH. REPORTS9(59),1 (2007), 21–34

(2)

the key difficulties in solving the multi-period mean variance problem is the non-separability in the associated stochastic control problem in the sense of dynamic programming. Due to that reason, a tractable (from the dynamic programming point of view) auxiliary problem is introduced.

In this paper we consider the multi-period mean variance stochastic op- timal control problem of discrete-time Markov jump with multiplicative noise linear systems. As in [5] we introduce a tractable auxiliary stochastic qua- dratic optimal control problem, where the performance criterion consists of a linear part and a quadratic cost of the state variable at the final time T. There is no penalty on the control variable, so that the standard techniques for the LQG problems cannot be used. It should be pointed out that problems with indefinite weighting matrices have been intensively studied lately as can be seen, for instance, in [8], [11], and for the case with Markov jumps and multiplicative noise in [3], [6], [7], [9].

The paper is organized as follows. In Section 2 we present the notation and some preliminary results that will be required for the solution of the auxil- iary stochastic quadratic optimal control problem. In Section 3 we present the three problems that we will consider. In Problem P1 it is asked to maximize the final expected value subject to a restriction on the final variance of the output of the system while in Problem P2 it is asked to minimize the final variance subject to a restriction on the final expected value of the output. In Problem P3 it is asked to minimize a performance criterion which is a linear combination of the final variance and expected value of the output of the sys- tem. An analytical optimal control policy for these problems can be obtained through an auxiliary stochastic quadratic optimal control problem, solved in Section 4 in terms of a set of interconnected Riccati difference equations. The paper is concluded in Section 5 with a solution for the 3 problems stated in Section 3, expressed in terms of some key parameters. These parameters are explicitly written as functions of the parameters of the system and in terms of the set of interconnected Riccati difference equations.

A possible application of the results in this paper would be in an asset liabilities management (ALM) model for defined-benefit (BD) pension funds with regime switching. We could assume that the market parameters depend on the market mode that switches according to a Markov chain among a finite number of states. The ALM for DB pension funds problem can then be written as a Markov jump with multiplicative noise LQ optimal control problem with linear and quadratic costs, so that the results presented here can be applied to solve the problem.

(3)

2. PRELIMINARIES

We denote by Rn the n-dimensional real Euclidean space and by B(Rn,Rm) the normed bounded linear space of all m×n real matrices, with B(Rn) := B(Rn,Rn). For a matrix A B(Rn,Rm), N(A) denotes the null space of A, R(A) the range of A and A the transpose of A. As usual, for A B(Rn), A 0 (A > 0 respectively) means that the matrix A is posi- tive semi-definite (positive definite), and tr(A) denotes the trace of A. The operator expected value will be denoted by E(.). Set Hn,m for the linear space made up of all N-sequences of real matrices V = (V1, . . . , VN) with Vi B(Rn,Rm), i = 1, . . . , N, and, for simplicity, set Hn := Hn,n. We say that V = (V1, . . . , VN) Hn+ (Hn) if V Hn and, for each i= 1, . . . , N, Vi

is a positive-semidefinite (symmetric) matrix. For V = (V1, . . . , VN) Hn, R= (R1, . . . , RN)Hn, we write V ≥R ifVi−Ri0 for each i= 1, . . . , N. We denote byB(Hn,Hm) the space of all bounded linear operators fromHnto Hm and, in particular, B(Hn) := B(Hn,Hn). We say that T ∈ B(Hn+,Hm+) ifT ∈ B(Hn,Hm) and is such that T(V) Hm+ whenever V Hn+. For a sequence ofndimensional square matricesA(0), . . . , A(T), we use the follow- ing notation: t

=sA() =A(t). . . A(s) for t≥s,I fort < s. We define 1S as the usual indicator function, that is, 1S(ω) = 1 if ω ∈S, zero elsewhere. We need the following definition (see [12], pages 12–13)).

Definition 1. For a matrix A B(Rn,Rm), the generalized inverse of A (or Moore-Penrose inverse of A) is defined to be the unique matrix A B(Rm,Rn) such that i) AAA = A, ii) AAA = A, iii) (AA) =AA, and iv) (AA) =AA.

We recall the result below (see [12], pages 12–13).

Proposition1 (Schur’s complement). The following assertion are equiv- alent.

a) Q=

Q11 Q12 Q12 Q22

0.

b) Q220, Q12=Q12Q22Q22 and Q11−Q12Q22Q120.

c) Q110, Q12=Q11Q11Q12 and Q22−Q12Q11Q120.

The following result will be useful in the sequel.

Proposition 2. Consider Y B(Rn) and M B(Rm) with Y 0.

Let A and B be stochastic matrices (that is, each entry of them is a random variable) in B(Rn) and B(Rm,Rn), respectively. Then

(1) E(AY A)−E(AY B)

E(BY B)

E(BY A)≥0

(4)

and

(2) E(AY B) =E(AY B)

E(BY B) +M

E(BY B) . Proof. Consider the stochastic matrix

(3) Q=

Q11 Q12 Q12 Q22

=

Y Y B BY R

,

where R = BY B. Clearly we have Q11 = Y 0, and Q12 = Y B = Q11Q11Q12 = Y YY B since Y YY = Y from Definition 1. Furthermore, using again Y YY = Y, we have Q22−Q12Q11Q12 = R−BY YY B = R−BY B= 0. Thus, by Schur’s complement (Proposition 1), Q≥0, hence

AY A AY B BY A R

=

A 0 0 I

Y Y B

BY R

A 0 0 I

0.

Taking the expected value of the above equation, we get Q11 Q12

Q12 Q22

=

E(AY A) E(AY B) E(BY A) E(R)

0.

By Schur’s complement again, we have

0≤Q11−Q12Q22Q12=E(AY A)−E(AY B)E(R)E(BY A) and Q12=E(AY B) =Q12Q22Q22=E(AY B)E(R)E(R), showing the result stated.

3. PROBLEM FORMULATION

On a probabilistic space (Ω,P,F) consider the Markov Jump Linear System with multiplicative noise

x(k+ 1) =

A¯θ(k)(k) +

εx

s=1

Aθ(k),s(k)wsx(k) x(k)+

(4)

+

B¯θ(k)(k) +

εu

s=1

Bθ(k),s(k)wsu(k) u(k), x(0) =x0, θ(0) =θ0,

where θ(k) is a time-varying Markov chain taking on values in {1, . . . , N}

with transition probability matrixP(k) = [pij(k)], {wsx(k); s = 1, . . . εx, k = 0,1, . . . , T 1} are zero-mean random variables independent of the Markov chain{θ(k)}, with variance equal to 1 and E(wxi(t)wjx(k)) = 0, for all t =k

(5)

and i = j. Similarly, {wus(k); s = 1, . . . εu, k = 0,1, . . . , T 1} are zero- mean random variables independent of the Markov chain{θ(k)}, with variance equal to 1 and E(wiu(t)wju(k)) = 0, for all t = k and i = j. The initial conditionsθ0 and x0 are assumed to be independent of {wsx(k)} and{wus(k)}, with x0 an n-dimensional random vector with finite second moments. Set µi(0) = E(x01{θ0=i}),µ(0)∈RN n asµ(0) = (µ1(0) · · · µN(0)), and Qi(0) = E(x(0)x(0)1{θ0=i}), Q(0) = (Q1(0), . . . , QN(0)) Hn+. The correlation of wsx1(k) andwus2(k) is denoted byE(wxs1(k)wus2(k)) =ρs1,s2(k). Without loss of generality, assume thatε=εx =εu. For eachk= 0,1, . . . , T1, we also have

A(k) = ( ¯¯ A1(k), . . . ,A¯N(k))Hn,

As(k) = (A1,s(k), . . . ,AN,s(k))Hn, s= 1, . . . , ε, B(k) = ( ¯¯ B1(k), . . . ,B¯N(k))Hm,n,

Bs(k) = (B1,s(k), . . . ,BN,s(k))Hm,n, s= 1, . . . , ε.

Set πi(k) = P(θ(k) = i), let Ft be the σ-field generated by {(θ(s), x(s));

s= 0, . . . , τ}, and write

U(τ) ={uτ = (u(τ), . . . , u(T1));u(k) is an m-dimensional random vector with finite second moments that isFk-measurable for each k=τ, . . . , T−1}.

Consider the scalar output

(5) y(t) =Lx(t)

of system (4), whereL∈B(Rn,R).

The multi-period mean-variance problem aims at selecting u U(0) which yields the greatest expected terminal value of the output E(yu(T)) given a maximal terminal output σ2 for the variance Var(yu(T)), or which produces the lesser terminal output variance Var(yu(T)) given a maximal ex- pected terminal value for the output E(yu(T)). Formally, these problems, caled respectively P1

σ2

and P2 ( ), can be stated as P1

σ2

: min

u∈U(0) −E(yu(T)) subject to Var (yu(T))≤σ2, (6)

P2 ( ) : min

u∈U(0) Var (yu(T)) subject to E(yu(T))≥ . (7)

Alternatively, an unconstrained form would be P3 (ν) : min

u∈U(0) νVar (y(T))−E(y(T)), (8)

where ν [0,) is a risk aversion coefficient, giving a trade-off preference between the expected terminal wealth and the associated risk level. Since problem P3(ν) involves a non-linear function of the expectation in Var(V(t)) =

(6)

E(V(t)2)−E(V(t))2, it cannot be directly solved by dynamic programming.

A solution procedure to seek an optimal dynamic control policy for problem P3(ν) based on a tractable auxiliary problem is proposed in [16]. We will adopt the same procedure in this paper, and consider the auxiliary problem

A(λ, ν) : min

u∈U(0) E

νy(T)2−λy(T) . (9)

4. SOLUTION OF THE AUXILIARY PROBLEM

Let us consider the following intermediate problems for problem (9). At each time k∈ {0, . . . , T 1} define

J(x(k), θ(k), k) = min

uk∈U(k)E

νy(T)2−λy(T)| Fk

.

Define next fork= 0, . . . , T1 the operatorsE(k, .)B(Hn),A(k, .) B(Hn), G(k, .) B(Hn,Hn,m), R(k, .) B(Hn,Hm), P(k, .) B(Hn), V(k, ., .) B(Hn×Hn,1,Hn,1),D(k, ., ., .)B(Hn×Hn,1×H1,H1), andH(k, .)B(Hn,1, Hn,m). ForX∈Hn,V Hn,1,γ H1, and i= 1, . . . , N, set

Ei(k, X) = N j=1

pij(k)Xj,

Ai(k, X) = ¯Ai(k)Ei(k, X) ¯Ai(k) + ε s=1

Ai,s(k)Ei(k, X)Ai,s(k),

Gi(k, X) =

A¯i(k)Ei(k, X) ¯Bi(k)+

+ ε s1=1

ε s2=1

ρs1,s2(k)Ai,s1(k)Ei(k, X)Bi,s2(k)

,

Ri(k, X) = ¯Bi(k)Ei(k, X) ¯Bi(k) + ε s=1

Bi,s(k)Ei(k, X)Bi,s(k), Pi(k, X) =Ai(k, X)− Gi(k, X)Ri(k, X)Gi(k, X),

Vi(k, X, V) =Ei(k, V)( ¯Ai(k)−B¯i(k)Ri(k, X)Gi(k, X)), Di(k, X, V, γ) =Ei(k, γ)1

4Ei(k, V) ¯Bi(k)Ri(k, X)B¯i(k)Ei(k, V), Hi(k, V) = ¯Bi(k)Ei(k, V).

(7)

It is easy to see that E(k, .) B(Hn+), A(k, .) B(Hn+) and R(k, .) B(Hn+,Hm+). The next proposition will be useful in the sequel. It justi- fies the definition of the operators above. Notice that this result is closely related to the optimality Bellman equation.

Proposition 3. Let P = (P1, . . . , PN)Hn+,V = (V1, . . . , VN)Hn,1 andγ H1. Then P(k, P)Hn+ and

Gi(k, P) =Gi(k, P)Ri(k, P)Ri(k, P).

(10)

Moreover, for anyuk U(k), u(k) =u, x(k) =x and θ(k) =i we have E

νx(k+ 1)Pθ(k+1)x(k+ 1)−λVθ(k+1)x(k+ 1) +λ2

ν γθ(k+1)|Fk = (11)

=ν[xAi(k, P)x+ 2xGi(k, P)u+uRi(k, P)u]

−λ[Ei(k, V)A¯i(k)x+ ¯Bi(k)u ] +λ2

ν Ei(k, γ), and, if for eachi,

(12) Ei(k, V) ¯Bi(k)R(Ri(k, P)), then (11) can be rewritten as

ν[xAi(k, P)x+ 2xGi(k, P)u+uRi(k, P)u] (13)

−λ[Ei(k, V)A¯i(k)x+ ¯Bi(k)u ] + λ2

ν Ei(k, γ) =

=ν[xPi(k, P)x+ (u+a(x))Ri(k, P)(u+a(x))]−

−λVi(k, P, V)x+λ2

ν Di(k, P, V, γ), where

a(x) =Ri(k, P)

Gi(k, P)x λ

B¯i(k)Ei(k, V) . (14)

Proof. SettingA= ¯Ai(k)+

ε s=1

Ai,s(k)wxs(k),B = ¯Bi(k)+

ε s=1

Bi,s(k)wsu(k), Y = Ei(P) in Proposition 2, from inequation (1), and the hypothesis on {wsx(k)} and {wus(k)} we have

Pi(k, P) = ¯Ai(k)Ei(P) ¯Ai(k) + ε s=1

Ai,s(k)Ei(P)Ai,s(k)

−Gi(k, P)Ri(k, P)Gi(k, P)0,

(8)

hencePi(k, P) 0. It also follows from Proposition 2 and equation (2) that (10) holds. We then have

E

x(k+ 1)Pθ(k+1)x(k+ 1)|Fk = (15)

=x

A¯i(k)Ei(k, P) ¯Ai(k) + ε s=1

Ai,s(k)Ei(k, P)Ai,s(k) x+

+2x

A¯i(k)Ei(k, P) ¯Bi(k) + ε s1=1

ε s2=1

ρs1,s2(k)Ai,s1(k)Ei(k, P)Bi,s2(k) u+

+u

B¯i(k)Ei(k, P) ¯Bi(k) + ε s=1

Bi,s(k)Ei(k, P)Bi,s(k) u

and

E

Vθ(k+1)x(k+ 1)|Fk =Ei(k, V)

A¯i(k)x+ ¯Bi(k)u , (16)

E

γθ(k+1)|Fk =Ei(k, γ).

(17)

Equations (15), (16) and (17) yield (11). Considering now on the right hand side of (11) only the terms dependent onu and calling themf(u), we have

f(u) =νuRi(k, P)u+ 2

νxGi(k, P) −λ

2Ei(k, V) ¯Bi(k) u.

(18)

It follows from (10) and (12) that (18) can be written as f(u) =νuRi(k, P)u+

(19)

+2ν

xGi(k, P) λ

Ei(k, V) ¯Bi(k) Ri(k, P)Ri(k, P)u.

Writing a(x) as in (14), equation (4) can be rewritten as f(u) =ν[uRi(k, P)u+ 2a(x)Ri(k, P)u]

=ν[(u+a(x))Ri(k, P)(u+a(x))−a(x)Ri(k, P)a(x)].

Notice now that

−a(x)Ri(k, P)a(x) =

xGi(k, P)Ri(k, P)Gi(k, P)x (20)

−λ

νEi(k, V) ¯Bi(k)Ri(k, P)Gi(k, P)x+

+λ2

2Ei(k, V) ¯Bi(k)Ri(k, P)(Ei(k, V) ¯Bi(k)) ,

(9)

where we have used the fact that Ri(k, P)Ri(k, P)Ri(k, P) = Ri(k, P). Thus we have

ν[xAi(k, P)x+ 2xGi(k, P)u+uRi(k, P)u] +λ2

ν Ei(k, γ) (21)

−λEi(k, V)[ ¯Ai(k)x+ ¯Bi(k)u] =

=νxAi(k, P)x−λEi(k, V) ¯Ai(k)x+λ2

ν Ei(k, γ) +f(u) =

=ν[x(Ai(k, P)− Gi(k, P)Ri(k, P)Gi(k, P))]x

−λ[Ei(k, V) ¯Ai(k)− Ei(k, V) ¯Bi(k)Ri(k, P)Gi(k, P)]x+

+ν(u+a(x))Ri(k, P)(u+a(x))+

+λ2 ν

Ei(k, γ)1

4Ei(k, V) ¯Bi(k)Ri(k, P)(Ei(k, V) ¯Bi(k)) =

=νxPi(k, P)x−λVi(k, P, V)x+λ2

ν Di(k, P, V, γ)+

+ν(u+a(x))Ri(k, P)(u+a(x)),

showing (13) and completing the proof of the proposition.

Fork=T, T 1, . . . ,0 define

P(k) =P(k, P(k+ 1)), P(T) = (LL, . . . , LL), (22)

V(k) =V(k, P(k+ 1), V(k+ 1)), V(T) = (L, . . . , L), (23)

γ(k) =D(k, P(k+ 1), V(k+ 1), γ(k+ 1)), γ(T) = 0.

(24)

Theorem 4. If

(25) Ei(k, V(k+ 1)) ¯Bi(k)R(Ri(k, P(k+ 1)))

for eachk= 0,1, . . . , T1, then the value functionJ(x(k), θ(k), k) is given by J(x(k), θ(k), k) =E

νx(k)Pθ(k)(k)x(k)−λVθ(k)(k)x(k) +λ2

ν γθ(k)(k), (26)

and an optimal control law is given by u(k) =−Rθ(k)(k, P(k+ 1))

Gθ(k)(k, P(k+ 1))x(k) (27)

−λ

Hθ(k)(k, V(k+ 1)) .

Proof. For k = T there is no control to take, and it follows that J(x(T), θ(T), T) =νE(x(T)LLx(T)−λLx(T)), showing (26) from the defi- nition of P(T), V(T) and γ(T) = 0. Suppose from the induction hypothesis

(10)

that (26)-(27) hold fork+ 1. From the Bellman equation, (11) and (13), for x(k) =x,θ(k) =i, we have

J(x, i, k) = inf

u∈Rm

E

J(x(k+ 1), θ(k+ 1), k+ 1)|Fk = (28)

= inf

u∈Rm

νE

x(k+ 1)Pθ(k+1)(k+ 1)x(k+ 1)−

−λVθ(k+1)(k+ 1)x(k+ 1) + λ2

ν γθ(k+1)(k+ 1)|Fk =

=νxPi(k, P(k+ 1))x−λVi(k, P(k+ 1), V(k+ 1))x+λ2

ν Di(k, P, V, γ), with a minimum value reached at u(k) as in (27). Now, (22), (23), (24) and (28) complete the proof.

5. SOLUTION OF PROBLEMS

In this section we solve the three mean-variance problems stated in Sec- tion 3. We assume throughout this section that (25) holds. Let Π

P1 σ2

, Π (P2 ( )), Π (P3 (ν)) and Π (A(λ, ν)) denote, respectively, the set of optimal solutions for problems P1

σ2

, P2 ( ), P3 (ν) and A(λ, ν). We recall the following results proved in [5].

Proposition 5. If u Π (P3 (ν)) and λ= 1 + 2νE(yu(T)), then u Π (A(λ, ν)). On the other hand, ifu∈Π (A(λ, ν))then a necessary condition for u∈Π (P3 (ν))is that λ= 1 + 2νE(yu(T)).

Proposition 6. Suppose that ν≥0 andu∈Π (P3 (ν)).

a) If Var (yu(T)) =σ2 then u∈Π P1

σ2 . b)If E(yu(T)) = then u∈Π (P2 ( )).

We shall next derive some expressions forλandνsuch that the conditions of Propositions 5 and 6 will be verified, yielding a solution of problemsP1

σ2 , P2 ( ) and P3 (ν). For i= 1, . . . , N define

Ki(k) =Ri(k, P(k+ 1))Gi(k, P(k+ 1)), (29)

Ui(k) =Ri(k, P(k+ 1))B¯i(k)Ei(k, V(k+ 1)),

Acli (k) = ¯Ai(k)−B¯i(k)Ki(k), Ci(k) =π(k) ¯Bi(k)Ui(k), A(k) =



p11(k)Acl1(k) . . . pN1(k)AclN(k) ... . . . ... p1N(k)Acl1(k) . . . pN N(k)AclN(k)

,

(11)

V(k) =







 N i=1

pi1(k)Ci(k) ... N

i=1

piN(k)Ci(k)









, I=

I . . . I ,

a=L

I

T−1 =0

A()

µ(0), b= 1 2L

T−1

t=0

I k−1

=t+1

A()

V(t)

,

c= N

i=1

tr(Pi(0)Qi(0)), d= N i=1

Vi(0)µi(0), e= N

i=1

πi(0)γi(0).

We present next an explicit formula for E(yu(T)) and Var(yu(T)) in terms of λ, a, b, c, d, ewhen the optimal control strategy (27) is applied to system (4).

Proposition7. Suppose that the optimal control strategy(27)is applied to system (4). Then

(30) E(yu(T)) =a+λ

νb (31) Var(yu(T)) =c−a2

λ ν

b 2

2(d−a) b +4a

λ

ν 2

1−e

b−b

. Proof. Using the control law (27) in (4), we get

xu(k+1) =

Aclθ(k)(k)+

ε s=1

Aθ(k),s(k)wxs(k)−Bθ(k),s(k)Kθ(k)(k)wsu(k)

xu(k)+

+ λ

B¯θ(k)(k) + ε s=1

Bθ(k),s(k)wus(k)

Uθ(k)(k).

(32)

Defining ziu(k) = xu(k)1{θ(t)=i}, and µi(k) = E(ziu(k)), µ(k) = (µ1(k)· · · µN(k)), from [2] we have

µj(k+ 1) = N

i=1

pij(k)Acli (k)µi(k) + λ

N i=1

pij(k)πi(k) ¯Bi(k)Ui(k), or, in other words,

µ(k+ 1) =A(k)µ(k) + λ 2νV(k).

(33)

(12)

Iterating (33) we get µ(k) =

k−1 =0

A()

µ(0) + λ

k−1

t=0

k−1 =t+1

A()

V(t).

(34)

From (34) we have

E(xu(T)) =E N

i=1

zui(T)

= N

i=1

E(ziu(T)) = (35)

= N i=1

µi(T) =Iµ(T) =I T−1

=0

A()

µ(0) + λ

T−1 t=0

I k−1

=t+1

A()

V(t).

Sinceyu(T) =Lxu(t), it follows that E(yu(T)) =L

I

T−1 =0

A()

µ(0)+ λL

T−1

t=0

I k−1

=t+1

A()

V(t)

=a+λ νb, showing (30). To show (31), notice that from (26) we have

E(νyu(T)2−λyu(T)) =E

νx(0)Pθ(0)(0)x(0)−λVθ(0)(0)x(0)+λ2

ν γθ(0)(0)

= (36)

=ν N

i=1

tr(Pi(0)Qi(0))−λ N

i=1

Vi(0)µi(0)+λ2 ν

N i=1

πi(0)γi(0) =νc−λd+λ2 ν e.

Therefore, it follows from (30) and (36) that νVar(yu(T)) =ν

E(yu(T)2)−E(yu(T))2

=

=νE(yu(T)2)−λE(yu(T)) +λE(yu(T))−νE(yu(T))2 =

=νc−λd+λ2 ν e+

λ−ν

a+λ

νb a+λ νb

, thus showing (31).

Next, we obtain the values of λ and ν such that the conditions in Propositions 5 and 6 hold in order to obtain a solution of problems P3 (ν), P1

σ2

andP2 ( ). First, we determine the value ofλsatisfying the equation λ= 1 + 2νE(yu(T)). From (30) we have

λ= 1 + 2νE(yu(T)) = 1 + 2ν

a+λ νb

,

(13)

hence an optimal strategyu for problem P3 (ν) is given by (27) with

(37) λ= 1 + 2νa

12b .

Next, we determine the value of ν such that E(yu(T)) = . From (30) and (37) we have

=E(yu(T)) =a+

1+2νa 1−2b

ν b, hence

(38) ν= b

(1−2b)−a.

Finally, we determine the value of ν such that Var (yu(T)) = σ2. Define f = c−a2, g = b

2 2(dba) + 4a , h = b

1eb −b

and υ = λν. It follows from (31) that

2−gυ+ (f −σ2) = 0, so thatυ= 2gh ±g

2h

2

fσ2

h . Butυ= λν =

1+2νa 1−2b

ν , so we have

(39) ν = 1

υ(1−2b)2a with the signal inυchosen such that ν >0.

Acknowledgements. This work was partially supported by CNPq (Brazilian Na- tional Research Council), Grant 304866/03-2, CAPES (Brazilian Ministry of Educa- tion Agency), FAPESP (Research Council of the State of S˜ao Paulo), Grant 03/06736- 7, IM-AGIMB, and PRONEX, Grant 015/98.

REFERENCES

[1] S. Chen, X. Li and X.Y. Zhou, Stochastic linear quadratic regulators with indefinite control weight costs.SIAM J. Control Optim.36(1998), 1685–1702.

[2] O.L.V. Costa, M.D. Fragoso and R.P. Marques, Discrete-Time Markov Jump Linear Systems.Springer-Verlag, 2005.

[3] V. Dr˘agan and T. Morozan, The linear quadratic optimization problems for a class of linear stochastic systems with multiplicative white noise and markovian jumping.IEEE Trans. Automat. Control49(2004), 665–675.

[4] Markus Leippold, Fabio Trojani and Paolo Vanini,A geometric approach to multiperiod mean variance optimization of assets and liabilities.J. Econom. Dynamics Control28 (2004), 1079–1113.

[5] Duan Li and Wan-Lung Ng, Optimal dynamic portfolio selection: Multiperiod mean- variance formulation. Math. Finance10(2000), 387–406.

[6] X. Li and X.Y. Zhou,Indefinite stochastic LQ controls with Markovian jumps in a finite time horizon.Comm. Inform. and Systems2(2002), 265–282.

(14)

[7] X. Li, X.Y. Zhou and M. Ait Rami, Indefinite stochastic linear quadratic control with markovian jumps in infinite time horizon.J. Global Optim.27(2003), 149–175.

[8] A. Lim and X.Y. Zhou, Stochastic optimal LQR control with integral quadratic con- straints and indefinite control weights.IEEE Trans. Automat. Control44(1999), 1359–

1369.

[9] Y. Liu, G. Yin, and X.Y. Zhou,Near-optimal controls of random-switching LQ problems with indefinite control weight costs.Automatica41(2005), 1063–1070.

[10] H. Markowitz.Portfolio Selection: Efficient Diversification of Investments.Wiley, New York, 1959.

[11] A.C.M. Ran and H.L. Trentelman,Linear quadratic problems with indefinite cost for discrete time systems.SIAM J. Matrix Anal. Appl.14(1993), 776–797.

[12] A. Saberi, P. Sannuti and B.M. Chen,H2-Optimal Control.Prentice Hall, 1995.

[13] U. C¸ akmak and S. ¨Ozeckici,Portfolio optimization in stochastic markets.Math. Meth- ods Oper. Res.63(2006), 151–168.

[14] G. Yin and X.Y. Zhou, Markowitz’s mean-variance portfolio selection with regime switching: From discrete-time models to their continuous-time limits.IEEE Trans. Au- tomat. Control49(2004), 349–360.

[15] X.Y. Zhou and D. Li,Continuous-time mean-variance portfolio selection: A stochastic LQ framework.Appl. Math. Optim.42(2000), 19–33.

[16] Shu-Shang Zhu, Duan Li and Shou-Yang Wang,Risk control over bankruptcy in dynamic portfolio selection: A generalized mean-variance formulation. IEEE Trans. Automat.

Control49(2004), 447–457.

Received 30 September 2006 Escola Polit´ecnica da

Universidade de S˜aoPaulo Departamento de Engenharia de

Telecomunica¸oes e Controle 05508-900 So Paulo SP, Brazil

oswaldo@lac.usp.br okimura@usp.br

Références

Documents relatifs

Due to the discrete nature of the charge carriers, the current presents some fluctuations (noise) known as shot noise , which we aim to characterize here (not to be confused with

According to this idea, it appears that the optimal linear quadratic control problem of singularly perturbed discrete systems can be solved in terms of completely independent

The class of discrete event dynamic systems involving only synchronization phenomena can be seen as linear time-invariant systems in a particular algebraic structure called ( min , +

In this paper the problem of reduced-order H ∞ filtering for a larger class of stochastic systems than those studied in the papers cited above is considered since the studied

We then provide a verification theorem which gives the optimal strategy for the mean-variance hedging using the solution of the previous system of BSDEs.. Finally, we prove that

We derive the Hamilton-Jacobi-Bellman equation and the associated verification theorem and prove a necessary and a sufficient maximum principles for such problems.. Explicit