• Aucun résultat trouvé

Positive invariance of polyhedrons and comparison of Markov reward models with different state spaces

N/A
N/A
Protected

Academic year: 2021

Partager "Positive invariance of polyhedrons and comparison of Markov reward models with different state spaces"

Copied!
9
0
0

Texte intégral

(1)

HAL Id: hal-00447075

https://hal.archives-ouvertes.fr/hal-00447075

Submitted on 26 Aug 2013

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Positive invariance of polyhedrons and comparison of

Markov reward models with different state spaces

Mourad Ahmane, James Ledoux, Laurent Truffet

To cite this version:

Mourad Ahmane, James Ledoux, Laurent Truffet. Positive invariance of polyhedrons and comparison

of Markov reward models with different state spaces. Lecture notes in control and information sciences,

Springer, 2006, 341, pp.153-160. �hal-00447075�

(2)

Positive invariance of polyhedrons and

comparison of Markov reward models with

different state spaces

Mourad Ahmane1, James Ledoux2, and Laurent Truffet3

1 IRCCyN, 1 rue de la No¨e BP 92101, 44321 Nantes Cedex 3, France.

Mourad.Ahmane@irccyn.ec-nantes.fr

2 INSA de Rennes & IRMAR, 20 avenue des Buttes de Coesmes, CS 14315, 35043

Rennes cedex, France.

James.Ledoux@insa-rennes.fr

3 Ecole des Mines de Nantes & IRCCyN, 4 rue Alfred Kastler, BP 20722, 44307´

Nantes Cedex 3, France. Laurent.Truffet@emn.fr

Abstract. In this paper, we discuss the comparison of expected rewards for discrete-time reward Markov chains with different state spaces. Necessary and suffi-cient conditions for such a comparison are derived. Due to the special nature of the introduced binary relation, a criterion may be formulated in terms of an inclusion of polyhedral sets. Then, algebraic and geometric forms are easily obtained from Haar’s Lemma. Our results allow us to discuss some earlier results on the stochastic comparison of functions of Markov chains.

Key words: Linear discrete-time systems; Invariant sets; Haar’s lemma.

Basic notation • For all m ∈ N, ≤

mis the component-wise ordering of Rm.

• Vectors are column vectors. ()> is the standard transpose operator. • 1k (resp. 0k) denotes the k-dimensional vector with all components equal

to 1 (resp. 0).

• ∀k ≥ 1, the set Sk denotes the set of all k-dimensional stochastic vectors.

• The set of real (resp. non-negative) k × n-matrices is denoted by Mk,n(R)

(resp. Mk,n(R+)). 0m,k is the element of Mm,k(R) with all coefficients

equal to 0. Ik is the identity element of Mk,k(R).

• If G1∈ Mm,n1(R), G2∈ Mm,n2(R), then (G1| G2) ∈ Mm,n1+n2(R).

Partially supported by the MATHSTIC CNRS project entitled Contrˆole de

syst`emes `a ´ev´enements discrets via des techniques de comparaison et de r´eduction de syst`emes stochastiques

(3)

50 Mourad Ahmane, James Ledoux, and Laurent Truffet

• If A, B ∈ Mm,n(R) then A ≤ B denotes the entry-wise comparison of

the matrices A and B.

1 Introduction

A discrete-time autonomous linear state-space model is the basic model in control theory specified by

 x(0) ∈ Rd

, x(n + 1) = Ax(n), n ∈ N ox(n) = Rx(n).

(1) where A ∈ Md,d(R) and R ∈ Ml,d(R). (x(n)) is the sequence of states of

the system and (ox(n)) the sequence of outputs. Such a system is said to be

positive [1] if, for any x(0) ∈ Rd, we have

0d≤dx(0) =⇒ ∀n ∈ N, 0d ≤dx(n) and 0l≤lox(n).

Then, it is easily seen that the autonomous linear state-space model (1) is positive iff 0d,d ≤ A and 0l,d ≤ R. We refer to [1] for a detailed account

for the theory and applications of such models. We only deal here with posi-tive autonomous linear state-space model for which A is a stochastic matrix. Specifically, the Markov reward models are considered.

A Markov reward model is a very general stochastic model. It is used, for instance, in assessing the performability of systems [2] which is a measure combining dependability and performance aspects of a system. Let us recall the basic setup for such a class of models. We consider a E -valued Markov chain (X(n)) whereE is a finite set and A its transition probabilities matrix, i.e. ∀i, j ∈E , Aj,i = P (X(n + 1) = j | X(n) = i). Suppose that (X(n)) is a

Markovian model of some system. Now, a cost or reward ri ≥ 0 is associated

with each visit of (X(n)) to the state i ∈E . Then, the random variable rX(n)is

the instantaneous reward “gained” at time n. The d-dimensional non-negative vector r corresponding to the family of rewards {ri| i ∈E } is often related to

a specific management/maintenance policy. With this simple Markov reward model, we associate the autonomous linear state-space model specified by

(d, A, 1, r>) : x(0) ∈ Sd, x(n + 1) = Ax(n), n ∈ N ox(n) = r>x(n).

(2) The scalar output ox(n) is the expected reward at time n for the vector of

rewards r. Then, we can be interested in comparing two policies which are identified to the two vectors r and r0. We define two autonomous linear state-space models (d, A, 1, r>) and (d, A, 1, r0>) as in (2). We can compare their output sequences through the usual ordering on R. A slightly more general situation is the comparison of two different weightings of the same family of policies. Indeed, let us identify the policies r1, r2, . . . , rl to a l × d-matrix R.

(4)

We consider two vectors of weights : c = (c1, . . . , cl) and c0= (c01, . . . , c0l). The

following autonomous linear state-space model is defined by (d, A, l, R) : x(0) ∈ Sd, x(n + 1) = Ax(n), n ∈ N

ox(n) = Rx(n).

(3) The vector ox(n) is the vector of expected reward “gained” at time n according

to the different policies of “rewarding”. Then, we analyze the output sequence (ox(n)) via

c>ox(n) ≤ c0>ox(n).

The two previous situations is easily adapted to the case where the underlying Markov chains are different, that is we compare two different Markov models. Thus, we can be interested in comparing various models of the same system through their corresponding maintenance policies. In other words, we deal with a decision making problem in a context of multi-criteria and multi-models of a system. Therefore, the general comparison of the output sequences of (multi-dimensional) Markov reward models should be investigated. This paper proposes criteria for two such sequences to be comparable with respect to the binary relation defined, for any (x, y) ∈ Rl

× Rl0, by

x ≤

C,C0

y ⇐⇒ Cx ≤def mC0y, (4)

where C ∈ Mm,l(R) and C0 ∈ Mm,l0(R). This binary relation should be thought of as an abstract version of an integral stochastic order for discrete random variables (e.g. see [3, Chap 2]). Let (ox(n)), (oy(n)) be the output

sequences of the systems (d, A, l, R) and (d0, A0, l0, R0) as defined in (3). We

investigate the conditions under which the assertion: ∀x(0) ∈ Rd

, ∀y(0) ∈ Rd0

x(0) ∈ Sd, y(0) ∈ Sd0, ox(0) ≤

C,C0oy(0) =⇒ ∀n ∈ N, ox(n) ≤C,C0oy(n), (5) is true. In this case, the two Markov reward models (d, A, l, R) and (d0, A0, l0, R0) are said to be (C, C0)-comparable.

The paper is organized as follows. Useful results on the polyhedrons and polyhedral sets inclusion are recalled below. In Section 2, necessary and suffi-cient conditions for (5) to be true are provided. In Section 3, the comparison of functions of Markov chains and related works are discussed. We conclude in Section 4.

The basic materials to our approach is now reported. Inclusion of polyhedrons and invariance sets

The polyhedron associated with P ∈ Mm,k(R) and p ∈ Rmis defined by

(5)

52 Mourad Ahmane, James Ledoux, and Laurent Truffet

Recall that Sk, Sk× Sk0 are polyhedrons and the intersect of two polyhedrons is a polyhedron.

We restate a result proved by Haar [4] that provides an algebraic charac-terization of the inclusion of two polyhedral sets.

Lemma 1 (Haar). Let P ∈ Mm,k(R), Q ∈ Mr,k(R) and p ∈ Rm, q ∈ Rr. If

P(P , p) 6= ∅ then

P(P , p) ⊆ P(Q, q) ⇐⇒ ∃H ∈ Mr,m(R+) : Q = HP and Hp ≤rq.

Now, we recall the connection between the inclusion of polyhedrons and the invariant sets. A reference for such material is [5] for instance.

Definition 1. A subset V of Rk is said to be positively invariant under M ∈

Mm,k(R) if M V ⊆ V, where M V := {M x | x ∈ V}.

The following corollary of Haar’s lemma is our basic tool for deriving our results.

Lemma 2. Let P ∈ Mm,k(R), M ∈ Mk,k(R) and p, q ∈ Rm.

If P(P , p) 6= ∅, then we have

M P(P , p) ⊆ P(P , q) ⇐⇒ P(P , p) ⊆ P(P M , q)

⇐⇒ ∃H ∈ Mm,m(R+) : P M = HP and Hp ≤mq.

2 Comparison of multi-dimensional Markov reward

models

LetE and F be two finite sets with respective cardinal d and d0. Introduce

two Markov chains (X(n)) and (Y (n)) with respective state spacesE , F and transition probabilities matrices A and A0. The probability distribution of the random variable X(n) (resp. Y (n)) is denoted by x(n) (resp. y(n)). Let us consider the corresponding autonomous linear state-space models (d, A, l, R) and (d0, A0, l0, R0) as in (3). The l × d (resp. l0× d0) matrix R (resp. R0)

is called the rewards matrix associated with (X(n)) (resp. (Y (n))). The au-tonomous linear state-space model (d, A, l, R) is called a multi-dimensional Markov reward model associated with the Markov chain (X(n)) (or the sto-chastic matrix A) and the rewards matrix R. The output ox(n) (resp. oy(n))

is the multi-dimensional expected reward at time n. Note that the sequence (ox(n)) does not have a linear dynamics in general.

Define RSd := {Rx, x ∈ Sd} and R0Sd0 := {R0y, y ∈ Sd0}. Let us introduce the following condition for C ∈ Mm,l(R) and C0∈ Mm,l0(R):

H : {(x, y) ∈ RSd× R0Sd0 | Cx ≤mC0y} 6= ∅. (7) Now, the main result of this section is proved by applying Lemma 2 to

(6)

P :=           CR −C0R0 −Id 0d,d0 0d0,d −Id0 1T d 0Td0 −1T d 0 T d0 0Td 1Td0 0Td −1T d0           , p :=       0m+d+d0 1 −1 1 −1      

and the following stochastic matrix as matrix M : M (A, A0) := 

A 0d,d0 0d0,d A0



. We omit the details for saving space.

Theorem 1. Assume H. The multi-dimensional Markov reward models (d, A, l, R) and (d0, A0, l0, R0) are said to be (C, C0)-comparable iff any of the following conditions is satisfied:

1. P (CR | −C0R0), 0m ∩Sd×Sd0 ⊆ P (CRA | −C0R0A0), 0m ∩Sd×Sd0. 2. Geometric criterion. The polyhedron P (CR | −C0R0), 0m ∩ Sd× Sd0

is positively invariant under the matrix M (A, A0).

3. Algebraic criterion. There exist H ∈ Mm,m(R+) and vectors u, v ∈

Rm such that

−u1>d ≤ HCR − CRA, HC0R0− C0R0A0≤ v1>d0, u + v ≤m0m.

The monotonicity property of a multi-dimensional Markov reward model is now introduced from (5).

Definition 2. A multi-dimensional Markov reward model (d, A, l, R) is said to be (C, C0)-monotone when we have, for any x1(0), x2(0) ∈ Rd,

x1(0), x2(0) ∈ Sd, ox1(0) ≤

C,C0ox2(0) =⇒ ∀n ∈ N, ox1(n) ≤C,C0 ox2(n). The following criteria for the (C, C0)-monotonicity of the Markov reward model (d, A, l, R) is deduced from Theorem 1 with A = A0.

Corollary 1. Assume H. The multi-dimensional Markov reward model (d, A, l, R) is (C, C0)-monotone iff any of the following conditions is satisfied: 1. P (CR | −C0R), 0m ∩ Sd× Sd0 ⊆ P (CRA | −C0RA), 0m ∩ Sd× Sd0. 2. Geometric criterion. The polyhedron P CR | −C0R), 0m ∩ Sd× Sd0

is positively invariant under the matrix M (A, A).

3. Algebraic criterion. There exist H ∈ Mm,m(R+) and vectors u, v ∈

Rm such that

−u1>d ≤ HCR − CRA, HC0R − C0RA ≤ v1>d, u + v ≤m0m.

These results provide comparison conditions of the marginal distributions of functions of Markov chains. Thus, earlier results on stochastic comparison of functions of Markov chains is discussed below.

(7)

54 Mourad Ahmane, James Ledoux, and Laurent Truffet

3 Functions of Markov chains

Let (X(n)) and (Y (n)) be two Markov chains with respective state spacesE := {e1, . . . , ed},F := {f1, . . . fd0} and transition probabilities matrices A, A0. Let G , G0be the finite sets {g

1, . . . , gl} and {g10, . . . , gl00} respectively. Consider two maps ϕ :E → G and ψ : F → G0. The matrices R and R0 are specified as follows

R(j, k) := 1 if ϕ(ek) = gj and R(j, k) := 0 otherwise

R0(j, k) := 1 if ψ(fk) = g0j and R

0(j, k) := 0 otherwise. (8)

The processes (ϕ(X(n))) and (ψ(Y (n))) are said to be (deterministic) func-tions of Markov chains. The sequence of outputs (ox(n)) does not have a linear

dynamics in general. Conditions on the matrix A under which (ox(n)) also

follows a linear dynamics may be found in [6]. Since R is stochastic, note that the condition H has the form P (C | −C0), 0m 6= ∅.

The stochastic comparison of one-dimensional distributions of functions of Markov chains is well-known. This kind of problem is motivated for instance by performance evaluation and/or optimization of telecommunications networks (e.g. see [7]), by reliability assessment (e.g. see [8]). It also appears when we look for Markovian bounds on the output stream of a queuing network (e.g. see[9]). Some contributions on the comparison of the one-dimensional distributions of functions of Markov chains are the following. We refer to the original reference for the full details.

Doisy’s work.

Doisy studied the comparison of functions of Markov chains using a coupling technique [8]. The setting is as follows

– G = G0 (and thus l = l0) and there exists a partial order ≺ on G ;

– C = C0 = UGst ∈ Mm,l(R), where UGst denotes the matrix associated

with the so-called strong ordering based on the partial order ≺ defined onG . It means that each row of matrix UGst corresponds to the indicator function of an ≺-increasing subset of G . Recall that a ≺-increasing set γ ⊆G is defined by: γ := {g ∈ G : ∀g0 ∈G (g ≺ g0 ⇒ g0 ∈ γ)}. Let us

note that when ≺ is a total order then UGst is the l × l matrix defined by: UGst := [1{g≺g0}]g,g0∈G where 1{·} is the indicator function of the set {·} (see also Kijima [10, Chap 3 and references therein]).

– two maps ϕ :E → G and ψ : F → G with associated matrices R and R0 defined in (8).

Doisy introduced the following definition of the (ϕ, ψ)-comparison of the two stochastic matrices A and A0 [8, Def 1, Section 3]:

A ≤ϕ,ψ A0 ⇔ ei∈E , fj ∈F , ϕ(ei) ≺ ψ(fj) ⇒ RAδi ≤ UGst,UGst

(8)

where (δi)i=1,...,d and (δ0j)j=1,...,d0 denote the canonical basis of Rd and Rd 0

, respectively. The strongest connected result of Doisy is reformulated as follows [8, Prop 2 and Th 4, Section 3].

Property 1. Let (d, A, l, R) and (d0, A0, l, R0) be two multi-dimensional Markov reward models as defined in (3). Then, A ≤ϕ,ψ A0 iff (d, A, l, R)

and (d0, A0, l, R0) are (UGst, UGst)-comparable.

The “only if” part was algebraically proved using the definition of the (UGst, UGst)-comparability of (d, A, l, R) and (d0, A0, l, R0). Doisy’s proof of the “if part” is based on the construction of a suitable coupling of the underlying Markov chains. An algebraic proof of the ‘if part” is provided from Theorem 1-(3). Indeed, we know the following criterion for the (UGst, UGst)-comparison to hold: there exist H ∈ Mm,m(R+) and u, v ∈ Rmsuch that

−u1>d ≤HUGstR − UGstRA (10a)

HUGstR0− UGstR0A0 ≤ v1>d0 (10b)

u + v ≤m0m. (10c)

We must prove that properties (10a-10c) implies that A ≤ϕ,ψA0. First, note

that for any ei∈E and fj ∈F , we have

ϕ(ei) ≺ ψ(fj) ⇐⇒ Rδi ≤ UGst,UGst R0δ0j ⇐⇒ UGstRδi ≤mUGstR 0δ0 j. Since H ∈ Mm,m(R+), we have HUGstRδi≤mHUGstR 0δ0 j. (11)

Second, suppose that the conditions (10a-10c) hold. We find that UGstRAδi ≤mHUGstRδi+ u from (10a)

≤mHUGstRδi− v from (10c) ≤mHUGstR 0δ0 j− v from (11) ≤mUGstR 0A0δ0 j from (10b).

Thus, we find from (9) that A ≤ϕ,ψA0. The proof is complete.

The comparison of the finite-dimensional distributions of functions of Markov chains (also called in this context coupling condition) may be achieved from condition (9) due to a special property of the strong ordering.

Abu-Amsha and Vincent’s work.

These authors considered the following setting in [11]: – E is assumed to be totally ordered finite set;

(9)

56 Mourad Ahmane, James Ledoux, and Laurent Truffet

– G = G0=F (thus l = l0= d0);

– R0= Id0, that is ψ = id;

– C = C0 (that implies m = d = d0) is assumed to be invertible.

The authors founded the following sufficient condition for the Markov reward models (d, A, l, R) and (d, A0, l, Id) to be comparable:

CRA ≤ CA0R and 0d,d≤ CA0C−1.

The conditions above are equivalent to : ∃H ∈ Md,d(R+) such that

CRA ≤ CA0R and CA0= HC.

It is easily checked that the algebraic criterion reported in Theorem 1 is ful-filled. However, it is also clear from Theorem 1 that these conditions are not necessary.

4 Conclusion

Criteria for the (C, C0)-comparison of multi-dimensional Markov rewards are provided using Haar’s lemma. In particular, this applied to the comparison of (deterministic) functions of Markov chains. Note that the probabilistic func-tions of Markov chains or the Hidden Markov models can also be considered. We do not go into further details here.

References

1. L. Farina and S. Rinaldi. Positive Linear Systems. Pure and Applid Mathemat-ics. Wiley, 2000.

2. J. F. Meyer. On evaluating the performability of degradable computing systems. IEEE Trans. Comput., 29:3720–7313, 1980.

3. A. Muller and D. Stoyan. Comparison Methods for Stochastic Models and Risks. J. Wiley & Sons, 2002.

4. A. Haar. Uber lineare ungleichungen (1918). In Reprinted in: A. Haar, Gesam-melte Arbeiten. Akademi Kiado, 1959.

5. A. A. ten Dam and J. W. Nieuwenhuis. A linear programming algorithm for invariant polyhedral sets of discrete-time linear systems. Systems Control Lett., 25:337–341, 1995.

6. J. Ledoux. Linear dynamics for the state vector of Markov chain functions. Adv. in Appl. Probab., 36(4):1198–1211, 2004.

7. N. Pekergin. Stochastic performance bounds by state space reduction. Perfor-mance Evaluation, 36-37:1–17, 1999.

8. M. Doisy. A coupling technique for comparison of functions of Markov processes. J. Appl. Math. Decis. Sci., 4:39–64, 2000.

9. L. Truffet. Geometrical bounds on an output stream of a queue in atm switch: Application to the dimensioning problem. In ATM Networks: Performance Mod-elling and Analysis, Vol. II. Chapman-Hall, 1996.

10. M. Kijima. Markov Processes for Stochastic Modeling. Chapman-Hall, 1997. 11. O. Abu-Amsha and J.-M. Vincent. An algorithm to bound functionals of Markov

Références

Documents relatifs

Finite mixture models (e.g. Fr uhwirth-Schnatter, 2006) assume the state space is finite with no temporal dependence in the ¨ hidden state process (e.g. the latent states are

And if one’s view is that something plays the role of S’s motivating reason just in case (a) S takes it to speak in favor of Φing, (b) is motivated by it to Φ, (c) is motivated by

Any collection of two-cycles, triangles, and at most one path with the same orientation as the the triangles; with in total T − 1 edges, is a decomposition of the state graph of

Les canaux ASIC3 pourraient être impliqués dans ces douleurs et les modifications sensorielles liées à l’alcalinisation mais ils pourraient également avoir pour

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

This paper introduces a new model for the SoC estimation using a Switching Markov State-Space Model (SMSSM): the battery behavior is described by a set of potential linear state-

We describe a quasi-Monte Carlo method for the simulation of discrete time Markov chains with continuous multi-dimensional state space.. The method simulates copies of the chain

If the transition matrix of the Markov chain (X n ) is ϕ-observable, then the Rogers-Pitman condition is essentially a criterion for weak lumpability prop- erty to hold.... It has