[ieee 2006 proceeding of the thrity-eighth southeastern symposium on system theory - cookeville, tn,...

5
A Measure-Theoretic Proof of the Markov Property for Hybrid Systems with Markovian Inputs Arturo Tejada, Oscar R. Gonz´ alez and W. Steven Gray Abstract— The behavior of a general hybrid system in discrete time can be represented by a non-linear difference equation x(k + 1) = F k (x(k), θ(k)), where θ(k) is assumed to be a finite state Markov chain. An important step in the stability analysis of these systems is to establish the Markov property of (x(k), θ(k)). There are, however, no complete proofs of this property which are simple to understand. This paper aims to correct this problem by presenting a complete and explicit proof, which uses only basic measure-theoretical concepts. I. I NTRODUCTION The last decade has witnessed a steady increase in the number of applications that combine logical decisions with continuous or discrete time dynamics. These so-called hybrid systems have been used in a broad range of contexts [1], [2]. Continuous time hybrid systems have been studied at a very high level of abstraction. These studies have led to models, like Piecewise Deterministic Markov Processes, whose stability properties are well understood [3]. The tools used to study these systems are, however, sophisticated and are not easily adaptable for applications. In discrete time, Markovian jump linear systems (MJLS) are, arguably, the most studied sub-class of hybrid systems. Their simple dynamics are given by x(k + 1) = A θ(k) x(k), where θ(k) is a finite-state Markov chain. The long-term behavior of these systems, i.e. their stability, has been studied by employing the well known fact that the state of a MJLS, (x(k), θ(k)), constitutes a Markov chain. This was possible because the long-term behavior of a Markov chain can be studied by analyzing its transition kernel. The dynamics of more general hybrid systems are complex because they are usually non-linear in nature. A good model for their behavior is x(k + 1) = F k (x(k), θ(k)), where F k are measurable functions. As in the previous case, the stability analysis of these systems utilizes the Markov property of the process (x(k), θ(k)). This fact is already well established in the literature [4], [5]. However, the typical proofs make use of either heuristic arguments or very sophisticated tools. Thus, the goal of this paper is to present a complete and explicit proof of the Markov property of (x(k), θ(k)) using basic measure-theoretical concepts and the properties of the random variables x(k) and θ(k). The authors are affiliated with the Department of Electrical and Computer Engineering, Old Dominion University, Norfolk, Virginia 23529-0246, USA. [email protected], {gonzalez,gray}@ece.odu.edu The rest of the paper is organized as follows. Section II states and proves the main result. The proof relies on two supporting results, which are proved in Sections III and IV. Finally, Section V provides the conclusions. II. MAIN RESULT A. Preliminaries The proofs that follow make frequent use of measurable and Borel functions. Recall that a function f : defined between two measurable spaces, (Ω, F ) and (Ω , F ), is called an F /F -measurable function if for every B F , the set {ω Ω: f (ω) B}∈ F . In particular, a B(R n )/B(R k )-measurable function is called a Borel function (B(R n ) denotes the Borel algebra over R n , and has the property B(R n )= B(R) B(R n1 ), for any n> 1, where denotes the direct product). When the fields are obvious from the context, f will be called, simply, measurable (or Borel). Random variables (F /B(R)-measurable functions), ran- dom vectors (F /B(R n )-measurable functions), and ran- dom elements (F /F -measurable functions) are denoted by lower case boldface letters x, y, and z. Throughout the paper, it is assumed that every random element (including random variables and vectors) and stochastic process is defined over the same underlying probability space (Ω, F , P). Finally, let G be any sub σ-algebra of F . A random element x is called G -measurable if for every B B(R n ), the set {ω Ω: x(ω) B}∈ G . As usual, the smallest σ-algebra with respect to which x is measurable is denoted by σ(x). B. Main Result This subsection states and proves our main theorem. Our main result is as follows. Theorem 2.1: Consider the system x(k + 1) = F k (x(k), θ(k)), x(0) = x 0 , (1) where x(k) R n , θ(k) is a discrete time Markov chain in R m , x 0 is an integrable random variable independent of θ(k), and F k : R n × R m R n are measurable functions for every k 0. Under these conditions, the process (x(k), θ(k)) is a Markov chain. The proof of Theorem 2.1 relies on the following three supporting results. Their proofs are relegated to Sections III and IV to simplify the presentation of the proof of Theorem 2.1. Proceedings of the 38th Southeastern Symposium on System Theory Tennessee Technological University Cookeville, TN, USA, March 5-7, 2006 MC4.1 0-7803-9457-7/06/$20.00 ©2006 IEEE. 328

Upload: ws

Post on 28-Feb-2017

215 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: [IEEE 2006 Proceeding of the Thrity-Eighth Southeastern Symposium on System Theory - Cookeville, TN, USA (March 5, 2006)] 2006 Proceeding of the Thrity-Eighth Southeastern Symposium

A Measure-Theoretic Proof of the Markov Property forHybrid Systems with Markovian Inputs

Arturo Tejada, Oscar R. Gonzalez and W. Steven Gray

Abstract— The behavior of a general hybrid system indiscrete time can be represented by a non-linear differenceequation x(k + 1) = Fk(x(k), θ(k)), where θ(k) is assumed tobe a finite state Markov chain. An important step in the stabilityanalysis of these systems is to establish the Markov property of(x(k), θ(k)). There are, however, no complete proofs of thisproperty which are simple to understand. This paper aimsto correct this problem by presenting a complete and explicitproof, which uses only basic measure-theoretical concepts.

I. INTRODUCTION

The last decade has witnessed a steady increase in thenumber of applications that combine logical decisions withcontinuous or discrete time dynamics. These so-called hybridsystems have been used in a broad range of contexts [1],[2]. Continuous time hybrid systems have been studied ata very high level of abstraction. These studies have ledto models, like Piecewise Deterministic Markov Processes,whose stability properties are well understood [3]. The toolsused to study these systems are, however, sophisticatedand are not easily adaptable for applications. In discretetime, Markovian jump linear systems (MJLS) are, arguably,the most studied sub-class of hybrid systems. Their simpledynamics are given by

x(k + 1) = Aθ(k)x(k),

where θ(k) is a finite-state Markov chain. The long-termbehavior of these systems, i.e. their stability, has been studiedby employing the well known fact that the state of a MJLS,(x(k), θ(k)), constitutes a Markov chain. This was possiblebecause the long-term behavior of a Markov chain can bestudied by analyzing its transition kernel.

The dynamics of more general hybrid systems are complexbecause they are usually non-linear in nature. A good modelfor their behavior is

x(k + 1) = Fk(x(k), θ(k)),

where Fk are measurable functions. As in the previous case,the stability analysis of these systems utilizes the Markovproperty of the process (x(k), θ(k)). This fact is alreadywell established in the literature [4], [5]. However, thetypical proofs make use of either heuristic arguments or verysophisticated tools. Thus, the goal of this paper is to presenta complete and explicit proof of the Markov property of(x(k), θ(k)) using basic measure-theoretical concepts andthe properties of the random variables x(k) and θ(k).

The authors are affiliated with the Department of Electrical and ComputerEngineering, Old Dominion University, Norfolk, Virginia 23529-0246, [email protected], gonzalez,[email protected]

The rest of the paper is organized as follows. Section IIstates and proves the main result. The proof relies on twosupporting results, which are proved in Sections III and IV.Finally, Section V provides the conclusions.

II. MAIN RESULT

A. Preliminaries

The proofs that follow make frequent use of measurableand Borel functions. Recall that a function f : Ω →

Ω′ defined between two measurable spaces, (Ω, F ) and(Ω′, F ′), is called an F/F ′-measurable function if forevery B ∈ F ′, the set ω ∈ Ω : f(ω) ∈ B ∈ F . Inparticular, a B(Rn)/B(Rk)-measurable function is calleda Borel function (B(Rn) denotes the Borel algebra overR

n, and has the property B(Rn) = B(R) ⊗ B(Rn−1), forany n > 1, where ⊗ denotes the direct product). When thefields are obvious from the context, f will be called, simply,measurable (or Borel).

Random variables (F/B(R)-measurable functions), ran-dom vectors (F/B(Rn)-measurable functions), and ran-dom elements (F/F ′′-measurable functions) are denoted bylower case boldface letters x, y, and z. Throughout the paper,it is assumed that every random element (including randomvariables and vectors) and stochastic process is defined overthe same underlying probability space (Ω, F , P).

Finally, let G be any sub σ-algebra of F . A randomelement x is called G -measurable if for every B ∈ B(Rn),the set ω ∈ Ω : x(ω) ∈ B ∈ G . As usual, the smallestσ-algebra with respect to which x is measurable is denotedby σ(x).

B. Main Result

This subsection states and proves our main theorem. Ourmain result is as follows.

Theorem 2.1: Consider the system

x(k + 1) = Fk(x(k), θ(k)), x(0) = x0, (1)

where x(k) ∈ Rn, θ(k) is a discrete time Markov chain

in Rm, x0 is an integrable random variable independent of

θ(k), and Fk : Rn× R

m→ R

n are measurable functionsfor every k ≥ 0. Under these conditions, the process(x(k), θ(k)) is a Markov chain.

The proof of Theorem 2.1 relies on the following threesupporting results. Their proofs are relegated to Sections IIIand IV to simplify the presentation of the proof of Theorem2.1.

Proceedings of the38th Southeastern Symposium on System TheoryTennessee Technological UniversityCookeville, TN, USA, March 5-7, 2006

MC4.1

0-7803-9457-7/06/$20.00 ©2006 IEEE. 328

Page 2: [IEEE 2006 Proceeding of the Thrity-Eighth Southeastern Symposium on System Theory - Cookeville, TN, USA (March 5, 2006)] 2006 Proceeding of the Thrity-Eighth Southeastern Symposium

Theorem 2.2: Consider system (1) under the conditions ofTheorem 2.1, and let f : R

m→ R be any bounded Borel

function. Then, it follows that

Ef(θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(θ(k + 1))|x(k), θ(k).1 (2)Theorem 2.3: [6] Let H be a linear space of bounded

real-valued functions defined on a set Φ. Assume that

(i) H contains the constant functions.(ii) If fn is a sequence in H and fn → f uniformly, then

f ∈ H.(iii) If fn is a sequence in H, 0 ≤ fn ≤ M , and fn is

a monotone sequence, then limn→∞ fn ∈ H.(iv) H has a subset C with the property: If c1 and c2 belong

to C, then their product, c1c2, also belongs to C.

Then H contains every bounded function g : Φ → R whichis measurable with respect to the σ-algebra generated by thesets: φ ∈ Φ : c(φ) ∈ B, c ∈ C, B ∈ B(R).

Theorem 2.4: Consider system (1) under the conditions ofTheorem 2.1, and let Φ = R

n× R

m× R

m. If H is the setof all bounded Borel functions f : Φ → R such that

Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k), (3)

then H satisfies conditions (i)-(iv) of Theorem 2.3.The fundamental step in the proof of Theorem 2.1 is to

establish that the set of bounded measurable functions whichsatisfy (3) contains all the bounded measurable functionson Φ. This is proven by combining Theorems 2.3 and 2.4.Theorem 2.2 is used to verify property (iv) in Theorem 2.4.

Proof of Theorem 2.1: First recall from [7, page 564] thata random process yk is a Markov chain if and only if forevery bounded, real-valued Borel function, h, it follows thatEh(yn+1)|yn, . . . ,y0 = Eh(yn+1)|yn. In particular, toshow that (x(k), θ(k)) is a Markov chain, it is sufficient toprove that for every bounded Borel function h : R

n×R

m→

R

Eh(x(k + 1), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Eh(x(k + 1), θ(k + 1))|x(k), θ(k).

or, equivalently,

Eh(Fk(x(k), θ(k)), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Eh(Fk(x(k), θ(k)), θ(k + 1))|x(k), θ(k).

Let H be the set of all the bounded Borel functions f : Φ →

R with Φ = Rn× R

m× R

m that satisfy the expression

Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k). (4)

It follows from Theorem 2.4 that H satisfies all the con-ditions of Theorem 2.3. Thus, H contains all the boundedfunctions g : Φ → R which are measurable with respect to

1Equalities and inequalities between random variables are taken in thealmost everywhere sense.

A , the σ-algebra generated by the sets φ ∈ Φ : c(φ) ∈ B,c ∈ C, B ∈ B(R).

As will be shown in the proof of Theorem 2.4, the set Cis composed of all separable bounded functions c : Φ → R

of the form c = c1c2, where c1 : Rn× R

m→ R and c2 :

Rm

→ R are bounded Borel functions.Now, let A ∈ A and observe that A must be of the form

A =⋃i,j

φ ∈ Φ : ci(φ) ∈ Bij,

for some ci ∈ C and Bij∈ B(R). But the functions ci are

Borel, thus φ ∈ Φ : ci(φ) ∈ Bij ∈ B(Rn) ⊗ B(Rm) ⊗

B(Rm), and consequently, A ∈ B(Rn)⊗B(Rm)⊗B(Rm).Moreover, since this is true for every A ∈ A , it follows thatA ⊆ B(Rn)⊗B(Rm)⊗B(Rm). Conversely, note that forevery B ∈ B(Rn) ⊗ B(Rm) ⊗ B(Rm), B = φ ∈ Φ :1B(φ) ∈ [0.5, 1.5) ∈ A , where 1B() is the indicatorfunction of the set B. This implies that A ⊇ B(Rn) ⊗B(Rm) ⊗ B(Rm), so A = B(Rn) ⊗ B(Rm) ⊗ B(Rm).

Hence, H contains every bounded Borel function of theform g : Φ → R, and each of these functions satisfies (4).In particular, let h : R

n× R

m→ R be any bounded Borel

function and define g(φ) = h(Fk(γ), λ), φ = (γ, λ) ∈ Φ.Clearly, g ∈ H, so in (4)

Eg(x(k), θ(k), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Eg(x(k), θ(k), θ(k + 1))|x(k), θ(k), (5)

which implies that

Eh(Fk(x(k), θ(k)), θ(k+1))|x(k), θ(k), . . . ,x(0), θ(0)

= Eh(Fk(x(k), θ(k)), θ(k + 1))|x(k), θ(k),

or

Eh(x(k + 1), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Eh(x(k + 1), θ(k + 1))|x(k), θ(k).

Since this is true for every bounded Borel function h :R

n× R

m→ R, we conclude that (x(k), θ(k)) is a Markov

chain.

III. PROOF OF THEOREM 2.2

The proof of Theorem 2.2 relies on three facts to be provenabout system (1):

Fact (i): The σ-algebra generated by the joint process(x(k), θ(k)) can be simplified as follows:

σ(x(k), θ(k), . . . ,x(0), θ(0))=σ(θ(k), . . . ,θ(0), x(0)).

Fact (ii): For any integrable random variable ξ,

Eξ|x(k), θ(k), . . . ,x(0), θ(0)

= Eξ|θ(k), . . . ,θ(0), x(0).

Fact (iii): Eθ(k + 1)|x(k), θ(k) = Eθ(k + 1)|θ(k).

Fact (i) is the core of the proof of Theorem 2.2. It estab-lishes that all the probabilistic information is contained inx(0) and θ(i), i = 0, . . . , k. That is, no new information can

329

Page 3: [IEEE 2006 Proceeding of the Thrity-Eighth Southeastern Symposium on System Theory - Cookeville, TN, USA (March 5, 2006)] 2006 Proceeding of the Thrity-Eighth Southeastern Symposium

be obtained by examining the variables x(i), i = 0, . . . , k,when x(0) and θ(i), i = 0, . . . , k, are available. Fact (i) isproven in Subsection A (Theorem 3.4). Its proof makes useof the binary operator (defined below) and standard resultson the structure of σ-algebras induced by random variables.Facts (ii) and (iii) are proven in Subsection B (Theorems 3.6and 3.7), after briefly reviewing some properties of expectedvalues and the basic definitions of independence. Finally, theproof of Theorem 2.2 is developed in Subsection C.

A. Some Properties of σ-Algebras

1) Structure:Theorem 3.1: [8, Theorem 20.1] Let x = (x1, . . . ,xn)

be a vector of random variables xi.

(i) The σ-algebra σ(x) = σ(x1, . . . ,xn) consists exactlyof the sets ω ∈ Ω : x(ω) ∈ B for B ∈ B(Rn).

(ii) The random variable y is σ(x)-measurable if and onlyif there exists a measurable function f : R

n→ R such

that y(ω) = f(x1(ω), . . . ,xn(ω)) for all ω ∈ Ω.2) Combination:Definition 3.1: Let C and D be two classes of subsets of

Ω. The binary operation combines the classes C and D toform a new class of subsets of Ω as follows:

C D E ⊂ Ω : E = C ∩ D; C ∈ C, D ∈ D.It follows readily from its definition that is commutative

and associative. It also follows that if C is an algebra (orσ-algebra) in Ω then C C = C . The operation can beused to combine σ-algebras induced by random variables asdescribed below.

Lemma 3.1: For any two random variables x1, x2,σ(x1) σ(x2) = σ(x1, x2).Proof : Let H ∈ σ(x1)σ(x2). Then H = H1 ∩H2, whereH1 ∈ σ(x1) and H2 ∈ σ(x2). That is, H1 = ω ∈ Ω :x1(ω) ∈ B1 and H2 = ω ∈ Ω : x2(ω) ∈ B2 for someB1, B2 ∈ B(R). Thus, it follows that

H = ω ∈ Ω : (x1(ω), x2(ω)) ∈ B1 × B2 ∈ σ(x1, x2),

since B1 × B2 ∈ B(R2). But H is arbitrary, so σ(x1)

σ(x2) ⊆ σ(x1, x2). Conversely, let H = ω ∈ Ω :(x1(ω), x2(ω)) ∈ B1 × B2 for some B1 × B2 ∈ B(R2)and let H1 = ω ∈ Ω : x1(ω) ∈ B1 and H2 = ω ∈ Ω :x2(ω) ∈ B2. Clearly, H = H1 ∩ H2 ∈ σ(x1) σ(x2).Again, H is arbitrary, which implies that σ(x1) σ(x2) ⊇σ(x1, x2). Thus, σ(x1) σ(x2) = σ(x1, x2).

Remark 3.1: Let x1, x2, and x3 be random variables andlet y1 = (y11, . . . ,y1n) and y2 = (y21, . . . ,y2n) be vectorsof random variables.

(i) Observe from Theorem 3.1, the definition of andLemma 3.1 that σ(x1, x2, x3) = σ(x1) σ(x2, x3) =σ(x1, x2) σ(x2, x3).

(ii) A similar derivation shows that σ(y1, y2) =σ(y11, . . . ,y1n, y21, . . . ,y2n) = σ(y1) σ(y2).

Both arguments can be extended inductively to any finitenumber of random variables or random vectors. This remarkplays an important role in the following collection of results

3) Simplification of σ-algebras:Theorem 3.2: Let x = (x1, . . . ,xn) be a vector of

random variables, and let f : Rn

→ R be a measurablefunction. If y = f(x) then σ(y, x) = σ(x).Proof : First, recall from Theorem 3.1 that σ(y, x) is com-posed of the sets ω ∈ Ω : (y(ω), x(ω)) ∈ B, B ∈

B(Rn+1). Fix any B ∈ B(Rn+1) and observe that B =B1 × B2, for some B1 ∈ B(R) and B2 ∈ B(Rn), sinceB(Rn+1) = B(R) ⊗ B(Rn). Then, it follows that

ω ∈ Ω :(y(ω), x(ω)) ∈ B

= ω ∈ Ω : y(ω) ∈ B1 ∩ ω ∈ Ω : x(ω) ∈ B2

= ω ∈ Ω : x(ω) ∈ f−1(B1) ∩ B2 ∈ σ(x).

The last inclusion above follows from the measurability of f ,which implies that f−1(B1) ∩ B2 ∈ B(Rn). Hence, ω ∈

Ω : (y(ω), x(ω)) ∈ B ∈ σ(x) and, since B is arbitrary,this implies that σ(y, x) ⊆ σ(x).

Conversely, note that σ(x) is composed of the sets ω ∈

Ω : x(ω) ∈ B, B ∈ B(Rn). Also note that y is a randomvariable (Theorem 3.1) and that

Ω = ω ∈ Ω : y(ω) ∈ Range(f) = ω ∈ Ω : y(ω) ∈ R.

Thus, fix any B ∈ B(Rn) and observe that

ω ∈ Ω : x(ω) ∈ B

= ω ∈ Ω : x(ω) ∈ B ∩ ω ∈ Ω : y(ω) ∈ R

= ω ∈ Ω : (y(ω), x(ω)) ∈ R × B.

Now, R × B ∈ B(Rn+1) so ω ∈ Ω : (y(ω), x(ω)) ∈

R × B ∈ σ(y, x). This and the arbitrariness of B yieldsσ(x) ⊆ σ(y, x). Hence, σ(x) = σ(y, x).

This result can be extended to the case when y is a vector-valued function of x as follows.

Theorem 3.3: Let x = (x1, . . . ,xn) be vector of randomvariables and F : R

n→ R

m be a measurable function. Ify = F (x) then σ(y, x) = σ(x).

4) Proof of Fact (i):Theorem 3.4: Consider system (1) under the conditions of

Theorem 2.1. It follows that

σ(x(k), θ(k), . . . ,x(0), θ(0)) = σ(θ(k), . . . ,θ(0), x(0)).Proof : First note that x(k) = Fk(x(k − 1), θ(k − 1)), fork ≥ 1. Thus, by Theorem 3.3, σ(x(k), x(k−1), θ(k−1)) =σ(x(k − 1), θ(k − 1)). It follows that

σ(x(k), θ(k), . . . ,x(0), θ(0))

= σ(θ(k)) σ(x(k), x(k − 1), θ(k − 1))

σ(x(k − 2), θ(k − 2), . . . ,x(0), θ(0))

= σ(θ(k)) σ(x(k − 1), θ(k − 1))

σ(x(k − 2), θ(k − 2), . . . ,x(0), θ(0))

= σ(θ(k)) σ(x(k − 1), θ(k − 1), . . . ,x(0), θ(0)).

Repeating the operation above k − 1 times yields

σ(x(k), θ(k), . . . ,x(0), θ(0))

= σ(θ(k)) . . . σ(θ(1)) σ(θ(0), x(0))

= σ(θ(k), . . . ,θ(0), x(0)).

330

Page 4: [IEEE 2006 Proceeding of the Thrity-Eighth Southeastern Symposium on System Theory - Cookeville, TN, USA (March 5, 2006)] 2006 Proceeding of the Thrity-Eighth Southeastern Symposium

B. On Independence and Expected Values

1) Independence:Definition 3.2: Let Ci, i = 1, . . . , n be a set of sub σ-

algebras of F . Also, let fi : (Ω, F ) → (Ωi, Fi), i =1, . . . , n be a set random elements.

(i) The σ-algebras C1, . . . ,Cn are independent if for everychoice of sets Ci ∈ Ci it follows that P(B1∩· · ·∩Bn) =P(B1) · · · · ·P(Bn) for all the possible 2n combinationsformed by taking Bi = Ci or Bi = Ω, i = 1, . . . , n.

(ii) The random elements f1, . . . , fn are independent ifσ(f1), . . . , σ(fn) are independent.

(iii) A random variable y and a stochastic process X =xk are independent if for any finite integer n ≥ 1 andevery sequence of integers 0 ≤ t1 < · · · < tn < ∞ itfollows that σ(y) and σ(xt1 , . . . ,xtn

) are independent.Remark 3.2: Let C and D be independent sub σ-algebras

of F . Clearly, if x and y are, respectively, C -measurableand D-measurable random elements, then x and y areindependent.

2) Expected Values:Lemma 3.2: The following are standard results from the

literature (cf. [7]).(i) Let x = (x1, . . . ,xn) be a vector of random variables.

If x is independent of a random variable y, then so iseach xi. Furthermore, Exy = ExEy.

(ii) Let F1 and F2 be σ-algebras such that F1 ⊆ F2. Ifx is an integrable random variable then

EEx|F1|F2 = EEx|F2|F1 = Ex|F1.

(iii) For any F -measurable random element x, Ex|F =x.

The following result is an extension of a basic result forrandom variables. The theorem and its corollary are offeredwithout proof for space limitations.

Theorem 3.5: Let x be a vector of random variables andlet y, z be two random elements such that z is independentof (x, y), i.e., σ(z) is independent from σ(z), σ(y), andσ(x, y). Then it follows that Ex|yz = Ex|y.

Corollary 3.1: Let θ(k) be a discrete time Markov chainin R

m and x0 be a second order random vector. If x0 is in-dependent from θ(k), then Eθ(k +1)|θ(k), . . . ,θ(0), x0=Eθ(k + 1)|θ(k), . . . ,θ(0).

3) Proofs of Facts (ii) and (iii):Theorem 3.6: Consider system (1) under the condition of

Theorem 2.1. Then, for any integrable random variable ξ,

Eξ|x(k), θ(k), . . . ,x(0), θ(0)=Eξ|θ(k), . . . ,θ(0), x(0).Proof : Observe that the integrability of ξ ensuresEξ|x(k), θ(k), . . . ,x(0), θ(0) is well defined, and∫

B

Eξ|x(k), θ(k), . . . ,x(0), θ(0) dP =

∫B

ξ dP,

for every B ∈ σ(x(k), θ(k), . . . ,x(0), θ(0)). Also, recallfrom Theorem 3.4 that σ(x(k), θ(k), . . . ,x(0), θ(0)) =σ(θ(k), . . . ,θ(0), x(0)), which implies that∫

B

ξ dP =

∫B

Eξ|θ(k), . . . ,θ(0), x0 dP

for every B ∈ σ(x(k), θ(k), . . . ,x(0), θ(0)), and the resultfollows.

Theorem 3.7: Consider system (1) under the conditionsof Theorem 2.1. Then Eθ(k + 1)|x(k), θ(k) = Eθ(k +1)|θ(k).Proof : First, note that σ(x(k), θ(k)) is comprised of setsof the form ω ∈ Ω : (x(k)(ω), θ(k)(ω)) ∈ H, H ∈

B(Rn) ⊗ B(Rm). Fix any H ∈ B(Rn) ⊗ B(Rm) andobserve that

ω ∈ Ω : (x(k)(ω), θ(k)(ω)) ∈ H

= ω : x(k)(ω) ∈ H1 ∩ ω : θ(k)(ω) ∈ H2, (6)

for some H1 ∈ B(Rn) and H2 ∈ B(Rm). But (1) yields

x(k) = Fk−1(x(k − 1), θ(k − 1))

= Fk−1(Fk−2(x(k − 2), θ(k − 2)), θ(k − 1))

...

= G(θ(k − 1), . . . ,θ(0), x(0)),

where G is the indicated composition of the functionsF0, . . . , Fk−1. Thus, (6) can be expressed as

ω ∈ Ω : (x(k)(ω), θ(k)(ω)) ∈ H

=ω : (θ(k)(ω), . . . ,θ(0)(ω), x(0)(ω)) ∈ H2 × G−1(H1).

Since G is measurable, it follows that

H2 × G−1(H1) ∈ B(Rm) ⊗ · · · ⊗ B(Rm)︸ ︷︷ ︸k+1 copies

⊗B(Rn),

which in turn implies that ω ∈ Ω : (x(k)(ω), θ(k)(ω)) ∈

H ∈ σ(θ(k), . . . ,θ(0), x(0)). But since H is arbitrary, itfollows that σ(x(k), θ(k)) ⊆ σ(θ(k), . . . ,θ(0), x0). Thisconclusion, Lemma 3.2, Corollary 3.1, and the Markovproperty of θ(k) yield

Eθ(k + 1)|x(k), θ(k)

= EEθ(k + 1)|θ(k), . . . ,θ(0), x(0)|x(k), θ(k)

= EEθ(k + 1)|θ(k), . . . ,θ(0)|x(k), θ(k)

= EEθ(k + 1)|θ(k)|x(k), θ(k).

Finally, note that Eθ(k + 1)|θ(k) is σ(θ(k))-measurable.Also note that σ(θ(k)) ⊆ σ(x(k), θ(k)), so Eθ(k +1)|θ(k) is also σ(x(k), θ(k))-measurable. Thus, Lemma3.2 yields EEθ(k + 1)|θ(k)|x(k), θ(k) = Eθ(k +1)|θ(k), and the result follows.

C. Proof of Theorem 2.2

Proof : First, observe that f(θ(k + 1)) is an integrablerandom variable, so Theorem 3.6 yields

Ef(θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(θ(k + 1))|θ(k), . . . ,θ(0), x(0). (7)

Next, note that f(θ(k + 1)) is σ(θ(k + 1))-measurable(Theorem 3.1), so σ(f(θ(k+1)), θ(k), . . . ,θ(0)) ⊆ σ(θ(k+

331

Page 5: [IEEE 2006 Proceeding of the Thrity-Eighth Southeastern Symposium on System Theory - Cookeville, TN, USA (March 5, 2006)] 2006 Proceeding of the Thrity-Eighth Southeastern Symposium

1), . . . ,θ(0)). Furthermore, this and the hypothesis of The-orem 2.2 imply that x(0) is independent of (f(θ(k +1)), θ(k), . . . ,θ(0)). Hence, (7), Theorem 3.5, and theMarkov nature of θ(k) yield

Ef(θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(θ(k + 1))|θ(k). (8)

Next, recall that σ(x(k), θ(k)) ⊆ σ(θ(k), . . . ,θ(0), x0).Thus, the argument in the proof of Theorem 3.7 yields

Ef(θ(k + 1))|x(k), θ(k) = Ef(θ(k + 1))|θ(k).

This last equation and (8)confirms identity (2).

IV. PROOF OF THEOREM 2.4

Proof : Observe that H is a linear space since any linearcombination of elements from H yields a bounded Borelfunction which satisfies (3) as a consequence of the linearityof the conditional expected value.

Now, to show (i) in Theorem 2.3, suppose f(φ) = a forall φ ∈ Φ and some a ∈ R. Note that f(x(k), θ(k), θ(k+1))is a constant random variable. Also, recall that for any σ-algebra G , Ef |G = a (a.e.) [7, page 215], which in turnimplies that f satisfies (3). Since this is true for any a ∈ R,then H contains all the constant functions.

To show (ii), suppose that fn ∈ H converges uniformlyto f , and let Ln be finite constants such that |fn(φ)| < Ln forall φ ∈ Φ. Next, fix ε > 0 and observe from the hypothesisthat there exists N(ε) such that |fn(φ)−f(φ)| < ε for all n ≥

N(ε) and for all φ ∈ Φ. In particular, |fN(ε)(φ)− f(φ)| < εimplies that |f(φ)| < LN(ε)+ε for every φ ∈ Φ. Note that theuniform convergence of fn implies that it also convergespointwise almost everywhere to f . This in turn shows f ismeasurable [9, Corollary 2.2.4]. Hence f is a bounded Borelfunction.

Set L = LN(ε) + ε and observe that |fn(φ) − f(φ)| <ε, n ≥ N(ε) implies that |fn(φ)| < L + ε for allφ ∈ Φ and all n ≥ N(ε). Next, define the functiong : Φ → R as g(φ) = maxL1, . . . , LN(ε)−1, L + ε, forall φ ∈ Φ, and let yn = fn(x(k), θ(k), θ(k + 1)) andy = g(x(k), θ(k), θ(k + 1)). Observe that yn convergespointwise to f(x(k), θ(k), θ(k+1)), that |yn| < y, and thatEy = maxL1, . . . , LN(ε)−1, L + ε < ∞. Under theseconditions, it follows that for any σ-algebra G , Eyn|G

converges to Ef(x(k), θ(k), θ(k + 1))|G [7, page 218].Thus,

limn→∞

Eyn|x(k), θ(k), . . . ,x(0), θ(0)

= Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0).

But fn ∈ H. Thus, (3) implies that

limn→∞

Eyn|x(k), θ(k), . . . ,x(0), θ(0)

= limn→∞

Eyn|x(k), θ(k)

= Ef(x(k), θ(k), θ(k + 1))|x(k), θ(k).

Hence, f ∈ H, and (ii) follows.To show (iii), let fn be a monotone sequence of func-

tions in H such that 0 ≤ fn ≤ M < ∞, and note that

fn converges pointwise to a function f , which is in turna bounded Borel function. Thus (iii) follows by using thesame argument as in (ii).

To show (iv), consider the set C composed of all theseparable functions c : Φ → R of the form c(γ, λ) =c1(γ)c2(λ), φ = (γ, λ), where c1 : R

n× R

m→ R and

c2 : Rm

→ R are bounded Borel functions. To prove (iv),it is necessary to prove that if c, d ∈ C, their product, cd, isalso in C, and that C ⊆ H.

Let c, d ∈ C and observe that c(γ, λ)d(γ, λ) =(c1(γ)d1(γ))(c2(λ)d2(λ)). But (c1(γ)d1(γ)) and(c2(λ)d2(λ)) are, respectively, real bounded Borel functionsin R

n× R

m and Rm. Thus cd ∈ C.

Finally, to see that C ∈ H, recall from [7, page 216] thatif ξ and η are random variables such that E|ξ| < ∞,E|ξη| < ∞, and η is G -measurable, then Eξη|G =ηEξ|G . Thus let c(γ, λ) = c1(γ)c2(λ) ∈ C and observefrom Theorem 2.2 that

Ec((x(k), θ(k)), θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

=c1(x(k), θ(k))Ec2(θ(k + 1))|x(k), θ(k), . . . ,x(0), θ(0)

=c1(x(k), θ(k))Ec2(θ(k + 1))|x(k), θ(k)

=Ec1(x(k), θ(k))c2(θ(k + 1))|x(k), θ(k)

=Ec((x(k), θ(k)), θ(k + 1))|x(k), θ(k).

Thus, c satisfies (3). Since this is true for every c ∈ C, thenC ∈ H and (iv) follows. This completes the proof.

V. CONCLUSIONS

A complete proof of the Markovian property of the stateof a general hybrid system, (x(k), θ(k)), has been presented.The proof is based on first principles and gives insight onthe underlying mechanism that produces the Markovianessof (x(k), θ(k)).

ACKNOWLEDGEMENTS

This research was supported by the NASA Langley ResearchCenter under grants NNL04AA03A and NCC-1-03026.

REFERENCES

[1] W. S. Gray, O. R. Gonzalez, and M. Dogan, “Stability analysis of digitallinear flight controllers subject to electromagnetic disturbances,” IEEETrans. Aerosp. Electron. Syst., no. 4, pp. 1204–1218, October 2000.

[2] H. Blom and J. Lygeros, “Hybridge final project re-port,” 2005. [Online]. Available: http://hosted.nlr.nl/public/hosted-sites/hybridge/documents.html

[3] M. H. A. Davis, Markov Models and Optimization. New York:Chapman & Hall/CRC, 1993.

[4] R. Srichander and K. Walker, “Stochastic stability analysis forcontinuous-time fault tolerant control systems,” Int. J. Control, vol. 57,no. 2, pp. 433–452, 1993.

[5] L. Arnold and W. Kleinmann, “Qualitative theory of stochastic sys-tems,” in Probabilistic Analysis and Related Topics, Vol 3, A. T.Bharucha-Reid, Ed. New York: Academic Press, 1982, pp. 1–79.

[6] T. Morozan, “On the stability of stochastic discrete systems,” in Controltheory and Topics in Functional Analysis: Lectures Presented at anInternational Seminar Course at Trieste from 11 Sept. To 29 Nov. 1974,International Atomic Energy Agency, Vienna, 1976, pp. 225–254.

[7] A. N. Shiryaev, Probability, Second Edition. New York: Springer,1995.

[8] P. Billingsley, Probability and Measure, Third Edition. New York:John Wiley & Sons, 1995.

[9] A. Friedman, Foundations of Modern Analysis. New York: DoverPublications Inc., 1982.

332