Markov jump linear systems - IMT School for Advanced...
Transcript of Markov jump linear systems - IMT School for Advanced...
Markov jump linear systems
Pantelis Sopasakis
IMT institute for Advanced Studies Lucca
March 13, 2017
Outline
1. Quick probability theory brush-up
2. Markov jump linear systems
3. Expectation and covariance dynamics
4. Mean square stability (MSS)
5. Conditions for MSS
6. Almost sure convergence
7. Stabilisation of MJLS
1 / 97
I. Probability theory
1. Random variables
2. Conditional Probability
3. Expectation and Conditional Expectation
4. Filtration
2 / 97
Random variable
Let (Ω,F ,P) be a probability space and (E, E) be a measurable space. Ameasurable function X : Ω→ E is called a random variable.
3 / 97
Expected value
The expected value E[X] of a random variable X is defined as
E[X] =
∫ΩXdP =
∫ΩX(ω) P(dω).
4 / 97
Conditional probability
On a probability space (Ω,F ,P) and for A,B ∈ F with P(B) > 0 wedefine the conditional probability of A given B as
P (A | B) =P (A ∩B)
P (B).
Let X be a random variable; then we define the conditional probabilityP(A | X) to be a random variable with
P(A | X)(ω) = P(A | X = X(ω)).
5 / 97
Conditional expectation
Given a random variables X on a prob. space (Ω,F ,P) and a H ∈ F , wedefine the conditional expecation E[X | H] to be
E[X | H] =
∫ΩX(ω) P(dω | H).
6 / 97
Conditional expectation
Given a random variables X on a prob. space (Ω,F ,P) and asub-σ-algebra H ⊆ F we call conditional expecation of X given H as ameasurable function E[X | H] such that∫
HE[X | H]dP =
∫HXdP, ∀H ∈ H.
The conditional expectation of X wrt another random variable Y is therandom variable
E[X | Y = y] =
∫ΩX(ω) P(dω | Y = y) = E[X | σ(Y )].
7 / 97
Stochastic process
A (discrete) stochastic process is a sequence of random variables from aprobability space (Ω,F ,P) to a measurable space (E, E), that is Xkk∈N.
8 / 97
Filtration
Formally, a filtration over a prob. space (Ω,F ,P) is a sequence ofsub-σ-algebras of F , Fkk∈N so that
F0 ⊆ F1 ⊆ . . . ⊆ F .
I The space (Ω,F , Fkk,P) is called a filtered prob. space.
I A filtration Fkk∈N represents the flow of information as theexperiment goes on in time,
I Fk encodes the available information up to time k,
9 / 97
Filtration
Formally, a filtration over a prob. space (Ω,F ,P) is a sequence ofsub-σ-algebras of F , Fkk∈N so that
F0 ⊆ F1 ⊆ . . . ⊆ F .
I The space (Ω,F , Fkk,P) is called a filtered prob. space.
I A filtration Fkk∈N represents the flow of information as theexperiment goes on in time,
I Fk encodes the available information up to time k,
9 / 97
Filtration
Formally, a filtration over a prob. space (Ω,F ,P) is a sequence ofsub-σ-algebras of F , Fkk∈N so that
F0 ⊆ F1 ⊆ . . . ⊆ F .
I The space (Ω,F , Fkk,P) is called a filtered prob. space.
I A filtration Fkk∈N represents the flow of information as theexperiment goes on in time,
I Fk encodes the available information up to time k,
9 / 97
Filtration
Formally, a filtration over a prob. space (Ω,F ,P) is a sequence ofsub-σ-algebras of F , Fkk∈N so that
F0 ⊆ F1 ⊆ . . . ⊆ F .
I The space (Ω,F , Fkk,P) is called a filtered prob. space.
I A filtration Fkk∈N represents the flow of information as theexperiment goes on in time,
I Fk encodes the available information up to time k,
9 / 97
Adapted processes
We say that a stochastic process Xkk∈N on a filtered probability space(Ω,F , Fkk,P) is adapted to the filtration Fkk if every Xk isFk-measurable. Equivalently:
E[Xk | Fk] = Xk.
14 / 97
Natural filtration
Given Xkk we can construct a filtration so that the given randomprocess is adapted to it. This is called the natural filtration and it isFk = σ(∪σ(Xj)j≤k).
15 / 97
Useful properties
I Markov’s inequality. If X is a positive, integrable random variable,
P[X ≥ α] ≤ E[X]
α.
I Chebychev’s inequality. If X a finite-variance random variable
P[|X − E[X]| ≥ α] ≤ Var[X]
α2.
I Jensen’s inequality. For ϕ convex and X integrable r.v.
ϕ(E[X]) ≤ E[ϕ(X)].
16 / 97
The Borel-Cantelli lemma
Let (Ω,F ,P) be a probability space and Eii a sequence of events. Wedenote by lim supiEi the following
lim supi→∞
Ei =⋂k≥1
⋃j≥k
Ej
Notice that
I this a set-theoretic entity and does assume/require a topology
I it is different from the limit-superior used to define thePainleve-Kuratowski limit
I x ∈ lim supi→∞Ei iff x is in every ∪j≥kEjI lim supi→∞Ei ∈ F
17 / 97
The Borel-Cantelli lemma
Let (Ω,F ,P) be a probability space and Eii a sequence of events. If
∞∑i=1
P[Ei] <∞,
thenP[lim sup
iEi] = 0.
18 / 97
Further reading
Excellent references:
1. S.R.S.Varadhan, Probability theory, Lecture notes, NY Univ., 2000.Available online at https://goo.gl/ygbqa3.
2. Erhan Cınlar, Probability and stochastics, Springer, 2011.
19 / 97
Summary of first section
We revised the notions
1. Probability and measure spaces, probability, conditional probability
2. Expectation and Conditional expectation
3. Stochastic process
4. Filtration and adaptation to a filtration
20 / 97
Markov processes
A Markov process is a stochastic process1 so that
P [w(k) = ω|w(0), w(1), . . . , w(k − 1)] = P [w(k) = ω|w(k − 1)] .
We assume our Markov processes are:
1. finite: w(k) is finite valued w(k) ∈ N = 1, . . . , N2. time-homogeneous: P [w(k) = j|w(k − 1) = i] = pij .
1We call stochastic process a sequence of random variables w(k) over a probabilityspace (Ω,F ,P).
22 / 97
Finite Markov processes
A
E
0.3
0.7
0.6
0.4
The transition matrix is
P =
[0.3 0.70.4 0.6
]with
N = 1, 2
23 / 97
Markov jump linear systems
A (autonomous) Markov jump linear system is a dynamical system of theform
x(k + 1) = Γθ(k)x(k),
where θ(k)k is a finite time-homogeneous Markov process, x(0) = x0
and θ(0) = θ0 follows an initial distribution with P[θ(0) = ωi] = vi fori = 1, . . . , N .
24 / 97
MJLS applications
Examples
1. Power systems [Li et al. 2007; Ugrinovskii, 2005]
2. Satellite control [Meskin & Khorasani,2009]
3. Flight control [Gray, Gonzalez & Dogan, 2000]
4. Air traffic mgmt [Zhou, 2011]
5. Solar thermal receiver [Sworder & Rogers, 1983]
6. Cell growth [Horowitz et al., 2010]
7. Samuelson’s macroeconomic model [Costa et al., 1999]
8. Printer over network with random time delays [Patrinos et al., 2011]
26 / 97
MJLS quirks
Assume that some or all Γi are unstable. How will the systemtrajectories look like? Example:
Γ1 =
[2 −1
0.1 0.1
],Γ2 =
[0.2 1−0.1 2
],
and
P =
[0.1 0.90.9 0.1
].
Take x(0) = (1, 1)′ and P [w(0) = 1] = 0.5.
27 / 97
Unstable + Unstable = Stable?!
Stable trajectories of a MJLS with unstable subsystems.
2 4 6 8 10 12 14 16 18 20
x1
-100
-50
0
50
time
2 4 6 8 10 12 14 16 18 20
x2
0
50
100
150
28 / 97
MJLS quirks continued
Assume that Γi are strictly Hurwitz for all i = 1, . . . , N (all stable) – howwill the MJLS Trajectories look like? Will they all converge to zero?Example:
Γ1 =
[−0.5 2−0.5 0.5
],Γ2 =
[0.5 01 0
],
for which ρ(Γ1) = 0.866 and ρ(Γ2) = 0.5, and
P =
[0.5 0.50.5 0.5
].
Take x(0) = (1, 1)′ and P [w(0) = 1] = 0.5.
29 / 97
Stable + Stable = Unstable?!
While the expected value E[‖x(k)‖] converges to 0, the covariancediverges to infinity!
2 4 6 8 10 12 14
x1
-20
-10
0
10
20
30
time
2 4 6 8 10 12 14
x2
-10
0
10
20
30
30 / 97
Is E[x(k)]→ 0 enough?
“Averages are funny things. Ask the statistician who drowned in a lakeof average depth 3 centimetres.” ∼ J. Barrow, 100 things you didn’t knowyou didn’t know.
31 / 97
End of second section
1. We learnt what a Markov process is
2. and the class of Markov jump linear systems
3. We then listed application of Markov jump systems
4. We realised there’s a lot to learn regarding MJLS...
32 / 97
Questions
1. How do we model the evolution of a MJLS?
2. What definition(s) of stability apply to MJLS?
3. How do we control such a system?
34 / 97
Some definitions
For a set F ∈ F we define its characteristic function χF : Ω→ 0, 1 as
χF (ω) :=
1, if ω ∈ F0, otherwise
Then, notice that
E[x(k)] =
N∑i=1
E[x(k)χθ(k)=i]
and
E[x(k)x(k)′] =
N∑i=1
E[x(k)x(k)′χθ(k)=i]
We are going to describe the dynamics of E[x(k)] and E[x(k)x(k)′].
35 / 97
Some useful Hilbert spaces
Spaces IRn and IRn×m are Hibert spaces with inner products
〈x, y〉 := x′y, and 〈X,Y 〉 := tr(X ′Y )
We define the Hilbert space Hn,m of N -tuples of m-by-n matrices, i.e.,H = (H1, . . . ,HN ) with Hi ∈ IRm×n with the following inner product:
〈H,V 〉 :=
N∑i=1
〈Hi, Vi〉
On Hn,m we define the norms
‖H‖1 :=
N∑i=1
‖Hi‖2, and ‖H‖2 :=
N∑i=1
〈Vi, Vi〉︸ ︷︷ ︸‖Vi‖f
12
36 / 97
Some useful Hilbert spaces
We denote by Cn = L2(ω,F ,P) the space of IRn-valued random variablesy with inner product 〈y, z〉 = E[y′z] so that ‖y‖ <∞.
We call `2(Cn) the space of sequences s = s(k)k with s(k) ∈ Cn and2
‖s‖`2(Cn) :=∑k∈N‖s(k)‖2Cn <∞.
2Note that ‖s(k)‖2Cn = 〈s(k), s(k)〉Cn = E[‖s(k)‖2].
37 / 97
Some useful Hilbert spaces
Now let (Ω,F , Fkk∈N,P) be the aforementioned probability space witha filtration Fkk. We define the space Cn ⊆ `2(Cn) if for every randomprocess r ∈ Cn, r(k) is Fk-measurable, i.e., r is adapted to the filtrationFkk .
38 / 97
MJLS dynamics
We are now going to describe the dynamics of E[x(k)] and E[x(k)x(k)′].Define
qi(k) = E[x(k)χθ(k)=i
]And defining q(k) := (q1(k), . . . , q2(k)) ∈ H1,n we have
µ(k) := E[x(k)] =
N∑i=1
qi(k).
Define alsoQi(k) := E
[x(k)x(k)′χθ(k)=i
],
and let Q(k) := (Q1(k), . . . , QN (k)) ∈ Hn,n, then
Σ(k) := E[x(k)x(k)′
]=
N∑i=1
Qi(k).
39 / 97
MJLS dynamics
We are now going to describe the dynamics of E[x(k)] and E[x(k)x(k)′].Define
qi(k) = E[x(k)χθ(k)=i
]And defining q(k) := (q1(k), . . . , q2(k)) ∈ H1,n we have
µ(k) := E[x(k)] =
N∑i=1
qi(k).
Define alsoQi(k) := E
[x(k)x(k)′χθ(k)=i
],
and let Q(k) := (Q1(k), . . . , QN (k)) ∈ Hn,n, then
Σ(k) := E[x(k)x(k)′
]=
N∑i=1
Qi(k).
39 / 97
MJLS dynamics
The mean and covariance dynamics are described by
qj(k + 1) =
N∑i=1
pijΓiqi(k)
and
Qj(k + 1) =
N∑i=1
pijΓiQi(k)Γ′i.
In what follows we will define linear operators B : q(k) 7→ q(k + 1) andT : Q(k) 7→ Q(k + 1).
40 / 97
A bound on ‖x(k)‖2
We can show that3
‖x(k)‖2 ≤ n‖Q(k)‖1,
where notice that the left hand side norm is the norm of Cn, whereas theright hand side is the 1-norm in Hn,n. What happens as Q(k)→ 0?
3In Cn = L2(Ω,F ,P, IRn) we defined the inner product 〈x, y〉 = E[x′y], so theinduced squared norm is ‖x‖2 = 〈x, x〉 = E[x′x] = E[‖x‖2].
41 / 97
Operators T and L = T ∗
We have Q(k + 1) = T [Q(k)] where T : Hn → Hn is the linear operatorT [Q] := (T1[Q], . . . , TN [Q]), with
Tj [Q] :=
N∑i=1
pijΓiQiΓ′i.
The adjoint of this operator is an operator L[Q] := (L1[Q], . . . ,LN [Q]),with
Li[Q] := ΓiEi[Q]Γi,
where Ei[Q] =∑N
i=j pijQj4.
4Use the definition, i.e., show that 〈T [Q], S〉 = 〈Q,L[S]〉 for all Q,S ∈ Hn. It isleft as an exercise to prove that T ∗ = L.
42 / 97
Operator B as a matrix
We have q(k + 1) = Bq(k), where B : IRNn → IRNn is a linear operator5
and is given by
B[q] = (P ′ ⊗ In) blkdiag(Γ1, . . . ,ΓN ) · q.
If ρ(B) < 1 then q(k)→ 0 and µ(k)→ 0, however, this does not implyQ(k)→ 0 or Σ(k)→ 0.
5Spaces IRNn and H1,n are homeomorphic (can be thought of as the same space).
43 / 97
Operator T as a matrix
For a matrix A ∈ IRm×n, A = [A1 ··· An ] we define
vec(A) := [A′1 ··· A′n ]′ ∈ IRmn.
For H ∈ Hn,m with H = (H1, . . . ,HN ), Hi ∈ IRm×n
vec(H) := [ vec(H1)′ ··· vec(Hn)′ ]′ ∈ IRNmn.
We are looking for a matrix T ∈ IRNn2×Nn2so that
vec(T [Q]) = T vec(Q)
We may verify that
T = (P ′ ⊗ In2) blkdiag(Γ′i ⊗ Γii).
44 / 97
If T is stable, then B is stable
We’ve seen that q(k)→ 0 does not imply Q(k)→ 0. However, theconverse holds true: If T is stable, then B is stable too!
ρ(T ) < 1⇒ ρ(B) < 1.
45 / 97
End of third section
1. We described the dynamics of E[x(k)] and E[x(k)x(k)′]
2. We introduced the Hilbert spaces Hn,m, Cn, `2(Cn) and Cn
3. We defined q(k) and Q(k) to study the dynamics of E[x(k)] andE[x(k)x(k)′]
4. We introduced the operators B and T using which we can study thestability properties of the dynamics of q(k) and Q(k).
46 / 97
Mean square stability
The MJLS x(k + 1) = Γθ(k)x(k) – where θ(k)k∈N is a Markov process –is mean square stable (MSS) if there is a µ and a Σ so that for anyinitial state x(0) = x0 and initial distribution of θ(0) it is
1. µ(k)→ µ,
2. Σ(k)→ Σ.
In our case, µ = 0 and Σ = 0 and #2 implies #1.
48 / 97
Why MSS?
1. It’s easy to test for
2. It implies stability of the expected state dynamics
3. It implies convergence in probability, i.e., for all ε > 0,
P[‖x(k)‖ ≥ ε]→ 0,
and further
4. It implies almost sure convergence, i.e.,
P[‖x(k)‖ → 0] = 1.
49 / 97
MSS ⇒ Convergence of E [x(k)]
MSS ⇒ Convergence of E [x(k)].
Already clear from the fact that ρ(T ) < 1⇒ ρ(B) < 1. Easy way to showit by Jensen’s inequality:
0 ≤ E [‖x(k)‖] ≤ E[‖x(k)‖2
] 12 .
BUT: Convergence of E [x(k)] ; MSS (Exercise).
50 / 97
MSS ⇔ T is stable
If we know T (or T ) we can tell whether our MJLS is stable.
If T is stable ⇔ MJLS is MSS.
Proof.Indeed, we have Q(k + 1) = T Q(k) and Q(k) = T kQ(0), so
Σ(k) =N∑i=1
Qi(k) =N∑i=0
T ki Qi(0)
Since T is stable, T k[Q]→ 0, thus Σ(k)→ 0.
The converse is very easy to show.
51 / 97
MSS ⇔ T is stable
If we know T (or T ) we can tell whether our MJLS is stable.
If T is stable ⇔ MJLS is MSS.
Proof.Indeed, we have Q(k + 1) = T Q(k) and Q(k) = T kQ(0), so
Σ(k) =
N∑i=1
Qi(k) =
N∑i=0
T ki Qi(0)
Since T is stable, T k[Q]→ 0, thus Σ(k)→ 0.
The converse is very easy to show.
51 / 97
MSS ⇔ MSES
MSS systems are mean square exponentially stable, i.e., there is a β ≥ 1and ζ ∈ (0, 1) so that
E[‖x(k)‖2
]≤ βζk‖x0‖2, ∀k ∈ N
Proof.To prove that an MSS system is also MSES we need the following resultfrom analysis: Let T be a linear operator with ρ(T ) < 1. Then there areβ ≥ 1 and ζ ∈ (0, 1) so that
‖T k‖ ≤ βζk,
where on the LHS we may have any operator norm. The rest is left as anexercise.
52 / 97
Unstable + Unstable = Stable?
Two unstable modes:
Γ1 =
[2 −1
0.1 0.1
],Γ2 =
[0.2 1−0.1 2
],
with ρ(Γ1) = 1.9458 and ρ(Γ2) = 1.9426 and
P =
[0.1 0.90.9 0.1
].
but ρ(T ) = ρ(T ) = 0.6798 < 1.
53 / 97
Stable + Stable = Unstable?
Two stable modes:
Γ1 =
[−0.5 2−0.5 0.5
],Γ2 =
[0.5 01 0
],
for which ρ(Γ1) = 0.866 and ρ(Γ2) = 0.5, and
P =
[0.5 0.50.5 0.5
].
but ρ(T ) = ρ(T ) = 1.0165 > 1.
54 / 97
Stochastic stability
A MJLS is said to be stochastically stable (SS) if for all x(0) and initialdistributions for θ(0) it is
∞∑k=0
E[‖x(k)‖2
]<∞.
This is equivalent to the random process x = x(k)k being in Cn.
55 / 97
MSS ⇔ SS
A MJLS is MSS iff it is SS6.
Proof.Since E
[‖x(k)‖2
]is a sequence of `2, it converges E
[‖x(k)‖2
]→ 0. But
we have
0 ≤ E[‖x(k)‖2
]= E
[tr(x(k)x(k)′)
]= trE
[(x(k)x(k)′)
]= trQ(k)
= Σ(k),
so Σ(k)→ 0 for any x(0) and θ(0).
6This is not true for infinite Markov jump linear systems.
56 / 97
Positive (semi)definiteness
An E ∈ IRn×n is positive semidefinite (PSD) if x′Ax ≥ 0, ∀x ∈ IRn; wedenote A ≥ 0. E is positive definite (PD) if it is PSD and x′Ax = 0 onlywhen x = 0. Space of PSD (PD) matrices: Sn+ (Sn++).
It is A < B whenever B −A > 0.
Hn+: the subset of Hn so that H ∈ Hn
+ if Hi ∈ Sn+ for all i.
57 / 97
Lyapunov-like stability conditions
1. A MJLS is MSS iff for every S ∈ Hn+, S > 0, there exists a uniqueV ∈ Hn+, V > 0 such that
V − T (V ) = S.
2. A MJLS is MSS iff for some V ∈ Hn+, V > 0 it is
T (V ) < V.
Notice that V − T (V ) = S is equivalent to (id−T )V = S, which usingvon Neuman’s identity yields V =
∑∞k=0 T k [S].
58 / 97
Lyapunov-like stability conditions
A MJLS is MSS iff for every S ∈ Hn+, S > 0, there exists a uniqueV ∈ Hn+, V > 0 such that V − T (V ) = S.
Proof.Sketch: we define the following system
Y (k + 1) = L(Y (k)); Y (0) ∈ Hn+.
We need to prove that this is stable using the following Lyapunov function:
Φ(Y ) = 〈V, Y 〉
Then T = L∗ will also be stable.
59 / 97
MSS conditions using L
The first condition can also be written as
V = L(V ) + S.
and now we have
V =
∞∑k=0
Lk [S]
60 / 97
MSS conditions using V and J
We have defined the operators T and L as
Tj [Q] :=
N∑i=1
pijΓiQiΓ′i, and Li[Q] := ΓiEi[Q]Γi.
We now define the (simpler) operators V and J 7 as
Vj [Q] :=
N∑i=1
pijΓjQiΓ′j , and Ji[Q] :=
N∑j=1
Γ′jQjΓj .
Then, the MSS conditions hold also if we replace T with either V or J .
7It is V∗ = J .
61 / 97
Outlook...
These stability conditions are useful:
1. For controller design using linear matrix inequalities (LMIs)
2. To devise stabilizability and detectability tests
3. For MJLS with additive uncertainty
4. Similar conditions apply to nonlinear Markovian switching systems
62 / 97
Understanding MSS
Let x(k)k be the response of a MSS MJLS – all x(k) are randomvariables, i.e., x(k)(ω). Let θ(k)k be the corresponding sequence ofrandom parameters. Can there be a realisation θ?k so that thecorresponding state trajectory
x(k;x(0), θ(0), . . . , θ(k − 1))→∞ as k →∞ ?
Short answer: Yes! (But the probability of running into one is zero...)
63 / 97
Understanding MSS
Let x(k)k be the response of a MSS MJLS – all x(k) are randomvariables, i.e., x(k)(ω). Let θ(k)k be the corresponding sequence ofrandom parameters. Can there be a realisation θ?k so that thecorresponding state trajectory
x(k;x(0), θ(0), . . . , θ(k − 1))→∞ as k →∞ ?
Short answer: Yes! (But the probability of running into one is zero...)
63 / 97
End of fourth section
Summary:
1. MSS is a “more meaningful” (practical) stability notion than MS
2. MSS: Both µ(k) and Σ(k) converge
3. MSS ⇔ T is stable
4. MSS ⇔ MSES
5. MSS ⇔ SS
6. MSS ⇔ T V < V , and V − T V = S ∈ Hn+
64 / 97
Zero probability for non converging x(k)
MSS ⇒ The probability that we find a realisation of x(k) that does notconverge is zero.
When is this probability zero?
If this probability is zero, do we have MSS?
66 / 97
Almost sure convergence
Under what conditions do we have x(k)→ 0 with probability one?
We say that x(k)k converges almost surely (or with probability 1) to 0if8
P(limk‖x(k)‖ = 0) = 1.
For a MJLS, MSS implies almost sure convergence – the converse is nottrue. Almost sure convergence is weaker than MSS9.
8This probability is over the space of all random processes x(k)k which satisfythe MJLS dynamics.
9This type of convergence is not induced by some metric and does not depend onany topology.
67 / 97
Almost sure convergence
MSS ⇒ Almost sure convergence
Proof.We have E
[‖x(k)‖2
]≤ βζk‖x0‖2 for all k ∈ N, so
∑∞k=0 E
[‖x(k)‖2
]is
finite and the rest follows from the Borel-Cantelli10 lemma using theChebychev-Markov inequality (Exercise).
10From the Borel-Cantelli lemma we have that z(k) → 0 almost surely (w.p.1)whenever
∑k∈N P [‖z(k)‖ ≥ ε] <∞ for all ε > 0.
68 / 97
ASC ; MSS
ASC does not imply MSS! Example with x ∈ IR, N = 2:
Γ1 = 2.5, and Γ2 = 0.1,
and
P =
[0.5 0.50.5 0.5
]Then this is not MSS, but it converges almost surely.11
Interestingly, ρ(T ) = 3.13 (not MSS), but also ρ(B) = 1.3 (not MS)!
11See: Fang et al. 1994.
69 / 97
Understanding ASC
There is zero probability that we find a realisation of x(k)kwhich does not converge to 0.
73 / 97
Interlude: Ergodicity
A finite Markov jump process θ(k) is called ergodic if from every modei ∈ N we can move to any mode j ∈ N (not necessarily in one step).
Let πi(k) := P [θ(k) = i]. Under the ergodicity assumption, there is aprobability distribution π = π1, . . . , πN, πi > 0, such that
πi(k)→ πi.
And we can compute π solving12
P ′π = π.
12For details, I recommend the lecture notes of J. Gravner (UC Davis) on “intro-ductory probability”.
83 / 97
Take ν pos. integer and let
θ(k) = (θ(kν + ν − 1), . . . , θ(kν)) ∈ N ν
be a Markov chain. Take ι = (iν−1, . . . , ι0) and τ = (jν−1, . . . , j0) ∈ N ν .The transition probabilities of θ(k) are
P[θ(k + 1) = τ | θ(k) = ι
]= piν−1j0pj0j1 . . . pjν−1jν−1 .
We have P[θ(k) = τ
]= P [θ(kν) = j0] pj0j1 . . . pjν−1jν−1 and as k →∞
this converges toπ := πj0pj0j1 . . . pjν−1jν−1 .
84 / 97
Conditions for ASC
TheoremLet Γι := Γiν−1 . . .Γi0 . If there is ν pos. integer. s.t.∏
ι∈N ν‖Γι‖πι < 1,
then x(k)→ 0 a.s.; ‖ · ‖ can be any matrix norm.
85 / 97
Exercises
Exercise 1. Determine a system with n = 2, N = 2 which is ASC but notMSS.
Exercise 2. Show that a system with n = 1, N = 2, and θ(k): i.i.d. withπ1 = π2 = 0.5 and Γ1 = α, Γ2 = β is ASC if and only if |αβ| < 1.
Exercise 3. For a stochastic process x(k)k we know that E[x(k)]→ 0.Is the process ASC?
86 / 97
End of fifth section
Summary:
1. Almost sure convergence is weaker than MSS
2. ASC is convergence w.p.1.
3. ASC implies convergence in probability
4. The notion of almost sure convergence is not induced by a topology
5. A system can be ASC but not MSS or even MS
6. When the Markov process is ergodic, πi(k) converges
87 / 97
Feedback stabilisation
We need to design a mean square stabilising controller for the system
x(k + 1) = Aθ(k)x(k) +Bθ(k)u(k).
We assume x(k), θ(k) are measured and the controller has the form
u(k) = κ(θ(k), x(k)) = Fθ(k)x(k),
so the closed-loop MJLS is
x(k + 1) = (Aθ(k) +Bθ(k)Fθ(k))x(k).
89 / 97
Feedback stabilisation
The closed-loop system is MSS if there is V ∈ Hn+ so that T (V ) < V , i.e.,
for all j ∈ N
Vj −N∑i=1
pij(Aθ(k) +Bθ(k)Fθ(k))Vi(Aθ(k) +Bθ(k)Fθ(k))′ > 0
which can be written as a LMI using the Schur complement lemma(Exercise).
90 / 97
Feedback stabilisation
Comments:
1. We have one controller for each mode.
2. For each i ∈ N , we didn’t select Fi so as to stabilise (Ai, Bi).
3. Had we done that, would the closed-loop MJLS be MSS?
4. In practice, θ(k) may not be known exactly.
91 / 97
Stabilisation with inexact knowledge of θ(k)
Assume θ(k) is not available at time k, but we have an estimateθ(k) ∈ N (which defines a random process). Then,
u(k) = Fθ(k)x(k),
and the closed-loop MJLS will be
x(k + 1) = (Aθ(k) +Bθ(k)Fθ(k))x(k).
92 / 97
Stabilisation with inexact knowledge of θ(k)
We define the following filtration Mkk∈N with M0 = F0 and Mk is theσ-algebra generated by
x(0), θ(0), x(1), θ(1), θ(0), . . . , x(k), θ(k), θ(k − 1).
We assume that
P[θ(k) = s |Mk
]= P
[θ(k) = s | θ(k)
]=: ρθ(k)s.
This means that
ρis = probability that θ(k) = i while we estimated θ(k) = s
andρis = probability of success: θ(k) = θ(k) = i
93 / 97
Stabilisation with inexact knowledge of θ(k)
Letp = min
iρii.
Let T : Hn → Hn be defined as before with Γi = Ai +BiFi. Since T isstable, there are β ≥ 1 and ζ ∈ (0, 1) s.t.13
‖T k‖1 ≤ βζk.
Finally, let
c0 := max‖Ai +BiFs‖2, i, s ∈ N, i 6= s, ρis > 0
.
13Question: How do we determine those parameters? Can you give an examplewhen N = 1 and N = 2?
94 / 97
Stabilisation with inexact knowledge of θ(k)
MSS condition:
p > 1− 1− ζc0β
.
Proof.The proof is a bit lengthy and can be found in the book of Costa et al.(Lemma 3.44). Sketch: We show that
0 ≤ Qj(k + 1) ≤ Tj [Q(k)] + Sj [Q(k)],
where S : Hn → Hn is such that ‖S‖1 ≤ c0(1− p).
95 / 97
Further reading
1. O.L.V. Costa, M.D. Fragoso and R.P. Marques, Discrete-time Markov Jump LinearSystems, Springer 2005.
2. Y. Fang and K.A. Loparo, Stochastic Stability of Jump Linear Systems, IEEETrans. Aut. Contr. 47(7), pp. 1204–1208, 2002.
96 / 97
References for ASC
1. Y. Fang, K.A. Loparo, and X. Feng. Almost sure and δ-moment stability of jumplinear systems. International Journal of Control, 59:1281–1307, 1994.
2. O. L. V. Costa and M. D. Fragoso. Stability results for discrete-time linear systemswith Markovian jumping parameters. Journal of Mathematical Analysis andApplications, 179:154–178, 1993.
3. J. B. R. do Val, C. Nespoli, and Y. R. Z. Caceres. Stochastic stability forMarkovian jump linear systems associated with a finite number of jump times.Jounal of Mathematical Analysis and Applications, 285:551–563, 2003.
4. Y. Fang. A new general sufficient condition for almost sure stability of jump linearsystems. IEEE Transactions on Automatic Control, 42:378–382, 1997.
5. Z. G. Li, Y. C. Soh, and C. Y. Wen. Sufficient conditions for almost sure stabilityof jump linear systems. IEEE Transactions on Automatic Control, 45:1325–1329,2000.
6. M. Mariton. Almost sure and moments stability of jump linear systems. Systemsand Control Letters, 30:1145–1147, 1985.
97 / 97