The Perron–Frobenius theory of nonnegative matrices has many useful dynamical consequences, in the field of Markov shifts in particular. The math in turn gives us insight into areas as diverse as Google page rank and virus dynamics, applications which will be discussed in this chapter.
![$$v \in \mathbb C^n, v \neq 0$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq1.png)
![$$p_A(t)= \det (A - t I)$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq2.png)
![$$\mathbb C^n$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq3.png)
![$$v \in \mathbb C^n$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq4.png)
![$$\lambda \in \mathbb C$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq5.png)
![$$\displaystyle \begin{aligned} (A - \lambda I)^k v = 0. \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ1.png)
![$$k \in \mathbb N$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq6.png)
7.1 Preliminary Background
Throughout this chapter, each matrix A = (a
ij), i, j = 1, …n is an n × n matrix. If each a
ij ≥ 0, we write A ≥ 0, and call A a nonnegative matrix. If a
ij > 0 for all i, j, we write A > 0, and say A is positive. For a vector , v ≥ 0 and v > 0 mean v
j ≥ 0 and v
j > 0 for all j, respectively. We use I to denote the n × n identity matrix. On finite dimensional spaces all norms are equivalent; we use the following norms on matrices and vectors, convenient for our setting. For
, ∥v∥∞ =max1≤j≤n|v
j|, and for A an n × n matrix,
We write ∥v∥ and ∥A∥ for these norms throughout this chapter; the reader is referred to Appendix B, Section B.1.2 for more about norms. We note that sometimes
is naturally a row vector, and other times a column vector (to make multiplication work). We assume it will be clear from the context, but when we need to change left multiplication of A by v to right multiplication of A
t, we use (vA)t = A
t
v
t and vice versa, where A
t = (a
ji) is the transpose of A = (a
ij). For two vectors
, by (v, w) we denote the usual inner product (dot product).
Of particular interest in dynamics are nonnegative matrices and vectors whose rows all sum to 1, stochastic matrices, introduced in Remark 6.18. The (ij)th entry in a stochastic matrix represents the probability of going from state i to state j in one time step. A stochastic vector gives the probabilities of initial states in the process being modelled. We draw on the next key result for the proof of the Perron–Frobenius Theorem.
Suppose A > 0 is a stochastic n × n matrix. If a vector is not of the form v = (t, t, …, t) for any
, then ∥Av∥ < ∥v∥.
By hypothesis, given v ≠ (0, …, 0), let ∥v∥ = α > 0; then for some j
o = 1, …, n. By using − v instead of v if needed, assume
.
![$$|(Av)_k | = |\sum _{i=1}^n a_{ki}v_i|$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq16.png)
![$$\alpha = \sum _{i=1}^n a_{ki}v_{j_o}.$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq17.png)
![$$v_m \neq v_{j_o}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq18.png)
![$$\displaystyle \begin{aligned}|(Av)_k| = \left| \sum_{i=1}^n a_{ki}v_i \right | \leq \left | \sum_{i=1, i \neq m}^n a_{ki}v_i \right | \leq \sum_{i=1, i \neq m}^n a_{ki}|v_i| < \sum_{i=1}^n a_{ki}v_{j_o}.\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equa.png)
![$$\displaystyle \begin{aligned}|(Av)_k| = \sum_{i=1}^n a_{ki}v_i \leq \sum_{i=1, i \neq m}^n a_{ki}v_i + a_{km}v_m < \sum_{i=1}^n a_{ki}v_{j_o},\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equb.png)
![$$v_m<v_{j_o}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq19.png)
Every n × n stochastic matrix A > 0 has 1 as an eigenvalue with a 1-dimensional eigenspace spanned by η = (1, 1, …, 1).
for each k. By Proposition 7.1 there is no other vector
such that Av = v apart from scalar multiples of η. □
For a square matrix A, the ℓ-fold product of A with itself is well-defined; we write it as . A nonnegative matrix A is irreducible if for each i, j, there exists some ℓ = ℓ(i, j) such that
. (This is consistent with Definition 6.15 and Remark 6.17). A stronger condition is that A is primitive, which is defined by the property that there is some
such that
for all i, j = 1, …, n, i.e., A
ℓ is positive.
After several intermediate results, we give a version of the Perron–Frobenius Theorem whose proof, and the proof of a stronger version, appears in Section 7.2 If r is an eigenvalue for A, then it is an eigenvalue for A t as well, and we say v is a left eigenvector for r if vA = rv, or equivalently if A t v t = rv t.
- 1.
A has an eigenvalue r > 0, which is a simple root of the characteristic polynomial of A. There exists an eigenvector v > 0 such that Av = rv.
- 2.
Every other eigenvalue λ of A satisfies |λ|≤ r.
- 3.
If A is primitive, then every eigenvalue λ ≠ r satisfies |λ| < r.
![$$\displaystyle \begin{aligned} A_1 = \left( \begin{array}{cc} 0 & 1 \\ 1 & 0 \end{array} \right) \quad \mbox{and} \quad A_2 = \left( \begin{array}{ccc} 0 & 1 & 0 \\ \frac 1 2 & 0 & \frac 1 2 \\ 0 & 1 & 0 \end{array} \right) \end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equc.png)
![$$\displaystyle \begin{aligned} e^A = \sum_{k=0}^\infty \frac{A^k}{k!}, \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ2.png)
![$$a_{ij}^{(k)}>0$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq26.png)
![$$k \in \mathbb N$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq27.png)
![$$\displaystyle \begin{aligned}e^A = I + A + \frac{ A^2}{2!} + \ldots \ \mbox{implies} \ e^A-I = A + \frac{A^2}{2!}+ \cdots + \frac{A^j}{j!} \cdots.\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equd.png)
We have the following lemma about the action of A and A t on nonnegative vectors when A is irreducible.
Given an irreducible n × n matrix A ≥ 0, let denote the positive cone of
. If v ∈ Q, then both Av ∈ Q and A
t
v ∈ Q, and if v ≠ 0, then both Av and A
t
v are nonzero.
Moreover if v ∈ Q ∖{0}, and Av = rv, r > 0, then v > 0. Similarly, if u is an eigenvector for A t in Q, for r > 0, then u > 0 (all coordinates of u and v are strictly positive).
![$$B=e^A-I = \sum _{k=1}^{\infty }\frac {A^k}{k!}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq30.png)
![$$\displaystyle \begin{aligned} B>0 \, \, \mathrm{and} \,\, v \in Q\setminus \{0\} \quad \Rightarrow \quad Bv > 0.\end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ3.png)
If v ∈ Q ∖{0} satisfies Av = rv for some r > 0, then using Eqn (7.3) (e A − I)v = (e r − 1)v > 0. Therefore v > 0.
Since A
t is irreducible if and only if A is, the remaining statements follow by the same argument replacing A with A
t, and using that . □
For A ≥ 0 irreducible, we consider the simplex in S ⊂ Q defined by S = {v ∈ Q : (v, η) = 1}, with η = (1, …, 1). Using the norm on given by
, we have S = {v ∈ Q : ∥v∥1 = 1}, and from Lemma 7.5, the map
defined by:
is well-defined and continuous on S.
The map defined by
has a fixed point.
![$$\mathbb R^n$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq38.png)
![$$\mathbb R^{n-1}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq39.png)
![$$\hat {A}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq40.png)
![$$\displaystyle \begin{aligned} \|tv+(1-t)w\|{}_1&=(tv+(1-t)w,\eta) \\& = t(v,\eta)+(1-t)(w,\eta) \\& = t+ (1-t)=1.\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Eque.png)
□
7.2 Spectrum and the Perron–Frobenius Theorem
We have now laid the groundwork to prove the first part of the Perron–Frobenius Theorem.
An irreducible n × n matrix A has an eigenvalue r > 0, with an eigenvector v ∗ > 0. Moreover, r is the only eigenvalue for A with an eigenvector in Q. Additionally the matrix A t has an eigenvector w ∗ > 0 for r, and r is the only eigenvalue for A t with an eigenvector in Q. Equivalently, there is a left eigenvector x > 0 such that xA = rx.
By Lemmas 7.5 and 7.6, A maps Q into Q, and there is a vector v ∗∈ S ⊂ Q such that v ∗ = Av ∗∕∥Av ∗∥1; therefore r = ∥Av ∗∥1 > 0 is an eigenvalue for A with eigenvector v ∗ > 0.
![$$\widehat {A^t}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq41.png)
![$$\displaystyle \begin{aligned} \lambda (v,w) =(v,A^tw)=(Av,w)=r (v,w). \end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equf.png)
![$$\lambda \in \mathbb C $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq42.png)
![$$\displaystyle \begin{aligned} \lambda (u,w^*) =(A u ,w^*)=(u, A^t w^*)=r(u,w^*). \end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equg.png)
We are interested primarily in irreducible matrices whose positive eigenvalue is 1, so we replace A by (1∕r)A if needed so that r = 1. Without loss of generality, for now we assume the eigenvalue from Proposition 7.7 is r = 1, and we continue to write the matrix as A. A classical result from linear algebra allows us to replace A by a similar matrix, i.e., one of the form C = D −1 AD, and it follows that Spec(C) = Spec(A). However, we make a strategic choice of similarity (the matrix D) that gives a more useful norm, since every matrix norm acts as an upper bound for Spec(A) (see Exercise 3). We consider the eigenvector for r = 1, v ∗ = (v 1, ⋯ , v n) > 0, and we define D = (d ij) to be the invertible diagonal matrix with d jj = v j > 0. It is not hard to see that the resulting matrix C = D −1 AD is stochastic (see the Exercises below). In the next proposition we show that these adjustments help simplify the proof of Theorem 7.3.
Recall that if A is an n × n matrix and λ is an eigenvalue, then λ is geometrically simple if the dimension of the eigenspace in for λ is one, and algebraically simple if λ is a root of multiplicity one of the characteristic polynomial p
A(t).
- 1.
r = 1 is an eigenvalue for C with η = (1, 1, ⋯ , 1) as an eigenvector.
- 2.
∥C∥ = ∥Cη∥∞ = 1.
- 3.
If λ is an eigenvalue for C, then |λ|≤ 1.
- 4.
The eigenvalue 1 is a simple root of the characteristic equation of C, so 1 is algebraically and geometrically simple.
- (1)
For each j = 1, …, n,
since each row of C sums to 1.
- (2)Let w = (w 1, ⋯ , w n) satisfy ∥w∥ = 1. Then since 0 ≤ c ij ≤ 1 for all i, j, it is clear that(7.4)
Therefore ∥C∥ = 1.
- (3)
Since Spec(C) ≤∥C∥, 1 ≥|λ| for every λ ∈Spec(C), and (3) follows immediately.
- (4)
It suffices to prove (4) for the matrix B = 1∕(e − 1)(e C − I); by Remark 7.4, B > 0 and proving (4) for B implies (4) holds for C as well (see Exercise 5). We first show that B is stochastic by showing all rows sum to 1. Because Cη = η and
, it follows that Bη = 1∕(e − 1)((e − 1)η) = η = (1, 1, …, 1). This shows that B is stochastic, since (Bη)j = 1 is the sum of the entries of the j th row of B.
To show (4) for B, we consider two B-invariant subspaces. First, is the one-dimensional eigenspace for 1 (by Corollary 7.2). We next consider an eigenvector u
∗ > 0 for B
t corresponding to the eigenvalue 1. We then define the second subspace
. If w ∈ W, then (Bw, u
∗) = (w, B
t
u
∗) = (w, u
∗) = 0, so Bw ∈ W. W is n − 1 dimensional, since it is the orthogonal complement to the eigenspace spanned by u
∗, and W ∩ V
1 = {0}. Since y∉V
1 implies By − y∉V
1 unless y is an eigenvector and By − y = 0, there are no generalized eigenvectors for the eigenvalue 1 except for vectors in V
1. This shows that the eigenvalue 1 is both geometrically and algebraically simple. □
Assume the matrix A ≥ 0 is irreducible; Proposition 7.7 implies that A has an eigenvalue r > 0 with an eigenvector ζ
∗ > 0. Set A
r = (1∕r)A so that A
r ≥ 0 is irreducible and has 1 as an eigenvalue, with eigenvector ζ
∗ > 0 by Lemma 7.5. Define D to be the diagonal matrix with so that the resulting matrix C
r = D
−1
A
r
D satisfies the hypotheses of Proposition 7.8. It then follows that 1 is an algebraically simple eigenvalue of C
r and therefore r is an algebraically simple eigenvalue of A, yielding (1) and (2) of the theorem.
To prove (3), assume that in addition A is primitive, so A
m > 0 for some , and set M = r
−m
A
m > 0. Now M has 1 as an eigenvalue with the eigenvector ζ
∗, and we claim that for (3) to hold for the original matrix A, it is enough to show that λ ∈Spec(M), λ ≠ 1 implies |λ| < 1. The claim is true because if β ∈Spec(A) and |β| = r, then β
m ∈Spec(A
m) and (β∕r)m ∈Spec(M); but |β∕r| = 1, so β = r if (3) holds for M.
Finally, to prove that (3) holds for M, denote by D the diagonal matrix defined above with . Then C = D
−1
MD is positive and stochastic, which yields a spectral radius of 1 with a geometrically and algebraically simple eigenvalue 1 for C with eigenspace V
1. Therefore
is the direct sum
, each a C invariant subspace, as in the proof of Proposition 7.8. In particular, there is a corresponding eigenvector u
∗ > 0 with C
t
u
∗ = u
∗, and
.
If λ ≠ 1 is an eigenvector for C, then λ ∈Spec(C|W). However by Proposition 7.1, if w ∈ W, and Cw = λw, then ∥Cw∥ < ∥w∥, which implies |λ| < 1. This proves (3) for C, and therefore for the similar matrix M, and hence for A.
The following version of the Perron–Frobenius Theorem, often used in applications, is a consequence of Proposition 7.8.
Suppose A is primitive and stochastic with left eigenvector q > 0 for the eigenvalue 1, normalized so that ∥q∥1 = 1. Let P denote the matrix where each row of P is identically q. Then for every
, A
k
x → Px exponentially fast. In particular, ∥A
k − P∥→ 0 as k →∞.
![$$m \in \mathbb N$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq55.png)
![$$V_1 =\{x \in \mathbb R^n : x=(a,a,\ldots ),\, a \in \mathbb R\} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq56.png)
![$$W = \{ w \in \mathbb R^n : (w, q) = 0 \}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq57.png)
![$$\mathbb R^n = V_1 \oplus W$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq58.png)
![$$\displaystyle \begin{aligned} \|A^m |{}_W \| = \beta<1 \,\, \mathrm{and} \, \|A\| = 1, \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ5.png)
![$$x \in \mathbb R^n$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq59.png)
The matrix P defined in Theorem 7.9 satisfies the following: (1) P is stochastic, (2) PA = AP = P, (3) an eigenvalue of A is an eigenvalue of P, and (4) P 2 = P.
In many applications, for example Section 7.4 below, the following consequence is used.
![$$\sum _{i=1}^nx_i=1$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq63.png)
![$$\displaystyle \begin{aligned}\lim_{k \rightarrow \infty} x A^k = q\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equi.png)
where q is as in Theorem 7.9.
We set v = x t to be a column vector, and then limk→∞(xA k)t =limk→∞(A k)t v = P t(v) = q. (See Exercise 8.) □
7.2.1 Application to Markov Shift Dynamics
Using the notation from Chapter 6.2, we let M denote an incidence matrix, and A a stochastic matrix with zero entries precisely where M has zeros. The matrix M = (m
ij) satisfies m
ij = 1 if and only if in one time step state i can go to state j. For each pair of states i and j, a
ij = 0 if and only if m
ij = 0; if a
ij > 0, its value represents the probability of going from state i to state j. We apply the above results to the dynamical system σ on to obtain proofs of the following results.
If A is an irreducible stochastic matrix, then there exists an invariant probability measure ρ
A satisfying all the conditions of Definition
6.12
, making into a finite measure-preserving Markov shift.
By Proposition 7.7, we have a left stochastic eigenvector q > 0 for the eigenvalue 1; this gives the existence of an invariant measure ρ A satisfying Definition 6.12. □
In order to show that ρ A is ergodic, or equivalently that the (q, A) Markov shift is ergodic, we need one more lemma.
Let A be an n × n irreducible stochastic matrix and q > 0 a stochastic vector in such that qA = q. Then
, where every row of P is q.
![$$(\Sigma _M,\mathcal B,\rho _A, \sigma )$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq68.png)
![$$C_0^j$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq69.png)
![$$\chi _j = \chi _{C_0^j}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq70.png)
![$$\displaystyle \begin{aligned}\lim_{N \rightarrow \infty} \frac 1 N \sum_{k=0}^{N-1} \chi_j (\sigma^k x) = \chi_j^* (x) \quad \mu\mathrm{-}\mathrm{a.e.}\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equj.png)
![$$\displaystyle \begin{aligned}\lim_{N \rightarrow \infty} \frac 1 N \sum_{k=0}^{N-1} q_i a_{i j}^{(k)} = \int \chi_j^* (x)\chi_i(x) d \rho_A.\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equk.png)
![$$\displaystyle \begin{aligned}b_{ij} = \frac{1}{q_i} \int \chi_j^* (x)\chi_i(x) d \rho_A\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equl.png)
If A is a primitive stochastic matrix, then is mixing (and hence weak mixing and ergodic).
![$$X=C_{k}^w$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq72.png)
![$$Y= C_\ell ^u$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq73.png)
![$$\displaystyle \begin{aligned}\lim_{n \rightarrow \infty} \rho(\sigma^{-n}X \cap Y) = \rho(X)\rho(Y).\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equm.png)
![$$\displaystyle \begin{aligned} \rho(\sigma^{-n} X \cap Y) = q_{u_1}a_{u_1u_2} \cdots a_{u_{s-1}u_{s}}a_{u_sw_1}^{(n+k-(\ell+s))}a_{w_1 w_2} \cdots a_{w_{r-1}w_r}, \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ8.png)
![$$q_j = \lim _{n \rightarrow \infty } a_{ij}^{(n)}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq74.png)
![$$\displaystyle \begin{aligned} \lim_{n \rightarrow \infty} \rho(\sigma^{-n}X \cap Y) &= q_{u_1}a_{u_1u_2} \cdots a_{u_{s-1}u_{s}}q_{w_1}a_{w_1 w_2} \cdots a_{w_{r-1}w_r} \\& = \rho(Y)\rho(X).\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equn.png)
□
If A is irreducible, then the (q, A) Markov shift is ergodic with respect to ρ.
7.3 An Application to Google’s PageRank
A search engine such as Google uses multiple ways to address an internet search query from a user, but PageRank is one of the best and oldest techniques. It was developed by Brin and Page in 1998 [22] at Stanford, and uses the Perron–Frobenius Theorem at its core. A PageRank is a score associated to a web page that assesses its importance in a particular search and is usually conducted only on web pages that contain data matching the original query. The list that appears in a Google search has as its first step finding pages that contain matches to the query words. Many details about PageRank can be found in sources such as [121] and [122]. We give a short account of the original method of Brin and Page to determine PageRank.
![$$\displaystyle \begin{aligned} \mathrm{{PR}}(V_t) = \sum_{V_j \, {\mathrm{links} \, \mathrm{to}} \, V_t} \frac {\mathrm{{PR}}(V_j)}{{\# \mathrm{links} \, \mathrm{out}\, \mathrm{of} \,}V_j}.\end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ9.png)
![../images/491076_1_En_7_Chapter/491076_1_En_7_Fig1_HTML.png](../images/491076_1_En_7_Chapter/491076_1_En_7_Fig1_HTML.png)
The links among 4 pages relevant to an internet search, where a directed arrow denotes a page link
Let A = (a ij) denote the (enormous) N × N matrix where a ij = 0 if page V i has no link to V j, and 1∕(#links out of Vi) if it does. We seek a left eigenvector for the eigenvalue 1 for A if it exists, since if vA = v, then writing v = (v 1, …, v t, …, v N), v t = PR(V t) as defined by (7.9). Even though A has row sums of 1, we do not necessarily have a stochastic matrix, since row sums of 0 also occur. Brin and Page made a few adjustments to A to yield an irreducible (positive) stochastic matrix to which the Perron–Frobenius applies, which we describe here.
The N × N matrix A contains many zeros, in particular, a dangling node is a page with no outgoing links and corresponds to a row of zeros in A. To take care of this issue, each row of zeros is replaced by a row of entries of 1∕N. We call the new matrix B, and B is a stochastic matrix. There is no obvious reason for it to be irreducible, but since there is always a very small probability that a browser might arrive at page V j by chance, we make one more adjustment which results in a primitive matrix C; in fact C > 0.
We assign a weight β < 1, but close to 1, to the matrix B, and the weight (1 − β) to the matrix with each entry 1∕N; denote this by F N. We define C = βB + (1 − β)F N. The PageRank vector v P is then defined to be the left eigenvector of C, for the eigenvalue 1, normalized so that ∥v P∥1 = 1. The entry of v P with the highest value determines the most important page, or, in a simple system, the page that is returned in a search.
As an example, we consider a tiny internet consisting of only 4 pages. They are interconnected via the graph shown in Figure 7.1.
![$$\displaystyle \begin{aligned} A = \left(\begin{array}{cccc} 0 & 0 & 1 & 0\\[.1 cm] \frac 12 & 0 & 0 & \frac 12\\[.1 cm] \frac 1 2 & 0 & 0& \frac 1 2\\[.1 cm] \frac 13& \frac 13&\frac 13&0\\ \end{array}\right). \end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equo.png)
![$$\displaystyle \begin{aligned} C = \left(\begin{array}{cccc} .0375 & .0375 & .8875&.0375\\ .4625 & .0375 & .0375 & .4625\\ .4625 & .0375 & .0375& .4625\\ .3208&.3208&.3208&.0375\\ \end{array}\right), \end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equp.png)
![$$\displaystyle \begin{aligned}v_P \approx (.3012, .1040, .3600, .2347)\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equq.png)
7.4 An Application to Virus Dynamics
Eradicating the human immunodeficiency virus (HIV) in an infected individual has remained out of reach of current medical practice; however antiretroviral therapy (ART) has changed HIV to a chronic disease. Studies going back to the 1990s reveal that even when HIV is undetectable in the blood, the virus is not completely gone. The cells infected by the virus are white blood cells that are called CD4 cells (these are a type of T cell) and form a critically important component of the immune system. A latent HIV reservoir is a group of CD4 cells in the body that are infected with HIV but are not actively producing new virus particles.
In a patient infected with HIV, although latently infected CD4 cells are relatively sparse, it is believed by many that they prevent HIV from being eradicated from an infected individual and therefore knowledge about their location (and source) is critical to the study of their elimination [36]. Following the discussion given in [87], and the numerous references therein, we construct a Markov model of the hidden dynamics of a latently infected CD4 cell in an HIV positive patient. We use the Perron–Frobenius Theorem to also show that a limiting distribution of the latently infected cells exists, which indicates in which anatomical compartments the HIV typically can be found, and for each, the relative likelihood of being in that compartment.
Although the cardiovascular system is the main part of the circulatory system, the lymphatic system is the second part. This system is connected to, but quite different from, the cardiovascular system insofar as the fluid, consisting of lymph and the immune cells called lymphocytes, circulates much more slowly. Lymph is primarily moved by surrounding muscles; it is of critical importance in the study of HIV. There are associated lymphatic tissue and organs that filter pathogens from the lymph before returning it to the bloodstream. With that in mind, we review the major repositories of latently infected cells, and number the corresponding compartment to give the states represented in our matrix model.
7.4.1 States of the Markov Process
- 0.
The peripheral blood system is the pool of circulating blood in the cardiovascular system; it contains red blood cells, white blood cells (including latently infected CD4 cells), and platelets suspended in plasma. The blood is rapidly circulated through a closed circulatory system pumped by the heart. Red blood cells do not escape the cardiovascular system in a healthy individual, though white blood cells, slightly smaller, do.
- 1.
Gut-associated lymphatic tissue (GALT) is the site of many CD4 cells and is the largest immune organ in the body. Clinical tests have shown the presence of significant numbers of latently infected cells in GALT (around 36 copies per million cells).
- 2.
Lymph nodes are (some of the) filtering compartments in the lymphatic system that have an elaborate structure through which lymph flows in order to attack pathogens. The lymph nodes in a human (there are likely to be 500–600 of them per person) also house a large percentage of CD4 cells, including a large percentage of the latently infected ones.
- 3.
Cerebrospinal fluid and the central nervous system (CSF) house reservoirs of HIV as well. Infected cells in CSF appear from the early stages of infection, and the infected cells harboring the viruses are believed to be CD4 cells, though the concentration is lower than in other parts of the body.
- 4
Bone marrow hosts HIV reservoirs, and clinical studies show that the HIV DNA is latently present in CD4 cells found in bone marrow.
We model the viral spread within an individual human patient using a Markov process on a finite state space , where k is the number of compartments of the body in which CD4 cells with latent virus have been found, and each state represents a compartment. We are interested in dynamics of latently infected resting CD4 cells in a sample taken from the peripheral blood; since this fluid circulates quickly, we choose a time increment for taking measurements, (assigning a time increment to be a day is a reasonable choice). We study the statistical properties of a generic latently infected cell located in one of the k compartments.
We define a sequence of -valued random variables {X
n : n ≥ 0} by first setting X
0 to be the compartment where the cell resides in the body at the beginning of our observations. Then X
1 is the compartment where the cell resides at the first (time 1) observation, and X
n is the compartment at the nth observation. We assume that X
n depends on X
n−1 and not on earlier X
j’s; this is a reasonable observation given that some of the compartments allow flow from one to another, but not all. We look at a sample case, using the states 0, 1, 2, 3, 4 described above.
![$$B^n=(b^{(n)}_{ij})$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq77.png)
![$$b^{(n)}_{ij}=1$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq78.png)
![../images/491076_1_En_7_Chapter/491076_1_En_7_Fig2_HTML.png](../images/491076_1_En_7_Chapter/491076_1_En_7_Fig2_HTML.png)
Labelled graph showing the fluid flow connections in one day among the compartments listed in 7.4.1
Using the matrix B and clinical data we construct a stochastic matrix P = (p ij) with the property that p ij = P(X n = j|X n−1 = i), where P denotes probability. Clearly p ij > 0 if and only if b ij > 0 for the corresponding incidence matrix B. We label the states as Lymph nodes (Node), Peripheral blood (Blood), Bone marrow (Bone), Cerebrospinal fluid and central nervous system(CSF), and Gut-associated lymphatic tissue (GALT), assigning numbers 0 − 4, respectively.
![$$\displaystyle \begin{aligned} B_{\mathrm{{day}}} =\left(\begin{array}{ccccc}1 & 1 & 1 & 1 & 1 \\ 1 & 1 & 1 & 0 & 1 \\ 1 & 1 & 1 & 0 & 0\\ 1 & 0 & 0 & 1 &0 \\ 1 & 1 & 0 & 0 & 1 \end{array} \right) \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ10.png)
![$$B_{\mathrm {{day}}}^2$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq79.png)
![$$p^{(2)}_{ij} >0$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq80.png)
![$$\displaystyle \begin{aligned} P_{\mathrm{{month}}}=\left(\begin{array}{ccccc}p_{00} & p_{01} & p_{02} & p_{03} & p_{04} \\ p_{10} & p_{11} & p_{12} & p_{13} & p_{14} \\ p_{20} & p_{21} & p_{22} & p_{23} & p_{24}\\ p_{30} & p_{31} & p_{32} & p_{33} &p_{34} \\ p_{40} & p_{41} & p_{42} & p_{43} & p_{44} \end{array} \right) \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ11.png)
![$$\displaystyle \begin{aligned} P_{\mathrm{{day}}}=\left(\begin{array}{ccccc}.5 & .19 & .005 & .005 & .3 \\ .35 & .4 & .05 & 0 & .2 \\ .3 & .2 & .5 & 0 & 0\\ .2 & 0 & 0 & .8 &0 \\ .4 & .1 & 0 & 0 & .5 \end{array} \right) \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ12.png)
![$$P_{\mathrm {{month}}} = P_{\mathrm {{day}}}^{28}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq81.png)
![$$\displaystyle \begin{aligned} q=(.428, .200, .024, .011, .337). \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ13.png)
![$$x \in \mathbb R^n$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_IEq82.png)
![$$\displaystyle \begin{aligned}\lim_{k \rightarrow \infty} x P_{\mathrm{{month}}}^k = \lim_{k \rightarrow \infty} x P_{\mathrm{{day}}}^{28 k}=q;\end{aligned}$$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equr.png)
![$$\displaystyle \begin{aligned} P_{\mathrm{{month}}}=\left(\begin{array}{ccccc}.428 & .200 & .024 & .011 & .337 \\ .428 & .200 & .024 & .011 & .337 \\ .428 & .200 & .024 & .011 & .337\\ .428 & .200 & .024 & .012 & .336 \\ .428 & .200 & .024 & .011 & .337\end{array} \right) \end{aligned} $$](../images/491076_1_En_7_Chapter/491076_1_En_7_Chapter_TeX_Equ14.png)
We can interpret the jth vector entry in q as the probability that a latently infected cell can be found in compartment j. With this assessment of the compartments in which latent CD4 cells are found in patients, the model shows that a generic latently infected CD4 cell is almost three times as likely to be found in the lymphatic system (GALT, 34%, or lymph nodes, 43%, giving 77%) as in the blood (20% likelihood) and the outcome is independent of where the latent infection started. The model also shows what is supported by clinical data, that there is a positive probability that there will be a steady state of latently infected cells found in the bone marrow and cerebral spinal fluid, despite the infrequency of its discovery.
The power of the Perron–Frobenius Theorem is that even though the assigned probabilities might be somewhat inaccurate, the existence of a left eigenvector q with positive entries is guaranteed, and the entries of q depend continuously on the entries of P; improvement in initial data will quickly accelerate the accuracy of the model. Due to Corollary 7.11, about four days after the original measurement in this example, we are quite close to the limiting distribution. The error is easily estimated [174], so at one month we can be assured we are (typically) extremely close to the limiting distribution, which could be used to make testing more efficient and informative.
- 1.
Prove that if A is an n × n matrix and B = D −1 AD for some invertible square matrix D, then λ is an eigenvalue for A if and only if λ is an eigenvalue for B.
- 2.
Show that if A is an n × n matrix and there exists a generalized eigenvector for A and λ, then λ is an eigenvalue for A.
- 3.
Prove that if A is an n × n matrix such that ∥A∥ = r for some norm for A, then Spec(A) ≤ r.
- 4.
Assume that A has an eigenvalue λ = 1, with eigenvector v ∗ = (v 1, ⋯ , v n) > 0. Define D = d ij to be the diagonal matrix with d jj = v j > 0. Show that C = D −1 AD is stochastic.
- 5.
If C ≥ 0 is an n × n matrix, show that if 1 is an algebraically simple eigenvalue for B = 1∕(e − 1)(e C − I), then 1 is an algebraically simple eigenvalue for C. Hint: Set f(z) = (e z − 1)∕(e − 1), and consider B = f(C). Use f(1) = 1 to write B − I = (C − I)g(C), where g(z) is entire on
.
- 6.
Assume that A is irreducible and has an eigenvalue λ = 1, with eigenvector v ∗ = (v 1, ⋯ , v n) > 0. Define D = d ij to be the diagonal matrix with d jj = v j > 0. Show that C = D −1 AD is irreducible.
- 7.
Prove that if M is a primitive incidence matrix, then the associated topological Markov shift σ on ΣM is topologically transitive.
- 8.
Assume A is a primitive stochastic matrix with left eigenvector q for the eigenvalue 1. Let P be the matrix with each row q. Show that for every vector
with ∥v∥1 = 1, P t v = q.
- 9.For k ≥ 2 an integer, find the Markov measure ρ A for the matrix
- 10.Find an invariant probability measure for the Markov shift determined by the stochastic matrix: