Eigenvalues and Differential Equations
Click Start Over menu at the left bottom to start Back to Contents
1. Eigenvalues and Eigenvectors
Let A be a n×n square matrix. If a complex number λ and a nonzero n×1 vector x satisfy
Ax=λx,
then we call λ an eigenvalue of A and x an eigenvector of A.
An eigenvalue λ is the solution of det(A−λI)=0, where det stands for determinant. If λ is an kth order of solution, then we say that λ has algebraic multiplicity αA(λ)=k. The dimension of the eigenspace EA(λ)≡{x∣Ax=λx}, or the number of independent eigenvectors associated with λ are called the geometric multiplicity γA(λ) of λ. If αA(λ)=1, λ is called a simple eigenvalue of A.
Example 1. I2×2 has one eigenvalue 1. Its algebraic multiplicity is 2. It has two independent eigenvectors [0,1]T and [1,0]T. Its geometric multiplicity is 2.
A1 <- matrix(c(1, 0, 0, 1), nrow = 2)
eigen(A1)
## eigen() decomposition
## $values
## [1] 1 1
##
## $vectors
## [,1] [,2]
## [1,] 0 -1
## [2,] 1 0
Example 2. A2 has one eigenvalue 1. Its algebraic multiplicity is 2. It has one independent eigenvector [1,0]T. So its geometric multiplicity is 1.
A2 <- matrix(c(1, 0, 1, 1), nrow = 2)
eigen(A2)
## eigen() decomposition
## $values
## [1] 1 1
##
## $vectors
## [,1] [,2]
## [1,] 1 -1.000000e+00
## [2,] 0 2.220446e-16
Example 3. A3 has two simple eigenvalues 2 and 1. Their corresponding eigenvectors are [0,1]T and [1,−1]T.
A3 <- matrix(c(1, 1, 0, 2), nrow = 2)
eigen(A3)
## eigen() decomposition
## $values
## [1] 2 1
##
## $vectors
## [,1] [,2]
## [1,] 0 0.7071068
## [2,] 1 -0.7071068
Theorem 1. Geometric multiplicity and algebraic multiplicity can be different. But geometric multiplicity can never exceed algebraic multiplicity.
Proof. Let x1,x2,⋯,xr be all the linearly independent eigenvectors associated with eigenvalue e. So the geometric multiplicity of A is r. Adding n−r independent vectors xr+1,xr+2,⋯,xn to form a basis of Rn, we can create an inversible matrix S=[x1,x2,⋯,xr,xr+1,xr+2,⋯,xn] such that
S−1AS=[eIr×r?0?]
Since det(S−1(A−λI)S)=det(A−λI), the characteristic polynomial det(A−λI) contains factor (λ−e)r. Therefore its algebraic multiplicity is not less than r.
2. Generalized Eigenspaces
Let A be a n×n square matrix and λ an eigenvalue of A. Let αA(λ) be the algebraic multiplicity of λ. A nonzero n×1 vector x satisfying
(A−λI)αA(λ)x=0
is called a generalized eigenvector of A associated with eigenvalue λ. The space GA(λ)={x∣(A−λI)αA(λ)x=0} is called the generalized eigenspace.
Theorem 2. Generalized Eigenspace Decomposition. If an n×n matrix A has m (m≤n) distinct eigenvalues λj, 1≤j≤m, then
Cn=GA(λ1)⨁GA(λ2)⨁⋯⨁GA(λm)
Corollary 1. Dimension of Generalized Eigenspaces. For any eigenvalue λ of a matrix A,
dim(GA(λ))=αA(λ)
Proofs of Theorem 2 and Corollary 1 are not given here. But we will explain the construction of a basis in a generalized eigenspace using the Jordan chain. Suppose that A has an eigenvalue λ whose geometricity is 1 and algebraic multiplicity 3. Let us compose three vectors:
(A−λI)x1=0(A−λI)x2=x1(A−λI)x3=x2
It is obvious that x1,x2,x3 are in GA(λ). If there is a constant vector [c1,c2,c3]T such that c1x1+c2x2+c3x3=0.
(A−λI)2(c1x1+c2x2+c3x3)=c3x1=0 yields c3=0;
(A−λI)(c1x1+c2x2)=c2x1=0 yields c2=0;
c1x1=0 yields c1=0.
So x1,x2,x3 are independent.
In R, we can first find a vector in the generalized eigenspace using nullspace((A−λI)αA(λ)) and then backward generate the Jordan chain (see Example 6).
3. Systems of Differential Equations
Consider the following first order, linear systems of ordinary differential equations with constant coefficients
dx(t)dt=Ax(t),t>0
Given initial value x(0)=x0, we want to solve Eq. (4).
Case 1. A has exactly n independent eigenvectors.
In this case, A has n independent eigenvectors e1,e2,⋯,en corresponding to n eigenvalues λ1,λ2,⋯,λn (possibly repeated). The initial value vector x0 can be uniquely expressed as x0=∑nj=1cjej, where [c1,c2,⋯,cn]T is the coeficient vector. Therefore the solution to Eq. (4) can be written as
x(t)=eAtx0=n∑j=1cjeAtej=n∑j=1cjeλjtej
Example 4. Find the solution to the following system
x′1=x1+2x2x1(0)=0x′2=3x1+2x2x2(0)=−4
We first convert the system into matrix form:
ddt[x1(t)x2(t)]=[1232][x1(t)x2(t)],
A <- matrix(c(1, 3, 2, 2), nrow = 2)
Aeig <- eigen(A)
c0 <- solve(Aeig$vectors, c(0, -4))
c0
## [1] 2.884441 -2.262742
Aeig$values
## [1] 4 -1
Aeig$vectors
## [,1] [,2]
## [1,] -0.5547002 -0.7071068
## [2,] -0.8320503 0.7071068
Solution
x(t)=2.884441e4t[−0.5547002−0.8320503]−2.262742e−t[−0.70710680.7071068]
Example 5. Solve the following initial value problem.
x′=[39−4−3]xx(0)=[2−4]
A <- matrix(c(3, -4, 9, -3), nrow = 2)
Aeig <- eigen(A)
c0 <- solve(Aeig$vectors, c(2, -4))
c0
## [1] 1.20185+3.469443i 1.20185-3.469443i
Aeig$values
## [1] 0+5.196152i 0-5.196152i
Aeig$vectors
## [,1] [,2]
## [1,] 0.8320503+0.0000000i 0.8320503+0.0000000i
## [2,] -0.2773501+0.4803845i -0.2773501-0.4803845i
Solution
x(t)=2.884441e+5.196152it[0.8320503+0.0000000i−0.2773501+0.4803845i]−2.262742e−5.196152it[0.8320503+0.0000000i−0.2773501−0.4803845i]
Case 2. A has less than n independent eigenvectors.
Lemma 1. If λ is an eigenvalue of A with algebraic multiplicity α, then for any generalized eigenvector x associated with λ
eAtx=α−1∑p=0tpppx.
Proof. Since
(At)kk!x=tkk!(λI+A−λI)kx=tkk!α−1∑p=0k!(k−p)!p!(λI)k−p(A−λI)px=α−1∑p=0tpp!tk−p(k−p)!(λI)k−p(A−λI)px,
eAtx=∞∑k=0(At)kk!x=∞∑k=0α−1∑p=0tpp!tk−p(k−p)!(λI)k−p(A−λI)px=α−1∑p=0tpppx.
In this case, A has at least one eigenvalue whose geometric multiplicity is less than its algebraic multiplicity. Let us consider a special case where the eigenvalue λ1 has algebraic multiplicity 3 and geometric multiplicity 1. Three independent vectors in the generalized eigenspace of λ1 are chosen as described in Section 2, i.e.
(A−λ1I)x1=0(A−λ1I)x2=x1(A−λ1I)x3=x2
So a basis can be formed as {x1,x2,x3,e4,⋯,en}$ corresponding to n eigenvalues λ1,λ1,λ1,λ4,⋯,λn (possibly repeated). The initial value vector x0 can be uniquely expressed as x0=c1x1+c2x2+c3x3+∑nj=4cjej, where [c1,c2,⋯,cn]T is the coeficient vector.
For x1, using Eq. (6), we obtain eAtx1=eλ1tx1.
For x2, using Eq. (6), we obtain eAtx2=eλ1tx2+teλ1tx1.
For x3, using Eq. (6), we obtain eAtx3=eλ1tx3+teλ1tx2+t22!eλ1tx1.
Therefore the solution to Eq. (4) can be written as
x(t)=eAtx0=c1eAtx1+c2eAtx2+c3eAtx3+n∑j=4cjeAtej=(c1+c2t+c3t22!)eλ1tx1+(c2+c3t)eλ1tx2+c3eλ1tx3+n∑j=4cjeλjtej
Example 6. Solve the following initial value problem.
x′=[71−43]xx(0)=[2−5]
Step 1. Find eigenvalues and eigenvectors
A <- matrix(c(7, -4, 1, 3), nrow = 2)
Aeig <- eigen(A)
Aeig$values
## [1] 5 5
Aeig$vectors
## [,1] [,2]
## [1,] -0.4472136 0.4472136
## [2,] 0.8944272 -0.8944272
Step 2. Find generalized eigenvectors
library(pracma)
M <- (A - Aeig$values[1] * diag(2)) %*% (A - Aeig$values[1] * diag(2))
K <- nullspace(M)
# Jordan chain
(A - Aeig$values[1] * diag(2)) %*% K
## [,1] [,2]
## [1,] 2 1
## [2,] -4 -2
(A - Aeig$values[1] * diag(2)) %*% (A - Aeig$values[1] * diag(2)) %*% K
## [,1] [,2]
## [1,] 0 0
## [2,] 0 0
Step 3. Construct a basis
B <- cbind(c(1, -2), K[,2])
B
## [,1] [,2]
## [1,] 1 0
## [2,] -2 1
# decompose initial value
C <- solve(B, c(2, -5))
C
## [1] 2 -1
Solution
(2−t)e5t[1−2]−e5t[01]
References
Kuang Z. and Yang M., On the index of the eigenvalue of a transport operator. Transport Theory and Statistical Physics 1995, 24 (9), 1411 - 1418.
https://www.statmethods.net/advstats/matrix.html