Processing math: 100%
Skip to Tutorial Content

Eigenvalues and Differential Equations

Click Start Over menu at the left bottom to start Back to Contents


1. Eigenvalues and Eigenvectors

Let A be a n×n square matrix. If a complex number λ and a nonzero n×1 vector x satisfy

Ax=λx,

then we call λ an eigenvalue of A and x an eigenvector of A.

An eigenvalue λ is the solution of det(AλI)=0, where det stands for determinant. If λ is an kth order of solution, then we say that λ has algebraic multiplicity αA(λ)=k. The dimension of the eigenspace EA(λ){xAx=λx}, or the number of independent eigenvectors associated with λ are called the geometric multiplicity γA(λ) of λ. If αA(λ)=1, λ is called a simple eigenvalue of A.

Example 1. I2×2 has one eigenvalue 1. Its algebraic multiplicity is 2. It has two independent eigenvectors [0,1]T and [1,0]T. Its geometric multiplicity is 2.

A1 <- matrix(c(1, 0, 0, 1), nrow = 2)
eigen(A1)
## eigen() decomposition
## $values
## [1] 1 1
## 
## $vectors
##      [,1] [,2]
## [1,]    0   -1
## [2,]    1    0

Example 2. A2 has one eigenvalue 1. Its algebraic multiplicity is 2. It has one independent eigenvector [1,0]T. So its geometric multiplicity is 1.

A2 <- matrix(c(1, 0, 1, 1), nrow = 2)
eigen(A2)
## eigen() decomposition
## $values
## [1] 1 1
## 
## $vectors
##      [,1]          [,2]
## [1,]    1 -1.000000e+00
## [2,]    0  2.220446e-16

Example 3. A3 has two simple eigenvalues 2 and 1. Their corresponding eigenvectors are [0,1]T and [1,1]T.

A3 <- matrix(c(1, 1, 0, 2), nrow = 2)
eigen(A3)
## eigen() decomposition
## $values
## [1] 2 1
## 
## $vectors
##      [,1]       [,2]
## [1,]    0  0.7071068
## [2,]    1 -0.7071068

Theorem 1. Geometric multiplicity and algebraic multiplicity can be different. But geometric multiplicity can never exceed algebraic multiplicity.

Proof. Let x1,x2,,xr be all the linearly independent eigenvectors associated with eigenvalue e. So the geometric multiplicity of A is r. Adding nr independent vectors xr+1,xr+2,,xn to form a basis of Rn, we can create an inversible matrix S=[x1,x2,,xr,xr+1,xr+2,,xn] such that

S1AS=[eIr×r?0?]

Since det(S1(AλI)S)=det(AλI), the characteristic polynomial det(AλI) contains factor (λe)r. Therefore its algebraic multiplicity is not less than r.

2. Generalized Eigenspaces

Let A be a n×n square matrix and λ an eigenvalue of A. Let αA(λ) be the algebraic multiplicity of λ. A nonzero n×1 vector x satisfying

(AλI)αA(λ)x=0

is called a generalized eigenvector of A associated with eigenvalue λ. The space GA(λ)={x(AλI)αA(λ)x=0} is called the generalized eigenspace.

Theorem 2. Generalized Eigenspace Decomposition. If an n×n matrix A has m (mn) distinct eigenvalues λj, 1jm, then

Cn=GA(λ1)GA(λ2)GA(λm)

Corollary 1. Dimension of Generalized Eigenspaces. For any eigenvalue λ of a matrix A,

dim(GA(λ))=αA(λ)

Proofs of Theorem 2 and Corollary 1 are not given here. But we will explain the construction of a basis in a generalized eigenspace using the Jordan chain. Suppose that A has an eigenvalue λ whose geometricity is 1 and algebraic multiplicity 3. Let us compose three vectors:

(AλI)x1=0(AλI)x2=x1(AλI)x3=x2

It is obvious that x1,x2,x3 are in GA(λ). If there is a constant vector [c1,c2,c3]T such that c1x1+c2x2+c3x3=0.

(AλI)2(c1x1+c2x2+c3x3)=c3x1=0 yields c3=0;

(AλI)(c1x1+c2x2)=c2x1=0 yields c2=0;

c1x1=0 yields c1=0.

So x1,x2,x3 are independent.

In R, we can first find a vector in the generalized eigenspace using nullspace((AλI)αA(λ)) and then backward generate the Jordan chain (see Example 6).

3. Systems of Differential Equations

Consider the following first order, linear systems of ordinary differential equations with constant coefficients

dx(t)dt=Ax(t),t>0

Given initial value x(0)=x0, we want to solve Eq. (4).

Case 1. A has exactly n independent eigenvectors.

In this case, A has n independent eigenvectors e1,e2,,en corresponding to n eigenvalues λ1,λ2,,λn (possibly repeated). The initial value vector x0 can be uniquely expressed as x0=nj=1cjej, where [c1,c2,,cn]T is the coeficient vector. Therefore the solution to Eq. (4) can be written as

x(t)=eAtx0=nj=1cjeAtej=nj=1cjeλjtej

Example 4. Find the solution to the following system

x1=x1+2x2x1(0)=0x2=3x1+2x2x2(0)=4

We first convert the system into matrix form:

ddt[x1(t)x2(t)]=[1232][x1(t)x2(t)],

A <- matrix(c(1, 3, 2, 2), nrow = 2)
Aeig <- eigen(A)
c0 <- solve(Aeig$vectors, c(0, -4))
c0
## [1]  2.884441 -2.262742
Aeig$values
## [1]  4 -1
Aeig$vectors
##            [,1]       [,2]
## [1,] -0.5547002 -0.7071068
## [2,] -0.8320503  0.7071068

Solution

x(t)=2.884441e4t[0.55470020.8320503]2.262742et[0.70710680.7071068]

Example 5. Solve the following initial value problem.

x=[3943]xx(0)=[24]

A <- matrix(c(3, -4, 9, -3), nrow = 2)
Aeig <- eigen(A)
c0 <- solve(Aeig$vectors, c(2, -4))
c0
## [1] 1.20185+3.469443i 1.20185-3.469443i
Aeig$values
## [1] 0+5.196152i 0-5.196152i
Aeig$vectors
##                       [,1]                  [,2]
## [1,]  0.8320503+0.0000000i  0.8320503+0.0000000i
## [2,] -0.2773501+0.4803845i -0.2773501-0.4803845i

Solution

x(t)=2.884441e+5.196152it[0.8320503+0.0000000i0.2773501+0.4803845i]2.262742e5.196152it[0.8320503+0.0000000i0.27735010.4803845i]



Case 2. A has less than n independent eigenvectors.

Lemma 1. If λ is an eigenvalue of A with algebraic multiplicity α, then for any generalized eigenvector x associated with λ

eAtx=α1p=0tpp![k=0tkp(kp)!(λI)kp](AλI)px.

Proof. Since

(At)kk!x=tkk!(λI+AλI)kx=tkk!α1p=0k!(kp)!p!(λI)kp(AλI)px=α1p=0tpp!tkp(kp)!(λI)kp(AλI)px,

eAtx=k=0(At)kk!x=k=0α1p=0tpp!tkp(kp)!(λI)kp(AλI)px=α1p=0tpp![k=0tkp(kp)!(λI)kp](AλI)px.

In this case, A has at least one eigenvalue whose geometric multiplicity is less than its algebraic multiplicity. Let us consider a special case where the eigenvalue λ1 has algebraic multiplicity 3 and geometric multiplicity 1. Three independent vectors in the generalized eigenspace of λ1 are chosen as described in Section 2, i.e.

(Aλ1I)x1=0(Aλ1I)x2=x1(Aλ1I)x3=x2

So a basis can be formed as {x1,x2,x3,e4,,en}$ corresponding to n eigenvalues λ1,λ1,λ1,λ4,,λn (possibly repeated). The initial value vector x0 can be uniquely expressed as x0=c1x1+c2x2+c3x3+nj=4cjej, where [c1,c2,,cn]T is the coeficient vector.

For x1, using Eq. (6), we obtain eAtx1=eλ1tx1.

For x2, using Eq. (6), we obtain eAtx2=eλ1tx2+teλ1tx1.

For x3, using Eq. (6), we obtain eAtx3=eλ1tx3+teλ1tx2+t22!eλ1tx1.

Therefore the solution to Eq. (4) can be written as

x(t)=eAtx0=c1eAtx1+c2eAtx2+c3eAtx3+nj=4cjeAtej=(c1+c2t+c3t22!)eλ1tx1+(c2+c3t)eλ1tx2+c3eλ1tx3+nj=4cjeλjtej

Example 6. Solve the following initial value problem.

x=[7143]xx(0)=[25]

Step 1. Find eigenvalues and eigenvectors

A <- matrix(c(7, -4, 1, 3), nrow = 2)
Aeig <- eigen(A)
Aeig$values
## [1] 5 5
Aeig$vectors
##            [,1]       [,2]
## [1,] -0.4472136  0.4472136
## [2,]  0.8944272 -0.8944272

Step 2. Find generalized eigenvectors

library(pracma)
M <- (A - Aeig$values[1] * diag(2)) %*% (A - Aeig$values[1] * diag(2))
K <- nullspace(M)
# Jordan chain
(A - Aeig$values[1] * diag(2)) %*% K
##      [,1] [,2]
## [1,]    2    1
## [2,]   -4   -2
(A - Aeig$values[1] * diag(2)) %*% (A - Aeig$values[1] * diag(2)) %*% K
##      [,1] [,2]
## [1,]    0    0
## [2,]    0    0

Step 3. Construct a basis

B <- cbind(c(1, -2), K[,2])
B
##      [,1] [,2]
## [1,]    1    0
## [2,]   -2    1
# decompose initial value
C <- solve(B, c(2, -5))
C
## [1]  2 -1

Solution

(2t)e5t[12]e5t[01]

References

Kuang Z. and Yang M., On the index of the eigenvalue of a transport operator. Transport Theory and Statistical Physics 1995, 24 (9), 1411 - 1418.

https://www.statmethods.net/advstats/matrix.html

http://linear.ups.edu/scla/html/frontmatter.html

https://tutorial.math.lamar.edu/Classes/DE/DE.aspx