Diagonalization converts a complicated matrix into the simplest possible form: a diagonal matrix. When you can diagonalize a matrix A, you’ve found the “eigenvector coordinate system” where A just stretches along axes. This makes everything easier,computing powers, solving differential equations, understanding long-term behavior. The key is finding enough eigenvectors to build a new basis.
What is Diagonalization?
(The Goal)
A square n×n matrix A is diagonalizable if it can be written as:
A=PDP−1
where:
D is a diagonal matrix (eigenvalues on the diagonal)
P is an invertible matrix (eigenvectors as columns)
Geometric meaning: In the eigenvector basis, A is just a scaling transformation. The columns of P define this special coordinate system, D tells you the scaling factors, and P−1 converts back.
(Why This Matters)
Once you have A=PDP−1, everything becomes easier:
Powers:
Ak=PDkP−1
Computing Dk is trivial,just raise each diagonal entry to the k-th power.
Matrix exponential:
eAt=PeDtP−1
where eDt has eλit on the diagonal.
Long-term behavior: The dominant eigenvalue (largest ∣λ∣) determines if Ak→0, explodes, or oscillates.
Eigenvalues and Eigenvectors
(Definition)
A scalar λ is an eigenvalue of A if there exists a nonzero vector v such that:
Av=λv
The vector v is an eigenvector corresponding to λ.
Interpretation: Eigenvectors are the special directions where A acts like pure scaling,no rotation, just stretch or compression by factor λ.
(Finding Eigenvalues)
Rewrite Av=λv as:
(A−λI)v=0
For a nontrivial solution to exist, A−λI must be singular:
det(A−λI)=0
This is the characteristic equation. Expanding it gives a polynomial of degree n in λ,the characteristic polynomial.
Two matrices A and B are similar if there exists an invertible matrix P such that:
B=P−1AP
Interpretation: Similar matrices represent the same linear transformation in different bases. They have the same intrinsic properties but different coordinate representations.
(Properties Preserved by Similarity)
If A and B are similar, they share:
Determinant:det(B)=det(A)
Trace:tr(B)=tr(A) (sum of diagonal entries)
Eigenvalues: Same characteristic polynomial, same eigenvalues (with multiplicity)
Rank:rank(B)=rank(A)
Invertibility:A invertible ⟺B invertible
Why? All these properties are basis-independent,they depend only on the transformation itself, not the coordinate system.
(Diagonalization as Similarity)
When A=PDP−1, we’re saying A is similar to a diagonal matrix D.
Key insight: Diagonalizable matrices are exactly those that are similar to diagonal matrices. The “nicest” matrices are those you can represent diagonally in some basis.
When Is a Matrix Diagonalizable?
(Sufficient Condition: Distinct Eigenvalues)
Theorem: If an n×n matrix has n distinct eigenvalues, it is diagonalizable.
Why? Eigenvectors from different eigenvalues are automatically linearly independent. So n distinct eigenvalues gives you n independent eigenvectors.
Example:A=[1221] has eigenvalues 3 and −1 (distinct), so it’s diagonalizable.
(Repeated Eigenvalues: It Depends)
When eigenvalues repeat, diagonalizability depends on whether there are enough eigenvectors.
Algebraic vs Geometric Multiplicity:
For an eigenvalue λ:
Algebraic multiplicity: How many times λ appears as a root of the characteristic polynomial
Geometric multiplicity:dim(ker(A−λI)),the number of linearly independent eigenvectors for λ
Computing this is far easier than multiplying A ten times!
(Long-Term Behavior)
As k→∞, Ak is dominated by the largest eigenvalue (in absolute value).
If ∣λmax∣<1: Ak→0 (everything decays)
If ∣λmax∣=1: Bounded behavior (might oscillate)
If ∣λmax∣>1: Ak explodes (growth along dominant eigenvector)
Application: In Markov chains, Leslie models, discrete dynamical systems,eigenvalues control the long-term fate.
Change of Basis Perspective
(What P Does)
The matrix P performs a change of basis from the standard basis to the eigenvector basis.
In the eigenvector basis:
Coordinates: [x]B=P−1x
Transformation: [A]B=D (diagonal!)
The transformation is just scaling along each eigenvector direction
In the standard basis:
Coordinates: x
Transformation: A
The transformation looks complicated because we’re using the “wrong” coordinates
The diagonalization formula:
A=PDP−1
can be read as:
P−1: Convert from standard to eigenvector basis
D: Apply the simple diagonal transformation
P: Convert back to standard basis
Applications
(Differential Equations)
The system dtdx=Ax has solution:
x(t)=eAtx(0)
If A=PDP−1:
eAt=PeDtP−1=Peλ1t⋱eλntP−1
The eigenvectors give you the “modes” of the system, and eigenvalues tell you whether each mode grows or decays.
(Fibonacci Numbers)
The Fibonacci recurrence Fn+1=Fn+Fn−1 can be written as:
[Fn+1Fn]=[1110][FnFn−1]
Diagonalizing A=[1110] gives eigenvalues ϕ=21+5 (golden ratio) and ϕ^=21−5.
This leads to Binet’s formula:
Fn=5ϕn−ϕ^n
Summary: The Diagonalization Playbook
To diagonalize an n×n matrix A:
Find eigenvalues: Solve det(A−λI)=0
Find eigenvectors: For each λ, solve (A−λI)v=0
Check linear independence: You need n independent eigenvectors
Build P and D:
P=[v1∣⋯∣vn] (eigenvectors as columns)
D=diag(λ1,…,λn) (eigenvalues in matching order)
Verify:AP=PD
If you can’t find n independent eigenvectors, the matrix is not diagonalizable,but you might still use the Jordan normal form (a nearly-diagonal form with 1’s above some diagonal entries).
Diagonalization reveals the coordinate system where a transformation is simplest,just scaling, no mixing. It’s the key to understanding matrix powers, exponentials, and the long-term behavior of linear systems.