12: Orthogonal Diagonalization and the Spectral Theorem
Symmetric matrices have a special property: they can always be diagonalized using orthogonal eigenvectors. This means you can write where has orthonormal columns. The “orthogonal” part is crucial,instead of computing a full matrix inverse (), you just transpose (). Geometrically, symmetric matrices can only stretch along perpendicular axes, never rotate. This is the Spectral Theorem, and it’s why symmetric matrices are the friendliest matrices in linear algebra.
The Spectral Theorem
(Statement)
Any symmetric matrix (meaning ) can be orthogonally diagonalized:
where:
- is diagonal (eigenvalues on the diagonal)
- is orthogonal (columns are orthonormal eigenvectors)
Key property: , so you transpose instead of inverting.
(Why This Matters)
Computational savings:
- Matrix inverse: operations
- Matrix transpose: operations
When , all the expensive inversion work disappears.
Geometric meaning: Rotate to the eigenvector basis (via ), scale along axes (via ), rotate back (via ).
Symmetric matrices cannot rotate,they only stretch along orthogonal axes. That’s why eigenvectors are automatically perpendicular.
Symmetric Matrices Are Special
(Why Eigenvectors Are Orthogonal)
Theorem: If is symmetric, eigenvectors from different eigenvalues are orthogonal.
Proof:
Let and where .
Since is symmetric, :
So:
Since , we must have . ✓
(Repeated Eigenvalues)
When eigenvalues repeat, the corresponding eigenspace might have dimension > 1. But you can always choose an orthonormal basis within that eigenspace using Gram-Schmidt.
Key insight: For symmetric matrices, there’s always enough “room” to find orthogonal eigenvectors, even when eigenvalues repeat. This is why the Spectral Theorem holds for all symmetric matrices.
(Symmetric Matrices Have Real Eigenvalues)
Theorem: All eigenvalues of a real symmetric matrix are real.
This is crucial,complex eigenvalues would break the geometric interpretation. The proof uses the fact that forces eigenvalues to equal their complex conjugates, making them real.
Orthogonal Matrices
(Definition)
A matrix is orthogonal if:
Equivalently: (transpose equals inverse).
In terms of columns: If , then the columns form an orthonormal set:
(Properties of Orthogonal Matrices)
1. Preserve lengths:
Proof:
2. Preserve angles:
3. Preserve dot products: Same as preserving angles.
4. Determinant is :
So .
(Geometric Interpretation)
Orthogonal matrices represent:
- Rotations:
- Reflections:
- Combinations: Rotations and reflections preserve lengths and angles,that’s exactly what orthogonal matrices do
Key distinction: Transpose reverses rotations/reflections, but doesn’t undo stretching. That’s why only works for orthogonal matrices (no stretching component).
The Spectral Decomposition
(The Formula)
For a symmetric matrix :
where are eigenvalues and are orthonormal eigenvectors.
Outer product form: Each term is a rank-1 projection matrix. The spectral decomposition says:
A symmetric matrix is a weighted sum of projections onto its eigenvectors, with weights equal to the eigenvalues.
(Example: Spectral Decomposition)
Find eigenvalues:
Eigenvalues: , .
Find eigenvectors:
For :
Eigenvector: , normalize:
For :
Eigenvector: , normalize:
Build and :
Verify:
Quadratic Forms and Principal Axes
(Quadratic Forms)
A quadratic form is an expression:
where is a symmetric matrix.
Example in :
(Geometric Interpretation: Ellipsoids)
The level set defines a quadric surface:
- Ellipsoid if all eigenvalues are positive
- Hyperboloid if eigenvalues have mixed signs
- Degenerate if any eigenvalue is zero
The eigenvectors of are the principal axes of this quadric,they point along the directions of maximum and minimum stretching.
(Diagonalizing Quadratic Forms)
Using , substitute :
In the eigenvector basis, the quadratic form has no cross terms,it’s just a sum of squares with weights .
Example: The ellipse (from our earlier ) becomes in the eigenvector coordinates, where axes are rotated by .
Why Transpose ≠ Inverse (Usually)
For a general matrix :
Why? Transpose reverses the direction of a transformation (rotations/reflections) but doesn’t undo stretching.
If stretches by factor in some direction, then:
- still stretches by in the corresponding direction (transpose doesn’t change singular values)
- compresses by factor (actually inverts the stretching)
When they’re equal: precisely when has no stretching component,when it’s a pure rotation or reflection. These are the orthogonal matrices.
Connection to SVD
(The Setup)
For a non-square matrix (), you can’t do eigendecomposition directly. But the matrix () is symmetric, so it has orthonormal eigenvectors.
These eigenvectors of are the right singular vectors of ,the optimal input directions.
(Singular Values from )
For an eigenvector of with eigenvalue :
If is normalized ():
The singular values are defined as:
where are the eigenvalues of .
(SVD Mechanics)
To build the Singular Value Decomposition :
- Compute (symmetric, )
- Find eigenvalues (rank )
- Compute singular values:
- Find eigenvectors of → normalize → columns of
- Apply to eigenvectors: → columns of
- Assemble:
Key formulas:
Special Matrices
(Involutions: )
Matrices satisfying are called involutions,applying them twice returns to the original.
Examples:
- Reflections across a line or plane
- Swap matrices (permutations that swap pairs)
For symmetric involutions, eigenvalues are .
(Finite Order: )
Matrices where for some integer have finite order.
Examples:
- Rotation by (order )
- Cyclic permutations
Eigenvalues are -th roots of unity: for .
Applications
(Principal Component Analysis)
PCA finds the directions of maximum variance in data. Given a covariance matrix (symmetric!), the eigenvectors are the principal components, and eigenvalues measure variance along each component.
The spectral decomposition directly gives you the optimal low-rank approximation.
(Vibrational Modes)
In physics, symmetric matrices appear as stiffness or inertia matrices. Eigenvectors represent vibrational modes (standing waves), and eigenvalues give frequencies.
(Stability Analysis)
For the system where is symmetric:
- Negative eigenvalues → stable (exponential decay)
- Positive eigenvalues → unstable (exponential growth)
- Zero eigenvalue → neutral stability
The orthogonal eigenvectors decouple the system into independent modes.
Summary: Why Symmetric Matrices Are Perfect
Symmetric matrices are the gold standard because:
- Always diagonalizable (Spectral Theorem)
- Real eigenvalues (no complex numbers)
- Orthogonal eigenvectors (automatic perpendicularity)
- Efficient inversion ( instead of )
- Geometric clarity (stretch along perpendicular axes, no rotation)
- Numerical stability (orthogonal transformations preserve conditioning)
The Spectral Theorem says symmetric matrices live in the simplest possible world: they’re diagonal in the right coordinate system, and finding that system is straightforward. Whenever you see , you know the geometry is clean, the computation is efficient, and the eigenvectors will behave.