Every linear transformation has two fundamental subspaces attached to it: the kernel (what gets destroyed) and the image (what gets produced). Understanding these reveals what a transformation does geometrically,where it collapses, where it stretches, and what it can reach.
The Image
(Image of a Linear Transformation)
Let T:V→W be a linear transformation. The image (or range) of T is the set of all outputs:
im(T)={T(v)∣v∈V}={w∈W∣w=T(v) for some v∈V}
Key fact: The image is always a subspace of W.
Proof:
T(0)=0, so 0∈im(T)
If w1=T(v1) and w2=T(v2), then w1+w2=T(v1)+T(v2)=T(v1+v2)∈im(T)
If w=T(v), then cw=cT(v)=T(cv)∈im(T)
(Image for Matrices: Column Space)
For a matrix transformation T(x)=Ax where A is m×n:
im(T)=col(A)=span{columns of A}
Why? Write A=[a1∣a2∣⋯∣an] in column form. Then:
Ax=x1a1+x2a2+⋯+xnan
Every output is a linear combination of the columns,so the image is exactly the span of the columns.
Interpretation: The column space tells you what vectors can be reached by the transformation. If b∈col(A), then Ax=b has a solution. If b∈/col(A), the system is inconsistent.
The Kernel
(Kernel of a Linear Transformation)
Let T:V→W be a linear transformation. The kernel (or null space) of T is the set of all inputs that map to zero:
ker(T)={v∈V∣T(v)=0}
Key fact: The kernel is always a subspace of V.
Proof:
T(0)=0, so 0∈ker(T)
If T(v1)=0 and T(v2)=0, then T(v1+v2)=T(v1)+T(v2)=0
If T(v)=0, then T(cv)=cT(v)=c0=0
(Kernel for Matrices: Null Space)
For a matrix transformation T(x)=Ax:
ker(T)=null(A)={x∈Rn∣Ax=0}
The null space is the solution set to the homogeneous system Ax=0.
Interpretation: The kernel captures all the redundancy in the transformation,the directions that get completely flattened to nothing.
(Computing the Kernel)
To find a basis for ker(A):
Row reduce A to rref(A)
Identify the free variables (columns without pivots)
For each free variable, set it to 1 and all others to 0, then solve
These solutions form a basis for the kernel
Example: Find ker(A) for
A=12124210−1022
Row reduce:
rref(A)=1002000101−10
Pivots in columns 1 and 3. Free variables: x2 and x4.
Setting x2=1,x4=0:
From row 2: x3=0
From row 1: x1=−2
Solution: (−2,1,0,0)
Setting x2=0,x4=1:
From row 2: x3=1
From row 1: x1=−1
Solution: (−1,0,1,1)
Basis for ker(A):
⎩⎨⎧−2100,−1011⎭⎬⎫
The kernel is a 2-dimensional subspace of R4.
The Rank-Nullity Theorem
(Rank and Nullity)
For a linear transformation T:V→W:
The rank of T is dim(im(T))
The nullity of T is dim(ker(T))
For a matrix A, rank(A)=dim(col(A)) equals the number of pivot columns.
(The Rank-Nullity Theorem)
For any linear transformation T:V→W where V is finite-dimensional:
dim(V)=rank(T)+nullity(T)
For an m×n matrix A:
n=rank(A)+dim(ker(A))
Interpretation: The dimension of the input space is split between:
Dimensions that survive and get mapped somewhere meaningful (the rank)
Dimensions that get crushed to zero (the nullity)
This is conservation of dimension,you can’t create or destroy dimensions, only redistribute them.
(Proof Sketch)
Let {v1,…,vk} be a basis for ker(T). Extend this to a basis {v1,…,vk,u1,…,ur} for V.
Claim: {T(u1),…,T(ur)} is a basis for im(T).
Spanning: Any w=T(v)∈im(T) can be written as:
v=c1v1+⋯+ckvk+d1u1+⋯+drur
so T(v)=d1T(u1)+⋯+drT(ur) (since T(vi)=0).
Linear independence: If d1T(u1)+⋯+drT(ur)=0, then T(d1u1+⋯+drur)=0, so d1u1+⋯+drur∈ker(T). This means it’s a combination of v1,…,vk, forcing all di=0 by independence.
Thus dim(V)=k+r=nullity(T)+rank(T).
Injectivity and Surjectivity Revisited
The kernel and image directly characterize injectivity and surjectivity:
(Kernel and Injectivity)
T is injective (one-to-one)⟺ker(T)={0}
Why? If ker(T)={0} and T(v1)=T(v2), then T(v1−v2)=0, so v1−v2∈ker(T), meaning v1=v2.
For matrices:A is injective iff null(A)={0} iff there’s a pivot in every column iff rank(A)=n.
(Image and Surjectivity)
T:V→W is surjective (onto)⟺im(T)=W
For matrices:A (m×n) is surjective iff col(A)=Rm iff there’s a pivot in every row iff rank(A)=m.
(The Counting Argument)
For an m×n matrix A:
Injective requires rank(A)=n (all columns are pivots)
Surjective requires rank(A)=m (all rows have pivots)
Bijective requires rank(A)=m=n (square and full rank)
Since rank(A)≤min(m,n):
If n>m: Cannot be injective (not enough room for pivots in every column)
If m>n: Cannot be surjective (not enough columns to pivot in every row)
The Row Space
There’s a fourth fundamental subspace hiding in every matrix.
(Row Space)
The row space of A, denoted row(A), is the span of the rows of A:
row(A)=span{rows of A}=col(AT)
This is a subspace of Rn (same as the domain).
(Key Property)
Row operations preserve the row space. So:
row(A)=row(rref(A))
This means the nonzero rows of rref(A) form a basis for row(A).
Note: This is different from column space! Row operations change column relationships but preserve row span.
(Row Space and Null Space are Orthogonal)
For any matrix A:
row(A)⊥ker(A)
Every vector in the row space is orthogonal to every vector in the null space.
Why? If x∈ker(A), then Ax=0. This means each row ri satisfies ri⋅x=0. So x is orthogonal to every row, hence to every linear combination of rows.
The Four Fundamental Subspaces
Every m×n matrix A defines four fundamental subspaces:
Subspace
Notation
Lives in
Dimension
Column space (image)
col(A)
Rm
r
Null space (kernel)
ker(A)
Rn
n−r
Row space
row(A)
Rn
r
Left null space
ker(AT)
Rm
m−r
where r=rank(A).
(Orthogonal Complements)
These subspaces pair up as orthogonal complements:
In Rn: row(A)⊥ker(A), and together they span Rn
In Rm: col(A)⊥ker(AT), and together they span Rm
Dimensions check:
dim(row(A))+dim(ker(A))=r+(n−r)=n ✓
dim(col(A))+dim(ker(AT))=r+(m−r)=m ✓
Geometric Summary
The kernel and image tell you exactly how a transformation reshapes space:
The image is the “shadow” of the domain in the codomain,all the places you can reach. Its dimension (the rank) is how many independent directions survive the transformation.
The kernel is the “blind spot” in the domain,all the inputs that become invisible. Its dimension (the nullity) is how many directions get flattened.