7: Kernel and Image

Every linear transformation has two fundamental subspaces attached to it: the kernel (what gets destroyed) and the image (what gets produced). Understanding these reveals what a transformation does geometrically,where it collapses, where it stretches, and what it can reach.

The Image

(Image of a Linear Transformation)

Let T:VWT: V \to W be a linear transformation. The image (or range) of TT is the set of all outputs:

im(T)={T(v)vV}={wWw=T(v) for some vV}\text{im}(T) = \{T(v) \mid v \in V\} = \{w \in W \mid w = T(v) \text{ for some } v \in V\}

Key fact: The image is always a subspace of WW.

Proof:

  1. T(0)=0T(\mathbf{0}) = \mathbf{0}, so 0im(T)\mathbf{0} \in \text{im}(T)
  2. If w1=T(v1)w_1 = T(v_1) and w2=T(v2)w_2 = T(v_2), then w1+w2=T(v1)+T(v2)=T(v1+v2)im(T)w_1 + w_2 = T(v_1) + T(v_2) = T(v_1 + v_2) \in \text{im}(T)
  3. If w=T(v)w = T(v), then cw=cT(v)=T(cv)im(T)cw = cT(v) = T(cv) \in \text{im}(T)

(Image for Matrices: Column Space)

For a matrix transformation T(x)=AxT(\mathbf{x}) = A\mathbf{x} where AA is m×nm \times n:

im(T)=col(A)=span{columns of A}\text{im}(T) = \text{col}(A) = \text{span}\{\text{columns of } A\}

Why? Write A=[a1a2an]A = [\mathbf{a}_1 \mid \mathbf{a}_2 \mid \cdots \mid \mathbf{a}_n] in column form. Then:

Ax=x1a1+x2a2++xnanA\mathbf{x} = x_1\mathbf{a}_1 + x_2\mathbf{a}_2 + \cdots + x_n\mathbf{a}_n

Every output is a linear combination of the columns,so the image is exactly the span of the columns.

Interpretation: The column space tells you what vectors can be reached by the transformation. If bcol(A)\mathbf{b} \in \text{col}(A), then Ax=bA\mathbf{x} = \mathbf{b} has a solution. If bcol(A)\mathbf{b} \notin \text{col}(A), the system is inconsistent.


The Kernel

(Kernel of a Linear Transformation)

Let T:VWT: V \to W be a linear transformation. The kernel (or null space) of TT is the set of all inputs that map to zero:

ker(T)={vVT(v)=0}\ker(T) = \{v \in V \mid T(v) = \mathbf{0}\}

Key fact: The kernel is always a subspace of VV.

Proof:

  1. T(0)=0T(\mathbf{0}) = \mathbf{0}, so 0ker(T)\mathbf{0} \in \ker(T)
  2. If T(v1)=0T(v_1) = \mathbf{0} and T(v2)=0T(v_2) = \mathbf{0}, then T(v1+v2)=T(v1)+T(v2)=0T(v_1 + v_2) = T(v_1) + T(v_2) = \mathbf{0}
  3. If T(v)=0T(v) = \mathbf{0}, then T(cv)=cT(v)=c0=0T(cv) = cT(v) = c\mathbf{0} = \mathbf{0}

(Kernel for Matrices: Null Space)

For a matrix transformation T(x)=AxT(\mathbf{x}) = A\mathbf{x}:

ker(T)=null(A)={xRnAx=0}\ker(T) = \text{null}(A) = \{\mathbf{x} \in \mathbb{R}^n \mid A\mathbf{x} = \mathbf{0}\}

The null space is the solution set to the homogeneous system Ax=0A\mathbf{x} = \mathbf{0}.

Interpretation: The kernel captures all the redundancy in the transformation,the directions that get completely flattened to nothing.


(Computing the Kernel)

To find a basis for ker(A)\ker(A):

  1. Row reduce AA to rref(A)\text{rref}(A)
  2. Identify the free variables (columns without pivots)
  3. For each free variable, set it to 11 and all others to 00, then solve
  4. These solutions form a basis for the kernel

Example: Find ker(A)\ker(A) for

A=[121024021212]A = \begin{bmatrix} 1 & 2 & 1 & 0 \\ 2 & 4 & 0 & 2 \\ 1 & 2 & -1 & 2 \end{bmatrix}

Row reduce:

rref(A)=[120100110000]\text{rref}(A) = \begin{bmatrix} 1 & 2 & 0 & 1 \\ 0 & 0 & 1 & -1 \\ 0 & 0 & 0 & 0 \end{bmatrix}

Pivots in columns 1 and 3. Free variables: x2x_2 and x4x_4.

Setting x2=1,x4=0x_2 = 1, x_4 = 0:

  • From row 2: x3=0x_3 = 0
  • From row 1: x1=2x_1 = -2
  • Solution: (2,1,0,0)(-2, 1, 0, 0)

Setting x2=0,x4=1x_2 = 0, x_4 = 1:

  • From row 2: x3=1x_3 = 1
  • From row 1: x1=1x_1 = -1
  • Solution: (1,0,1,1)(-1, 0, 1, 1)

Basis for ker(A)\ker(A):

{[2100],[1011]}\left\{ \begin{bmatrix} -2 \\ 1 \\ 0 \\ 0 \end{bmatrix}, \begin{bmatrix} -1 \\ 0 \\ 1 \\ 1 \end{bmatrix} \right\}

The kernel is a 2-dimensional subspace of R4\mathbb{R}^4.


The Rank-Nullity Theorem

(Rank and Nullity)

For a linear transformation T:VWT: V \to W:

  • The rank of TT is dim(im(T))\dim(\text{im}(T))
  • The nullity of TT is dim(ker(T))\dim(\ker(T))

For a matrix AA, rank(A)=dim(col(A))(A) = \dim(\text{col}(A)) equals the number of pivot columns.


(The Rank-Nullity Theorem)

For any linear transformation T:VWT: V \to W where VV is finite-dimensional:

dim(V)=rank(T)+nullity(T)\dim(V) = \text{rank}(T) + \text{nullity}(T)

For an m×nm \times n matrix AA:

n=rank(A)+dim(ker(A))n = \text{rank}(A) + \dim(\ker(A))

Interpretation: The dimension of the input space is split between:

  • Dimensions that survive and get mapped somewhere meaningful (the rank)
  • Dimensions that get crushed to zero (the nullity)

This is conservation of dimension,you can’t create or destroy dimensions, only redistribute them.


(Proof Sketch)

Let {v1,,vk}\{v_1, \ldots, v_k\} be a basis for ker(T)\ker(T). Extend this to a basis {v1,,vk,u1,,ur}\{v_1, \ldots, v_k, u_1, \ldots, u_r\} for VV.

Claim: {T(u1),,T(ur)}\{T(u_1), \ldots, T(u_r)\} is a basis for im(T)\text{im}(T).

Spanning: Any w=T(v)im(T)w = T(v) \in \text{im}(T) can be written as:

v=c1v1++ckvk+d1u1++drurv = c_1v_1 + \cdots + c_kv_k + d_1u_1 + \cdots + d_ru_r

so T(v)=d1T(u1)++drT(ur)T(v) = d_1T(u_1) + \cdots + d_rT(u_r) (since T(vi)=0T(v_i) = \mathbf{0}).

Linear independence: If d1T(u1)++drT(ur)=0d_1T(u_1) + \cdots + d_rT(u_r) = \mathbf{0}, then T(d1u1++drur)=0T(d_1u_1 + \cdots + d_ru_r) = \mathbf{0}, so d1u1++drurker(T)d_1u_1 + \cdots + d_ru_r \in \ker(T). This means it’s a combination of v1,,vkv_1, \ldots, v_k, forcing all di=0d_i = 0 by independence.

Thus dim(V)=k+r=nullity(T)+rank(T)\dim(V) = k + r = \text{nullity}(T) + \text{rank}(T).


Injectivity and Surjectivity Revisited

The kernel and image directly characterize injectivity and surjectivity:

(Kernel and Injectivity)

T is injective (one-to-one)    ker(T)={0}T \text{ is injective (one-to-one)} \iff \ker(T) = \{\mathbf{0}\}

Why? If ker(T)={0}\ker(T) = \{\mathbf{0}\} and T(v1)=T(v2)T(v_1) = T(v_2), then T(v1v2)=0T(v_1 - v_2) = \mathbf{0}, so v1v2ker(T)v_1 - v_2 \in \ker(T), meaning v1=v2v_1 = v_2.

For matrices: AA is injective iff null(A)={0}\text{null}(A) = \{\mathbf{0}\} iff there’s a pivot in every column iff rank(A)=n(A) = n.


(Image and Surjectivity)

T:VW is surjective (onto)    im(T)=WT: V \to W \text{ is surjective (onto)} \iff \text{im}(T) = W

For matrices: AA (m×nm \times n) is surjective iff col(A)=Rm\text{col}(A) = \mathbb{R}^m iff there’s a pivot in every row iff rank(A)=m(A) = m.


(The Counting Argument)

For an m×nm \times n matrix AA:

  • Injective requires rank(A)=n(A) = n (all columns are pivots)
  • Surjective requires rank(A)=m(A) = m (all rows have pivots)
  • Bijective requires rank(A)=m=n(A) = m = n (square and full rank)

Since rank(A)min(m,n)(A) \leq \min(m, n):

  • If n>mn > m: Cannot be injective (not enough room for pivots in every column)
  • If m>nm > n: Cannot be surjective (not enough columns to pivot in every row)

The Row Space

There’s a fourth fundamental subspace hiding in every matrix.

(Row Space)

The row space of AA, denoted row(A)\text{row}(A), is the span of the rows of AA:

row(A)=span{rows of A}=col(AT)\text{row}(A) = \text{span}\{\text{rows of } A\} = \text{col}(A^T)

This is a subspace of Rn\mathbb{R}^n (same as the domain).


(Key Property)

Row operations preserve the row space. So:

row(A)=row(rref(A))\text{row}(A) = \text{row}(\text{rref}(A))

This means the nonzero rows of rref(A)\text{rref}(A) form a basis for row(A)\text{row}(A).

Note: This is different from column space! Row operations change column relationships but preserve row span.


(Row Space and Null Space are Orthogonal)

For any matrix AA:

row(A)ker(A)\text{row}(A) \perp \ker(A)

Every vector in the row space is orthogonal to every vector in the null space.

Why? If xker(A)\mathbf{x} \in \ker(A), then Ax=0A\mathbf{x} = \mathbf{0}. This means each row ri\mathbf{r}_i satisfies rix=0\mathbf{r}_i \cdot \mathbf{x} = 0. So x\mathbf{x} is orthogonal to every row, hence to every linear combination of rows.


The Four Fundamental Subspaces

Every m×nm \times n matrix AA defines four fundamental subspaces:

SubspaceNotationLives inDimension
Column space (image)col(A)\text{col}(A)Rm\mathbb{R}^mrr
Null space (kernel)ker(A)\ker(A)Rn\mathbb{R}^nnrn - r
Row spacerow(A)\text{row}(A)Rn\mathbb{R}^nrr
Left null spaceker(AT)\ker(A^T)Rm\mathbb{R}^mmrm - r

where r=rank(A)r = \text{rank}(A).


(Orthogonal Complements)

These subspaces pair up as orthogonal complements:

  • In Rn\mathbb{R}^n: row(A)ker(A)\text{row}(A) \perp \ker(A), and together they span Rn\mathbb{R}^n
  • In Rm\mathbb{R}^m: col(A)ker(AT)\text{col}(A) \perp \ker(A^T), and together they span Rm\mathbb{R}^m

Dimensions check:

  • dim(row(A))+dim(ker(A))=r+(nr)=n\dim(\text{row}(A)) + \dim(\ker(A)) = r + (n - r) = n
  • dim(col(A))+dim(ker(AT))=r+(mr)=m\dim(\text{col}(A)) + \dim(\ker(A^T)) = r + (m - r) = m

Geometric Summary

The kernel and image tell you exactly how a transformation reshapes space:

The image is the “shadow” of the domain in the codomain,all the places you can reach. Its dimension (the rank) is how many independent directions survive the transformation.

The kernel is the “blind spot” in the domain,all the inputs that become invisible. Its dimension (the nullity) is how many directions get flattened.

Domaindim=n=KernelcrushedComplementsurvivesImage\underbrace{\text{Domain}}_{\dim = n} = \underbrace{\text{Kernel}}_{\text{crushed}} \oplus \underbrace{\text{Complement}}_{\text{survives} \to \text{Image}}

The rank-nullity theorem says the domain splits cleanly: some directions die, some survive, and nothing is lost in the accounting.