| |

7: Kernel and Image

Every linear transformation has two fundamental subspaces attached to it: the kernel (what gets destroyed) and the image (what gets produced). Understanding these reveals what a transformation does geometrically,where it collapses, where it stretches, and what it can reach.

The Image

(Image of a Linear Transformation)

Let T:Vโ†’WT: V \to W be a linear transformation. The image (or range) of TT is the set of all outputs:

im(T)={T(v)โˆฃvโˆˆV}={wโˆˆWโˆฃw=T(v)ย forย someย vโˆˆV}\text{im}(T) = \{T(v) \mid v \in V\} = \{w \in W \mid w = T(v) \text{ for some } v \in V\}

Key fact: The image is always a subspace of WW.

Proof:

  1. T(0)=0T(\mathbf{0}) = \mathbf{0}, so 0โˆˆim(T)\mathbf{0} \in \text{im}(T)
  2. If w1=T(v1)w_1 = T(v_1) and w2=T(v2)w_2 = T(v_2), then w1+w2=T(v1)+T(v2)=T(v1+v2)โˆˆim(T)w_1 + w_2 = T(v_1) + T(v_2) = T(v_1 + v_2) \in \text{im}(T)
  3. If w=T(v)w = T(v), then cw=cT(v)=T(cv)โˆˆim(T)cw = cT(v) = T(cv) \in \text{im}(T)

(Image for Matrices: Column Space)

For a matrix transformation T(x)=AxT(\mathbf{x}) = A\mathbf{x} where AA is mร—nm \times n:

im(T)=col(A)=span{columnsย ofย A}\text{im}(T) = \text{col}(A) = \text{span}\{\text{columns of } A\}

Why? Write A=[a1โˆฃa2โˆฃโ‹ฏโˆฃan]A = [\mathbf{a}_1 \mid \mathbf{a}_2 \mid \cdots \mid \mathbf{a}_n] in column form. Then:

Ax=x1a1+x2a2+โ‹ฏ+xnanA\mathbf{x} = x_1\mathbf{a}_1 + x_2\mathbf{a}_2 + \cdots + x_n\mathbf{a}_n

Every output is a linear combination of the columns,so the image is exactly the span of the columns.

Interpretation: The column space tells you what vectors can be reached by the transformation. If bโˆˆcol(A)\mathbf{b} \in \text{col}(A), then Ax=bA\mathbf{x} = \mathbf{b} has a solution. If bโˆ‰col(A)\mathbf{b} \notin \text{col}(A), the system is inconsistent.


The Kernel

(Kernel of a Linear Transformation)

Let T:Vโ†’WT: V \to W be a linear transformation. The kernel (or null space) of TT is the set of all inputs that map to zero:

kerโก(T)={vโˆˆVโˆฃT(v)=0}\ker(T) = \{v \in V \mid T(v) = \mathbf{0}\}

Key fact: The kernel is always a subspace of VV.

Proof:

  1. T(0)=0T(\mathbf{0}) = \mathbf{0}, so 0โˆˆkerโก(T)\mathbf{0} \in \ker(T)
  2. If T(v1)=0T(v_1) = \mathbf{0} and T(v2)=0T(v_2) = \mathbf{0}, then T(v1+v2)=T(v1)+T(v2)=0T(v_1 + v_2) = T(v_1) + T(v_2) = \mathbf{0}
  3. If T(v)=0T(v) = \mathbf{0}, then T(cv)=cT(v)=c0=0T(cv) = cT(v) = c\mathbf{0} = \mathbf{0}

(Kernel for Matrices: Null Space)

For a matrix transformation T(x)=AxT(\mathbf{x}) = A\mathbf{x}:

kerโก(T)=null(A)={xโˆˆRnโˆฃAx=0}\ker(T) = \text{null}(A) = \{\mathbf{x} \in \mathbb{R}^n \mid A\mathbf{x} = \mathbf{0}\}

The null space is the solution set to the homogeneous system Ax=0A\mathbf{x} = \mathbf{0}.

Interpretation: The kernel captures all the redundancy in the transformation,the directions that get completely flattened to nothing.


(Computing the Kernel)

To find a basis for kerโก(A)\ker(A):

  1. Row reduce AA to rref(A)\text{rref}(A)
  2. Identify the free variables (columns without pivots)
  3. For each free variable, set it to 11 and all others to 00, then solve
  4. These solutions form a basis for the kernel

Example: Find kerโก(A)\ker(A) for

A=[1210240212โˆ’12]A = \begin{bmatrix} 1 & 2 & 1 & 0 \\ 2 & 4 & 0 & 2 \\ 1 & 2 & -1 & 2 \end{bmatrix}

Row reduce:

rref(A)=[1201001โˆ’10000]\text{rref}(A) = \begin{bmatrix} 1 & 2 & 0 & 1 \\ 0 & 0 & 1 & -1 \\ 0 & 0 & 0 & 0 \end{bmatrix}

Pivots in columns 1 and 3. Free variables: x2x_2 and x4x_4.

Setting x2=1,x4=0x_2 = 1, x_4 = 0:

  • From row 2: x3=0x_3 = 0
  • From row 1: x1=โˆ’2x_1 = -2
  • Solution: (โˆ’2,1,0,0)(-2, 1, 0, 0)

Setting x2=0,x4=1x_2 = 0, x_4 = 1:

  • From row 2: x3=1x_3 = 1
  • From row 1: x1=โˆ’1x_1 = -1
  • Solution: (โˆ’1,0,1,1)(-1, 0, 1, 1)

Basis for kerโก(A)\ker(A):

{[โˆ’2100],[โˆ’1011]}\left\{ \begin{bmatrix} -2 \\ 1 \\ 0 \\ 0 \end{bmatrix}, \begin{bmatrix} -1 \\ 0 \\ 1 \\ 1 \end{bmatrix} \right\}

The kernel is a 2-dimensional subspace of R4\mathbb{R}^4.


The Rank-Nullity Theorem

(Rank and Nullity)

For a linear transformation T:Vโ†’WT: V \to W:

  • The rank of TT is dimโก(im(T))\dim(\text{im}(T))
  • The nullity of TT is dimโก(kerโก(T))\dim(\ker(T))

For a matrix AA, rank(A)=dimโก(col(A))(A) = \dim(\text{col}(A)) equals the number of pivot columns.


(The Rank-Nullity Theorem)

For any linear transformation T:Vโ†’WT: V \to W where VV is finite-dimensional:

dimโก(V)=rank(T)+nullity(T)\dim(V) = \text{rank}(T) + \text{nullity}(T)

For an mร—nm \times n matrix AA:

n=rank(A)+dimโก(kerโก(A))n = \text{rank}(A) + \dim(\ker(A))

Interpretation: The dimension of the input space is split between:

  • Dimensions that survive and get mapped somewhere meaningful (the rank)
  • Dimensions that get crushed to zero (the nullity)

This is conservation of dimension,you canโ€™t create or destroy dimensions, only redistribute them.


(Proof Sketch)

Let {v1,โ€ฆ,vk}\{v_1, \ldots, v_k\} be a basis for kerโก(T)\ker(T). Extend this to a basis {v1,โ€ฆ,vk,u1,โ€ฆ,ur}\{v_1, \ldots, v_k, u_1, \ldots, u_r\} for VV.

Claim: {T(u1),โ€ฆ,T(ur)}\{T(u_1), \ldots, T(u_r)\} is a basis for im(T)\text{im}(T).

Spanning: Any w=T(v)โˆˆim(T)w = T(v) \in \text{im}(T) can be written as:

v=c1v1+โ‹ฏ+ckvk+d1u1+โ‹ฏ+drurv = c_1v_1 + \cdots + c_kv_k + d_1u_1 + \cdots + d_ru_r

so T(v)=d1T(u1)+โ‹ฏ+drT(ur)T(v) = d_1T(u_1) + \cdots + d_rT(u_r) (since T(vi)=0T(v_i) = \mathbf{0}).

Linear independence: If d1T(u1)+โ‹ฏ+drT(ur)=0d_1T(u_1) + \cdots + d_rT(u_r) = \mathbf{0}, then T(d1u1+โ‹ฏ+drur)=0T(d_1u_1 + \cdots + d_ru_r) = \mathbf{0}, so d1u1+โ‹ฏ+drurโˆˆkerโก(T)d_1u_1 + \cdots + d_ru_r \in \ker(T). This means itโ€™s a combination of v1,โ€ฆ,vkv_1, \ldots, v_k, forcing all di=0d_i = 0 by independence.

Thus dimโก(V)=k+r=nullity(T)+rank(T)\dim(V) = k + r = \text{nullity}(T) + \text{rank}(T).


Injectivity and Surjectivity Revisited

The kernel and image directly characterize injectivity and surjectivity:

(Kernel and Injectivity)

Tย isย injectiveย (one-to-one)โ€…โ€ŠโŸบโ€…โ€Škerโก(T)={0}T \text{ is injective (one-to-one)} \iff \ker(T) = \{\mathbf{0}\}

Why? If kerโก(T)={0}\ker(T) = \{\mathbf{0}\} and T(v1)=T(v2)T(v_1) = T(v_2), then T(v1โˆ’v2)=0T(v_1 - v_2) = \mathbf{0}, so v1โˆ’v2โˆˆkerโก(T)v_1 - v_2 \in \ker(T), meaning v1=v2v_1 = v_2.

For matrices: AA is injective iff null(A)={0}\text{null}(A) = \{\mathbf{0}\} iff thereโ€™s a pivot in every column iff rank(A)=n(A) = n.


(Image and Surjectivity)

T:Vโ†’Wย isย surjectiveย (onto)โ€…โ€ŠโŸบโ€…โ€Šim(T)=WT: V \to W \text{ is surjective (onto)} \iff \text{im}(T) = W

For matrices: AA (mร—nm \times n) is surjective iff col(A)=Rm\text{col}(A) = \mathbb{R}^m iff thereโ€™s a pivot in every row iff rank(A)=m(A) = m.


(The Counting Argument)

For an mร—nm \times n matrix AA:

  • Injective requires rank(A)=n(A) = n (all columns are pivots)
  • Surjective requires rank(A)=m(A) = m (all rows have pivots)
  • Bijective requires rank(A)=m=n(A) = m = n (square and full rank)

Since rank(A)โ‰คminโก(m,n)(A) \leq \min(m, n):

  • If n>mn > m: Cannot be injective (not enough room for pivots in every column)
  • If m>nm > n: Cannot be surjective (not enough columns to pivot in every row)

The Row Space

Thereโ€™s a fourth fundamental subspace hiding in every matrix.

(Row Space)

The row space of AA, denoted row(A)\text{row}(A), is the span of the rows of AA:

row(A)=span{rowsย ofย A}=col(AT)\text{row}(A) = \text{span}\{\text{rows of } A\} = \text{col}(A^T)

This is a subspace of Rn\mathbb{R}^n (same as the domain).


(Key Property)

Row operations preserve the row space. So:

row(A)=row(rref(A))\text{row}(A) = \text{row}(\text{rref}(A))

This means the nonzero rows of rref(A)\text{rref}(A) form a basis for row(A)\text{row}(A).

Note: This is different from column space! Row operations change column relationships but preserve row span.


(Row Space and Null Space are Orthogonal)

For any matrix AA:

row(A)โŠฅkerโก(A)\text{row}(A) \perp \ker(A)

Every vector in the row space is orthogonal to every vector in the null space.

Why? If xโˆˆkerโก(A)\mathbf{x} \in \ker(A), then Ax=0A\mathbf{x} = \mathbf{0}. This means each row ri\mathbf{r}_i satisfies riโ‹…x=0\mathbf{r}_i \cdot \mathbf{x} = 0. So x\mathbf{x} is orthogonal to every row, hence to every linear combination of rows.


The Four Fundamental Subspaces

Every mร—nm \times n matrix AA defines four fundamental subspaces:

SubspaceNotationLives inDimension
Column space (image)col(A)\text{col}(A)Rm\mathbb{R}^mrr
Null space (kernel)kerโก(A)\ker(A)Rn\mathbb{R}^nnโˆ’rn - r
Row spacerow(A)\text{row}(A)Rn\mathbb{R}^nrr
Left null spacekerโก(AT)\ker(A^T)Rm\mathbb{R}^mmโˆ’rm - r

where r=rank(A)r = \text{rank}(A).


(Orthogonal Complements)

These subspaces pair up as orthogonal complements:

  • In Rn\mathbb{R}^n: row(A)โŠฅkerโก(A)\text{row}(A) \perp \ker(A), and together they span Rn\mathbb{R}^n
  • In Rm\mathbb{R}^m: col(A)โŠฅkerโก(AT)\text{col}(A) \perp \ker(A^T), and together they span Rm\mathbb{R}^m

Dimensions check:

  • dimโก(row(A))+dimโก(kerโก(A))=r+(nโˆ’r)=n\dim(\text{row}(A)) + \dim(\ker(A)) = r + (n - r) = n โœ“
  • dimโก(col(A))+dimโก(kerโก(AT))=r+(mโˆ’r)=m\dim(\text{col}(A)) + \dim(\ker(A^T)) = r + (m - r) = m โœ“

Geometric Summary

The kernel and image tell you exactly how a transformation reshapes space:

The image is the โ€œshadowโ€ of the domain in the codomain,all the places you can reach. Its dimension (the rank) is how many independent directions survive the transformation.

The kernel is the โ€œblind spotโ€ in the domain,all the inputs that become invisible. Its dimension (the nullity) is how many directions get flattened.

DomainโŸdimโก=n=KernelโŸcrushedโŠ•ComplementโŸsurvivesโ†’Image\underbrace{\text{Domain}}_{\dim = n} = \underbrace{\text{Kernel}}_{\text{crushed}} \oplus \underbrace{\text{Complement}}_{\text{survives} \to \text{Image}}

The rank-nullity theorem says the domain splits cleanly: some directions die, some survive, and nothing is lost in the accounting.