Matrices aren’t just grids of numbers,they’re objects with their own arithmetic. Matrix addition and scalar multiplication work entry-by-entry, but matrix multiplication is something stranger and more powerful: it’s function composition in disguise.
Matrix Addition
(Matrix Addition)
Two matrices A and B can be added if and only if they have the same dimensions. If A and B are both m×n, their sum is defined entry-by-entry:
(A+B)ij=Aij+Bij
Example:
[1324]+[5768]=[610812]
(Properties of Matrix Addition)
For matrices A, B, C of the same size:
Commutativity: A+B=B+A
Associativity: (A+B)+C=A+(B+C)
Identity: A+O=A where O is the zero matrix
Inverse: A+(−A)=O
Matrix addition inherits all the nice properties of real number addition.
Scalar Multiplication
(Scalar Multiplication)
For a scalar c∈R and matrix A:
(cA)ij=c⋅Aij
Every entry gets multiplied by c.
Example:
3[1324]=[39612]
(Properties of Scalar Multiplication)
For scalars c,d and matrices A,B:
Associativity: c(dA)=(cd)A
Distributivity over matrix addition: c(A+B)=cA+cB
Distributivity over scalar addition: (c+d)A=cA+dA
Identity: 1⋅A=A
Matrix Multiplication
Matrix multiplication is where things get interesting. Unlike addition, it’s not entry-by-entry,it encodes something deeper.
(When Can You Multiply?)
You can compute AB if and only if:
(columns of A)=(rows of B)
If A is m×n and B is n×p, then AB is m×p.
m×nA⋅n×pB=m×pAB
The inner dimensions must match; the outer dimensions give the result size.
Intuition: “Rotate then reflect” is not the same as “reflect then rotate.”
Properties of Matrix Multiplication
(What DOES Hold)
For matrices of compatible sizes:
Associativity: (AB)C=A(BC)
Distributivity: A(B+C)=AB+AC and (A+B)C=AC+BC
Scalar compatibility: c(AB)=(cA)B=A(cB)
Identity: ImA=A=AIn for A of size m×n
Associativity is remarkable,it says we can chain transformations without worrying about grouping. This is why we can write ABC without parentheses.
(The Identity Matrix)
The identity matrixIn is the n×n matrix with 1s on the diagonal and 0s elsewhere:
I3=100010001
Property:Inx=x for all x∈Rn.
The identity matrix is the “do nothing” transformation.
The Transpose
(Transpose)
The transpose of A, denoted AT, swaps rows and columns:
(AT)ij=Aji
If A is m×n, then AT is n×m.
Example:
[142536]T=123456
(Properties of Transpose)
(AT)T=A
(A+B)T=AT+BT
(cA)T=cAT
(AB)T=BTAT (note the order reversal!)
The transpose reversal (AB)T=BTAT mirrors how function composition reverses: to undo “first B, then A,” you undo A first, then B.
(Symmetric Matrices)
A matrix is symmetric if AT=A.
This means Aij=Aji,the matrix equals its mirror across the diagonal.
Example:
123245356
Symmetric matrices have special properties: their eigenvalues are real, and they can be orthogonally diagonalized.
Matrix Powers
(Powers of Square Matrices)
For a square matrix A and positive integer k:
Ak=k timesA⋅A⋯A
By convention, A0=I.
Interpretation:Ak represents applying the transformation A a total of k times.
(Example: Powers Reveal Structure)
A=[0010]
Then A2=[0000].
This matrix is nilpotent,some power of it is zero. Geometrically, applying it twice collapses everything.
Matrix Inverses
(Inverse Matrix)
An n×n matrix A is invertible (or nonsingular) if there exists a matrix A−1 such that:
AA−1=A−1A=In
The matrix A−1 is called the inverse of A.
Interpretation: If A represents a transformation, then A−1 is the transformation that undoes it. Applying A then A−1 (or vice versa) returns you to where you started.
(Uniqueness)
If A is invertible, its inverse is unique.
Proof: Suppose B and C are both inverses of A. Then:
B=BI=B(AC)=(BA)C=IC=C
(2×2 Inverse Formula)
For A=[acbd] with det(A)=ad−bc=0:
A−1=ad−bc1[d−c−ba]
Swap the diagonal entries, negate the off-diagonal entries, and divide by the determinant.
Example:
[2513]−1=6−51[3−5−12]=[3−5−12]
(Computing Inverses via Row Reduction)
For larger matrices, use the augmented matrix method:
Form the augmented matrix [A∣I]
Row reduce until the left side becomes I
The right side becomes A−1
[A∣I]row ops[I∣A−1]
If A cannot be reduced to I (a row of zeros appears on the left), then A is not invertible.
For an n×n matrix A, the following are equivalent:
A is invertible
det(A)=0
rank(A)=n
rref(A)=In
The columns of A are linearly independent
The columns of A span Rn
Ax=b has exactly one solution for every b
Ax=0 has only the trivial solution
ker(A)={0}
These are all ways of saying A doesn’t collapse any dimension.
(Properties of Inverses)
For invertible matrices A and B:
(A−1)−1=A
(AB)−1=B−1A−1 (note the order reversal!)
(AT)−1=(A−1)T
(cA)−1=c1A−1 for c=0
(Ak)−1=(A−1)k
The reversal in (AB)−1=B−1A−1 makes sense: to undo “first B, then A,” you must undo A first, then undo B.
(Solving Systems with Inverses)
If A is invertible, the system Ax=b has the unique solution:
x=A−1b
Proof: Multiply both sides of Ax=b by A−1:
A−1(Ax)=A−1b⟹Ix=A−1b⟹x=A−1b
Note: In practice, row reduction is more efficient than computing A−1 explicitly.
(Singular Matrices)
A matrix that is not invertible is called singular.
Singular matrices have det(A)=0, meaning they collapse at least one dimension. There’s no way to “uncollapse”,information is lost, so no inverse exists.
Example:
A=[1224]
The columns are linearly dependent (second is twice the first), det(A)=0, and A maps all of R2 onto a line. No inverse exists.
Why Matrix Multiplication Works This Way
The definition of matrix multiplication seems arbitrary until you realize it’s forced by the requirement that matrices represent linear transformations.
If we want (AB)x=A(Bx) to hold for all x, there’s only one possible definition for AB. The entry formula, the column view, and all the properties follow inevitably.
Matrix multiplication is function composition,the rest is just computing what that means entry-by-entry.