Mascot image.
#Math#Vectors#Geometry

Matrices

The general m×nm \times n system written out at the end of Lesson 3AM already arranges its coefficients aija_{ij} in mm rows and nn columns on the page, and its right-hand sides bib_i in a single column of length mm. Up to now those two rectangles were only typographic scaffolding for the equations. Once they are named as objects of their own, every question about existence, uniqueness, and structure of the system becomes a question about them, and the individual unknowns xjx_j drop out of the bookkeeping. The rest of the course is essentially the study of those rectangles.

The Rectangular Array

Definition 36 (Rectangular Matrix)

An m×nm \times n matrix over a field F\mathbb{F} is an ordered array of mnmn scalars aijFa_{ij} \in \mathbb{F}, indexed by 1im1 \le i \le m and 1jn1 \le j \le n, arranged in mm rows and nn columns:

A=[a11a12a1na21a22a2nam1am2amn].A = \begin{bmatrix} a_{11} & a_{12} & \cdots & a_{1n} \\ a_{21} & a_{22} & \cdots & a_{2n} \\ \vdots & \vdots & & \vdots \\ a_{m1} & a_{m2} & \cdots & a_{mn} \end{bmatrix}.

The scalar aija_{ij} is the entry of AA at row ii and column jj, and the pair (m,n)(m, n) is the size of AA. We shorten the display to

A=[aij]i,j=1m,n,orA=[aij]A = [a_{ij}]_{i,\, j = 1}^{m,\, n}, \qquad \text{or} \qquad A = [a_{ij}]

when the index ranges are clear from context. The set of all m×nm \times n matrices over F\mathbb{F} is written Fm×n\mathbb{F}^{m \times n}; in particular, when F=R\mathbb{F} = \mathbb{R} we write Rm×n\mathbb{R}^{m \times n}.

The field F\mathbb{F} is the generic field introduced at the close of Lesson 3AM; unless stated otherwise the default reading F=R\mathbb{F} = \mathbb{R} remains in force.

Definition 37 (Equality of Matrices)

Matrices A=[aij]A = [a_{ij}] and B=[bij]B = [b_{ij}] are equal, written A=BA = B, if and only if they have the same size (m,n)(m, n) and aij=bija_{ij} = b_{ij} for every 1im1 \le i \le m and 1jn1 \le j \le n.

The size requirement is not redundant. A 2×32 \times 3 and a 3×23 \times 2 matrix with the same multiset of entries are never equal, because the index pairs on either side of the proposed equality do not even match up.

Example 47 (The Coefficient Matrix of a System)

The system at the end of Lesson 3AM,

{x+y+z=1,xy+2z=0,\begin{cases} x + y + z = 1, \\ x - y + 2z = 0, \end{cases}

has m=2m = 2 equations in n=3n = 3 unknowns, so its coefficient table is the 2×32 \times 3 matrix

A=[111112]R2×3,A = \begin{bmatrix} 1 & 1 & 1 \\ 1 & -1 & 2 \end{bmatrix} \in \mathbb{R}^{2 \times 3},

while its right-hand sides form the 2×12 \times 1 column

b=[10]R2×1.\mathbf{b} = \begin{bmatrix} 1 \\ 0 \end{bmatrix} \in \mathbb{R}^{2 \times 1}.

The pair (A,b)(A, \mathbf{b}) determines the system: the unknowns x,y,zx, y, z never appear once the entries have been read off. Row reduction and every later test for existence, uniqueness, and structure will be carried out on (A,b)(A, \mathbf{b}) without rewriting the equations.

Square Matrices and Diagonals

The coefficient matrix of a system is square precisely when the number of equations equals the number of unknowns, the case where geometric intuition in R2\mathbb{R}^2 and R3\mathbb{R}^3 from the two earlier lessons was sharpest: two non-parallel lines meeting in a point, three generic planes meeting in a point. A dedicated vocabulary for the square case is worth pinning down before we have to exercise it.

Definition 38 (Square Matrix and Its Diagonals)

A matrix A=[aij]Fn×nA = [a_{ij}] \in \mathbb{F}^{n \times n} is square, and its common side length nn is the order of AA. The entries a11,a22,,anna_{11}, a_{22}, \ldots, a_{nn} form the main diagonal of AA; the entries a1n,a2,n1,,an1a_{1n}, a_{2,\, n-1}, \ldots, a_{n1} form the secondary diagonal:

A=[a11a12a1na21a22a2nan1an2ann],A = \begin{bmatrix} \color{#c0392b}{a_{11}} & a_{12} & \cdots & \color{#2874a6}{a_{1n}} \\ a_{21} & \color{#c0392b}{a_{22}} & \cdots & a_{2n} \\ \vdots & \vdots & \ddots & \vdots \\ \color{#2874a6}{a_{n1}} & a_{n2} & \cdots & \color{#c0392b}{a_{nn}} \end{bmatrix},

with the main diagonal picked out in red and the secondary diagonal in blue.

The main diagonal is where self-interaction sits: in a square system the coefficient aiia_{ii} measures how strongly the iith equation depends on the iith unknown, and the structure of the remainder governs how quickly an elimination schedule can clean that dependence up.

Triangular and Diagonal Matrices

The square systems easiest to solve by hand are those whose coefficient matrix already has half of its off-diagonal entries equal to zero.

Definition 39 (Triangular Matrix)

A square matrix A=[aij]Fn×nA = [a_{ij}] \in \mathbb{F}^{n \times n} is upper triangular when aij=0a_{ij} = 0 for every i>ji > j, and lower triangular when aij=0a_{ij} = 0 for every i<ji < j. Schematically,

Aupper=[a11a12a1n0a22a2n00ann],Alower=[a1100a21a220an1an2ann].A_{\text{upper}} = \begin{bmatrix} a_{11} & a_{12} & \cdots & a_{1n} \\ 0 & a_{22} & \cdots & a_{2n} \\ \vdots & & \ddots & \vdots \\ 0 & \cdots & 0 & a_{nn} \end{bmatrix}, \qquad A_{\text{lower}} = \begin{bmatrix} a_{11} & 0 & \cdots & 0 \\ a_{21} & a_{22} & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ a_{n1} & a_{n2} & \cdots & a_{nn} \end{bmatrix}.

An upper triangular system is solved by back-substitution: the last row fixes xnx_n, the penultimate row fixes xn1x_{n-1} once xnx_n is known, and so on up. Lower triangular systems are solved the same way from the top down. The whole apparatus of row reduction, developed from the next lesson on, aims at one of these two shapes.

Definition 40 (Diagonal, Scalar, Identity, and Zero Matrices)

A square matrix A=[aij]Fn×nA = [a_{ij}] \in \mathbb{F}^{n \times n} is diagonal when aij=0a_{ij} = 0 for every iji \neq j; equivalently, when it is both upper and lower triangular:

diag(a11,a22,,ann)=[a11000a22000ann].\operatorname{diag}(a_{11}, a_{22}, \ldots, a_{nn}) = \begin{bmatrix} a_{11} & 0 & \cdots & 0 \\ 0 & a_{22} & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & a_{nn} \end{bmatrix}.

When all those diagonal entries are the same scalar aFa \in \mathbb{F}, the matrix is scalar; the case a=1a = 1 is the identity matrix II, and the case a=0a = 0 is the square zero matrix 0\mathbf{0}:

aI=[a000a000a],I=[100010001],0=[000000000].aI = \begin{bmatrix} a & 0 & \cdots & 0 \\ 0 & a & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & a \end{bmatrix}, \qquad I = \begin{bmatrix} 1 & 0 & \cdots & 0 \\ 0 & 1 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & 1 \end{bmatrix}, \qquad \mathbf{0} = \begin{bmatrix} 0 & 0 & \cdots & 0 \\ 0 & 0 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & 0 \end{bmatrix}.

A rectangular matrix of any size whose every entry is zero is likewise called a zero matrix.

The hierarchy is cumulative: scalar matrices are a special case of diagonal matrices, which are a special case of both triangular families. The identity II will turn out to behave like the multiplicative 11 in F\mathbb{F} once matrix multiplication is on the table, and a diagonal system diag(d1,,dn)x=b\operatorname{diag}(d_1, \ldots, d_n)\mathbf{x} = \mathbf{b} is essentially nn uncoupled one-variable equations dixi=bid_i x_i = b_i.

Example 48 (Spotting the Shape)

Consider

M1=[300210547],M2=[200020002],M3=[100040009].M_1 = \begin{bmatrix} 3 & 0 & 0 \\ 2 & -1 & 0 \\ 5 & 4 & 7 \end{bmatrix}, \qquad M_2 = \begin{bmatrix} 2 & 0 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 2 \end{bmatrix}, \qquad M_3 = \begin{bmatrix} 1 & 0 & 0 \\ 0 & 4 & 0 \\ 0 & 0 & 9 \end{bmatrix}.

Every entry strictly above the main diagonal of M1M_1 is zero, so M1M_1 is lower triangular; the matrix M2M_2 has every off-diagonal entry zero and every diagonal entry equal to 22, so it is scalar, with M2=2IM_2 = 2I when II is read as the 3×33 \times 3 identity; and M3M_3 is diagonal but not scalar, because its diagonal entries differ. The main diagonal of M1M_1 is (3,1,7)(3, -1, 7) and its secondary diagonal is (0,1,5)(0, -1, 5).

Hermitian and Symmetric Matrices

Here C\mathbb{C} denotes the complex numbers. If z=a+biz = a + bi with a,bRa, b \in \mathbb{R} and i2=1i^2 = -1, then its complex conjugate is z=abi\overline{z} = a - bi.

The next two definitions record an internal symmetry of the entries of a square matrix. Over C\mathbb{C} the symmetry is combined with conjugation; over R\mathbb{R} conjugation is invisible and the two families coincide.

Definition 41 (Hermitian Matrix)

A square matrix A=[aij]Cn×nA = [a_{ij}] \in \mathbb{C}^{n \times n} is Hermitian, or self-adjoint, when

aji=aijfor every 1i,jn.a_{ji} = \overline{a_{ij}} \qquad \text{for every } 1 \le i, j \le n.

Setting i=ji = j forces aii=aiia_{ii} = \overline{a_{ii}}, so every diagonal entry of a Hermitian matrix is real.

Definition 42 (Symmetric Matrix)

A square matrix A=[aij]Fn×nA = [a_{ij}] \in \mathbb{F}^{n \times n} is symmetric when

aji=aijfor every 1i,jn.a_{ji} = a_{ij} \qquad \text{for every } 1 \le i, j \le n.

When F=R\mathbb{F} = \mathbb{R} conjugation acts as the identity, so aji=aija_{ji} = \overline{a_{ij}} and aji=aija_{ji} = a_{ij} say the same thing: a real matrix is Hermitian precisely when it is symmetric. Over C\mathbb{C} the Hermitian condition is strictly stronger than symmetry, since it insists the diagonal be real and the off-diagonal pairs be conjugate rather than equal.

Example 49 (Hermitian Without Being Symmetric)

The 2×22 \times 2 matrix

H=[1ii2]C2×2H = \begin{bmatrix} 1 & i \\ -i & 2 \end{bmatrix} \in \mathbb{C}^{2 \times 2}

is Hermitian: its diagonal entries 11 and 22 are real, and i=i\overline{i} = -i gives a21=a12a_{21} = \overline{a_{12}}. It is not symmetric, because iii \neq -i. Over R\mathbb{R} no analogous separation is possible: for a real matrix, matching entries across the diagonal already matches their conjugates, which is why lesson series restricted to R\mathbb{R} often speak only of symmetric matrices and drop the Hermitian label entirely.

Rows, Columns, and Position Vectors

The thinnest matrices, those with m=1m = 1 or n=1n = 1, are the column and row arrangements drawn since Lesson 1AM, now relabelled inside the framework of this lesson.

Definition 43 (Row and Column Matrices)

A column matrix of length nn is an n×1n \times 1 matrix

b=[b1b2bn]Fn×1,\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \\ \vdots \\ b_n \end{bmatrix} \in \mathbb{F}^{n \times 1},

and a row matrix of length nn is a 1×n1 \times n matrix

c=[c1c2cn]F1×n.\mathbf{c} = \begin{bmatrix} c_1 & c_2 & \cdots & c_n \end{bmatrix} \in \mathbb{F}^{1 \times n}.

Both are called vectors, or ordered nn-tuples, extending the usage of Lesson 1AM and Lesson 2AM to arbitrary length.

Matrix equality refuses to identify a column matrix in Fn×1\mathbb{F}^{n \times 1} with the row matrix in F1×n\mathbb{F}^{1 \times n} carrying the same entries, because their sizes disagree. A formal bridge between the two, the transpose, is introduced later in this lesson; until then we keep the column arrangement as the default avatar of a vector, as in every previous lesson of the course.

One-, two-, and three-dimensional coordinate systems showing a point P together with its position vector OP drawn as a directed arrow from the origin.

For n=1,2,3n = 1, 2, 3 the geometric reading is the one the figure illustrates. A point PP in three-dimensional Euclidean space with Cartesian coordinates (x0,y0,z0)(x_0, y_0, z_0) is associated with the 3×13 \times 1 column

[x0y0z0]R3×1,\begin{bmatrix} x_0 \\ y_0 \\ z_0 \end{bmatrix} \in \mathbb{R}^{3 \times 1},

and this column is exactly the position vector OP\overrightarrow{OP} fixed to PP in Lesson 1PM and extended to R3\mathbb{R}^3 in Lesson 2AM. The arrow from the origin to PP and the column of its coordinates carry the same content; the two points of view will be used interchangeably. For n4n \ge 4 the arrow picture stops being available, but the column matrix still keeps the bookkeeping running, and that is the only reason the higher-dimensional arguments ahead remain tractable.

Length is imported without change. The magnitude b\|\mathbf{b}\| of a real column vector was defined from Pythagoras in Lesson 1AM and inherited by R3\mathbb{R}^3 in Lesson 2AM through the iterated-Pythagoras argument; the same Euclidean formula defines b\|\mathbf{b}\| for every bRn×1\mathbf{b} \in \mathbb{R}^{n \times 1}, no alteration required.

Problem 31

Take the system

{2x1+x23x3+x4=5,x1x2+4x3+2x4=1,3x1+2x2+x3x4=0.\begin{cases} 2x_1 + x_2 - 3x_3 + x_4 = 5, \\ x_1 - x_2 + 4x_3 + 2x_4 = -1, \\ 3x_1 + 2x_2 + x_3 - x_4 = 0. \end{cases}

Write down its coefficient matrix AA and right-hand side column b\mathbf{b}, name the field F\mathbb{F} for which the system lives in Fm×n\mathbb{F}^{m \times n} and Fm×1\mathbb{F}^{m \times 1}, and decide which (if any) of the labels “square”, “upper triangular”, “lower triangular”, “diagonal”, “symmetric” apply to AA. Justify each rejection by producing an entry that violates the corresponding condition.

Addition and Scalar Multiplication

A vector of length nn is a matrix with one row or one column, and its addition and scalar multiplication were fixed entry by entry in Lesson 1AM and inherited by R3\mathbb{R}^3 without change in Lesson 2AM. Widening the rectangle from one row or column to m×nm \times n requires no new idea: we define addition and scaling on matrices the same way, entry by entry, and the previous geometric readings, parallelogram law and directed rescaling, survive untouched on the thin-matrix cases.

Definition 44 (Matrix Addition)

Let A=[aij]A = [a_{ij}] and B=[bij]B = [b_{ij}] both lie in Fm×n\mathbb{F}^{m \times n}. Their sum is the m×nm \times n matrix

A+B=[aij+bij]i,j=1m,n.A + B = [a_{ij} + b_{ij}]_{i,\, j = 1}^{m,\, n}.

The operation is defined only when AA and BB have the same size; mismatched sizes have no sum.

Restricted to column or row matrices the formula reproduces the componentwise vector addition of Lesson 1AM, together with the parallelogram-law reading illustrated there. Subtraction is introduced by the same device as in F\mathbb{F}: the difference ABA - B is the unique XFm×nX \in \mathbb{F}^{m \times n} for which X+B=AX + B = A, namely X=[aijbij]X = [a_{ij} - b_{ij}]. The m×nm \times n zero matrix 0\mathbf{0} is the additive identity: A+0=AA + \mathbf{0} = A and A0=AA - \mathbf{0} = A for every AFm×nA \in \mathbb{F}^{m \times n}.

Theorem 46 (Algebra of Matrix Addition)

For A,B,CFm×nA, B, C \in \mathbb{F}^{m \times n},

A+B=B+A,(A+B)+C=A+(B+C).A + B = B + A, \qquad (A + B) + C = A + (B + C).
Proof

Every entry of A+BA + B is the scalar aij+bija_{ij} + b_{ij}, and every entry of B+AB + A is bij+aijb_{ij} + a_{ij}; these coincide because addition in F\mathbb{F} is commutative, one of the nine field axioms carried over from R\mathbb{R} to any F\mathbb{F} at the close of Lesson 3AM. The same argument applied to (aij+bij)+cij=aij+(bij+cij)(a_{ij} + b_{ij}) + c_{ij} = a_{ij} + (b_{ij} + c_{ij}) handles associativity.

Shape is preserved under addition: the sum of two upper (respectively lower) triangular matrices of the same order is upper (respectively lower) triangular, and the sum of two diagonal matrices of the same order is diagonal, since the zero pattern on the relevant side of the main diagonal is carried through the entrywise sum unchanged.

Definition 45 (Scalar Multiplication of a Matrix)

For A=[aij]Fm×nA = [a_{ij}] \in \mathbb{F}^{m \times n} and αF\alpha \in \mathbb{F}, the scalar multiple αA\alpha A is the m×nm \times n matrix

αA=[αaij]i,j=1m,n.\alpha A = [\alpha\, a_{ij}]_{i,\, j = 1}^{m,\, n}.

We write A-A for (1)A(-1)A, so that AB=A+(B)A - B = A + (-B).

Restricted to columns or rows the formula reproduces the scalar multiplication of vectors from Lesson 1AM, and for real α\alpha the geometry illustrated there applies unchanged: the length of αa\alpha\mathbf{a} is α|\alpha| times the length of a\mathbf{a}, with the direction preserved when α>0\alpha > 0 and reversed when α<0\alpha < 0.

Theorem 47 (Algebra of Scalar Multiplication)

For A,BFm×nA, B \in \mathbb{F}^{m \times n} and α,βF\alpha, \beta \in \mathbb{F},

0A=0,(α+β)A=αA+βA,α(A+B)=αA+αB,α(βA)=(αβ)A.0 A = \mathbf{0}, \qquad (\alpha + \beta) A = \alpha A + \beta A, \qquad \alpha(A + B) = \alpha A + \alpha B, \qquad \alpha(\beta A) = (\alpha\beta) A.
Proof

Each identity holds entrywise from a single field axiom applied to the scalars α,β,aij,bij\alpha, \beta, a_{ij}, b_{ij}: annihilation 0aij=00 \cdot a_{ij} = 0 for the first, distributivity of F\mathbb{F} for the second and third, and associativity of multiplication in F\mathbb{F} for the fourth.

Example 50 (Adding and Scaling Matrices)

Let

A=[210341],B=[123015].A = \begin{bmatrix} 2 & -1 & 0 \\ 3 & 4 & 1 \end{bmatrix}, \qquad B = \begin{bmatrix} 1 & 2 & -3 \\ 0 & 1 & 5 \end{bmatrix}.

Then

A+B=[313356],2A3B=[1896513],A + B = \begin{bmatrix} 3 & 1 & -3 \\ 3 & 5 & 6 \end{bmatrix}, \qquad 2A - 3B = \begin{bmatrix} 1 & -8 & 9 \\ 6 & 5 & -13 \end{bmatrix},

while pairing AA with a 3×23 \times 2 matrix would leave A+A + \cdot undefined, the sizes not agreeing.

Matrix Multiplication

The m×nm \times n system packaged at the close of Lesson 3AM writes the iith equation as

ai1x1+ai2x2++ainxn=k=1naikxk=bi,a_{i1} x_1 + a_{i2} x_2 + \cdots + a_{in} x_n = \sum_{k = 1}^{n} a_{ik} x_k = b_i,

a scalar entirely determined by the iith row of the coefficient matrix AA and the column x\mathbf{x} of unknowns. Collecting all mm scalars into a single column recovers the right-hand side b\mathbf{b}. Any operation that packages AA and x\mathbf{x} into a product AxA\mathbf{x} must reproduce exactly this “row into column” pairing, and the definition below is the unique extension of that rule to arbitrary conformable matrices.

Definition 46 (Matrix Product)

Let A=[aij]Fm×lA = [a_{ij}] \in \mathbb{F}^{m \times l} and B=[bij]Fl×nB = [b_{ij}] \in \mathbb{F}^{l \times n}; here the column count of AA and the row count of BB agree, and A,BA, B are called conformable in this order. Their product is the m×nm \times n matrix AB=[cij]AB = [c_{ij}] with entries

cij=ai1b1j+ai2b2j++ailblj=k=1laikbkj,1im,1jn.c_{ij} = a_{i1} b_{1j} + a_{i2} b_{2j} + \cdots + a_{il} b_{lj} = \sum_{k = 1}^{l} a_{ik}\, b_{kj}, \qquad 1 \le i \le m, \quad 1 \le j \le n.

When the column count of AA does not match the row count of BB, the product ABAB is not defined.

Remark (Row-Into-Column as a Special Case)

If

r=[a1a2al]F1×l,c=[b1b2bl]Fl×1,\mathbf{r} = \begin{bmatrix} a_1 & a_2 & \cdots & a_l \end{bmatrix} \in \mathbb{F}^{1 \times l}, \qquad \mathbf{c} = \begin{bmatrix} b_1 \\ b_2 \\ \vdots \\ b_l \end{bmatrix} \in \mathbb{F}^{l \times 1},

then rc\mathbf{r}\mathbf{c} is a 1×11 \times 1 matrix whose single entry is

a1b1+a2b2++albl.a_1 b_1 + a_2 b_2 + \cdots + a_l b_l.

Over R\mathbb{R} this is exactly the familiar dot product of the corresponding vectors.

Computationally, cijc_{ij} is obtained by pairing the iith row of AA with the jjth column of BB and summing the products one index at a time. For the system of Lesson 3AM the product AxA\mathbf{x} is precisely the column of left-hand sides of the equations, so the entire m×nm \times n system collapses to the single identity

Ax=b.A\mathbf{x} = \mathbf{b}.

This reformulation is the whole reason matrix multiplication is fixed the way it is. Every existence, uniqueness, and structure question from Lesson 3AM is now a question about a single equation in the matrix variable x\mathbf{x}.

Example 51 (A First Row-Into-Column Product)

Let

A=[210031],B=[302110].A = \begin{bmatrix} 2 & -1 & 0 \\ 0 & 3 & 1 \end{bmatrix}, \qquad B = \begin{bmatrix} 3 & 0 \\ 2 & 1 \\ 1 & 0 \end{bmatrix}.

The column count of AA and the row count of BB both equal 33, so ABAB lies in R2×2\mathbb{R}^{2 \times 2}. Pairing each row of AA with each column of BB,

(AB)11=2(3)+(1)(2)+0(1)=4,(AB)12=2(0)+(1)(1)+0(0)=1,(AB)_{11} = 2(3) + (-1)(2) + 0(1) = 4, \quad (AB)_{12} = 2(0) + (-1)(1) + 0(0) = -1,(AB)21=0(3)+3(2)+1(1)=7,(AB)22=0(0)+3(1)+1(0)=3,(AB)_{21} = 0(3) + 3(2) + 1(1) = 7, \quad (AB)_{22} = 0(0) + 3(1) + 1(0) = 3,

so

AB=[4173].AB = \begin{bmatrix} 4 & -1 \\ 7 & 3 \end{bmatrix}.

The reverse product BABA is also conformable, since the column count of BB and the row count of AA both equal 22, but BABA lands in R3×3\mathbb{R}^{3 \times 3}: the two products live in different spaces, and the question of whether AB=BAAB = BA is, here, meaningless.

Theorem 48 (The Identity Acts as a Unit)

Let Im,InI_m, I_n denote the identity matrices of orders mm and nn. For every AFm×nA \in \mathbb{F}^{m \times n},

ImA=A,AIn=A.I_m A = A, \qquad A I_n = A.

In particular, IA=AI=AIA = AI = A whenever AA is square of the same order as II.

Proof

The (i,k)(i,k) entry of ImI_m is 11 when i=ki = k and 00 otherwise. Therefore

(ImA)ij=k=1m(Im)ikakj=aij,(I_m A)_{ij} = \sum_{k = 1}^{m} (I_m)_{ik}\, a_{kj} = a_{ij},

since only the term k=ik = i survives. The argument for AInA I_n is identical: in

(AIn)ij=k=1naik(In)kj,(A I_n)_{ij} = \sum_{k = 1}^{n} a_{ik}\, (I_n)_{kj},

only the term k=jk = j survives.

Theorem 49 (Distributive and Associative Laws)

For matrices of sizes such that every product and sum below is defined,

A(B+C)=AB+AC,(B+C)D=BD+CD,A(BC)=(AB)C.A(B + C) = AB + AC, \qquad (B + C) D = BD + CD, \qquad A(BC) = (AB)C.
Proof

Fix AFm×lA \in \mathbb{F}^{m \times l} and B,CFl×nB, C \in \mathbb{F}^{l \times n}. For the left distributive law,

(A(B+C))ij=k=1laik(bkj+ckj)=k=1laikbkj+k=1laikckj=(AB)ij+(AC)ij,\bigl(A(B + C)\bigr)_{ij} = \sum_{k = 1}^{l} a_{ik}(b_{kj} + c_{kj}) = \sum_{k = 1}^{l} a_{ik} b_{kj} + \sum_{k = 1}^{l} a_{ik} c_{kj} = (AB)_{ij} + (AC)_{ij},

using distributivity of multiplication over addition in F\mathbb{F}. The right distributive law is symmetric.

For associativity take AFm×lA \in \mathbb{F}^{m \times l}, BFl×pB \in \mathbb{F}^{l \times p}, CFp×nC \in \mathbb{F}^{p \times n}, so that all three products ABAB, BCBC, (AB)C(AB)C, A(BC)A(BC) are defined. Then

((AB)C)ij=r=1p(AB)ircrj=r=1p(k=1laikbkr)crj=k=1laik(r=1pbkrcrj)=(A(BC))ij,\bigl((AB)C\bigr)_{ij} = \sum_{r = 1}^{p} (AB)_{ir}\, c_{rj} = \sum_{r = 1}^{p} \Bigl(\sum_{k = 1}^{l} a_{ik} b_{kr}\Bigr) c_{rj} = \sum_{k = 1}^{l} a_{ik} \Bigl(\sum_{r = 1}^{p} b_{kr} c_{rj}\Bigr) = \bigl(A(BC)\bigr)_{ij},

the middle equality swapping the order of the two finite sums, legitimate in any field.

Matrix multiplication loses several properties that scalar multiplication enjoyed. The failures are cumulative, and the sharpest of them is the following.

Theorem 50 (Matrix Multiplication is Not Commutative)

For every n2n \ge 2, there exist square matrices A,BFn×nA, B \in \mathbb{F}^{n \times n} with ABBAAB \neq BA.

Proof

Take n=2n = 2 and

A=[2100],B=[1010].A = \begin{bmatrix} 2 & 1 \\ 0 & 0 \end{bmatrix}, \qquad B = \begin{bmatrix} 1 & 0 \\ -1 & 0 \end{bmatrix}.

Direct computation gives

AB=[1000],BA=[2121],AB = \begin{bmatrix} 1 & 0 \\ 0 & 0 \end{bmatrix}, \qquad BA = \begin{bmatrix} 2 & 1 \\ -2 & -1 \end{bmatrix},

which disagree at every entry outside the (1,1)(1, 1) position. Padding AA and BB with zeros embeds the same witness into Fn×n\mathbb{F}^{n \times n} for every n2n \ge 2.

Remark

Two further failures are worth recording. The product of two non-zero matrices may vanish: the choice

A=[1000],B=[0001]A = \begin{bmatrix} 1 & 0 \\ 0 & 0 \end{bmatrix}, \qquad B = \begin{bmatrix} 0 & 0 \\ 0 & 1 \end{bmatrix}

gives AB=0AB = \mathbf{0}, so non-zero matrices can multiply to zero. Cancellation fails too: if AB=ACAB = AC with A0A \neq \mathbf{0}, it need not follow that B=CB = C. With the same AA as above, any two matrices B,CB, C that agree on their first row but differ on their second satisfy AB=ACAB = AC, because the zero second row of AA annihilates every disagreement in the second row of BB and CC.

Having a well-behaved notion of product, and an identity II that acts as a unit, permits iterating the operation.

Definition 47 (Powers of a Square Matrix)

Let AFn×nA \in \mathbb{F}^{n \times n} be square. For every positive integer pp set

Ap=AAAp factors,A^p = \underbrace{A A \cdots A}_{p \text{ factors}},

and define A0=IA^0 = I and A1=AA^1 = A.

Theorem 51 (Exponent Laws for Matrix Powers)

For AFn×nA \in \mathbb{F}^{n \times n} and non-negative integers p,qp, q,

ApAq=Ap+q,(Ap)q=Apq.A^p A^q = A^{p + q}, \qquad (A^p)^q = A^{pq}.
Remark (A Word on Induction)

A few arguments in this lesson depend on a positive integer parameter, such as an exponent pp. In such cases we will sometimes use mathematical induction, an important proof technique that we will study properly later in MCEA and again in MA1B.

For now, the version we need is simple:

  1. first verify the statement at an initial value, usually p=0p = 0 or p=1p = 1;
  2. then show that if the statement holds for one value p=kp = k, it must also hold for the next value p=k+1p = k + 1.

Once these two steps are established, the statement follows for every integer from the starting point onward. In the present lesson we use induction only in this basic step-by-step form.

Proof

Fix pp and induct on qq for the first identity. The base case q=0q = 0 reads ApI=ApA^p I = A^p, which is the identity acting as a unit. For the inductive step,

ApAq+1=Ap(AqA)=(ApAq)A=Ap+qA=A(p+q)+1,A^p A^{q + 1} = A^p (A^q A) = (A^p A^q) A = A^{p + q} A = A^{(p + q) + 1},

by associativity of matrix multiplication and the inductive hypothesis. Induct on qq for the second identity. The base case q=0q = 0 reads (Ap)0=I=A0=Ap0(A^p)^0 = I = A^0 = A^{p \cdot 0}. For the inductive step,

(Ap)q+1=(Ap)qAp=ApqAp=Apq+p=Ap(q+1),(A^p)^{q + 1} = (A^p)^q A^p = A^{pq} A^p = A^{pq + p} = A^{p(q + 1)},

by the inductive hypothesis and the first exponent law.

Example 52 (Cubing a Small Upper-Triangular Matrix)

Let A=[2102]A = \begin{bmatrix} 2 & -1 \\ 0 & 2 \end{bmatrix}. Then

A2=[2102][2102]=[4404],A3=AA2=[81208].A^2 = \begin{bmatrix} 2 & -1 \\ 0 & 2 \end{bmatrix} \begin{bmatrix} 2 & -1 \\ 0 & 2 \end{bmatrix} = \begin{bmatrix} 4 & -4 \\ 0 & 4 \end{bmatrix}, \qquad A^3 = A \cdot A^2 = \begin{bmatrix} 8 & -12 \\ 0 & 8 \end{bmatrix}.

The upper-triangular shape of AA persists through every power. For p1p \ge 1 and indices i>ji > j, the entry (Ap)ij(A^p)_{ij} is a sum of products aik1ak1k2akp1ja_{i k_1} a_{k_1 k_2} \cdots a_{k_{p-1} j}; a non-zero such product would force ik1k2kp1ji \le k_1 \le k_2 \le \cdots \le k_{p-1} \le j, which contradicts i>ji > j.

Problem 32

Return to the system at the close of Lesson 3AM,

{x+y+z=1,xy+2z=0,\begin{cases} x + y + z = 1, \\ x - y + 2z = 0, \end{cases}

and write it in the single-line form Ax=bA\mathbf{x} = \mathbf{b} by identifying the coefficient matrix AA and the right-hand side column b\mathbf{b}. Produce any particular column x0R3×1\mathbf{x}_0 \in \mathbb{R}^{3 \times 1} that satisfies the system and verify by direct row-into-column multiplication that Ax0=bA \mathbf{x}_0 = \mathbf{b}.

Problem 33

Let aFm×1\mathbf{a} \in \mathbb{F}^{m \times 1} be a column matrix and cF1×n\mathbf{c} \in \mathbb{F}^{1 \times n} a row matrix. Show that the product ac\mathbf{a}\mathbf{c}, in this order, is defined and lies in Fm×n\mathbb{F}^{m \times n}, and write out the (i,j)(i, j) entry of ac\mathbf{a}\mathbf{c} in terms of the components of a\mathbf{a} and c\mathbf{c}. This matrix is called an outer product of a\mathbf{a} and c\mathbf{c}.

Problem 34

Let AFm×lA \in \mathbb{F}^{m \times l} have columns a1,a2,,alFm×1\mathbf{a}_1, \mathbf{a}_2, \ldots, \mathbf{a}_l \in \mathbb{F}^{m \times 1}, and let BFl×nB \in \mathbb{F}^{l \times n} have rows r1,r2,,rlF1×n\mathbf{r}_1, \mathbf{r}_2, \ldots, \mathbf{r}_l \in \mathbb{F}^{1 \times n}. Using the outer-product construction of the previous problem, prove the column-row decomposition

AB=k=1lakrk,AB = \sum_{k = 1}^{l} \mathbf{a}_k \mathbf{r}_k,

where each summand akrk\mathbf{a}_k \mathbf{r}_k is an m×nm \times n outer product. Compare this expression with the row-into-column formula that defines ABAB directly, and convince yourself that the two points of view compute the same entries.

The failure of commutativity in the earlier noncommutativity theorem is the source of most of the subtlety in matrix algebra, so it pays to name the cases where commutation is restored and the shapes whose self-multiplication behaves in an unusually rigid way. Four labels below cover the patterns that will reappear throughout the rest of the course.

Commuting Matrices

Definition 48 (Commuting Matrices)

Square matrices A,BFn×nA, B \in \mathbb{F}^{n \times n} commute when AB=BAAB = BA. The difference [A,B]=ABBA[A, B] = AB - BA is the commutator of AA and BB, and AA and BB commute exactly when [A,B]=0[A, B] = 0.

Commutation is the exception, not the rule; the earlier noncommutativity theorem already supplied a 2×22 \times 2 counterexample, and every computation with two generic matrices should start from the assumption that order matters. The results below are the standard reasons commutation is recovered in practice.

Theorem 52

For any AFn×nA \in \mathbb{F}^{n \times n} and any scalar λF\lambda \in \mathbb{F}, the matrix λIn\lambda I_n commutes with AA:

A(λIn)=λA=(λIn)A.A(\lambda I_n) = \lambda A = (\lambda I_n) A.
Proof

The (i,j)(i, j) entry of A(λIn)A(\lambda I_n) is

k=1naik(λIn)kj,\sum_{k=1}^n a_{ik}(\lambda I_n)_{kj},

and all terms vanish except the one with k=jk=j, because the only non-zero entry in column jj of λIn\lambda I_n is the diagonal entry λ\lambda. Hence (A(λIn))ij=λaij(A(\lambda I_n))_{ij} = \lambda a_{ij}. Likewise,

((λIn)A)ij=k=1n(λIn)ikakj,((\lambda I_n)A)_{ij} = \sum_{k=1}^n (\lambda I_n)_{ik} a_{kj},

and all terms vanish except the one with k=ik=i, so ((λIn)A)ij=λaij((\lambda I_n)A)_{ij} = \lambda a_{ij}. Both therefore agree with the (i,j)(i,j) entry of λA\lambda A.

Problem 35

Show that a matrix SFn×nS \in \mathbb{F}^{n \times n} commutes with every square matrix in Fn×n\mathbb{F}^{n \times n} if and only if SS is a scalar matrix aInaI_n for some aFa \in \mathbb{F}.

Theorem 53 (Diagonal Matrices Commute)

If

D=diag(d1,d2,,dn),E=diag(e1,e2,,en),D = \operatorname{diag}(d_1, d_2, \ldots, d_n), \qquad E = \operatorname{diag}(e_1, e_2, \ldots, e_n),

then

DE=ED=diag(d1e1,d2e2,,dnen).DE = ED = \operatorname{diag}(d_1 e_1, d_2 e_2, \ldots, d_n e_n).
Proof

Since every off-diagonal entry of DD and EE is zero, the product of row ii with column jj vanishes when iji \neq j, so both DEDE and EDED are diagonal. On the diagonal,

(DE)ii=diei=eidi=(ED)ii,(DE)_{ii} = d_i e_i = e_i d_i = (ED)_{ii},

because multiplication in F\mathbb{F} is commutative.

Theorem 54 (Commuting with a Diagonal Matrix of Distinct Entries)

Let

D=diag(d1,d2,,dn)Fn×nD = \operatorname{diag}(d_1, d_2, \ldots, d_n) \in \mathbb{F}^{n \times n}

with didjd_i \neq d_j whenever iji \neq j. If A=[aij]Fn×nA = [a_{ij}] \in \mathbb{F}^{n \times n} satisfies AD=DAAD = DA, then AA is diagonal.

Proof

The (i,j)(i,j) entry of ADAD is aijdja_{ij} d_j, while the (i,j)(i,j) entry of DADA is diaijd_i a_{ij}. Hence

aijdj=diaij,a_{ij} d_j = d_i a_{ij},

so

(djdi)aij=0.(d_j - d_i)a_{ij} = 0.

If iji \neq j, then djdi0d_j - d_i \neq 0, and since we are working in a field this forces aij=0a_{ij} = 0. Thus every off-diagonal entry of AA is zero, so AA is diagonal.

Theorem 55

If A,BFn×nA, B \in \mathbb{F}^{n \times n} commute, then ApB=BApA^p B = B A^p for every p0p \ge 0, and more generally ApBq=BqApA^p B^q = B^q A^p for every p,q0p, q \ge 0.

Proof

Fix BB and induct on pp. The base p=0p = 0 reads IB=B=BIIB = B = BI. For the step,

Ap+1B=A(ApB)=A(BAp)=(AB)Ap=(BA)Ap=BAp+1,A^{p+1} B = A(A^p B) = A(B A^p) = (AB) A^p = (BA) A^p = B A^{p+1},

using the inductive hypothesis at the second equality and commutation at the fourth. A symmetric induction on qq gives ABq=BqAA B^q = B^q A, and combining the two yields ApBq=BqApA^p B^q = B^q A^p.

Theorem 56 (The Commutant Is Closed)

Fix AFn×nA \in \mathbb{F}^{n \times n} and let C(A)={XFn×n:AX=XA}C(A) = \{X \in \mathbb{F}^{n \times n} : AX = XA\} be its commutant. If X,YC(A)X, Y \in C(A) and αF\alpha \in \mathbb{F}, then X+YX + Y, αX\alpha X, and XYXY all lie in C(A)C(A).

Proof

From AX=XAAX = XA and AY=YAAY = YA: A(X+Y)=AX+AY=XA+YA=(X+Y)AA(X + Y) = AX + AY = XA + YA = (X + Y)A; A(αX)=αAX=αXA=(αX)AA(\alpha X) = \alpha AX = \alpha XA = (\alpha X) A; and A(XY)=(AX)Y=(XA)Y=X(AY)=X(YA)=(XY)AA(XY) = (AX)Y = (XA)Y = X(AY) = X(YA) = (XY) A. The first two computations use distributivity and scalar compatibility, and the last uses associativity of matrix multiplication together with the commutation hypotheses.

Example 53 (Commutant of a Diagonal Matrix)

Find every BR2×2B \in \mathbb{R}^{2 \times 2} that commutes with

A=[2003].A = \begin{bmatrix} 2 & 0 \\ 0 & 3 \end{bmatrix}.

Write B=[pqrs]B = \begin{bmatrix} p & q \\ r & s \end{bmatrix}. Then

AB=[2p2q3r3s],BA=[2p3q2r3s],AB = \begin{bmatrix} 2p & 2q \\ 3r & 3s \end{bmatrix}, \qquad BA = \begin{bmatrix} 2p & 3q \\ 2r & 3s \end{bmatrix},

so AB=BAAB = BA forces 2q=3q2q = 3q and 3r=2r3r = 2r, that is q=r=0q = r = 0, while p,sp, s remain free. The commutant is

C(A)={[p00s]:p,sR}.C(A) = \left\{ \begin{bmatrix} p & 0 \\ 0 & s \end{bmatrix} : p, s \in \mathbb{R} \right\}.

Were the two diagonal entries of AA equal, the calculation would instead return every 2×22 \times 2 matrix, in agreement with the scalar-matrix commuting theorem above.

Problem 36

Determine every BR2×2B \in \mathbb{R}^{2 \times 2} that commutes with

A=[1101],A = \begin{bmatrix} 1 & 1 \\ 0 & 1 \end{bmatrix},

and describe what changes, compared with the diagonal example just above, once the off-diagonal entry is introduced.

Idempotent Matrices

A matrix whose square returns itself is the algebraic shape every projection in the course will turn out to have.

Definition 49 (Idempotent Matrix)

A square matrix AFn×nA \in \mathbb{F}^{n \times n} is idempotent when A2=AA^2 = A. Both InI_n and 0n\mathbf{0}_n are idempotent by inspection, and induction on pp gives Ap=AA^p = A for every p1p \ge 1.

Example 54 (A Non-Trivial Idempotent)

The matrix

A=[122001001]A = \begin{bmatrix} 1 & 2 & 2 \\ 0 & 0 & -1 \\ 0 & 0 & 1 \end{bmatrix}

satisfies A2=AA^2 = A. The rows of A2A^2 are read off row-into-column: row one of AA times the columns of AA gives (1,2,12+2(1)+21)=(1,2,2)(1, 2, 1 \cdot 2 + 2 \cdot (-1) + 2 \cdot 1) = (1, 2, 2), row two gives (0,0,02+0(1)+(1)1)=(0,0,1)(0, 0, 0 \cdot 2 + 0 \cdot (-1) + (-1) \cdot 1) = (0, 0, -1), and row three gives (0,0,02+0(1)+11)=(0,0,1)(0, 0, 0 \cdot 2 + 0 \cdot (-1) + 1 \cdot 1) = (0, 0, 1). Stacking these reproduces AA. The matrix is neither 0\mathbf{0} nor II, so idempotence is genuinely richer than the two trivial endpoints.

Theorem 57 (The Complementary Idempotent)

If AFn×nA \in \mathbb{F}^{n \times n} is idempotent, so is InAI_n - A, and the two matrices annihilate each other:

A(InA)=(InA)A=0.A(I_n - A) = (I_n - A) A = \mathbf{0}.
Proof

Using the unit and distributive laws established earlier,

(IA)2=I2A+A2=I2A+A=IA,(I - A)^2 = I - 2A + A^2 = I - 2A + A = I - A,

so IAI - A is idempotent. For the orthogonality, A(IA)=AA2=AA=0A(I - A) = A - A^2 = A - A = \mathbf{0}, and (IA)A=AA2=0(I - A)A = A - A^2 = \mathbf{0} by the same cancellation.

Example 55 (A Projection and Its Complement)

Let

P=[1000].P = \begin{bmatrix} 1 & 0 \\ 0 & 0 \end{bmatrix}.

Then P2=PP^2 = P, so PP is idempotent. Its complementary idempotent is

I2P=[0001],I_2 - P = \begin{bmatrix} 0 & 0 \\ 0 & 1 \end{bmatrix},

and indeed

P(I2P)=(I2P)P=0.P(I_2-P) = (I_2-P)P = \mathbf{0}.

These two matrices project onto the coordinate axes separately: left multiplication by PP keeps the first coordinate and kills the second, while left multiplication by I2PI_2-P does the opposite.

Problem 37

Let A,BFn×nA, B \in \mathbb{F}^{n \times n} be idempotent and commute. Prove that ABAB is idempotent, and exhibit two idempotents in R2×2\mathbb{R}^{2 \times 2} that do not commute and whose product is not idempotent, showing that the hypothesis AB=BAAB = BA is load-bearing rather than cosmetic.

Problem 38

Classify all idempotent matrices in F2×2\mathbb{F}^{2 \times 2}. In other words, determine all

A=[abcd]A = \begin{bmatrix} a & b \\ c & d \end{bmatrix}

for which A2=AA^2 = A.

Nilpotent Matrices

Opposite to idempotents, for which repeated multiplication stabilises, are matrices whose repeated multiplication annihilates them outright.

Definition 50 (Nilpotent Matrix)

A square matrix AFn×nA \in \mathbb{F}^{n \times n} is nilpotent when there is an integer p1p \ge 1 with Ap=0A^p = \mathbf{0}. The smallest such pp is the index of nilpotence of AA.

Example 56 (Strictly Upper-Triangular Nilpotent)

The matrix

N=[010001000]N = \begin{bmatrix} 0 & 1 & 0 \\ 0 & 0 & 1 \\ 0 & 0 & 0 \end{bmatrix}

satisfies

N2=[001000000],N3=03,N^2 = \begin{bmatrix} 0 & 0 & 1 \\ 0 & 0 & 0 \\ 0 & 0 & 0 \end{bmatrix}, \qquad N^3 = \mathbf{0}_3,

so NN is nilpotent with index 33. Each successive power pushes the non-zero band one step further above the main diagonal until it exits the matrix.

Theorem 58 (Strictly Triangular Matrices Are Nilpotent)

Every strictly upper-triangular AFn×nA \in \mathbb{F}^{n \times n}, meaning aij=0a_{ij} = 0 whenever iji \ge j, satisfies An=0A^n = \mathbf{0}.

Proof

We prove by induction on p1p \ge 1 that

(Ap)ij=0whenever ji<p.(A^p)_{ij} = 0 \qquad \text{whenever } j-i < p.

For p=1p=1 this is exactly the assumption that AA is strictly upper-triangular: if ji<1j-i<1, then jij \le i, so aij=0a_{ij}=0.

Now assume the claim for some pp. Then

(Ap+1)ij=k=1n(Ap)ikakj.(A^{p+1})_{ij} = \sum_{k=1}^n (A^p)_{ik} a_{kj}.

For a term (Ap)ikakj(A^p)_{ik} a_{kj} to be non-zero, the induction hypothesis forces kipk-i \ge p, and strict upper-triangularity forces jk1j-k \ge 1. Adding gives

ji=(jk)+(ki)1+p=p+1.j-i = (j-k) + (k-i) \ge 1 + p = p+1.

So if ji<p+1j-i < p+1, every term in the sum is zero, and hence (Ap+1)ij=0(A^{p+1})_{ij}=0. This proves the induction step.

Finally, in an n×nn \times n matrix we always have jin1j-i \le n-1. Therefore when p=np=n the inequality ji<nj-i < n holds for every pair (i,j)(i,j), so every entry of AnA^n is zero. Hence An=0A^n = \mathbf{0}.

The same argument, with the inequalities reversed, handles the strictly lower-triangular case.

Problem 39

Show that if AFn×nA \in \mathbb{F}^{n \times n} is nilpotent with Ap=0A^p = \mathbf{0}, then

(InA)(In+A+A2++Ap1)=In.(I_n - A)(I_n + A + A^2 + \cdots + A^{p - 1}) = I_n.

Deduce that a nilpotent matrix can never equal InI_n, and exhibit a non-zero BR2×2B \in \mathbb{R}^{2 \times 2} with B2=0B^2 = \mathbf{0} to confirm that the nilpotent class is non-empty beyond the zero matrix.

Problem 40

Classify all nilpotent matrices in F2×2\mathbb{F}^{2 \times 2}. Determine separately which ones have index of nilpotence 11 and which have index of nilpotence 22.

Involutory Matrices

Definition 51 (Involutory Matrix)

A square matrix AFn×nA \in \mathbb{F}^{n \times n} is involutory when A2=InA^2 = I_n. Geometrically, multiplication by AA is its own undoing; algebraically, AA will turn out to be its own inverse once the inverse is defined in a later lesson.

Example 57 (The Swap Matrix Is Involutory)

The matrix

J=[0110]J = \begin{bmatrix} 0 & 1 \\ 1 & 0 \end{bmatrix}

has

J2=[00+1101+1010+0111+00]=[1001]=I2,J^2 = \begin{bmatrix} 0 \cdot 0 + 1 \cdot 1 & 0 \cdot 1 + 1 \cdot 0 \\ 1 \cdot 0 + 0 \cdot 1 & 1 \cdot 1 + 0 \cdot 0 \end{bmatrix} = \begin{bmatrix} 1 & 0 \\ 0 & 1 \end{bmatrix} = I_2,

so JJ is involutory. Acting on a column by left multiplication, JJ swaps the two entries, and performing the swap twice returns the original column.

Theorem 59 (Powers of an Involutory Matrix)

If AFn×nA \in \mathbb{F}^{n \times n} is involutory, then A2k=InA^{2k} = I_n and A2k+1=AA^{2k + 1} = A for every k0k \ge 0.

Proof

Induct on kk, with A0=IA^0 = I and A1=AA^1 = A as bases. For the even step, A2(k+1)=A2kA2=II=IA^{2(k+1)} = A^{2k} A^2 = I \cdot I = I, and for the odd step, A2(k+1)+1=A2k+1A2=AI=AA^{2(k+1)+1} = A^{2k+1} A^2 = A \cdot I = A, using A2=IA^2 = I at each closure.

Theorem 60 (Idempotent Splitting of an Involution)

Let AFn×nA \in \mathbb{F}^{n \times n} be involutory and suppose 12F\tfrac{1}{2} \in \mathbb{F}, which holds for F=R\mathbb{F} = \mathbb{R} and F=C\mathbb{F} = \mathbb{C}. The matrices

P+=12(In+A),P=12(InA)P_+ = \tfrac{1}{2}(I_n + A), \qquad P_- = \tfrac{1}{2}(I_n - A)

are idempotent and satisfy P++P=InP_+ + P_- = I_n, P+P=AP_+ - P_- = A, and P+P=PP+=0P_+ P_- = P_- P_+ = \mathbf{0}.

Proof

Using A2=IA^2 = I,

P+2=14(I+A)2=14(I+2A+A2)=14(2I+2A)=12(I+A)=P+,P_+^2 = \tfrac{1}{4}(I + A)^2 = \tfrac{1}{4}(I + 2A + A^2) = \tfrac{1}{4}(2I + 2A) = \tfrac{1}{2}(I + A) = P_+,

and the symmetric calculation gives P2=PP_-^2 = P_-. The sum and difference are by direct expansion, and

P+P=14(I+A)(IA)=14(IA2)=0,P_+ P_- = \tfrac{1}{4}(I + A)(I - A) = \tfrac{1}{4}(I - A^2) = \mathbf{0},

with PP+=0P_- P_+ = \mathbf{0} by the same cancellation.

Example 58 (A Reflection Splits into Two Projections)

Take

A=[1001].A = \begin{bmatrix} 1 & 0 \\ 0 & -1 \end{bmatrix}.

Then A2=I2A^2 = I_2, so AA is involutory. The associated idempotents are

P+=12(I2+A)=[1000],P=12(I2A)=[0001].P_+ = \tfrac{1}{2}(I_2 + A) = \begin{bmatrix} 1 & 0 \\ 0 & 0 \end{bmatrix}, \qquad P_- = \tfrac{1}{2}(I_2 - A) = \begin{bmatrix} 0 & 0 \\ 0 & 1 \end{bmatrix}.

So this involution is exactly the difference of two complementary coordinate projections:

A=P+P.A = P_+ - P_-.

Geometrically, AA fixes the xx-axis and flips the sign on the yy-axis.

Remark

An involution can also be characterized by the factorization

A2=In(InA)(In+A)=0.A^2 = I_n \qquad\Longleftrightarrow\qquad (I_n - A)(I_n + A) = \mathbf{0}.

Indeed,

(InA)(In+A)=In+AAA2=InA2,(I_n - A)(I_n + A) = I_n + A - A - A^2 = I_n - A^2,

since InI_n commutes with AA. So either condition is exactly the statement InA2=0I_n - A^2 = \mathbf{0}.

Problem 41

Verify that

A=[4354]A = \begin{bmatrix} 4 & 3 \\ -5 & -4 \end{bmatrix}

is involutory, and write down the idempotents P±=12(I2±A)P_\pm = \tfrac{1}{2}(I_2 \pm A) from the idempotent-splitting theorem above for this AA.

Problem 42

For any real angle θ\theta, show that

A=[cosθsinθsinθcosθ]A = \begin{bmatrix} \cos \theta & \sin \theta \\ \sin \theta & -\cos \theta \end{bmatrix}

is involutory. Reminder:

cos2θ+sin2θ=1.\cos^2 \theta + \sin^2 \theta = 1.
Problem 43

Let JnJ_n be the n×nn \times n matrix with 11 in every position on the secondary diagonal and 00 elsewhere. Show that JnJ_n is involutory, and conclude that Jn-J_n is also involutory.

Problem 44

Prove that if A,BFn×nA, B \in \mathbb{F}^{n \times n} are involutory and commute, then ABAB is involutory. Exhibit a 2×22 \times 2 example showing that the conclusion fails once commutation is dropped.

Problem 45

Assume 12F\tfrac{1}{2} \in \mathbb{F}, for example F=R\mathbb{F} = \mathbb{R} or C\mathbb{C}. Classify all involutory matrices in F2×2\mathbb{F}^{2 \times 2}. That is, determine all

A=[abcd]A = \begin{bmatrix} a & b \\ c & d \end{bmatrix}

for which A2=I2A^2 = I_2.

Transpose and Conjugate Transpose

The previous section left one asymmetry untouched: a column of length nn and a row carrying the same entries are genuinely distinct objects in Fn×1\mathbb{F}^{n \times 1} and F1×n\mathbb{F}^{1 \times n} because their sizes differ. The transpose is the explicit bridge between the two, and upgraded to complex matrices it becomes the conjugate transpose, the operation that the Hermitian and symmetric conditions introduced earlier in this lesson were secretly measuring.

Definition 52 (Transpose)

The transpose of A=[aij]Fm×nA = [a_{ij}] \in \mathbb{F}^{m \times n} is the n×mn \times m matrix

A=[aji]i,j=1n,mA^{\top} = [a_{ji}]_{i,\, j = 1}^{n,\, m}

obtained by swapping rows and columns: the (i,j)(i, j) entry of AA^{\top} is the (j,i)(j, i) entry of AA. Thus, if

A=[a11a12a1na21a22a2nam1am2amn],A = \begin{bmatrix} a_{11} & a_{12} & \cdots & a_{1n} \\ a_{21} & a_{22} & \cdots & a_{2n} \\ \vdots & \vdots & & \vdots \\ a_{m1} & a_{m2} & \cdots & a_{mn} \end{bmatrix},

then

A=[a11a21am1a12a22am2a1na2namn].A^{\top} = \begin{bmatrix} a_{11} & a_{21} & \cdots & a_{m1} \\ a_{12} & a_{22} & \cdots & a_{m2} \\ \vdots & \vdots & & \vdots \\ a_{1n} & a_{2n} & \cdots & a_{mn} \end{bmatrix}.

Two readings of the transpose are worth keeping side by side. The rows of AA^{\top}, read left to right, are the columns of AA, read top to bottom; and the columns of AA^{\top} are the rows of AA. Geometrically the operation is a reflection of the rectangle across its main diagonal.

The symmetric condition introduced earlier asks aji=aija_{ji} = a_{ij} for every i,ji, j, which is precisely A=AA^{\top} = A, so a matrix is symmetric if and only if it equals its own transpose. The Hermitian condition aji=aija_{ji} = \overline{a_{ij}} mixes transposition with conjugation, and it is restated cleanly only after the conjugate transpose is defined below.

Example 59 (Transposes of Small Matrices)

For

A=[123456],b=[123],A = \begin{bmatrix} 1 & 2 & 3 \\ 4 & 5 & 6 \end{bmatrix}, \qquad \mathbf{b} = \begin{bmatrix} 1 \\ -2 \\ 3 \end{bmatrix},

direct reading gives

A=[142536]R3×2,b=[123]R1×3.A^{\top} = \begin{bmatrix} 1 & 4 \\ 2 & 5 \\ 3 & 6 \end{bmatrix} \in \mathbb{R}^{3 \times 2}, \qquad \mathbf{b}^{\top} = \begin{bmatrix} 1 & -2 & 3 \end{bmatrix} \in \mathbb{R}^{1 \times 3}.

The transpose of a column is the row carrying the same entries, fulfilling the bridge promised in the earlier discussion of rows and columns.

Theorem 61 (Transpose Is an Involution)

For every AFm×nA \in \mathbb{F}^{m \times n},

(A)=A.(A^{\top})^{\top} = A.
Proof

The (i,j)(i, j) entry of AA^{\top} is ajia_{ji}, so the (i,j)(i, j) entry of (A)(A^{\top})^{\top} is aija_{ij}, which is the (i,j)(i, j) entry of AA.

Theorem 62 (Transpose of Sum and Scalar Multiple)

For A,BFm×nA, B \in \mathbb{F}^{m \times n} and αF\alpha \in \mathbb{F},

(A+B)=A+B,(αA)=αA.(A + B)^{\top} = A^{\top} + B^{\top}, \qquad (\alpha A)^{\top} = \alpha A^{\top}.
Proof

The (i,j)(i, j) entry of (A+B)(A + B)^{\top} is (A+B)ji=aji+bji(A + B)_{ji} = a_{ji} + b_{ji}, which is the (i,j)(i, j) entry of A+BA^{\top} + B^{\top}. The (i,j)(i, j) entry of (αA)(\alpha A)^{\top} is (αA)ji=αaji(\alpha A)_{ji} = \alpha a_{ji}, which is the (i,j)(i, j) entry of αA\alpha A^{\top}.

Problem 46

If A,BFn×nA, B \in \mathbb{F}^{n \times n} are symmetric, show that A+BA + B is symmetric.

Theorem 63 (Transpose Reverses Products)

For AFm×lA \in \mathbb{F}^{m \times l} and BFl×nB \in \mathbb{F}^{l \times n},

(AB)=BA.(AB)^{\top} = B^{\top} A^{\top}.

More generally, for matrices A1,A2,,AkA_1, A_2, \ldots, A_k whose sizes make the product A1A2AkA_1 A_2 \cdots A_k defined,

(A1A2Ak)=AkA2A1.(A_1 A_2 \cdots A_k)^{\top} = A_k^{\top} \cdots A_2^{\top} A_1^{\top}.
Proof

The (i,j)(i, j) entry of (AB)(AB)^{\top} is (AB)ji=k=1lajkbki(AB)_{ji} = \sum_{k = 1}^{l} a_{j k} b_{k i}, while the (i,j)(i, j) entry of BAB^{\top} A^{\top} is

k=1l(B)ik(A)kj=k=1lbkiajk,\sum_{k = 1}^{l} (B^{\top})_{ik} (A^{\top})_{kj} = \sum_{k = 1}^{l} b_{k i} a_{j k},

and the two sums agree termwise by commutativity of multiplication in F\mathbb{F}. The kk-fold identity follows by induction on kk, using the two-fold version together with the associativity of matrix multiplication established earlier.

Example 60 (Transposing a Product)

Let

A=[120134],B=[210152].A = \begin{bmatrix} 1 & 2 & 0 \\ -1 & 3 & 4 \end{bmatrix}, \qquad B = \begin{bmatrix} 2 & 1 \\ 0 & -1 \\ 5 & 2 \end{bmatrix}.

Then

AB=[21184],so(AB)=[21814].AB = \begin{bmatrix} 2 & -1 \\ 18 & 4 \end{bmatrix}, \qquad\text{so}\qquad (AB)^{\top} = \begin{bmatrix} 2 & 18 \\ -1 & 4 \end{bmatrix}.

On the other hand,

B=[205112],A=[112304],B^{\top} = \begin{bmatrix} 2 & 0 & 5 \\ 1 & -1 & 2 \end{bmatrix}, \qquad A^{\top} = \begin{bmatrix} 1 & -1 \\ 2 & 3 \\ 0 & 4 \end{bmatrix},

and direct multiplication gives

BA=[21814]=(AB).B^{\top}A^{\top} = \begin{bmatrix} 2 & 18 \\ -1 & 4 \end{bmatrix} = (AB)^{\top}.

The order reversal is visible in the calculation: the transpose does not preserve the order of the factors.

Skew-Symmetric Matrices

Definition 53 (Skew-Symmetric Matrix)

A square matrix AFn×nA \in \mathbb{F}^{n \times n} is skew-symmetric when A=AA^{\top} = -A, equivalently aji=aija_{ji} = -a_{ij} for every 1i,jn1 \le i, j \le n. Setting i=ji = j gives 2aii=02 a_{ii} = 0, so whenever 12F\tfrac{1}{2} \in \mathbb{F} every diagonal entry of a skew-symmetric matrix is zero.

Example 61 (A Skew-Symmetric Matrix)

The matrix

S=[023205350]R3×3S = \begin{bmatrix} 0 & 2 & -3 \\ -2 & 0 & 5 \\ 3 & -5 & 0 \end{bmatrix} \in \mathbb{R}^{3 \times 3}

is skew-symmetric: its diagonal is zero, and reflecting each off-diagonal entry across the main diagonal negates it.

Problem 47

Let AFn×nA \in \mathbb{F}^{n \times n}. Show that the matrices

AA,AA,A+AAA^{\top}, \qquad A^{\top}A, \qquad A^{\top} + A

are symmetric, and that

AAA - A^{\top}

is skew-symmetric.

Theorem 64 (Symmetric-Skew Decomposition)

Assume 12F\tfrac{1}{2} \in \mathbb{F}. Every AFn×nA \in \mathbb{F}^{n \times n} admits a unique decomposition

A=S+K,S=S,K=K,A = S + K, \qquad S^{\top} = S, \quad K^{\top} = -K,

with SS symmetric and KK skew-symmetric, given by

S=12(A+A),K=12(AA).S = \tfrac{1}{2}(A + A^{\top}), \qquad K = \tfrac{1}{2}(A - A^{\top}).
Proof

Existence. Using the transpose involution and the transpose sum rule,

S=12(A+A)=S,K=12(AA)=K,S^{\top} = \tfrac{1}{2}(A^{\top} + A) = S, \qquad K^{\top} = \tfrac{1}{2}(A^{\top} - A) = -K,

and S+K=12(A+A)+12(AA)=AS + K = \tfrac{1}{2}(A + A^{\top}) + \tfrac{1}{2}(A - A^{\top}) = A.

Uniqueness. Suppose A=S+KA = S' + K' with SS' symmetric and KK' skew-symmetric. Transposing gives A=SKA^{\top} = S' - K', and solving the two equations for SS' and KK' returns the formulas above, forcing (S,K)=(S,K)(S', K') = (S, K).

Problem 48

For

A=[132045712],A = \begin{bmatrix} 1 & 3 & -2 \\ 0 & 4 & 5 \\ 7 & -1 & 2 \end{bmatrix},

compute the symmetric matrix

S=12(A+A)S = \tfrac{1}{2}(A + A^{\top})

and the skew-symmetric matrix

K=12(AA),K = \tfrac{1}{2}(A - A^{\top}),

and verify directly that A=S+KA = S + K.

Complex Conjugate and Conjugate Transpose

Over C\mathbb{C} the transpose by itself is not quite the right operation: the Hermitian condition introduced earlier mixes transposition with complex conjugation, and the pair needs both halves introduced explicitly before it can be used cleanly.

Definition 54 (Conjugate of a Matrix)

For A=[aij]Cm×nA = [a_{ij}] \in \mathbb{C}^{m \times n} the conjugate A\overline{A} is the m×nm \times n matrix whose (i,j)(i, j) entry is aij\overline{a_{ij}}. Conjugation is the identity on Rm×n\mathbb{R}^{m \times n}, since real scalars are fixed by conjugation in C\mathbb{C}.

Definition 55 (Conjugate Transpose)

For ACm×nA \in \mathbb{C}^{m \times n} the conjugate transpose, written AA^{*} (also AHA^{H} in some texts), is the n×mn \times m matrix

A=A=(A),A^{*} = \overline{A^{\top}} = (\overline{A})^{\top},

that is, (A)ij=aji(A^{*})_{ij} = \overline{a_{ji}}. Explicitly,

A=[a11a21am1a12a22am2a1na2namn].A^{*} = \begin{bmatrix} \overline{a_{11}} & \overline{a_{21}} & \cdots & \overline{a_{m1}} \\ \overline{a_{12}} & \overline{a_{22}} & \cdots & \overline{a_{m2}} \\ \vdots & \vdots & & \vdots \\ \overline{a_{1n}} & \overline{a_{2n}} & \cdots & \overline{a_{mn}} \end{bmatrix}.

Over R\mathbb{R} conjugation is trivial, so AA^{*} collapses to AA^{\top}.

The Hermitian condition introduced earlier now reads A=AA^{*} = A: a matrix is Hermitian precisely when it equals its own conjugate transpose, and over R\mathbb{R} this reduces further to A=AA^{\top} = A, the symmetric condition of the previous subsection.

Remark

The 1×11 \times 1 complex matrices are just the complex numbers in matrix clothing:

[z]zC.\begin{bmatrix} z \end{bmatrix} \longleftrightarrow z \in \mathbb{C}.

Under this identification, the Hermitian condition becomes

[z]=[z]z=z,\begin{bmatrix} z \end{bmatrix}^{*} = \begin{bmatrix} z \end{bmatrix} \qquad\Longleftrightarrow\qquad \overline{z} = z,

so the 1×11 \times 1 Hermitian matrices are exactly the real numbers.

Example 62 (Conjugate Transpose of a Complex Matrix)

For

A=[1+i23i4i]C2×2,A = \begin{bmatrix} 1 + i & 2 \\ -3 i & 4 - i \end{bmatrix} \in \mathbb{C}^{2 \times 2},

transposing then conjugating gives

A=[1+i3i24i],A=[1i3i24+i].A^{\top} = \begin{bmatrix} 1 + i & -3 i \\ 2 & 4 - i \end{bmatrix}, \qquad A^{*} = \begin{bmatrix} 1 - i & 3 i \\ 2 & 4 + i \end{bmatrix}.

Conjugating first and then transposing produces the same result, illustrating the commutativity of the two operations built into the definition of conjugate transpose.

Theorem 65 (Conjugate Transpose Is an Involution)

For every ACm×nA \in \mathbb{C}^{m \times n},

(A)=A.(A^{*})^{*} = A.
Proof

The (i,j)(i, j) entry of AA^{*} is aji\overline{a_{ji}}, so the (i,j)(i, j) entry of (A)(A^{*})^{*} is aij=aij\overline{\overline{a_{ij}}} = a_{ij}, using that conjugation in C\mathbb{C} is involutive.

Theorem 66 (Conjugate Transpose of Sum and Scalar Multiple)

For A,BCm×nA, B \in \mathbb{C}^{m \times n} and αC\alpha \in \mathbb{C},

(A+B)=A+B,(αA)=αA.(A + B)^{*} = A^{*} + B^{*}, \qquad (\alpha A)^{*} = \overline{\alpha}\, A^{*}.
Proof

Entrywise, ((A+B))ij=aji+bji=aji+bji((A + B)^{*})_{ij} = \overline{a_{ji} + b_{ji}} = \overline{a_{ji}} + \overline{b_{ji}}, matching A+BA^{*} + B^{*}. For the scalar, ((αA))ij=αaji=αaji((\alpha A)^{*})_{ij} = \overline{\alpha a_{ji}} = \overline{\alpha}\, \overline{a_{ji}}, matching αA\overline{\alpha} A^{*}. The appearance of α\overline{\alpha} in place of α\alpha is the genuinely new feature once conjugation joins the transpose.

Theorem 67 (Conjugate Transpose Reverses Products)

For ACm×lA \in \mathbb{C}^{m \times l} and BCl×nB \in \mathbb{C}^{l \times n},

(AB)=BA,(AB)^{*} = B^{*} A^{*},

and more generally (A1A2Ak)=AkA2A1(A_1 A_2 \cdots A_k)^{*} = A_k^{*} \cdots A_2^{*} A_1^{*} whenever the product is defined.

Proof

Combine the transpose-of-products theorem with entrywise conjugation:

(AB)=(AB)=BA=BA=BA,(AB)^{*} = \overline{(AB)^{\top}} = \overline{B^{\top} A^{\top}} = \overline{B^{\top}}\, \overline{A^{\top}} = B^{*} A^{*},

where the third equality is checked entrywise:

(BA)ij=k=1l(B)ik(A)kj=k=1l(B)ik(A)kj=(BA)ij.\bigl(\overline{B^{\top} A^{\top}}\bigr)_{ij} = \overline{\sum_{k=1}^{l} (B^{\top})_{ik} (A^{\top})_{kj}} = \sum_{k=1}^{l} \overline{(B^{\top})_{ik}}\, \overline{(A^{\top})_{kj}} = \bigl(\overline{B^{\top}}\, \overline{A^{\top}}\bigr)_{ij}.

Thus conjugation commutes with transposition and respects matrix products because it respects sums and products of scalars in C\mathbb{C}. The kk-fold identity is by induction on kk.

Theorem 68 (Hermitian Combinations from a Complex Matrix)

For every ACn×nA \in \mathbb{C}^{n \times n}, the four matrices

A+A,i(AA),AA,AAA + A^{*}, \qquad i(A - A^{*}), \qquad A A^{*}, \qquad A^{*} A

are all Hermitian.

Proof

Use the involution rule for ^{*} together with the sum, scalar, and product rules above. First, (A+A)=A+(A)=A+A(A + A^{*})^{*} = A^{*} + (A^{*})^{*} = A^{*} + A. Second, (i(AA))=i(AA)=i(AA)=i(AA)(i(A - A^{*}))^{*} = \overline{i}(A - A^{*})^{*} = -i(A^{*} - A) = i(A - A^{*}). For the last two, (AA)=(A)A=AA(A A^{*})^{*} = (A^{*})^{*} A^{*} = A A^{*}, and the same argument with the roles of AA and AA^{*} swapped handles AAA^{*} A.

Example 63 (Hermitian Combinations in Practice)

Let

A=[1+i23i4i].A = \begin{bmatrix} 1+i & 2 \\ -3i & 4-i \end{bmatrix}.

From the earlier computation,

A=[1i3i24+i].A^{*} = \begin{bmatrix} 1-i & 3i \\ 2 & 4+i \end{bmatrix}.

Then

A+A=[22+3i23i8],i(AA)=[23+2i32i2].A + A^{*} = \begin{bmatrix} 2 & 2+3i \\ 2-3i & 8 \end{bmatrix}, \qquad i(A-A^{*}) = \begin{bmatrix} -2 & 3+2i \\ 3-2i & 2 \end{bmatrix}.

Both are Hermitian: each equals its own conjugate transpose. The first captures the Hermitian part of AA, and the second captures the Hermitian matrix obtained from its skew-Hermitian part by multiplication by ii.

Skew-Hermitian Matrices

Definition 56 (Skew-Hermitian Matrix)

A square matrix ACn×nA \in \mathbb{C}^{n \times n} is skew-Hermitian when A=AA^{*} = -A, equivalently aji=aija_{ji} = -\overline{a_{ij}} for every 1i,jn1 \le i, j \le n. Setting i=ji = j forces aii=aiia_{ii} = -\overline{a_{ii}}, so every diagonal entry of a skew-Hermitian matrix is purely imaginary.

Example 64 (A Skew-Hermitian Matrix)

The matrix

K=[i2+i2+i3i]C2×2K = \begin{bmatrix} i & 2 + i \\ -2 + i & -3 i \end{bmatrix} \in \mathbb{C}^{2 \times 2}

is skew-Hermitian: the diagonal entries ii and 3i-3 i are purely imaginary, and a21=2+i=2+i=a12a_{21} = -2 + i = -\overline{2 + i} = -\overline{a_{12}}. Equivalently, iKiK is Hermitian, since (iK)=iK=i(K)=iK(iK)^{*} = \overline{i}\, K^{*} = -i(-K) = iK. Multiplication by ii therefore exchanges the Hermitian and skew-Hermitian classes.

Example 65 (A Tiny Skew-Hermitian Check)

The matrix

[iiii]\begin{bmatrix} i & i \\ i & i \end{bmatrix}

is skew-Hermitian because

[iiii]=[iiii]=[iiii].\begin{bmatrix} i & i \\ i & i \end{bmatrix}^{*} = \begin{bmatrix} -i & -i \\ -i & -i \end{bmatrix} = -\begin{bmatrix} i & i \\ i & i \end{bmatrix}.

This is the smallest possible example in which every entry is the same and the whole matrix is still skew-Hermitian.

Theorem 69 (Hermitian-Skew Decomposition)

Every ACn×nA \in \mathbb{C}^{n \times n} admits a unique decomposition

A=H+K,H=H,K=K,A = H + K, \qquad H^{*} = H, \quad K^{*} = -K,

with HH Hermitian and KK skew-Hermitian, given by

H=12(A+A),K=12(AA).H = \tfrac{1}{2}(A + A^{*}), \qquad K = \tfrac{1}{2}(A - A^{*}).
Proof

Existence. By the sum and scalar rule for conjugate transpose, H=12(A+A)=HH^{*} = \tfrac{1}{2}(A^{*} + A) = H and K=12(AA)=KK^{*} = \tfrac{1}{2}(A^{*} - A) = -K, while H+K=AH + K = A by direct expansion.

Uniqueness. If A=H+KA = H' + K' with HH' Hermitian and KK' skew-Hermitian, applying ^{*} gives A=HKA^{*} = H' - K', and solving the pair of equations returns the formulas above, so (H,K)=(H,K)(H', K') = (H, K).

Remark

Equivalently, every complex square matrix can be written uniquely in the form

A=H1+iH2,A = H_1 + i H_2,

where both H1H_1 and H2H_2 are Hermitian. Indeed, taking

H1=12(A+A),H2=12i(AA)H_1 = \tfrac{1}{2}(A + A^{*}), \qquad H_2 = \tfrac{1}{2i}(A - A^{*})

gives Hermitian matrices H1,H2H_1, H_2 with A=H1+iH2A = H_1 + i H_2.

Problem 49

Let

A=[1+i23i4+i2i].A = \begin{bmatrix} 1 + i & 2 - 3i \\ 4 + i & -2i \end{bmatrix}.

Compute AA^{*} and then form

H=12(A+A),K=12(AA).H = \tfrac{1}{2}(A + A^{*}), \qquad K = \tfrac{1}{2}(A - A^{*}).

Verify directly that HH is Hermitian, KK is skew-Hermitian, and A=H+KA = H + K.