Mascot image.
#Math#Vectors

Vectors

Vectors in the Plane

In elementary physics, a vector is often described as a quantity possessing both magnitude and direction (such as displacement, velocity, or force), distinguished from scalar quantities like mass or temperature. While it is possible to formalise vectors strictly through magnitude and direction, generalising “direction” to higher dimensions proves cumbersome. Instead, we adopt an algebraic definition based on ordered sets of numbers, deriving the geometric properties of magnitude and direction as consequences.

Definition 1 (Vector)

A vector in the plane is an ordered pair of real numbers, written as a column. The numbers comprising the pair are called the components of the vector. We denote vectors by bold-faced lower-case letters. Thus, if a1,a2Ra_1, a_2 \in \mathbb{R}, the vector a\mathbf{a} is given by:

a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}

The component a1a_1 is called the first component and a2a_2 the second component.

Remark

In handwritten work, where bold typeface is impractical, it is customary to indicate a vector by placing an arrow over the letter (a\vec{a}) or a bar beneath it (a\underline{a}).

Note

The symbol R\mathbb{R} denotes the real numbers: the entire number line, including integers, fractions, and irrationals such as 2\sqrt{2} and π\pi. The notation aRa \in \mathbb{R} reads ”aa belongs to R\mathbb{R}” and simply means that aa is a real number. This notation and its formal foundations will be introduced properly in the EECS lectures (see Lesson 2, MCE A); for now, treat R\mathbb{R} as a shorthand for “the real numbers”.

The word “ordered” is doing essential work here. The vectors [35]\begin{bmatrix} 3 \\ 5 \end{bmatrix} and [53]\begin{bmatrix} 5 \\ 3 \end{bmatrix} contain the same two real numbers, but they are different because the order of the components matters. This raises an immediate question: when are two vectors the same?

Equality between vectors is defined component-wise. Two vectors a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} and b=[b1b2]\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} are equal, written a=b\mathbf{a} = \mathbf{b}, if and only if a1=b1a_1 = b_1 and a2=b2a_2 = b_2.

Example 1

Let a=[27]\mathbf{a} = \begin{bmatrix} 2 \\ 7 \end{bmatrix}, b=[27]\mathbf{b} = \begin{bmatrix} 2 \\ 7 \end{bmatrix}, and c=[72]\mathbf{c} = \begin{bmatrix} 7 \\ 2 \end{bmatrix}. Then a=b\mathbf{a} = \mathbf{b} since 2=22 = 2 and 7=77 = 7. However, ac\mathbf{a} \neq \mathbf{c}: the second components differ (727 \neq 2).

One vector deserves immediate recognition.

Definition 2 (The Zero Vector)

The vector 0=[00]\mathbf{0} = \begin{bmatrix} 0 \\ 0 \end{bmatrix} is called the zero vector. Both of its components are zero.

Visualisation

Vectors in the plane may be visualised as arrows in the Cartesian plane. To represent the vector a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} graphically, we draw a directed line segment from the origin O=(0,0)O = (0, 0) to the point with coordinates (a1,a2)(a_1, a_2). Every vector corresponds to a unique arrow originating at OO, and conversely, every arrow starting at OO determines a vector.

Graphical representation of vectors b, c, and d as directed segments from the origin to the points (1,3), (-2,4), and (4,2)
Example 2

The vectors b=[13]\mathbf{b} = \begin{bmatrix} 1 \\ 3 \end{bmatrix}, c=[24]\mathbf{c} = \begin{bmatrix} -2 \\ 4 \end{bmatrix}, and d=[42]\mathbf{d} = \begin{bmatrix} 4 \\ 2 \end{bmatrix} are represented in the figure above as directed arrows from the origin to the points (1,3)(1, 3), (2,4)(-2, 4), and (4,2)(4, 2) respectively. The components of each vector are precisely the coordinates of the arrowhead.

An ordered pair of real numbers can fundamentally represent two distinct concepts: a geometric point or an algebraic vector. Points are fixed locations in the plane; vectors are algebraic quantities that typically represent displacements or shifts. We adopt the following convention to distinguish them.

Remark (Point vs Vector)
  • A point is denoted by a capital letter with coordinates in parentheses: A=(a1,a2)A = (a_1, a_2).
  • A vector is denoted by a bold lower-case letter as a column: a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}.

This distinction is not merely cosmetic. The point A=(3,5)A = (3, 5) is a fixed location in the plane. The vector a=[35]\mathbf{a} = \begin{bmatrix} 3 \\ 5 \end{bmatrix} represents a displacement of 33 units horizontally and 55 units vertically. When we draw a\mathbf{a} as an arrow from the origin to AA, we are anchoring the displacement at a specific starting point, but the displacement itself is independent of where it begins. This will become clearer when we discuss vector addition and translation in subsequent sections.

Example 3

Consider a city grid where the origin is placed at the town hall. The police station sits at the point P=(2,3)P = (2, 3), a fixed location. The instruction “walk 22 blocks east and 33 blocks north” is the vector p=[23]\mathbf{p} = \begin{bmatrix} 2 \\ 3 \end{bmatrix}, a displacement that makes sense from any starting position. If you follow p\mathbf{p} from the town hall, you arrive at PP. If you follow p\mathbf{p} from the library at (1,1)(1, 1), you arrive at (3,4)(3, 4) instead. The vector is the same; the destination changes with the starting point.

Problem 1

Let u=[x+13]\mathbf{u} = \begin{bmatrix} x + 1 \\ 3 \end{bmatrix} and v=[4y2]\mathbf{v} = \begin{bmatrix} 4 \\ y - 2 \end{bmatrix}. Determine the values of xx and yy for which u=v\mathbf{u} = \mathbf{v}.

Problem 2

A particle starts at the origin and undergoes two successive displacements: first p=[31]\mathbf{p} = \begin{bmatrix} 3 \\ -1 \end{bmatrix}, then q=[14]\mathbf{q} = \begin{bmatrix} -1 \\ 4 \end{bmatrix}. Without defining vector addition formally, determine the coordinates of the particle’s final position by applying each displacement component-wise. What single vector from the origin would land the particle at the same final position?

Magnitude

Given a vector a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}, the directed segment from the origin OO to the point A=(a1,a2)A = (a_1, a_2) has a length determined by Pythagoras’ theorem. We define this length as the magnitude of the vector.

Definition 3 (Magnitude)

The magnitude (or norm) of a vector a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}, denoted a|\mathbf{a}|, is the non-negative real number given by:

a=a12+a22|\mathbf{a}| = \sqrt{a_1^2 + a_2^2}
Example 4

For the vectors from the figure above, we compute:

b=12+32=10,c=(2)2+42=20,d=42+22=20.|\mathbf{b}| = \sqrt{1^2 + 3^2} = \sqrt{10}, \qquad |\mathbf{c}| = \sqrt{(-2)^2 + 4^2} = \sqrt{20}, \qquad |\mathbf{d}| = \sqrt{4^2 + 2^2} = \sqrt{20}.

Observe that c\mathbf{c} and d\mathbf{d} have different components but equal magnitude. Magnitude alone does not determine a vector.

Vectors with magnitude 11 are of particular importance.

Definition 4 (Unit Vector)

A vector u\mathbf{u} is called a unit vector if u=1|\mathbf{u}| = 1.

Examples include [1/21/2]\begin{bmatrix} 1/\sqrt{2} \\ -1/\sqrt{2} \end{bmatrix} and, more generally, [cosθsinθ]\begin{bmatrix} \cos\theta \\ \sin\theta \end{bmatrix} for any angle θ\theta. The standard unit coordinate vectors are denoted

e1=[10],e2=[01].\mathbf{e}_1 = \begin{bmatrix} 1 \\ 0 \end{bmatrix}, \qquad \mathbf{e}_2 = \begin{bmatrix} 0 \\ 1 \end{bmatrix}.

Every vector in the plane decomposes uniquely in terms of these two: if a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}, then a=a1e1+a2e2\mathbf{a} = a_1\mathbf{e}_1 + a_2\mathbf{e}_2, as one verifies directly from the definitions. The components of a\mathbf{a} are therefore precisely the coefficients of e1\mathbf{e}_1 and e2\mathbf{e}_2 in this decomposition.

Theorem 1 (Zero Vector Magnitude)

Let a\mathbf{a} be a vector. Then a=0|\mathbf{a}| = 0 if and only if a=0\mathbf{a} = \mathbf{0}.

Proof

Since a120a_1^2 \geq 0 and a220a_2^2 \geq 0 for any real numbers, the sum a12+a22=0a_1^2 + a_2^2 = 0 implies a1=0a_1 = 0 and a2=0a_2 = 0. Conversely, if a1=a2=0a_1 = a_2 = 0, the magnitude is clearly zero.

This characterises the zero vector (Definition 2) as the unique vector with zero magnitude. Geometrically, it corresponds to a degenerate segment where the initial and terminal points coincide at the origin. All other vectors are non-zero, possessing a well-defined positive magnitude and a specific direction.

Problem 3

For what values of tRt \in \mathbb{R} is [t1t]\begin{bmatrix} t \\ 1 - t \end{bmatrix} a unit vector?

The Vector Space R2\mathbb{R}^2

The primary advantage of vector geometry lies in the algebraic system formed by vectors, which mirrors properties of complex numbers. We define two fundamental operations: addition and scalar multiplication.

Definition 5 (Vector Addition)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} and b=[b1b2]\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} be vectors. The sum a+b\mathbf{a} + \mathbf{b} is defined as:

a+b=[a1+b1a2+b2]\mathbf{a} + \mathbf{b} = \begin{bmatrix} a_1 + b_1 \\ a_2 + b_2 \end{bmatrix}

The reader who solved Problem 2 has already performed this operation: combining the displacements p\mathbf{p} and q\mathbf{q} component-wise is precisely vector addition.

Definition 6 (Scalar Multiplication)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} be a vector and rRr \in \mathbb{R} a scalar. The scalar multiple rar\mathbf{a} is defined as:

ra=[ra1ra2]r\mathbf{a} = \begin{bmatrix} ra_1 \\ ra_2 \end{bmatrix}

We also define the negative of a vector as a=(1)a=[a1a2]-\mathbf{a} = (-1)\mathbf{a} = \begin{bmatrix} -a_1 \\ -a_2 \end{bmatrix}, and vector subtraction as ba=b+(a)\mathbf{b} - \mathbf{a} = \mathbf{b} + (-\mathbf{a}).

Definition 7 (Linear Combination)

A linear combination of vectors a\mathbf{a} and b\mathbf{b} is any expression of the form ca+dbc\mathbf{a} + d\mathbf{b}, where c,dRc, d \in \mathbb{R}.

Vector addition, subtraction, and scalar multiples are all special cases: a+b=1a+1b\mathbf{a} + \mathbf{b} = 1\cdot\mathbf{a} + 1\cdot\mathbf{b}, the difference is 1a+(1)b1\cdot\mathbf{a} + (-1)\cdot\mathbf{b}, and ca=ca+0bc\mathbf{a} = c\cdot\mathbf{a} + 0\cdot\mathbf{b}. The zero vector itself is the linear combination 0a+0b0\cdot\mathbf{a} + 0\cdot\mathbf{b}.

Example 5

Let a=[31]\mathbf{a} = \begin{bmatrix} 3 \\ -1 \end{bmatrix} and b=[24]\mathbf{b} = \begin{bmatrix} 2 \\ 4 \end{bmatrix}. Then:

a+b=[53],3a=[93],ab=[15].\mathbf{a} + \mathbf{b} = \begin{bmatrix} 5 \\ 3 \end{bmatrix}, \qquad 3\mathbf{a} = \begin{bmatrix} 9 \\ -3 \end{bmatrix}, \qquad \mathbf{a} - \mathbf{b} = \begin{bmatrix} 1 \\ -5 \end{bmatrix}.

Geometric Interpretation

Vector addition adheres to the Parallelogram Law. If vectors a\mathbf{a} and b\mathbf{b} are represented by directed segments OA\overrightarrow{OA} and OB\overrightarrow{OB}, their sum c=a+b\mathbf{c} = \mathbf{a} + \mathbf{b} corresponds to OC\overrightarrow{OC}, where OACBOACB forms a parallelogram.

The Parallelogram Law of vector addition

An equivalent description is the head-to-tail rule: place the start of b\mathbf{b} at the tip of a\mathbf{a}; the endpoint of b\mathbf{b} then lands at the tip of a+b\mathbf{a} + \mathbf{b}. Traversing a\mathbf{a} then b\mathbf{b}, or b\mathbf{b} then a\mathbf{a}, traces opposite sides of the same parallelogram, confirming commutativity (Theorem 2, Property 1) geometrically.

Scalar multiplication rar\mathbf{a} corresponds to scaling the length of the vector by a factor of r|r|. If r>0r > 0, the direction remains unchanged; if r<0r < 0, the direction is reversed. If r=0r = 0, the result is the zero vector (Definition 2).

Scalar multiples of a vector
Definition 8 (Vector Space R2\mathbb{R}^2)

The collection of all vectors in the plane, together with the operations of vector addition and scalar multiplication, is denoted R2\mathbb{R}^2 and is called the vector space of the plane.

Remark

These definitions extend naturally to any number of components. A vector in R3\mathbb{R}^3 has three components [a1a2a3]\begin{bmatrix} a_1 \\ a_2 \\ a_3 \end{bmatrix}, and addition and scalar multiplication proceed component-wise as before. More generally, a vector in Rn\mathbb{R}^n is an ordered nn-tuple of real numbers, and the same axioms continue to hold. When writing vectors with many components inline, we sometimes use the row shorthand v=(v1,v2,v3)\mathbf{v} = (v_1, v_2, v_3); this is still understood as a column vector temporarily lying on its side to save space, not as a row vector.

The algebraic structure of R2\mathbb{R}^2 is governed by the following fundamental properties.

Theorem 2 (Properties of R2\mathbb{R}^2)

Let a,b,c\mathbf{a}, \mathbf{b}, \mathbf{c} be vectors in R2\mathbb{R}^2 and let r,sr, s be scalars. Then:

  1. a+b=b+a\mathbf{a} + \mathbf{b} = \mathbf{b} + \mathbf{a} (Commutativity)
  2. (a+b)+c=a+(b+c)(\mathbf{a} + \mathbf{b}) + \mathbf{c} = \mathbf{a} + (\mathbf{b} + \mathbf{c}) (Associativity)
  3. There exists a unique vector 0\mathbf{0} such that a+0=a\mathbf{a} + \mathbf{0} = \mathbf{a}. (Additive Identity)
  4. For every a\mathbf{a}, there exists a unique vector a-\mathbf{a} such that a+(a)=0\mathbf{a} + (-\mathbf{a}) = \mathbf{0}. (Additive Inverse)
  5. (rs)a=r(sa)(rs)\mathbf{a} = r(s\mathbf{a})
  6. (r+s)a=ra+sa(r + s)\mathbf{a} = r\mathbf{a} + s\mathbf{a}
  7. r(a+b)=ra+rbr(\mathbf{a} + \mathbf{b}) = r\mathbf{a} + r\mathbf{b}
  8. 1a=a1\mathbf{a} = \mathbf{a}

The proof follows directly from the properties of real numbers applied to the components. For example, a+b=[a1+b1a2+b2]=[b1+a1b2+a2]=b+a\mathbf{a} + \mathbf{b} = \begin{bmatrix} a_1 + b_1 \\ a_2 + b_2 \end{bmatrix} = \begin{bmatrix} b_1 + a_1 \\ b_2 + a_2 \end{bmatrix} = \mathbf{b} + \mathbf{a}.

Remark

The correspondence between complex numbers and vectors is notable. Identifying a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} with z=a1+ia2z = a_1 + ia_2, vector addition corresponds to complex addition, and scalar multiplication corresponds to multiplying a complex number by a real number. However, unlike complex numbers, R2\mathbb{R}^2 does not inherently possess a vector-by-vector multiplication operation in this context.

We now demonstrate the power of these axioms by proving algebraic theorems in two ways: first using components (concrete), and second using only the axioms (abstract). This dual approach highlights that the results hold for any system satisfying the properties of Theorem 2, not just R2\mathbb{R}^2.

Theorem 3 (Zero Product Law)

Let a\mathbf{a} be a vector and rr a scalar. Then ra=0r\mathbf{a} = \mathbf{0} if and only if r=0r = 0 or a=0\mathbf{a} = \mathbf{0}.

Proof 1 (Components)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}. Then ra=[ra1ra2]r\mathbf{a} = \begin{bmatrix} ra_1 \\ ra_2 \end{bmatrix}.

If r=0r = 0, then ra=[00]=0r\mathbf{a} = \begin{bmatrix} 0 \\ 0 \end{bmatrix} = \mathbf{0}. If a=0\mathbf{a} = \mathbf{0}, then a1=a2=0a_1 = a_2 = 0, so ra=[00]=0r\mathbf{a} = \begin{bmatrix} 0 \\ 0 \end{bmatrix} = \mathbf{0}.

Conversely, suppose ra=0r\mathbf{a} = \mathbf{0}. Then ra1=0ra_1 = 0 and ra2=0ra_2 = 0. If r0r \neq 0, we must have a1=0a_1 = 0 and a2=0a_2 = 0, so a=0\mathbf{a} = \mathbf{0}.

Proof 2 (Axiomatic)

We use the properties from Theorem 2.

To show 0a=00\mathbf{a} = \mathbf{0}:

0a=(0+0)a=(6)0a+0a.0\mathbf{a} = (0 + 0)\mathbf{a} \overset{(6)}{=} 0\mathbf{a} + 0\mathbf{a}.

Adding (0a)-(0\mathbf{a}) to both sides gives 0=0a\mathbf{0} = 0\mathbf{a}.

To show r0=0r\mathbf{0} = \mathbf{0}:

r0=r(0+0)=(7)r0+r0.r\mathbf{0} = r(\mathbf{0} + \mathbf{0}) \overset{(7)}{=} r\mathbf{0} + r\mathbf{0}.

Adding (r0)-(r\mathbf{0}) to both sides gives 0=r0\mathbf{0} = r\mathbf{0}.

Conversely, suppose ra=0r\mathbf{a} = \mathbf{0} with r0r \neq 0. Then 1/r1/r exists and

a=(8)1a=(1rr)a=(5)1r(ra)=1r0=0.\mathbf{a} \overset{(8)}{=} 1\mathbf{a} = \left(\tfrac{1}{r} \cdot r\right)\mathbf{a} \overset{(5)}{=} \tfrac{1}{r}(r\mathbf{a}) = \tfrac{1}{r}\mathbf{0} = \mathbf{0}.
Theorem 4 (Negative Scalars)

Let a\mathbf{a} be a vector and rr a scalar. Then:

(r)a=r(a)=(ra)(-r)\mathbf{a} = r(-\mathbf{a}) = -(r\mathbf{a})
Proof 1 (Components)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}. Then:

(r)a=[ra1ra2],r(a)=r[a1a2]=[ra1ra2],(ra)=[ra1ra2]=[ra1ra2].(-r)\mathbf{a} = \begin{bmatrix} -ra_1 \\ -ra_2 \end{bmatrix}, \qquad r(-\mathbf{a}) = r\begin{bmatrix} -a_1 \\ -a_2 \end{bmatrix} = \begin{bmatrix} -ra_1 \\ -ra_2 \end{bmatrix}, \qquad -(r\mathbf{a}) = -\begin{bmatrix} ra_1 \\ ra_2 \end{bmatrix} = \begin{bmatrix} -ra_1 \\ -ra_2 \end{bmatrix}.

All three expressions yield identical components.

Proof 2 (Axiomatic)

The vector (ra)-(r\mathbf{a}) is defined as the unique additive inverse of rar\mathbf{a}. It suffices to show that both (r)a(-r)\mathbf{a} and r(a)r(-\mathbf{a}) act as inverses of rar\mathbf{a}.

For (r)a(-r)\mathbf{a}:

ra+(r)a=(6)(r+(r))a=0a=0.r\mathbf{a} + (-r)\mathbf{a} \overset{(6)}{=} (r + (-r))\mathbf{a} = 0\mathbf{a} = \mathbf{0}.

For r(a)r(-\mathbf{a}):

ra+r(a)=(7)r(a+(a))=r0=0.r\mathbf{a} + r(-\mathbf{a}) \overset{(7)}{=} r(\mathbf{a} + (-\mathbf{a})) = r\mathbf{0} = \mathbf{0}.

Since additive inverses are unique (Property 4), (r)a=r(a)=(ra)(-r)\mathbf{a} = r(-\mathbf{a}) = -(r\mathbf{a}).

Problem 4

Using only the properties from Theorem 2, prove that a+b=a+c\mathbf{a} + \mathbf{b} = \mathbf{a} + \mathbf{c} implies b=c\mathbf{b} = \mathbf{c} (the cancellation law).

Linear Independence

While e1\mathbf{e}_1 and e2\mathbf{e}_2 generate the plane, they also satisfy a property of non-redundancy: neither is a scalar multiple of the other. If e1=re2\mathbf{e}_1 = r\mathbf{e}_2, then [10]=[0r]\begin{bmatrix} 1 \\ 0 \end{bmatrix} = \begin{bmatrix} 0 \\ r \end{bmatrix}, implying 1=01 = 0, a contradiction. This concept is formalised as linear independence.

Definition 9 (Linear Dependence and Independence)

Two vectors a\mathbf{a} and b\mathbf{b} in R2\mathbb{R}^2 are said to be linearly dependent if one is a scalar multiple of the other. That is, either a=rb\mathbf{a} = r\mathbf{b} or b=sa\mathbf{b} = s\mathbf{a} for some scalars r,sr, s.

If neither vector is a scalar multiple of the other, they are said to be linearly independent.

Remark

If a=0\mathbf{a} = \mathbf{0}, then a=0b\mathbf{a} = 0\mathbf{b}, making the pair linearly dependent. Similarly, any vector is linearly dependent with itself. Thus, linear independence is a property relevant to distinct, non-zero vectors.

Geometrically, two non-zero vectors are linearly dependent if and only if they are collinear with the origin; that is, the directed segments OA\overrightarrow{OA} and OB\overrightarrow{OB} lie on the same straight line passing through OO.

This has a striking consequence for linear combinations. If a\mathbf{a} and b\mathbf{b} are linearly dependent and non-zero, every combination ca+dbc\mathbf{a} + d\mathbf{b} is a scalar multiple of a\mathbf{a} (since b\mathbf{b} itself is), so the combinations fill only the line through the origin in the direction of a\mathbf{a}. If a\mathbf{a} and b\mathbf{b} are linearly independent, the opposite holds: for any target vector t=[t1t2]\mathbf{t} = \begin{bmatrix} t_1 \\ t_2 \end{bmatrix}, the system ca+db=tc\mathbf{a} + d\mathbf{b} = \mathbf{t} amounts to two equations in two unknowns, and when a1b2a2b10a_1 b_2 - a_2 b_1 \neq 0 this system always has a unique solution. The combinations therefore fill the entire plane R2\mathbb{R}^2. Determining which situation holds is a question of fundamental importance.

Theorem 5 (Determinant Criterion for Dependence)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} and b=[b1b2]\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix}. The vectors a\mathbf{a} and b\mathbf{b} are linearly dependent if and only if

a1b2a2b1=0.a_1 b_2 - a_2 b_1 = 0.
Proof

Suppose a1b2a2b1=0a_1 b_2 - a_2 b_1 = 0. If a=0\mathbf{a} = \mathbf{0}, the vectors are dependent. Assume a0\mathbf{a} \neq \mathbf{0}; then at least one component, say a1a_1, is non-zero. From a1b2=a2b1a_1 b_2 = a_2 b_1, we have b2=a2a1b1b_2 = \frac{a_2}{a_1}b_1. We can express b\mathbf{b} as:

b=[b1b2]=[b1a2a1b1]=b1a1[a1a2]=b1a1a.\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} = \begin{bmatrix} b_1 \\ \frac{a_2}{a_1}b_1 \end{bmatrix} = \frac{b_1}{a_1} \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} = \frac{b_1}{a_1}\,\mathbf{a}.

Thus b\mathbf{b} is a scalar multiple of a\mathbf{a}. A similar argument holds if a20a_2 \neq 0.

Conversely, suppose a\mathbf{a} and b\mathbf{b} are linearly dependent. If a=rb\mathbf{a} = r\mathbf{b}, then a1=rb1a_1 = rb_1 and a2=rb2a_2 = rb_2. Substituting:

a1b2a2b1=(rb1)b2(rb2)b1=r(b1b2b2b1)=0.a_1 b_2 - a_2 b_1 = (rb_1)b_2 - (rb_2)b_1 = r(b_1 b_2 - b_2 b_1) = 0.

The same result follows if b=sa\mathbf{b} = s\mathbf{a}.

The expression a1b2a2b1a_1 b_2 - a_2 b_1 is called the determinant of the pair of vectors. It plays a role analogous to the discriminant in quadratic equations, providing a purely algebraic test for a geometric property. We will see more of this in subsequent notes.

Example 6

Consider a=[24]\mathbf{a} = \begin{bmatrix} 2 \\ 4 \end{bmatrix} and b=[12]\mathbf{b} = \begin{bmatrix} 1 \\ 2 \end{bmatrix}. The determinant is 2241=02 \cdot 2 - 4 \cdot 1 = 0, so the pair is linearly dependent. Indeed, a=2b\mathbf{a} = 2\mathbf{b}.

Now consider a=[31]\mathbf{a} = \begin{bmatrix} 3 \\ 1 \end{bmatrix} and b=[12]\mathbf{b} = \begin{bmatrix} 1 \\ 2 \end{bmatrix}. The determinant is 3211=503 \cdot 2 - 1 \cdot 1 = 5 \neq 0, so the pair is linearly independent: neither is a scalar multiple of the other.

We may also characterise linear dependence through linear combinations yielding the zero vector.

Theorem 6 (Algebraic Criterion for Dependence)

Two vectors a\mathbf{a} and b\mathbf{b} are linearly dependent if and only if there exist scalars rr and ss, not both zero, such that

ra+sb=0.r\mathbf{a} + s\mathbf{b} = \mathbf{0}.
Proof

Suppose such scalars exist. Without loss of generality, assume r0r \neq 0. Then we can rearrange:

ra=sb    a=(sr)b.r\mathbf{a} = -s\mathbf{b} \implies \mathbf{a} = \left(-\frac{s}{r}\right)\mathbf{b}.

Thus a\mathbf{a} is a scalar multiple of b\mathbf{b}, implying dependence.

Conversely, if a\mathbf{a} and b\mathbf{b} are dependent, then either a=kb\mathbf{a} = k\mathbf{b} or b=ka\mathbf{b} = k\mathbf{a}. In the first case, 1a+(k)b=01\cdot\mathbf{a} + (-k)\mathbf{b} = \mathbf{0} (where r=10r = 1 \neq 0). In the second, (k)a+1b=0(-k)\mathbf{a} + 1\cdot\mathbf{b} = \mathbf{0} (where s=10s = 1 \neq 0).

We summarise these findings in two equivalent statements, distinguishing the dependent and independent cases.

Theorem 7 (Conditions for Linear Dependence)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} and b=[b1b2]\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} be vectors in R2\mathbb{R}^2. The following statements are equivalent:

  1. a\mathbf{a} and b\mathbf{b} are linearly dependent.
  2. The points O=(0,0)O = (0,0), A=(a1,a2)A = (a_1, a_2), and B=(b1,b2)B = (b_1, b_2) are collinear.
  3. a1b2a2b1=0a_1 b_2 - a_2 b_1 = 0.
  4. There exist scalars r,sr, s, not both zero, such that ra+sb=0r\mathbf{a} + s\mathbf{b} = \mathbf{0}.
Theorem 8 (Conditions for Linear Independence)

Let a=[a1a2]\mathbf{a} = \begin{bmatrix} a_1 \\ a_2 \end{bmatrix} and b=[b1b2]\mathbf{b} = \begin{bmatrix} b_1 \\ b_2 \end{bmatrix} be vectors in R2\mathbb{R}^2. The following statements are equivalent:

  1. a\mathbf{a} and b\mathbf{b} are linearly independent.
  2. a\mathbf{a} and b\mathbf{b} are non-zero, and the points OO, AA, BB are distinct and not collinear.
  3. a1b2a2b10a_1 b_2 - a_2 b_1 \neq 0.
  4. The equation ra+sb=0r\mathbf{a} + s\mathbf{b} = \mathbf{0} implies r=0r = 0 and s=0s = 0.
Remark

One should observe the distinction between the two types of proofs presented in this chapter.

  1. Component-based proofs (e.g., Theorem 2, Theorem 5) rely on the explicit definition of a vector as an ordered pair of numbers. These are specific to R2\mathbb{R}^2.
  2. Axiomatic proofs (e.g., Theorem 3, Theorem 6) rely only on the algebraic properties of vector addition and scalar multiplication. These proofs are more powerful as they apply to any vector space, regardless of dimension or the nature of the vectors involved.
Problem 5

Let a=[23]\mathbf{a} = \begin{bmatrix} 2 \\ -3 \end{bmatrix} and b=[4t]\mathbf{b} = \begin{bmatrix} -4 \\ t \end{bmatrix}. Determine the value of tt for which a\mathbf{a} and b\mathbf{b} are linearly dependent. For this value of tt, express b\mathbf{b} as a scalar multiple of a\mathbf{a}.