Vector Operations:
- Addition
- Scalar multiplication
Linear Combinations (c, d, e $\in \mathbb{R}$):
- cu + dv
- cu + dv + ew
Properties of Dot Product:
- Distributive: $u^T(v + w) = u^Tv+u^Tw$
- Non-Associative: $u^T(v^Tw) \ne (u^Tv)^Tw$
- Commutative: $u^Tv = v^Tu$
Cosine Formula for Dot Product: $v^Tw = |v| . |w| . cos(θ)$ (note: cos(θ) = 0 when θ = 90°)
Length of a vector: $\left| \mathbf{v} \right| = \sqrt{v_1^2 + v_2^2 + v_3^2 + …}$
Matrix operations:
- Addition (both matrices should have the same dimensions)
- Scalar multiplication
- Transpose: mapping A ∈ $\mathbb{R}^{n \times m}$ to B ∈ $\mathbb{R}^{m \times n}$ with $a_{ij} = b_{ji}$
- Trace: sum of all the elements along the diagonal
- Matrix multiplication (reminder: A∈$\mathbb{R}^{n \times m}$ and B∈$\mathbb{R}^{m \times k}$ then C=AB∈$\mathbb{R}^{n \times k}$)
Multiplication Properties:
- Distributivity: $(A +B)C = AC+BC$ and $A(C+D) =AC+AD$
- Associativity: $(AB)C = A(BC)$
- Non-commutative: $AB \ne BA$ (because of dimensions)
- Multiplication with the identity matrix results in the matrix itself
Properties of Transpose:
- $(A+B)^T =A^T +B^T$ and $(A - B)^T =A^T - B^T$
- $(AB)^T =B^T . A^T$
Systems of Linear Equations Recap:
- $Ax = b$
- Gaussian Elimination
- Row Echelon/Reduced Row Echelon Form
A matrix is in row echelon form if
- all rows having only zero entries are at the bottom
- The pivot (the leftmost non-zero entry) of every non-zero row, called the pivot, is to the right of the leading entry of every row above
A matrix is in reduced row echelon form if:
- it is in row echelon form
- the leading entry in each nonzero row is 1 (called a leading one)
- each column containing a leading 1 has zeros in all its other entries (or in other words, above the pivot if condition one is achieved)
Properties of Determinants:
- Matrix must be square
- The determinant of the identity matrix is 1.
- The exchange of two rows multiplies the determinant by −1.
- Multiplying a row or a column by a number multiplies the determinant by this number.
- Adding a multiple of one row to another row does not change the determinant.
- If two rows of matrix A are equal, then $det(A) = 0$.
- A matrix with a row of zeros has $det(A) = 0$.
- If A is triangular then $det(A)$ is the product of diagonal entries.
- $det(AB) = det(A).det(B)$
- $det(A^T) = det(A)$
Laplace Expansion Example:

Invertibility:
- Matrix must be square
- If A is singular, then $det(A) = 0$. If A is invertible, then $det(A) \ne 0$.
- Can calculate the inverse using Gaussian Elimination
- $[A \vert I]$ -> $[I \vert A^-1]$
- Inverse for a 2x2:

Vector Spaces and Subspaces
A subspace is defined as a set of all vectors that can be created by taking linear combinations of some vectors or a set of vectors.
Formally, a subspace is the set of all vectors that satisfy the following conditions:
- Must be closed under addition and multiplication
- Must contain the zero vector
Note
A vector space needs to contain all linear combinations of its vectors.
Example 1: Possible subspaces
| Dimension | Subspaces in $R^2$: | Subspaces in $R^3$: |
|---|---|---|
| 0 | The zero vector | The zero vector |
| 1 | Lines that pass through the origin | Lines that pass through the origin |
| 2 | All of $R^2$ | Planes that pass through the origin |
| 3 | - | All of $R^3$ |
Dimension of a subspace
The dimension of a subspace is always ≤ the dimension of the space it lives in.
Example 2: What does NOT qualify as a subspace?
Think about a line that doesn’t pass through the origin - say, all points where $y = x + 1$. If we pick a vector on that line, like (0, 1), and we multiply it by the scalar 0, we get (0, 0) - which is NOT on the line $y = x + 1$. So that line isn’t closed under scalar multiplication, and therefore can’t be a subspace.
The distinction:
- While $y = x + 1$ is contained in ℝ² (every point on the line is in ℝ² - (0,1), (1,2), (2,3) etc.)
- It is NOT a subspace of ℝ² (doesn’t contain zero vector, not closed under operations)
- Therefore, it does not qualify as (does not form) a subspace.
The four fundamental subspaces of $A_{mxn}$
| Name | Dim | Note |
|---|---|---|
| Column Space | C(A)∈$R^m$ | Pivot columns of the original matrix |
| Null Space | N(A)∈$R^n$ | - Solve $Ax=0$, set free column variables to 1 respectively - # of free columns = dimension of null space |
| Row Space | C($A^T$)∈$R^n$ | Pivot rows |
| Left Null Space | N($A^T$)∈$R^m$ | Solve for y where $y^T.A=0$ |
Dimensions
For matrix A with dimensions mxn:
- $dim(C(A)) + dim(N(A)) = n$ (# of columns)
- $dim(C(A^T)) + dim(N(A^T)) = m$ (# of rows)
If the null space is {(0,0,0)} then (only consists of the zero vector),
- Dimension of null space = 0
- Number of free columns = 0
- All columns are pivot columns
Complete Solution System of Linear Equations
$x_{particular}$ : Set all free variables to 0, then solve $Ax=b$ $x_{null space}$ : Set free columns to 1 (if there are more than one, set them respectively)
\[x = x_{particular} + x_{null space}\]A system’s solution set has three possibilities:
- unique solution
- infinite solutions
- no solution
Rank
- Matrix rank = # of pivots
- In a square matrix, if the rank < # of columns that means there are linearly dependent rows, thus the matrix is not invertible ($det(A) = 0$)
- The rank of A = # of independent rows = # of independent columns (very important)
Basis
Tip for Finding the Span
If we want to find that \(\vec{v}=\begin{bmatrix} 1 \\ 2 \\ 0 \end{bmatrix}\) in the span of \(S = \{\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}, \begin{bmatrix} 1 \\ 1 \\ 7 \end{bmatrix}\}\), we can just look at the rank of \(\begin{bmatrix} 1 && 1 && 1 \\ 1 && 1 && 2 \\ 0 && 7 && 0 \end{bmatrix}\). If the rank is smaller than the number of columns, can be expressed as a linear combination of the other vectors. In this case, the rank is <3 so yes, $\vec{v}$ is in the span of S.
Why use different bases?
- Simplification: Some problems become way easier in a different basis
- Example: In physics, choosing a basis aligned with forces makes calculations simpler
- Natural coordinates: Sometimes a problem has a natural coordinate system
- Example: If you’re studying oscillations, sine and cosine functions form a natural basis
- Revealing structure: Different bases can reveal hidden patterns in data
- Example: Principal Component Analysis (PCA) in data science finds a basis that shows the most important directions in your data
Changing Basis
Solve for $\vec{w}$ where the basis A, B and $\vec{v}$ is known: $A.\vec{v} = B.\vec{w}$ You get $\vec{w}$ represented in basis B.