Vector Operations:

  • Addition
  • Scalar multiplication

Linear Combinations (c, d, e $\in \mathbb{R}$):

  • cu + dv
  • cu + dv + ew

    Properties of Dot Product:

  • Distributive: $u^T(v + w) = u^Tv+u^Tw$
  • Non-Associative: $u^T(v^Tw) \ne (u^Tv)^Tw$
  • Commutative: $u^Tv = v^Tu$

Cosine Formula for Dot Product: $v^Tw = |v| . |w| . cos(θ)$ (note: cos(θ) = 0 when θ = 90°)

Length of a vector: $\left| \mathbf{v} \right| = \sqrt{v_1^2 + v_2^2 + v_3^2 + …}$

Matrix operations:

  • Addition (both matrices should have the same dimensions)
  • Scalar multiplication
  • Transpose: mapping A ∈ $\mathbb{R}^{n \times m}$ to B ∈ $\mathbb{R}^{m \times n}$ with $a_{ij} = b_{ji}$
  • Trace: sum of all the elements along the diagonal
  • Matrix multiplication (reminder: A∈$\mathbb{R}^{n \times m}$ and B∈$\mathbb{R}^{m \times k}$ then C=AB∈$\mathbb{R}^{n \times k}$)

Multiplication Properties:

  • Distributivity: $(A +B)C = AC+BC$ and $A(C+D) =AC+AD$
  • Associativity: $(AB)C = A(BC)$
  • Non-commutative: $AB \ne BA$ (because of dimensions)
  • Multiplication with the identity matrix results in the matrix itself

Properties of Transpose:

  • $(A+B)^T =A^T +B^T$ and $(A - B)^T =A^T - B^T$
  • $(AB)^T =B^TA^T$

Symmetric Matrices:

  • $A^TB = B^TA$ if $A^TB$ is symmetric

    Systems of Linear Equations Recap:

  • $Ax = b$
  • Gaussian Elimination
  • Row Echelon/Reduced Row Echelon Form

A matrix is in row echelon form if

  • all rows having only zero entries are at the bottom
  • The pivot (the leftmost non-zero entry) of every non-zero row, called the pivot, is to the right of the leading entry of every row above

A matrix is in reduced row echelon form if:

  • it is in row echelon form
  • the leading entry in each nonzero row is 1 (called a leading one)
  • each column containing a leading 1 has zeros in all its other entries (or in other words, above the pivot if condition one is achieved)

    Properties of Determinants:

  • Matrix must be square
  • The determinant of the identity matrix is 1.
  • The exchange of two rows multiplies the determinant by −1.
  • Multiplying a row or a column by a number multiplies the determinant by this number.
  • Adding a multiple of one row to another row does not change the determinant.
  • If two rows of matrix A are equal, then $det(A) = 0$.
  • A matrix with a row of zeros has $det(A) = 0$.
  • If A is triangular then $det(A)$ is the product of diagonal entries.
  • $det(AB) = det(A).det(B)$
  • $det(A^T) = det(A)$

Laplace Expansion Example:

Invertibility:

  • Matrix must be square
  • If A is singular, then $det(A) = 0$. If A is invertible, then $det(A) \ne 0$.
  • Can calculate the inverse using Gaussian Elimination
    • $[A \vert I]$ -> $[I \vert A^-1]$
  • Inverse for a 2x2:

Vector Spaces and Subspaces

A subspace is defined as a set of all vectors that can be created by taking linear combinations of some vectors or a set of vectors.

Formally, a subspace is the set of all vectors that satisfy the following conditions:

  • Must be closed under addition and multiplication
  • Must contain the zero vector

Note

A vector space needs to contain all linear combinations of its vectors.

Example 1: Possible subspaces

Dimension Subspaces in $R^2$: Subspaces in $R^3$:
0 The zero vector The zero vector
1 Lines that pass through the origin Lines that pass through the origin
2 All of $R^2$ Planes that pass through the origin
3 - All of $R^3$

Dimension of a subspace

The dimension of a subspace is always ≤ the dimension of the space it lives in.

Example 2: What does NOT qualify as a subspace? Think about a line that doesn’t pass through the origin - say, all points where $y = x + 1$. If we pick a vector on that line, like (0, 1), and we multiply it by the scalar 0, we get (0, 0) - which is NOT on the line $y = x + 1$. So that line isn’t closed under scalar multiplication, and therefore can’t be a subspace.

The distinction:

  • While $y = x + 1$ is contained in ℝ² (every point on the line is in ℝ² - (0,1), (1,2), (2,3) etc.)
  • It is NOT a subspace of ℝ² (doesn’t contain zero vector, not closed under operations)
  • Therefore, it does not qualify as (does not form) a subspace.

The four fundamental subspaces of $A_{mxn}$

Name Dim Note
Column Space C(A)∈$R^m$ Pivot columns of the original matrix
Null Space N(A)∈$R^n$ - Solve $Ax=0$, set free column variables to 1 respectively
- # of free columns = dimension of null space
Row Space C($A^T$)∈$R^n$ Pivot rows
Left Null Space N($A^T$)∈$R^m$ Solve for y where $y^T.A=0$

Dimensions

For matrix A with dimensions $m$x$n$:

  • $dim(C(A)) + dim(N(A)) = n$ (# of columns)
  • $dim(C(A^T)) + dim(N(A^T)) = m$ (# of rows)

Orthogonal Subspaces

  • $N(A) \bot C(A^T)$
  • $N(A^T) \bot C(A)$

If the null space is {(0,0,0)} then (only consists of the zero vector),

  • Dimension of null space = 0
  • Number of free columns = 0
  • All columns are pivot columns

Null space of A

If vector $\vec{v}$ is in the null space of matrix A, then A$\vec{v}=0$.

Complete Solution System of Linear Equations $x_{particular}$ : Set all free variables to 0, then solve $Ax=b$ (or $Rx=b$ where R is the echelon/reduced echelon form)

$x_{null space}$ : Set free columns to 1 and solve $Rx=0$. If there are more than one, set them respectively. For example, if there are two free variables (let’s say $x_2$ and $x_3$) first set $x_2 = 1$, $x_3 = 0$ and solve for $Rx=0$. Then set $x_2 = 0$, $x_3 = 1$ and solve for $Rx=0$. Two free variables mean there will be two special solutions.

x=xparticular+xnullspacex = x_{particular} + x_{null space}

A system’s solution set has three possibilities:

  1. unique solution
  2. infinite solutions
  3. no solution

If $det(A) = 0$ the system has either,

  • no solution OR
  • infinitely many solutions

Rank

  • Matrix rank = # of pivots
  • In a square matrix, if the rank < # of columns that means there are linearly dependent rows, thus the matrix is not invertible ($det(A) = 0$)
  • The rank of A = # of independent rows = # of independent columns (very important)

Basis

Tip for Finding the Span

If we want to find that v=[120]\vec{v}=\begin{bmatrix} 1 \\ 2 \\ 0 \end{bmatrix} in the span of S={[110],[117]}S = \{\begin{bmatrix} 1 \\ 1 \\ 0 \end{bmatrix}, \begin{bmatrix} 1 \\ 1 \\ 7 \end{bmatrix}\}, we can just look at the rank of [111112070]\begin{bmatrix} 1 && 1 && 1 \\ 1 && 1 && 2 \\ 0 && 7 && 0 \end{bmatrix} If the rank is smaller than the number of columns, can be expressed as a linear combination of the other vectors. In this case, the rank is <3 so yes, $\vec{v}$ is in the span of S.

Why use different bases?

  1. Simplification: Some problems become way easier in a different basis
    • Example: In physics, choosing a basis aligned with forces makes calculations simpler
  2. Natural coordinates: Sometimes a problem has a natural coordinate system
    • Example: If you’re studying oscillations, sine and cosine functions form a natural basis
  3. Revealing structure: Different bases can reveal hidden patterns in data
    • Example: Principal Component Analysis (PCA) in data science finds a basis that shows the most important directions in your data

Changing Basis

Solve for $\vec{w}$ where the basis A, B and $\vec{v}$ is known: $A.\vec{v} = B.\vec{w}$ You get $\vec{w}$ represented in basis B.

Orthogonality

  • Two vectors are orthogonal if their dot product is 0.
  • Two subspaces $V$ and $W$ of a vector space are orthogonal if every vector in $V$ is perpendicular to every vector in $W$.
  • The null space $N(A)$ and the row space $C(A^T)$ are orthogonal subspaces of $R^n$.
  • The left null space $N(A^T)$ and the column space $C(A)$ are orthogonal subspaces of $R^m$.

Projections

The projection of $b$ onto a subspace $S$ is the closest vector $p$ in $S$. A projection matrix $P$ is a symmetric matrix with $P^2=P$. The projection of is $b$ is given by $Pb$.

Projection onto a line

p=xap = xa

x=aTbaTax = {a^Tb \over a^Ta}

p=aTbaTaap = {a^Tb \over a^Ta}a

P=aaTaTaP = {aa^T \over a^Ta}

  • x is a scalar.

    Projection onto a Subspace

p=Axp = Ax

x=(ATA)1ATbx = (A^TA)^{-1}A^Tb

p=A(ATA)1ATbp = A(A^TA)^{-1}A^Tb

P=A(ATA)1ATP = A(A^TA)^{-1}A^T

  • x is a scalar.

Theorem: If $A$ has linearly independent columns then $A^TA$ is invertible.

Least Squares Approximation

  • Solve $A^TAx = A^Tb$ where

x=(cm)x = \begin{pmatrix} c \\ m \end{pmatrix}

and $y=c+mx$

Orthonormal Vectors

Vectors $q_1, …, q_n$ are orthonormal if $q_i^Tq_j=$

  • 0 when $i\neq j$
  • 1 when $i=j$

Assume A has orthonormal column, then we can find the projection matrix using

x=(QTQ)1QTbx = (Q^TQ)^{-1}Q^Tb

x=QTbx = Q^Tb

P=Q(QTQ)1QTP = Q(Q^TQ)^{-1}Q^T

P=QI1QTP = QI^{-1}Q^T

P=QQTP = QQ^T

Gram-Schmidt Process

Eigenvectors & Eigenvalues

If, Av=λvAv = \lambda vthen $v$ is an eigenvector of A with eigenvalue $\lambda$. In other words, $Av$ lies in the same one-dimensional subspace as $\lambda v$.

Finding eigenvectors:

Av=λvAv = \lambda v

Avλv=0Av - \lambda v = 0

(AλI)v=0(A - \lambda I)v = 0

$(A - \lambda I)$ should have a non-trivial null space. Which means it is singular and $det(A - \lambda I)=0$.

Diagonalization

$AV = VΛ$ where $Λ$ is the diagonal eigenvalue matrix and $V$ is the eigenvector matrix.

Spectral Theorem

For a symmetric matrix $S$, the following diagonalization can be written:

S=QΛQ1=QΛQTS = QΛQ^{-1} = QΛQ^T

where $Q$ is the orthonormal eigenvector matrix. i.e. eigenvectors of a real symmetric matrix are always perpendicular.

Singular Value Decomposition

Tip: SVD can be used for dimensionality reduction or compression.

A=UεVTA = U\varepsilon V^T

ATA=(UεVT)UεVTA^TA = (U\varepsilon V^T)U\varepsilon V^T

ATA=VεTUTUεVTA^TA = V\varepsilon^T U^TU\varepsilon V^T

ATA=VεTεVTA^TA = V\varepsilon^T \varepsilon V^T

ATA=Vε2VTA^TA = V\varepsilon^2 V^T

or,

AAT=Uε2UTAA^T = U\varepsilon^2 U^T

Once you find $V$ and $\varepsilon$, find $U$ by calculating $Av_i = \sigma_i u_i$ where $v$ and $u$ are eigenvectors of $V$ and $U$. $\sigma_i$ is the square root of eigenvalues obtained from $\varepsilon^2$.

Important note: $V$ and $U$ should have unit eigenvectors. Also, sort eigenvectors from in descending order of corresponding eigenvalues.