Linear Algebra
MATH 355
fall, sophomore year
existence
uniqueness
coefficient matrix
augmented matrix
Linear algebra formalizes the procedure to find all solutions to a linear system of equations when they exist, or to prove that they don’t exist.
Linear Combinations and Span
Linear Independence
T : V → W T : V \rightarrow W T : V → W is a linear transformation if:
For u , v ∈ V u, v \in V u , v ∈ V , T ( u + v ) = T ( u ) + T ( v ) T(u + v) = T(u) + T(v) T ( u + v ) = T ( u ) + T ( v )
For v ∈ V v \in V v ∈ V and scalar a a a , T ( a v ) = a ⋅ T ( v ) T(av) = a \cdot T(v) T ( a v ) = a ⋅ T ( v )
Properties:
T ( 0 ) = 0 T(0) = 0 T ( 0 ) = 0
any linear transformations respects linear combinations
Onto and Into
The image of a linear transformation T : V → W T : V \rightarrow W T : V → W is the set { T ( v ) : v ∈ V } \{T(v) : v \in V\} { T ( v ) : v ∈ V } . The image is also the span of the columns of A A A , the std matrix of T T T .
A linear transformation T : V → W T : V \rightarrow W T : V → W is onto if the image of T T T is W W W .
A linear transformation T : V → W T : V \rightarrow W T : V → W is into if for every u , v ∈ V u, v \in V u , v ∈ V , T ( u ) = T ( V ) ⟹ u = v T(u) = T(V) \implies u = v T ( u ) = T ( V ) ⟹ u = v .
T T T is onto
T T T is into
A A A 's RREF has a pivot in each row
A A A 's RREF has a pivot in each column
The columns of A A A span R m \mathbb{R}^m R m
The columns of A A A are LI
A x = w Ax = w A x = w has a solution for each w ∈ R m w \in \mathbb{R}^m w ∈ R m
A x = 0 Ax = 0 A x = 0 has only the trivial solution x = 0 x = 0 x = 0
The image of T T T is R m \mathbb{R}^m R m
The kernel of T T T is { 0 } \{0\} { 0 }
Existence
Uniqueness
Matrix Multiplication
If S : R p → R n S : \mathbb{R}^p \rightarrow \mathbb{R}^n S : R p → R n and T : R n → R m T : \mathbb{R}^n \rightarrow \mathbb{R}^m T : R n → R m are linear transformations, T ∘ S : R p → R m T \circ S : \mathbb{R}^p \rightarrow \mathbb{R}^m T ∘ S : R p → R m is a linear transformation.
If the m × n m \times n m × n matrix A A A is the std matrix for T : R n → R m T : \mathbb{R}^n \rightarrow \mathbb{R}^m T : R n → R m and the n × p n \times p n × p matrix B B B is the std matrix for S : R p → R n S : \mathbb{R}^p \rightarrow \mathbb{R}^n S : R p → R n , then A B AB A B will be a m × p m \times p m × p matrix which is the std matrix for T ∘ S : R p → R m T \circ S : \mathbb{R}^p \rightarrow \mathbb{R}^m T ∘ S : R p → R m .
For a m × n m \times n m × n matrix A A A , A I n = A A I_n = A A I n = A and I m A = A I_m A = A I m A = A .
Inverses
Consider T : V → W T: V \rightarrow W T : V → W and S : W → V S : W \rightarrow V S : W → V . S S S and T T T are inverses of each other if S ( T ( V ) ) = v ∀ v ∈ V S(T(V)) = v \forall v \in V S ( T ( V ) ) = v ∀ v ∈ V and T ( S ( w ) ) = w ∀ w ∈ W T(S(w)) = w \forall w \in W T ( S ( w ) ) = w ∀ w ∈ W .
If S = T − 1 S = T^{-1} S = T − 1 exists, T T T is invertible.
A linear transformation is invertible iff it is both into and onto.
Consider a m × n m \times n m × n matrix A A A and a n × m n \times m n × m matrix B B B . A A A and B B B are inverses of each other if A B = I m AB = I_m A B = I m and B A = I n BA = I_n B A = I n .
If B = A − 1 B = A^{-1} B = A − 1 exists, A A A is invertible.
Transposes and Inverses
A = ( A t ) t A = {(A^t)}^t A = ( A t ) t
A matrix A A A is symmetric if A = A t A = A^t A = A t
If A A A and B B B are matrices such that A B AB A B is defined, ( A B ) t = B t A t {(AB)}^t = B^t A^t ( A B ) t = B t A t .
If A A A is invertible, A t A^t A t is invertible. ( A t ) − 1 = ( A − 1 ) t {(A^t)}^{-1} = {(A^{-1})}^t ( A t ) − 1 = ( A − 1 ) t .
Invertible Matrix Theorem
Consider a linear transformation T : R n → R n T : \mathbb{R}^n \rightarrow \mathbb{R}^n T : R n → R n with a n × n n \times n n × n standard matrix A A A . The following statements are equivalent:
T T T is into
ker ( T ) = { 0 } \ker(T) = \{0\} ker ( T ) = { 0 }
The columns of A A A are linearly independent
A x = 0 Ax = 0 A x = 0 has only the trivial solution
A A A 's RREF has a pivot in every column (no free vars)
A A A 's RREF is the identity matrix
A A A 's RREF has a pivot in each row
For every v ∈ R n v \in \mathbb{R}^n v ∈ R n , A x = v Ax = v A x = v is consistent
The columns of A A A span R n \mathbb{R}^n R n
T T T is onto
T T T is invertible
A A A is invertible
A t A^t A t is invertible
The rows of A A A span R n \mathbb{R}^n R n
The rows of A A A are linearly independent
To calculate the matrix of a matrix, row reduce it with the identity matrix on the right side.
Elementary Matrices
Scaling a row
E 1 = [ 1 ⋱ 1 c 1 ⋱ 1 ] E_1 = \left[ \begin{array}{ccccccc} 1 & & & & & & \\ & \ddots & & & & & \\ & & 1& & & & \\ & & & c & & & \\ & & & & 1 & & \\ & & & & & \ddots & \\ & & & & & & 1 \end{array} \right]
E 1 = ⎣ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎡ 1 ⋱ 1 c 1 ⋱ 1 ⎦ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎤
Adding c c c times a row to another row
E 2 = [ 1 ⋱ 1 ⋱ c 1 ⋱ 1 ] E_2 = \left[ \begin{array}{ccccccc} 1 & & & & & & \\ & \ddots & & & & & \\ & & 1& & & & \\ & & & \ddots & & & \\ & & c & & 1 & & \\ & & & & & \ddots & \\ & & & & & & 1 \end{array} \right]
E 2 = ⎣ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎡ 1 ⋱ 1 c ⋱ 1 ⋱ 1 ⎦ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎤
Swapping two rows
E 3 = [ 1 ⋱ 0 1 1 ⋱ 1 1 0 ⋱ 1 ] E_3 = \left[ \begin{array}{ccccccccccc} 1 & & & & & & & & & & \\ & \ddots & & & & & & & & & \\ & & & 0 & & & & 1 & & & \\ & & & & 1 & & & & & & \\ & & & & & \ddots & & & & & \\ & & & & & & 1 & & & & \\ & & & 1 & & & & 0 & & & \\ & & & & & & & & & \ddots & \\ & & & & & & & & & & 1 \end{array} \right]
E 3 = ⎣ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎡ 1 ⋱ 0 1 1 ⋱ 1 1 0 ⋱ 1 ⎦ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎥ ⎤
Dimension
If V V V is spanned by a finite list of vectors, the dimension of V V V is the number of vectors in a basis. If V V V is nto spanned by a finite set, V V V is infinite-dimensional.
Any two bases of V V V have the same size.
Any set of more than n n n vectors is linearly dependent.
Consider a n n n -dimensional vector space V V V :
if { v 1 , ⋯ , v n } \{v_1, \cdots, v_n\} { v 1 , ⋯ , v n } are LI, they also span V V V and hence form a basis
if { v 1 , ⋯ , v n } \{v_1, \cdots, v_n\} { v 1 , ⋯ , v n } span V V V , they are also LI and hence form a basis
Matrices Relative to Bases
Let T : V → W T: V \rightarrow W T : V → W . B = { b 1 , … , b n } B = \{b_1, \dots, b_n\} B = { b 1 , … , b n } is a basis for V V V and C C C is a basis for W W W . If A A A is the matrix for T T T relative to B B B and C C C , A [ v ] B = [ T ( v ) ] C A {[v]}_B = {[T(v)]}_C A [ v ] B = [ T ( v ) ] C for all v ∈ V v \in V v ∈ V .
A A A has the columns: [ T ( b 1 ) ] C , … , [ T ( b n ) ] C {[T(b_1)]}_C, \dots, {[T(b_n)]}_C [ T ( b 1 ) ] C , … , [ T ( b n ) ] C .
T T T is invertible iff A A A is invertible.
A − 1 A^{-1} A − 1 is the matrix for T − 1 T^{-1} T − 1 relative to C C C and B B B .
Similarity and Diagonalization
A square n × n n \times n n × n matrix A A A is diagonalizable if it is similar to some diagonal matrix D D D .
If A A A has n n n linearly independent eigenvectors { v 1 , … , v n } \{v_1, \dots, v_n\} { v 1 , … , v n } , A A A is diagonalizable and A = P D P − 1 A = P D P^{-1} A = P D P − 1 , where P P P has columns v 1 , … , v n v_1, \dots, v_n v 1 , … , v n and D D D has λ 1 , … , λ n \lambda_1, \dots, \lambda_n λ 1 , … , λ n along its diagonal.
Eigentheory
Any list of eigenvectors for a matrix A A A with distinct eigenvalues must be linearly independent.
A n × n n \times n n × n matrix has at most n n n distinct eigenvalues.
If a n × n n \times n n × n matrix has n n n distinct eigenvalues, it has an eigenbasis and is therefore diagonalizable.
If A A A is a square matrix and λ \lambda λ is a scalar, ker ( A − λ I ) \ker (A - \lambda I) ker ( A − λ I ) (plus 0 ⃗ \vec 0 0 ) is the λ \lambda λ -eigenspace of A A A .
The characteristic polynomial of A A A in λ \lambda λ is ∣ A − λ I ∣ |A - \lambda I| ∣ A − λ I ∣ . A scalar is an eigenvalue of A A A iff it is a solution to ∣ A − λ I = 0 |A - \lambda I = 0 ∣ A − λ I = 0 .
If λ \lambda λ is an eigenvalue for A A A , the algebraic multiplicity of A A A is the number of times λ \lambda λ appears as a root of the characteristic poly (aka the exponent of the corresponding term). The geometric multiplicity of λ \lambda λ is the dimension of the λ \lambda λ -eigenspace (dim A − λ I \dim A - \lambda I dim A − λ I ).
Complex Numbers
Let z = a + b i z = a + bi z = a + b i . ℜ ( z ) = a = z + z ‾ 2 \Re(z) = a = \frac{z + \overline{z}}{2} ℜ ( z ) = a = 2 z + z and ℑ ( z ) = b = z − z ‾ 2 i \Im(z) = b = \frac{z - \overline{z}}{2i} ℑ ( z ) = b = 2 i z − z .
Fundamental theorem of algebra: Any non-constant polynomial with real or complex coefficients factors completely into linear factors over C \mathbb{C} C .
If f f f is a polynomial with real coefficients, and f ( z ) = 0 f(z) = 0 f ( z ) = 0 , f ( z ‾ ) f(\overline{z}) f ( z ) is also zero.
Norm: ∣ z ∣ = a 2 + b 2 |z| = \sqrt{a^2 + b^2} ∣ z ∣ = a 2 + b 2
Argument: tan − 1 b a \tan^{-1} \frac{b}{a} tan − 1 a b
∣ z w ∣ = ∣ z ∣ ⋅ ∣ w ∣ |zw| = |z| \cdot |w| ∣ z w ∣ = ∣ z ∣ ⋅ ∣ w ∣ and arg ( z w ) = arg ( z ) + arg ( w ) \arg(zw) = \arg(z) + \arg(w) arg ( z w ) = arg ( z ) + arg ( w )
Complex Eigenvalues
If v v v is an eigenvector ofor a real matrix with eigenvalue λ \lambda λ , then v ‾ \overline v v is also an eigenvector with eigenvalue λ ‾ \overline \lambda λ .
Suppose A A A is a real 2 × 2 2 \times 2 2 × 2 matrix. Suppose that λ = a + b i \lambda = a + bi λ = a + b i is a non-real complex eigenvalue for A A A , with a corresponding eigenvector v v v . Let x = ℜ ( v ) x = \Re(v) x = ℜ ( v ) and y = − ℑ ( v ) y = -\Im(v) y = − ℑ ( v ) . Then B = { x ; y } B = \{x; y\} B = { x ; y } is a basis of R 2 \mathbb{R}^2 R 2 , and the matrix for A A A with respect to B B B is [ a − b b a ] \begin{bmatrix}a & -b \\ b & a\end{bmatrix} [ a b − b a ] .
Inner Products
For two vectors v = [ a 1 , … , a n ] v = [a_1, \dots, a_n] v = [ a 1 , … , a n ] and w = [ b 1 , … , b n ] w = [b_1, \dots, b_n] w = [ b 1 , … , b n ] in R n \mathbb{R}^n R n , v w ˙ = ⟨ v , w ⟩ = a 1 b 1 + ⋯ + a n b n v \dot w = \langle v, w \rangle = a_1 b_1 + \dots + a_n b_n v w ˙ = ⟨ v , w ⟩ = a 1 b 1 + ⋯ + a n b n .
⟨ v , w ⟩ = ⟨ w , v ⟩ \langle v, w \rangle = \langle w, v \rangle ⟨ v , w ⟩ = ⟨ w , v ⟩
⟨ a v , w ⟩ = a ⋅ ⟨ v , w ⟩ \langle av, w \rangle = a \cdot \langle v, w \rangle ⟨ a v , w ⟩ = a ⋅ ⟨ v , w ⟩
⟨ v , v ⟩ ≥ 0 \langle v, v \rangle \geq 0 ⟨ v , v ⟩ ≥ 0
⟨ u + v , w ⟩ = ⟨ u , w ⟩ + ⟨ v , w ⟩ \langle u + v, w \rangle = \langle u, w \rangle + \langle v, w \rangle ⟨ u + v , w ⟩ = ⟨ u , w ⟩ + ⟨ v , w ⟩
Norm /Magnitude ∣ ∣ v ∣ ∣ = ⟨ v , v ⟩ ||v|| = \sqrt{\langle v, v \rangle} ∣ ∣ v ∣ ∣ = ⟨ v , v ⟩
The distance between two vectors v v v and w w w is ∣ ∣ v − w ∣ ∣ ||v - w|| ∣ ∣ v − w ∣ ∣ .
For two vectors v v v , w w w in a vector space V V V that has an inner product, v v v and w w w are orthogonal if ⟨ v , w ⟩ = 0 \langle v, w \rangle = 0 ⟨ v , w ⟩ = 0 . Additionally, v v v and w w w are orthogonal iff ∣ ∣ v + w ∣ ∣ 2 = ∣ ∣ v ∣ ∣ 2 + ∣ ∣ w ∣ ∣ 2 {||v + w||}^2 = {||v||}^2 + {||w||}^2 ∣ ∣ v + w ∣ ∣ 2 = ∣ ∣ v ∣ ∣ 2 + ∣ ∣ w ∣ ∣ 2 .
A list { v 1 , … , v n } \{v_1, \dots, v_n\} { v 1 , … , v n } is orthogonal iff ⟨ v i , v j ⟩ = 0 \langle v_i, v_j \rangle = 0 ⟨ v i , v j ⟩ = 0 for i ≠ j i \neq j i = j . If the list is orthogonal and ∣ ∣ v i ∣ ∣ = 1 ||v_i|| = 1 ∣ ∣ v i ∣ ∣ = 1 for all i i i , it is also orthonormal .
A list of non-zero orthogonal vectors is linearly independent.
Orthogonal Projections
For a v ∈ V v \in V v ∈ V ; W W W , a subspace of V V V ; { w 1 , … , w k } \{w_1, \dots, w_k\} { w 1 , … , w k } , an orthonormal basis of W W W ; v ^ = ∑ i = 1 k ⟨ v , w i ⟩ w i \hat v = \sum_{i = 1}^k \langle v, w_i \rangle w_i v ^ = ∑ i = 1 k ⟨ v , w i ⟩ w i .
v ^ ∈ W \hat v \in W v ^ ∈ W
v − v ^ v - \hat v v − v ^ is perpendicular to all vectors in W W W
v ^ \hat v v ^ is the only vector that satisfies (1) and (2)
v ^ \hat v v ^ is the closest vector to v v v in W W W
If W W W has an orthogonal basis { u 1 , … , u k } \{u_1, \dots, u_k\} { u 1 , … , u k } , v ^ = ∑ i = 1 k ⟨ v , w i ⟩ w i ⟨ w i , w i ⟩ \hat v = \sum_{i = 1}^k \langle v, w_i \rangle \frac{w_i}{\langle w_i, w_i \rangle} v ^ = ∑ i = 1 k ⟨ v , w i ⟩ ⟨ w i , w i ⟩ w i .
⟨ v , w ⟩ = v t w \langle v, w \rangle = v^t w ⟨ v , w ⟩ = v t w
If A A A is an n × n n \times n n × n matrix and the columns of A A A form an ONB of R n \mathbb{R}^n R n , A t A = I n A^t A = I_n A t A = I n and A t = A − 1 A^t = A^{-1} A t = A − 1 .
Orthogonal Decomposition
V V V is the direct sum of U U U and W W W if U + W = V U + W = V U + W = V and U ∩ W = 0 U \cap W = 0 U ∩ W = 0 . V = U ⊕ W V = U \oplus W V = U ⊕ W .
Let W W W be a subspace of V V V . W ⊥ W^\perp W ⊥ , the orthogonal complement of W W W , is { v ∈ V : ⟨ v , w ⟩ = 0 ∀ w ∈ W } \{v \in V : \langle v, w \rangle = 0 \forall w \in W\} { v ∈ V : ⟨ v , w ⟩ = 0 ∀ w ∈ W } . W ⊥ W^\perp W ⊥ is a subspace of V V V . Additionally, V = W ⊕ W ⊥ V = W \oplus W^\perp V = W ⊕ W ⊥ .
( W ⊥ ) ⊥ = W {(W^\perp)}^\perp = W ( W ⊥ ) ⊥ = W
( ker A ) ⊥ = i m A t {(\ker A)}^\perp = im A^t ( ker A ) ⊥ = i m A t and ( i m A ) ⊥ = ker A t {(im A)}^\perp = \ker A^t ( i m A ) ⊥ = ker A t
Least Squares
Let A A A be a m × n m \times n m × n matrix and v ∈ R m v \in \mathbb{R}^m v ∈ R m . A t A x = A t v A^t A x = A^t v A t A x = A t v is consistent, and its solutions are least-squares solutions to A x = v Ax = v A x = v .
Orthogonal Matrices
If A A A is symmetric, eigenvectors with distinct eigenvalues are orthogonal.
If A A A is a real, symmetric matrix, A A A is orthogonally diagonalizable.
For an n × n n \times n n × n matrix A A A , the following are equivalent:
A A A is an orthogonal matrix
The columns form an orthonormal basis for R n \mathbb{R}^n R n
A t = A − 1 A^t = A^{-1} A t = A − 1
A A A preserves inner products: ⟨ A v , A w ⟩ = ⟨ v , w ⟩ \langle Av, Aw \rangle = \langle v, w \rangle ⟨ A v , A w ⟩ = ⟨ v , w ⟩
A A A preserves norms: ∣ ∣ A v ∣ ∣ = ∣ ∣ v ∣ ∣ ||Av|| = ||v|| ∣ ∣ A v ∣ ∣ = ∣ ∣ v ∣ ∣
Spectral Theorem
A A A is symmetric if and only if A A A is orthogonally diagonalizable.
Spectral theorem : For a symmetric matrix A A A :
All roots of ∣ A − λ I ∣ |A - \lambda I| ∣ A − λ I ∣ are real
The geometric multiplicity of each eigenvalue equals its algebraic multiplicity
Eigenspaces with distinct eigenvalues are mutually orthogonal
A A A is orthogonally diagonalizable