Cheat Sheet Fionaw - Linear Algebra Math 232
Cheat Sheet Fionaw - Linear Algebra Math 232
Basic Equations Vectors, Norm, Dot Product (cont) Matrix Algebra, Identity and Inverse Matrix
Network Flows u·v = u1v1 + u2v2 dot product (A + B)ij = (A)ij + (B)ij (A - B)ij = (A)ij -
1. the flow in an arc is only in one directions ...+unvn (B)ij
||u|| ||v|| cos(θ) (cA)ij = c(A)ij (AT)ij = (A)ji
2. flow into a node = flow out of a node
u and v are orthogonal if u·v = 0 (cos(θ) = 0) (AB)ij = ai1b1j + ai2b2j + ... aikbkj
3. flow into the network = flow out of the
network a set of vectors is an orthogonal set iff vi·vj Inner Product (number) is u Tv = u·v , u and v
= 0,if i≠j same size
Balancing Chemical Equations
a set of vectors is an orthonormal set iff vi·vj Outer Product (matrix) is uvT, u and v can
1. add x's before each combo and both side
= 0,if i≠j, and ||vi|| = 1 for all i be any size
2. carbo = x1 + 2(x3), set as system, solve
(u·v)2 ≤ ||u||2||v||2 or Cauchy-Schwarz (AT) T = A (kA)T = k(A)T
Matrix
|u·v| ≤ ||u|| ||v|| Inequality
(A+B)T = AT + BT (AB)T = BTAT
augmented variables and soluti‐
d(uv) ≤ d(u,w) + Triangle Inequality
matrix on(rhs) tr(A T) = tr(A) tr(AB) = tr(BA)
d(w,v)
T T
coefficient coefficients only, no rhs u v = tr(uv ) tr(uvT) = tr(vuT)
||u+v|| ≤ ||u|| + ||v||
matrix tr(A) = a11 + a22 ... + (AT)ij = Aji
||v1 + v2 ... + vk|| = ||v1|| + ||v2|| ... + ||vk||
ann
Vectors, Norm, Dot Product Identity matrix is square matrix with 1 along
Lines and Planes
maginitude (norm) of vector v is ||v||; ||v|| ≥ 0 diagonals
a vector equation with x = x0 + tv,
if k>0, kv same direction magnitude = parameter t -∞ < t < +∞ If A is m x n, A n = A and mA = A
as v k||v|| a square matrix is AB = = BA
solutin set for 3 dimension linear equation is
if k<0, kv opposite magnitude = |k| a plane invertible(nonsingular)
direction to v ||v|| if:
if x is a point on this plane n·(x-x0) = 0
vectors in Rn (n = v = (v1, v2, ..., (point-normal equation) B is the inverse of A B = A-1
dimension) vn) if A has no inverse, A is not invertible
A(x-x0)+B(y-y0)+C(z-z0) = x0 =
v = P1P2 = OP2 - OP1 displacement 0 (x0,y0,z0), (singular)
vector n = (A, B, C) det(A) = ad - bc ≠ 0 is invertible
2
norm/magnitude of vector sqrt( (v1) + general/algebraic equation Ax+By+Cz = if A is invertible: (AB)-1 = B-1A-1
2
||v|| (v2) ...) D (An) -1 = A-n = (A-1) n (AT) -1 = (A-1) T
||v|| = 0 iff v =0 ||kv|| = |k| ||v|| two planes are parallel if n1 = kn2, (kA)-1 1/k(A -1), k≠0
unit vector u in same u = (1/ ||v||) v orthogonal if n1·n2 = 0
direct as v Elementary Matrix and Unifying Theorem
e1 = (1,0...) ... en = standard unit elementary matrices are invertible
(0,...1) in Rn vector
A-1 = Ek Ek-1 ... E2 E1
d(u,v) = sqrt((u1-v1)2 + (u2-v2)2 ... (un-vn)2)
[A| ] -> [ | A-1 ]
= ||u-v||
(how to find inverse of A)
d(u,v) = 0 iff u = v
Ax = b; x = A-1b
Elementary Matrix and Unifying Theorem Examples of Subspaces Linear Independent (cont)
(cont)
IF: w1, w2 are then w1+w2 are - list the vectors as the columns of a matrix,
- A -> RREF = within S within S row reduce it, if many solution, then it is
- A can be express as a product of E and kw1 is within S dependent
- A is invertible - the zero vector 0 it self is a subspace - after RREF, the columns with leading 1's
- Ax = 0 has only the trivial solution n are a maxmially linearly independent subset
- R is a subspace of all vectors
- Ax = b is consistent for every vector b in according to Pivot Theorem
- Lines and planes through the origin are
Rn
subspaces
- Ax = b has eactly 1 solution for every b in Diagonal, Triangular, Symmetric Matrices
Rn - The set of all vectors b such that Ax = b is
Diagonal all zeros along the
- colum and rowvectors of A are linealy consistent, is a subspace
Matrices diagonal
independent - If {v1, v2, ...vk} is any set of vectors in R n,
- det(A) ≠ 0 Lower zeros above diagonal
then the set W of all linear combinations of
- λ = 0 is not an eigenvalue of A Triangular
these vector is a subspace
- TA is one to one and onto Upper zeros below the
W = {c1v1 + c2v2 + ... ckvk}; c are within
If not, then all no. Triangular diagonal
real numbers
Symmetric if: AT = A
Consistency
Span Skew-Symm‐ AT = -A
EAx = Eb -> Rx = b' , where b' = Eb etric if:
- the span of a set of vectors { v1, v2, ... vk}
(Ax=b) [ A | b ] -> [ EA | Eb ] (Rx = b') is the set of all linear combinations of these
(but treat b as unknown: b1, b2...) vectors Determinants
For it to be consistent, if R has zero rows at span { v1, v2, ... vk} = { v11t, t2v2, ... , tkvk} det(A) = a1jC1j + expansion along
the bottom, b' that row must equal to zero a2jC2j ... + anjCnj jth column
If S = { v1, v2, ... vk}, then W = span(S) is a
subspace det(A) = ai1Ci1 + expansion along
Homogeneous Systems ai2Ci2 ... + ainCin the ith row
Ax = b is consistent if and only if b is a
Linear Combination of the vectors: i+j
linear combination of col(A) Cij = (-1) Mij
v = c1v1 + c2v2 ... + cnvn
Mij = deleted ith row and jth column matrix
(use matrix to find c) Linear Independent
- pick the one with most zeros to calculate
Ax = 0 Homogeneous
- if unique solution for a set of vectors, then easier
Ax = b Non-homog‐ it is linearly independent det(AT) = det(A) det(A-1) =
enous
c1v1 + c2v2 ... + cnvn = 0; all the c = 0 1/det(A)
x = x0 + t1v1 + t2v2 ... + Homogeneous
- for dependent, not all the c = 0 det(AB) = det(A)det(B) det(kA) = kndet(A)
tkvk
Dependent if: - A is invertible iff det(A) not equal 0
x = t1v1 + t2v2 ... + tkvk Non-homog‐
- a linear combination of the other vectors - det of triangular or diagonal matrix is the
eneous
- a scalar multiple of the other product of the diagonal entries
xp is any solution of NH x = xp + xh - a set of more than n vectors in Rn
det(A) for 2x2 matrix ad - bc
system
Independent if:
and xh is a solution of H
- the span of these two vectors form a plane
system
u x v = -v x k(u x v) = (ku) x v = u x (kv) Eigenvalues and Eigenvectors compression in y-direction [x, y] -> [x, ky]
u Ax= λx shear in x-direction T(x,y) = (x+ky, y);
uxu=0 parallel vectors has 0 for n
det(λI - A) = (-1) det(A - λI) [x+ky (1, k), y( 0, 1)]
c.p. shear in y-direction T(x,y) = (x, y+kx);
pa(λ) = 3x3: det(A - λI); 2x2: det(λI - A)
u (u x v) = 0 v (u x v) = 0 [x (1, 0), y (k, 1)]
- solve for (λI - A)x = 0 for eigenvectors
u x v is perpendicular to span {u, v} orthogonal projection on the xy-plane: [x, y ,
Work Flow:
||u x v|| = ||u|| ||v|| sin(theta), where theta is 0]
- form matrix
the angle between vectors - compute pa(λ) = det(λI - A) orthogonal projection on the xz-plane: [x, 0 ,
- find roots of pa(λ) -> eigenvalues of A z]
- plug in roots then solve for the equation orthogonal projection on the yz-plane: [0, y ,
z]
reflection about the xy-plane: [x, y, -z]
reflection about the xz-plane: [x, -y, z]
reflection about the yz-plane: [-x, y, z]
Kernel, Range, Composition row(A) = rows with leading ones after RREF
ker(T) is the set of all vectors x such that col(A) = columns with leading ones from
T(x) = 0, RREF matrix, find the vector, original A
ker(T) = span{(v)} null(A) = free variable's vectors
the solution space of Ax = 0 is the null null(AT) = after transform, the free variable
space; vector
null(A) = ker(A) The Rank Theorem: rank(A) = rank(AT) for
range of T, ran(T) is the set of vectors y any matrix have the same dimension
such that rank(A) = # of free vectors in span
y = T(x) for some x
dim(row(A)) = dim(col(A)) = rank(A)
ran(T) = col([T]) = span{ [col1], [col2] ...}; Ax
dim(null(A)) = nullity(A)
=b
Important Facts: Orthogonal Compliment, DImention
1. T is one to one iff ker(T) = {0} Theorem
2. Ax = b, if consistent, has a unique
S⟂ = {v ∈ Rn | v · w = 0 for all w ∈ S}
solution
iff null(A) = {0}; Ax = 0 has only the trivial S⟂ is a subspace of Rn; S⟂ = span(S)⟂ = W⟂
solution iff null(A) = {0} row(A)⟂ = null(A) null(A)⟂ = row(A)
Important facts 2: ((S ⟂) ⟂ = S iff S is
1.T:Rn -> Rm is onto iff the system Tx = y subspace
n m T
has a solution x in R for every y in R ⟂
col(A) = null(A ) null(AT) ⟂ = col(A)
2. Ax = b is consistent for every b in R m(A is The Dimension rank(A) + nullity(A) =
onto) iff col(A) = Rm Theorem n
The composition of T2 with T1 is: T2 ◦ T1 A is m x n matrix (k + (n-k) = n)
n m
(T2 ◦ T1)(x) = T2(T1(x)); T2 ◦ T1: R -> R if W is a subspace dim(W) + dim(W⟂) =
compostion of linear transformations corres‐ of Rn n
ponds to matrix application: [T2 ◦ T1] = [T1]
[T2]