100% found this document useful (1 vote)
512 views4 pages

Cheat Sheet Fionaw - Linear Algebra Math 232

This document provides a cheat sheet on key concepts in linear algebra, including: 1. Definitions and properties of vectors, norms, dot products, and distance between vectors. 2. Descriptions of matrix algebra concepts such as addition, multiplication, transposes, and inverses of matrices. 3. Explanations of elementary matrix operations and the unifying theorem for finding the inverse of a matrix. 4. Discussions of subspaces, linear independence, diagonal matrices, and triangular matrices.

Uploaded by

SATHE S
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
512 views4 pages

Cheat Sheet Fionaw - Linear Algebra Math 232

This document provides a cheat sheet on key concepts in linear algebra, including: 1. Definitions and properties of vectors, norms, dot products, and distance between vectors. 2. Descriptions of matrix algebra concepts such as addition, multiplication, transposes, and inverses of matrices. 3. Explanations of elementary matrix operations and the unifying theorem for finding the inverse of a matrix. 4. Discussions of subspaces, linear independence, diagonal matrices, and triangular matrices.

Uploaded by

SATHE S
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Linear Algebra - MATH 232 Cheat Sheet

by fionaw via cheatography.com/124375/cs/23750/

Basic Equations Vectors, Norm, Dot Product (cont) Matrix Algebra, Identity and Inverse Matrix

Network Flows u·v = u1v1 + u2v2 dot product (A + B)ij = (A)ij + (B)ij (A - B)ij = (A)ij -
1. the flow in an arc is only in one directions ...+unvn (B)ij
||u|| ||v|| cos(θ) (cA)ij = c(A)ij (AT)ij = (A)ji
2. flow into a node = flow out of a node
u and v are orthogonal if u·v = 0 (cos(θ) = 0) (AB)ij = ai1b1j + ai2b2j + ... aikbkj
3. flow into the network = flow out of the
network a set of vectors is an orthogonal set iff vi·vj Inner Product (number) is u Tv = u·v , u and v
= 0,if i≠j same size
Balancing Chemical Equations
a set of vectors is an orthon​ormal set iff vi·vj Outer Product (matrix) is uvT, u and v can
1. add x's before each combo and both side
= 0,if i≠j, and ||vi|| = 1 for all i be any size
2. carbo = x1 + 2(x3), set as system, solve
(u·v)2 ≤ ||u||2||v||2 or Cauchy​-Sc​hwarz (AT) T = A (kA)T = k(A)T
Matrix
|u·v| ≤ ||u|| ||v|| Inequality
(A+B)T = AT + BT (AB)T = BTAT
augmented variables and soluti​‐
d(uv) ≤ d(u,w) + Triangle Inequality
matrix on(rhs) tr(A T) = tr(A) tr(AB) = tr(BA)
d(w,v)
T T
coeffi​cient coeffi​cients only, no rhs u v = tr(uv ) tr(uvT) = tr(vuT)
||u+v|| ≤ ||u|| + ||v||
matrix tr(A) = a11 + a22 ... + (AT)ij = Aji
||v1 + v2 ... + vk|| = ||v1|| + ||v2|| ... + ||vk||
ann
Vectors, Norm, Dot Product Identity matrix is square matrix with 1 along
Lines and Planes
maginitude (norm) of vector v is ||v||; ||v|| ≥ 0 diagonals
a vector equation with x = x0 + tv,
if k>0, kv same direction magnitude = parameter t -∞ < t < +∞ If A is m x n, A n = A and mA = A
as v k||v|| a square matrix is AB = = BA
solutin set for 3 dimension linear equation is
if k<0, kv opposite magnitude = |k| a plane invert​ibl​e(n​ons​ing​ular)
direction to v ||v|| if:
if x is a point on this plane n·(x-x0) = 0
vectors in Rn (n = v = (v1, v2, ..., (point​-normal equation) B is the inverse of A B = A-1
dimension) vn) if A has no inverse, A is not invertible
A(x-x0​)+B​(y-​y0)​+C(​z-z0) = x0 =
v = P1P2 = OP2 - OP1 displa​cement 0 (x0,y0​,z0), (singular)
vector n = (A, B, C) det(A) = ad - bc ≠ 0 is invertible
2
norm/m​agn​itude of vector sqrt( (v1) + genera​l/a​lge​braic equation Ax+By+Cz = if A is invert​ible: (AB)-1 = B-1A-1
2
||v|| (v2) ...) D (An) -1 = A-n = (A-1) n (AT) -1 = (A-1) T
||v|| = 0 iff v =0 ||kv|| = |k| ||v|| two planes are parallel if n1 = kn2, (kA)-1 1/k(A -1), k≠0
unit vector u in same u = (1/ ||v||) v orthogonal if n1·n2 = 0
direct as v Elementary Matrix and Unifying Theorem
e1 = (1,0...) ... en = standard unit elementary matrices are invertible
(0,...1) in Rn vector
A-1 = Ek Ek-1 ... E2 E1
d(u,v) = sqrt((​u1-v1)2 + (u2-v2)2 ... (un-vn)2)
[A| ] -> [ | A-1 ]
= ||u-v||
(how to find inverse of A)
d(u,v) = 0 iff u = v
Ax = b; x = A-1b

By fionaw Published 16th July, 2020. Sponsored by ApolloPad.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Everyone has a novel in them. Finish
Page 1 of 4. Yours!
https://apollopad.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Elementary Matrix and Unifying Theorem Examples of Subspaces Linear Indepe​ndent (cont)
(cont)
IF: w1, w2 are then w1+w2 are - list the vectors as the columns of a matrix,
- A -> RREF = within S within S row reduce it, if many solution, then it is
- A can be express as a product of E and kw1 is within S dependent
- A is invertible - the zero vector 0 it self is a subspace - after RREF, the columns with leading 1's
- Ax = 0 has only the trivial solution n are a maxmially linearly indepe​ndent subset
- R is a subspace of all vectors
- Ax = b is consistent for every vector b in according to Pivot Theorem
- Lines and planes through the origin are
Rn
subspaces
- Ax = b has eactly 1 solution for every b in Diagonal, Triang​ular, Symmetric Matrices
Rn - The set of all vectors b such that Ax = b is
Diagonal all zeros along the
- colum and rowvectors of A are linealy consis​tent, is a subspace
Matrices diagonal
indepe​ndent - If {v1, v2, ...vk} is any set of vectors in R n,
- det(A) ≠ 0 Lower zeros above diagonal
then the set W of all linear combin​ations of
- λ = 0 is not an eigenvalue of A Triangular
these vector is a subspace
- TA is one to one and onto Upper zeros below the
W = {c1v1 + c2v2 + ... ckvk}; c are within
If not, then all no. Triangular diagonal
real numbers
Symmetric if: AT = A
Consis​tency
Span Skew-S​ymm​‐ AT = -A
EAx = Eb -> Rx = b' , where b' = Eb etric if:
- the span of a set of vectors { v1, v2, ... vk}
(Ax=b) [ A | b ] -> [ EA | Eb ] (Rx = b') is the set of all linear combin​ations of these
(but treat b as unknown: b1, b2...) vectors Determ​inants

For it to be consis​tent, if R has zero rows at span { v1, v2, ... vk} = { v11t, t2v2, ... , tkvk} det(A) = a1jC1j + expansion along
the bottom, b' that row must equal to zero a2jC2j ... + anjCnj jth column
If S = { v1, v2, ... vk}, then W = span(S) is a
subspace det(A) = ai1Ci1 + expansion along
Homoge​neous Systems ai2Ci2 ... + ainCin the ith row
Ax = b is consistent if and only if b is a
Linear Combin​ation of the vectors: i+j
linear combin​ation of col(A) Cij = (-1) Mij
v = c1v1 + c2v2 ... + cnvn
Mij = deleted ith row and jth column matrix
(use matrix to find c) Linear Indepe​ndent
- pick the one with most zeros to calculate
Ax = 0 Homoge​neous
- if unique solution for a set of vectors, then easier
Ax = b Non-ho​mog​‐ it is linearly indepe​ndent det(AT) = det(A) det(A-1) =
enous
c1v1 + c2v2 ... + cnvn = 0; all the c = 0 1/det(A)
x = x0 + t1v1 + t2v2 ... + Homoge​neous
- for dependent, not all the c = 0 det(AB) = det(A)​det(B) det(kA) = kndet(A)
tkvk
Dependent if: - A is invertible iff det(A) not equal 0
x = t1v1 + t2v2 ... + tkvk Non-ho​mog​‐
- a linear combin​ation of the other vectors - det of triangular or diagonal matrix is the
eneous
- a scalar multiple of the other product of the diagonal entries
xp is any solution of NH x = xp + xh - a set of more than n vectors in Rn
det(A) for 2x2 matrix ad - bc
system
Indepe​ndent if:
and xh is a solution of H
- the span of these two vectors form a plane
system

By fionaw Published 16th July, 2020. Sponsored by ApolloPad.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Everyone has a novel in them. Finish
Page 2 of 4. Yours!
https://apollopad.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Adjoint and Cramer's Rule Complex Number Linear Transf​orm​ation

adj(A) = CT CT = matrix confactor complex number a + ib f: Rn -> Rm, n = domain, m = co-domain


of A (a + ib) + (c + id) = (a + c) + i(b + d) f(x1, x2, ...xn) = (y1, ...ym)
A-1 = (1/det(A)) adj(A)A = det(A) I (a + ib) - (c + id) = (a - c) + i(b - d) T: Rn -> Rm is a linear transf​ormatin if
adj(A) 1. T(cu) = cT(u)
(a + ib) (c + id) = (ac + bd) + i(ad + bc)
x1 = det(A1) / x2 = det(A2) / det(A) 2. T(u +v) = T(u)+ T(v)
(a + bx) (c + dx) = (ac + bdx2) + x(ad + bc)
det(A) for any linear transf​orm​ation, T(0) = 0
i2 = -1
xn = det(An) / det(A) not equal 0 Rθ = [T(e1) T(e2)] = [cosθ matrix for
z = a + ib z bar = a - ib
det(A) −sinθ] rotation
the length​(ma​gni​tude) of |z| = sqrt(z x z ​[sinθ
An is the matrix when the nth column is
vector z bar) cosθ]
replaced by b
= sqrt(a2 + b2)
reflection across y-axis: T(x, y) = (-x, y)
Hyperp​lane, Area/V​olume z -1 = 1/z = z bar / |z|2
reflection across x-axis: T(x, y) = (y, -x)
a hyperplane in a1x1 + a2x2 ... + anxn = z1 / z2 = z1z2-1
reflection across diagonal y = x, T(x, y) = (y,
Rn b z = |z| (cos(θ) + i (sin(θ)) polar form (r = x)
- can also written as ax = b |z|)
orthogonal projection onto the x-axis: T(x, y)
to find aperp
ax = 0, find the span z1z2 = |z1| |z2| (cos(θ1 + θ2) + i (sin(θ1 + = (x, 0)
θ2))
if A is 2x2 matrix: orthogonal projection onto the y-axis: T(x, y)
- |det(A)| is the area of parall​elogram z1/z2 = |z1| / |z2| (cos(θ1 - θ2) + i (sin(θ1 - = (0, y)
θ2))
if A is 3x3 matrix: u = (1/ ||v||)v; express it vertically as u1 and
- |det(A)| is the volume of parall​ele​piped z n = rn(cos(n θ) + i sin(n r = |z| u2
θ))
- subtract points to get three vectors, then A = [(u1)2 u2u1] projection
i theta
make it to a matrix to find the area/v​olume e = cos(θ) + i sin(θ) ​[u1u2 (u2)2] matrix
i pi i pi
e = -1 e +1 = 0 contra​ction with 0 ≤ k < 1 (shrink), k > 1
Cross Product z1z2 = r1r2 e i (θ1 + θ2) n
z =r e n i nθ (stretch)
u x v = (u2v3 - u3v2, u3v1 - u1v3, u1v2 - i (θ1 - θ2) [x, y] -> [kx, ky]
z1 /z2 = r1 / r2 e
u2v1) compre​ssion in x-dire​ction [x, y] -> [kx, y]

u x v = -v x k(u x v) = (ku) x v = u x (kv) Eigenv​alues and Eigenv​ectors compre​ssion in y-dire​ction [x, y] -> [x, ky]
u Ax= λx shear in x-dire​ction T(x,y) = (x+ky, y);
uxu=0 parallel vectors has 0 for n
det(λI - A) = (-1) det(A - λI) [x+ky (1, k), y( 0, 1)]
c.p. shear in y-dire​ction T(x,y) = (x, y+kx);
pa(λ) = 3x3: det(A - λI); 2x2: det(λI - A)
u (u x v) = 0 v (u x v) = 0 [x (1, 0), y (k, 1)]
- solve for (λI - A)x = 0 for eigenv​ectors
u x v is perpen​dicular to span {u, v} orthogonal projection on the xy-plane: [x, y ,
Work Flow:
||u x v|| = ||u|| ||v|| sin(th​eta), where theta is 0]
- form matrix
the angle between vectors - compute pa(λ) = det(λI - A) orthogonal projection on the xz-plane: [x, 0 ,
- find roots of pa(λ) -> eigenv​alues of A z]
- plug in roots then solve for the equation orthogonal projection on the yz-plane: [0, y ,
z]
reflection about the xy-plane: [x, y, -z]
reflection about the xz-plane: [x, -y, z]
reflection about the yz-plane: [-x, y, z]

By fionaw Published 16th July, 2020. Sponsored by ApolloPad.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Everyone has a novel in them. Finish
Page 3 of 4. Yours!
https://apollopad.com
Linear Algebra - MATH 232 Cheat Sheet
by fionaw via cheatography.com/124375/cs/23750/

Orthogonal Transf​orm​ation Kernel, Range, Compos​ition (cont)

an orthogonal transf​orm​ation is a linear [T(θ1+θ2)] = [Tθ2] ◦ [Tθ1];


transf​orm​ation T; Rn -> Rn that preserves rotate then shear ≠ shear then rotate
lengths; ||T(u)|| = ||u|| linear trans T: Rn->Rm has an inverse iff T
||T(u)|| = ||u|| <=> T(x)·T(y) = x·y for all x,y in is one to one, T -1: Rm -> Rn, Tx = y <=> x =
n
R T -1y
orthogonal matrix is square matrix A such for Rn to Rn, [T-1] = [T]-1; [T]-1◦T = 1n <=>
that AT = A-1 [T-1][T]= n
T
1. if A is orthog​onal, then so is A and A -1 1n is identity transf​orm​ation; n is identity
matrix
2. a product of orthonal matrices is
orthogonal
Basis, Dimension, Rank
3. if A is orthog​onal, then det(A) = 1 or -1
S is a basis for the subspace V of Rn if:
4. if A is orthog​onal, then rows and columns
S is linearly idenpe​ndent and span(S) = V
of A are each orthon​ormal sets of vectors
dim(V) = k, k is the # of vectors

Kernel, Range, Compos​ition row(A) = rows with leading ones after RREF

ker(T) is the set of all vectors x such that col(A) = columns with leading ones from
T(x) = 0, RREF matrix, find the vector, original A
ker(T) = span{(v)} null(A) = free variable's vectors
the solution space of Ax = 0 is the null null(AT) = after transform, the free variable
space; vector
null(A) = ker(A) The Rank Theorem: rank(A) = rank(AT) for
range of T, ran(T) is the set of vectors y any matrix have the same dimension
such that rank(A) = # of free vectors in span
y = T(x) for some x
dim(ro​w(A)) = dim(co​l(A)) = rank(A)
ran(T) = col([T]) = span{ [col1], [col2] ...}; Ax
dim(nu​ll(A)) = nullity(A)
=b
Important Facts: Orthogonal Compli​ment, DImention
1. T is one to one iff ker(T) = {0} Theorem
2. Ax = b, if consis​tent, has a unique
S⟂ = {v ∈ Rn | v · w = 0 for all w ∈ S}
solution
iff null(A) = {0}; Ax = 0 has only the trivial S⟂ is a subspace of Rn; S⟂ = span(S)⟂ = W⟂
solution iff null(A) = {0} row(A)⟂ = null(A) null(A)⟂ = row(A)
Important facts 2: ((S ⟂) ⟂ = S iff S is
1.T:Rn -> Rm is onto iff the system Tx = y subspace
n m T
has a solution x in R for every y in R ⟂
col(A) = null(A ) null(AT) ⟂ = col(A)
2. Ax = b is consistent for every b in R m(A is The Dimension rank(A) + nullity(A) =
onto) iff col(A) = Rm Theorem n
The compos​ition of T2 with T1 is: T2 ◦ T1 A is m x n matrix (k + (n-k) = n)
n m
(T2 ◦ T1)(x) = T2(T1(x)); T2 ◦ T1: R -> R if W is a subspace dim(W) + dim(W⟂) =
compostion of linear transf​orm​ations corres​‐ of Rn n
ponds to matrix applic​ation: [T2 ◦ T1] = [T1]
[T2]

By fionaw Published 16th July, 2020. Sponsored by ApolloPad.com


cheatography.com/fionaw/ Last updated 10th August, 2020. Everyone has a novel in them. Finish
Page 4 of 4. Yours!
https://apollopad.com

You might also like