6.036 Cheat Sheet
6.036 Cheat Sheet
654 Go X11611 s O XX ly y
Jridgel6Oo s
t y
É bridge XX'tnXI Xy
Jridge 6,60 x x6 y
Z Egg g
2 6
g
TrainingError En h IE had y
Test Error Ecu
Ej hoco yo
GRADIENT DESCENT
to
to be convex toconverge globalmin
f needs convex localminimum
I if f nonupdate ruleforleastsquares
I
µ Is stepsizeisfun
Point fi is pickedrandomly
saveruntimeandmemory
likelihood Loss
Negativelog
Etangoyay
iterations ToJw II gli gli x xp Tooptimize
stopwhen Iget get e I 6
stopwhen
311
Tooptimize
s lg y o
Direct soli vs AD
Advantages no more matrix inversion seelab
Trae off have to look at u datapoints every step
is low we should stilluse direct sola
maybewhen d
can be linear butinput size can be huge
if res as 1
yet
function
go guys
if oct a o s l
hypothesis is sigmoid
x
T.co a D
tÉk
fEsIYt y
FEATURE.gg zq scale so that s are closetogether
l oooo 01000
One notencoding when there's no order
1 230
d
2 pm
gg k i o Eso
o o
ft
i z o
z Rew function to 230
ing
input
final
prediction
zj AM
s activation
forprod Ecgivation
Ictivation
Yj output layer
y