the final video for the @weights_biases Math4ML series, on probability, is now up on YouTube!
@_ScottCondron and I talk entropies, divergence, and loss functions
๐:
this is the final video in a four-part series of "exercise" videos, where Scott and I work through a collection of Jupyter notebooks with automatically-graded Python coding exercises on math concepts
each exercise notebook has a corresponding lecture video.
the focus of the lectures is on intuition, and in particular on intuition that i think programmers trying to get better at ML will grok
for linear algebra, we take a "programmer's view": arrays are functions that operate on arrays -- loops, parallelization, and higher-order functions all appear
for calculus, the focus is on approximation -- on using gradients as a "good enough" answer -- rather than on dynamics (as in physics) or limits (as in analysis)
for probability, we cover the importance of unlikely events and the utility of log-probabilities, which can be understood as a quantification of the commonsense notion of "surprise"
these are my own, hard-won intuitions for these topics, honed by trying to program learning machines for a decade with more chutzpah than formal mathematical training
i hope they are as useful for others as they have been for me!
โข โข โข
Missing some Tweet in this thread? You can try to
force a refresh
New video series out this week (and into next!) on the @weights_biases YouTube channel.
They're Socratic livecoding sessions where @_ScottCondron and I work through the exercise notebooks for the Math4ML class.
Details in ๐งตโคต๏ธ
Socratic: following an ancient academic tradition, I try to trick @_ScottCondron into being wrong, so that students can learn from mistakes and see their learning process reflected in the content.
(i was inspired to try this style out by the @PyTorchLightnin Master Class series, in which @_willfalcon and @alfcnz talk nitty-gritty of DL with PyTorch+Lightning while writing code. strong recommend!)
tl;dr: the basic idea of the SVD works for _any_ function.
it's a three step decomposition:
- throw away the useless bits โคต
- rename what remains ๐
- insert yourself into the right context โคด
also, if you're more of a "YouTube talk" than a "tweet wall" kinda person, check out the video version, given as part of the @weights_biases Deep Learning Salon webinar series