In this unit, after reviewing where we’ve been, we push towards state-of-the-art models (still focusing on computer vision). We’ll first show how our work last 2 weeks connects to the pre-trained models we used in the opening weeks. Then, we’ll introduce or revisit tools that allow our models to achieve high performance, such as data augmentation and regularization. Finally, we’ll get more practice with how neural networks work from the ground up as we implement our own simple neural net image classifier from scratch (definitely something to mention in an interview!). Students who complete this unit will demonstrate that they can:
The fastai course videos are still a bit disorganized, sorry about that.
A nice intuition about why layers matter: Why depth matters in a neural network (Deep Learning / AI) - YouTube
Strategies for getting state-of-the-art performance:
We’ll be doing some automatic differentiation this week:
autograd-for-dummies: A minimal autograd engine and neural network library for machine learning students.Finally, I sometimes remark that “machine learning is lazy” (in that it tends to focus on superficial easy features). Here’s a more precise statement of a related claim: What do deep networks learn and when do they learn it. A recent paper describes what to do about it: Learning an Invertible Output Mapping Can Mitigate Simplicity Bias in Neural Networks | Abstract
If anyone uncovers a pit or digs one and fails to cover it and an ox or a donkey falls into it, the one who opened the pit must pay the owner for the loss and take the dead animal in exchange.
If anyone’s bull injures someone else’s bull and it dies, the two parties are to sell the live one and divide both the money and the dead animal equally. However, if it was known that the bull had the habit of goring, yet the owner did not keep it penned up, the owner must pay, animal for animal, and take the dead animal in exchange.
Exodus 21:33-36 (NIV)
See Slides for the precision-recall chart I was trying to draw in class. (I almost had it right: it’s True Positive vs False Positive.)
Linear layers