Unit 6: Recap and Regularization

In this unit, after reviewing where we’ve been, we push towards state-of-the-art models (still focusing on computer vision). We’ll first show how our work last 2 weeks connects to the pre-trained models we used in the opening weeks. Then, we’ll introduce or revisit tools that allow our models to achieve high performance, such as data augmentation and regularization. Finally, we’ll get more practice with how neural networks work from the ground up as we implement our own simple neural net image classifier from scratch (definitely something to mention in an interview!). Students who complete this unit will demonstrate that they can:

Preparation

The fastai course videos are still a bit disorganized, sorry about that.

Supplemental Materials

A nice intuition about why layers matter: Why depth matters in a neural network (Deep Learning / AI) - YouTube

Strategies for getting state-of-the-art performance:

We’ll be doing some automatic differentiation this week:

Finally, I sometimes remark that “machine learning is lazy” (in that it tends to focus on superficial easy features). Here’s a more precise statement of a related claim: What do deep networks learn and when do they learn it. A recent paper describes what to do about it: Learning an Invertible Output Mapping Can Mitigate Simplicity Bias in Neural Networks | Abstract

Class Meetings

Monday class

If anyone uncovers a pit or digs one and fails to cover it and an ox or a donkey falls into it, the one who opened the pit must pay the owner for the loss and take the dead animal in exchange.

If anyone’s bull injures someone else’s bull and it dies, the two parties are to sell the live one and divide both the money and the dead animal equally. However, if it was known that the bull had the habit of goring, yet the owner did not keep it penned up, the owner must pay, animal for animal, and take the dead animal in exchange.

Exodus 21:33-36 (NIV)

Wednesday in lab

See Slides for the precision-recall chart I was trying to draw in class. (I almost had it right: it’s True Positive vs False Positive.)

Friday in lab

Contents