Unit 11: Generation

We’ve seen models that classify images and text, then more recently models that can generate one single token. What if we want to generate whole articles? Or images? Music? Programs? We can adapt the same basic approaches that we used already, but with interesting twists… and, I must admit, the results are fun.

By the end of this week you should be able to:

Preparation

We’ll also discuss GANs and Diffusion Models. I found the Foreward to this book on Deep Generative Modeling (Available through Calvin library) to be reasonably accessible, but you may prefer the author’s blog posts. (github).

Now, how do you control what gets generated? Choose your favorite modality and skim a very recent paper:

Supplemental Material

Class Meetings

Monday

We worked through the decoding activity (PDF) using Translation as Language Modeling notebook (preview, Colab)

Wednesday

Friday: Guest Lecture

Contents

Due this Week