Unit 11: Generation

We’ve seen models that classify images and text, then more recently models that can generate one single token. What if we want to generate whole articles? Or images? Music? Programs? We can adapt the same basic approaches that we used already, but with interesting twists… and, I must admit, the results are fun.

By the end of this week you should be able to:

Preparation

We’ll also discuss GANs and Diffusion Models. I found the Foreward to this book on Deep Generative Modeling (Available through Calvin library) to be reasonably accessible, but you may prefer the author’s blog posts. (github).

Now, how do you control what gets generated? Choose your favorite modality and skim a very recent paper:

Supplemental Material

Class Meetings

Monday

We reviewed and extended Lab 9. Summary in the following:

Wednesday

We worked through:

Programming with Self-Attention (name: u11n1-self-attention.ipynb; show preview, open in Colab)

Friday: More transformers.

Contents