376 Preparation 4

Warning: This content has not yet been fully revised for this year.

Find answers to the following questions; the articles below should be helpful.

  1. How can a causal language model be used to power a dialog agent? (What does a “document” look like?)
  2. What is “few-shot learning”, aka “in-context learning”, and how is it helpful for getting a LM to do what you want?
  3. How does “chain of thought” prompting help a model reason better? (What does that have to do with autoregressive generation?)
  4. How does “tool use” work in LMs?
  5. How could you get a model to give an output in a specific structure that you could use in a program?
  6. In general, how can you use a “chat” API to do useful things?
  7. Are LM outputs always accurate? How can you tell?

Some resources:

Supplemental Material

We probably won’t get to this until next week, but:

The Illustrated Stable Diffusion – Jay Alammar – Visualizing machine learning one concept at a time.

I found the Foreward to this book on Deep Generative Modeling (Available through Calvin library) to be reasonably accessible, but you may prefer the author’s blog posts. (github).

Now, how do you control what gets generated?

Lab 376.4: Dialogue Agents, Prompt Engineering, Retrieval-Augmented Generation, and Tool Use