376 Unit 5: Multimodal Models and Diffusion

Contents

Discussion 376.4: Fans and Skeptics, Optimists and Pessimists

You’ve spent several weeks learning how generative AI systems actually work — tokenization, attention, training pipelines, tool use, failure modes. You’re now much more qualified than average to answer the questions people will ask you: Where is AI going? Is it good or bad?

But even experts disagree. Some are impressed by what AI can do (“fans”). Others doubt the claims (“skeptics”). Some are optimistic about societal benefits (“optimists”). Others worry about serious harm (“concerned”). Many thoughtful people hold several of these views at once. To be wise, we need to engage honestly with perspectives we don’t naturally hold.

This Discussion addresses the course objectives Overall-PhilNarrative and Overall-Impact.

We’ll share our findings in class on the last day and compare with the results of a national survey.

Instructions

Step 1: Take the survey. The Moodle forum includes a link to a brief survey about your current views on AI. Fill it out first.

Step 2: Find two articles that represent genuinely different perspectives on the future of AI. Your two articles should pull in different directions — not two versions of the same take. Read with hospitality: you’ll need to articulate the other side’s view convincingly.

For each article:

Step 3: Articulate your own position (~150-250 words), drawing substantively on both articles. Where do you land, and why?

Ground your position in something beyond personal preference. You might draw on:

The best posts will show that you’ve genuinely wrestled with a view you don’t naturally hold.

Finding Sources

The landscape changes fast, so find your own sources rather than relying on a list. Some places to look:

Where to search:

Kinds of voices to look for:

  • Researchers building these systems and explaining what excites them
  • Researchers studying these systems and explaining what worries them
  • Economists analyzing labor market effects
  • Legal scholars on copyright, liability, regulation
  • Philosophers on intelligence, consciousness, agency
  • Journalists investigating real-world impacts on specific communities
  • Skeptics who think current AI is overhyped (Gary Marcus, AI Snake Oil, Emily Bender, Melanie Mitchell)
  • Optimists who think AI will transform society for the better (Dario Amodei, Marc Andreessen)
  • People trying to hold both views at once (Arvind Narayanan, Nicholas Carlini)

Replies

Read several classmates’ posts. Reply to at least one (~75-150 words):

Rubric

Q&A Week 12

This covers both the RL unit and Human-Centered AI part 1.

RL

Different Approaches to RL

Main difference is what functions we learn:

What’s the “loss” (or target) in RL?

That’s what makes it hard! e.g,. in Q-learning we try to minimize the temporal difference: how much the reward we get differs from the reward we predict (by subtracting the next-state value function from the current-state value function). But that’s a difference of two predictions; if we were wrong, which of those two predictions was wrong?

In general, we’re hoping to learn something about all possible things that could happen and things we could do, given data about only a fraction of what happened and things we did.

They’re good at different things:

So, unsurprisingly, the state of the art often combines both! See, e.g., MuZero.

Can an agent trained in simulation be trusted in the real world?

Hm. Pro:

Con:

Do human newborns learn by RL?

Maybe somewhat, but not really:

Interpretable AI

Why can we ever trust a model if we can’t see how it’s making its decisions?

We trust human doctors routinely, even though, despite decades of effort by cognitive scientists, we have very limited knowledge about the process by which people make their decisions.

Is there always a trade-off between understandability and accuracy?

No.

So why are we only learning about this now? Good question…

What’s CART?

The classic algorithm for learning decision trees.

Other

What’s dropout?

How do you get bitwise determinism?

Lab 376.6: Stable Diffusion (draft!)
The content may not be revised for this year. If you really want to see it, click the link above.