flowchart LR
A[Input] --> B[Feature Extractor]
B --> C[Linear Classifier]
C --> D[Output]
The internal data structure of neural networks.
flowchart LR
A[Input] --> B[Feature Extractor]
B --> C[Linear Classifier]
C --> D[Output]
Example:
flowchart LR
A[Input] --> B["Pre-trained CNN"]
B --> C["Linear layer with 3 outputs"]
C --> D["Softmax"]
D --> E["Predicted probabilities"]
style B stroke-width:4px
The feature extractor constructs a representation of the input that’s useful for classification.
Source: Pinecone blog
Source: Understanding RL Vision
a, b, c, and d as outputs of some linear layers, then computes max(a, b, c, d). Can you backpropagate through the max? Why or why not?ML researchers have been very creative!
https://poloclub.github.io/cnn-explainer/
Activity:
Some figures from Understanding Deep Learning, by Simon J.D. Prince, used with permission
In your notes, write a one-sentence response to each of these questions by using specific concepts that we’ve learned in class. Share your responses with table-mates.

Why: Often more efficient to process several items at once; gradients less noisy. (But some noise in gradients is useful to avoid sharp minima, which don’t generalize well.)
How:
Two big improvements over fully-connected layers:
Example: Conv2D(filters=16, kernel_size=(3, 3), input_shape=(32, 32, 3))
Simonyan, 2016, Very Deep Convolutional Networks for Large-Scale Image Recognition
Butlin et al. 2023 Consciousness in Artificial Intelligence argues that (1) no current AI system is conscious but we could build one that at least satisfies these indicators.
What issues might come up?
What issues might come up?
What issues might come up?
What issues might come up?
Source: Shneiderman, Human-Centered AI, pp 246-247
Did this classifier successfully learn to recognize a “dumbbell”?
Source: Google Blog