In this unit we extend our modeling skills to encompass classification models, and start to build the tools that will let us represent complex functions by using hidden layers. Both of these objectives require us to learn about nonlinear operations. We’ll focus on the two most commonly used ones: the softmax operator (which converts scores to probabilities) and the rectifier (“ReLU”, which clips negative values).
Neural Models
Students who complete this unit will demonstrate that they can:
Describe and compute cross-entropy loss
Explain the purpose and mathematical properties of the softmax operation.
Explain the role of nonlinearities in a neural network (e.g., why they are used between linear layers)
Implement a logistic regression model using basic numerical computing primitives
Implement numerical computing operations on multidimensional arrays using PyTorch
Explain the purpose of automatic differentiation in machine learning
Automatic Differentiation
We’ll be doing some automatic differentiation this week: