Discuss with neighbors:
Training set:
Prompt:
The capital of the United Kingdom is:
The capital of the United States is:
The capital of Canada is:
The capital of Australia is:
The capital of France is:
The capital of the United Kingdom is: London
The capital of the United States is: Washington, D.C.
The capital of Canada is: Ottawa
The capital of Australia is: Canberra
The capital of France is: Paris
Title: My visit to Paris
The capital of France is a city of art and culture. It is a city of romance and love. …
“The capital of France is”
What tokens would come next if this document were a:
At first, the model doesn’t know what kind of document it’s trying to predict. With additional context, it can make a better guess.
“The capital of France is”
“The capital of France is:”
The capital of France is a city of contrasts. It is a city of history, culture, and art, but it is also a city of modern life. The city is home to some of the world’s most famous landmarks, including the Eiffel Tower, the Louvre, and the Arc de Triomphe. The city is also home to some of the world’s most famous museums, including the Musee d’Orsay and the Musee du Louvre. The city is also home to some of the world’s most famous monuments, including the Arc de Triomphe and the Eiffel Tower. The city is also home to some of the world’s most famous monuments, including the Arc de Triomphe and the Eiffel Tower.
The capital of France is a city of contrasts. It is a city of history, culture, and art, but it is also a city of modern life. The city is home to some of the world’s most famous landmarks, including the Eiffel Tower, the Louvre, and the Arc de Triomphe. The city is also home to some of the world’s most famous museums, including the Musee d’Orsay and the Musee du Louvre. The city is also home to some of the world’s most famous monuments, including the
prompt engineering (n): the process of designing prompts to elicit desired responses from a model
Example of prompt engineering: I typed (n) after “prompt engineering” to get Copilot to suggest a definition.
Works by narrowing down the range of possible contexts we might be in.
Explicitly give examples of the kind of thing you want.
Frame the prompt as a conversation with a particular person (e.g., an expert, a child, …
Minimizing next-word surprisal is a powerful objective: models learn about:
After mimicry training, do two kinds of fine-tuning:
Prompt: “User: What is the capital of France?:”
Model’s completions (from temperature sampling)
A. Paris\nB. London\nC. Berlin\nD. RomeWhat is the capital of England?\nWhat is the capital of Belgium?\nWhat is the capital of Italy?"The capital of France is Paris.""Paris"Human chooses 3
Goal: model should be more likely to generate completions like 3 in the future. A simple approach:
OpenAI’s approach is more sophisticated, but the basic idea is the same.