Prompting & Current LMs#
Stay tuned for slides from the fourth lecture!
Additional materials#
If you want to dig a bit deeper, here are (optional!) supplementary readings introducing the prompting strategies and the mentioned LMs that were covered in class:
- Wei et al. (2023) Chain-of-Thought Prompting Elicits Reasoning in Large Language Models 
- Kojima et al. (2023) Large Language Models are Zero-Shot Reasoners 
- Webson & Pavlick (2022) Do Prompt-Based Models Really Understand the Meaning of Their Prompts? 
- Nye et al. (2022) Show Your Work: Scratchpads for Intermediate Computation with Language Models 
- Wang et al. (2023) Self-Consistency Improves Chain of Thought Reasoning in Language Models 
- Liu et al. (2022) Generated Knowledge Prompting for Commonsense Reasoning 
- Yao et al. (2023) Tree of Thoughts: Deliberate Problem Solving with Large Language Models 
- Xie et al. (2022) An Explanation of In-context Learning as Implicit Bayesian Inference 
- Min et al. (2022) Rethinking the Role of Demonstrations: What Makes In-Context Learning Work? 
- Lampinen et al. (2022) Can language models learn from explanations in context? 
- Touvron et al. (2023) LLaMA: Open and Efficient Foundation Language Models 
- Touvron et al. (2023) Llama 2: Open Foundation and Fine-Tuned Chat Models 
