Authors: Vaswani et al., Google BrainKey Contribution: Introduced the Transformer architecture that powers modern LLMsWhy Read: Understanding transformers is fundamental to understanding how prompts are processedLink:arXiv:1706.03762
Language Models are Few-Shot Learners (2020)
Authors: Brown et al., OpenAI (GPT-3 Paper)Key Contribution: Demonstrated that large language models can perform tasks with just a few examples (few-shot learning)Why Read: Foundational paper on in-context learning and prompt-based task solvingLink:arXiv:2005.14165
Training language models to follow instructions (2022)
Authors: Ouyang et al., OpenAI (InstructGPT Paper)Key Contribution: Showed how RLHF (Reinforcement Learning from Human Feedback) improves instruction followingWhy Read: Explains why modern models are better at following promptsLink:arXiv:2203.02155
Authors: Wei et al., Google ResearchKey Contribution: Introduced Chain of Thought prompting, showing 30-50% accuracy improvements on reasoning tasksWhy Read: The definitive paper on CoT promptingLink:arXiv:2201.11903
Large Language Models are Zero-Shot Reasoners (2022)
Authors: Kojima et al., University of TokyoKey Contribution: Showed that simply adding “Let’s think step by step” dramatically improves reasoningWhy Read: Demonstrates the power of simple prompting modificationsLink:arXiv:2205.11916
Self-Consistency Improves Chain of Thought Reasoning (2022)
Authors: Wang et al., Google ResearchKey Contribution: Introduced self-consistency (ensembling multiple reasoning paths)Why Read: Shows how to improve CoT reliability through multiple samplesLink:arXiv:2203.11171
Tree of Thoughts: Deliberate Problem Solving (2023)
Authors: Yao et al., Princeton UniversityKey Contribution: Extended CoT to explore multiple reasoning branches like a search treeWhy Read: Advanced technique for complex problem-solvingLink:arXiv:2305.10601
Retrieval-Augmented Generation for Knowledge-Intensive NLP (2020)
Authors: Lewis et al., Facebook AI ResearchKey Contribution: Introduced RAG, combining retrieval with generationWhy Read: Foundational paper on grounding LLM outputs in external knowledgeLink:arXiv:2005.11401
In-Context Retrieval-Augmented Language Models (2023)
Authors: Ram et al., AI21 LabsKey Contribution: Showed how to effectively integrate retrieved documents into promptsWhy Read: Practical techniques for implementing RAGLink:arXiv:2302.00083
Authors: Liu et al., Carnegie Mellon UniversityKey Contribution: Comprehensive survey of prompting methodsWhy Read: Excellent overview of the prompting landscapeLink:arXiv:2107.13586
A Survey of Large Language Models (2023)
Authors: Zhao et al., Renmin University of ChinaKey Contribution: Comprehensive survey covering LLM architectures, training, and promptingWhy Read: Up-to-date overview of the entire LLM fieldLink:arXiv:2303.18223
Focus: Training AI systems to be helpful, harmless, and honestKey Paper: “Constitutional AI: Harmlessness from AI Feedback” (Anthropic, 2022)Why Important: Addresses AI safety and alignmentLink:arXiv:2212.08073
Focus: Prompting with images, audio, and textKey Papers:
“Flamingo: a Visual Language Model” (DeepMind, 2022)
Note: This field evolves rapidly. Links and resources are current as of 2025, but new papers and tools emerge frequently. Check the communities and newsletters above to stay updated.