This is a reading list intended to give a flavor of the types of problems in deep learning and AI that practicing (and erstwhile) physicists have worked on.
Papers bracketed by ** are required reading, and appear in the schedule.
Please note, that this is a selected literature review. A large amount of high quality science is not listed, but can be found if you follow citations presented in the papers below.
Part 1: Deep Learning Primer
Concepts in Deep Learning
Inspiration:
A call to arms for physicsts to tackle just one of many existential risks to humanity
Review papers
Philosophy
Overparameterization and Double Descent
Feature vs Lazy Learning (how to scale deep nets)
Yang, Simon, & Bernstein, A spectral conditioning for feature learning (2024)
Karkada, The lazy (NTK) and rich (μP) regimes: a gentle tutorial (2024)
Pehlevan and Bordelon, Lecture Notes on Infinite-Width Limits, Sec. 4
Deep Learning Architectures
For the most part, neural network architectures are described well in theory papers. This one is devoted exclusively to describing transformers, and is useful once you penetrate the pseudocode notation.
Statistical Mechanics of Learning
Language modeling
Signal Propagation
Signal prop in feedforward MLPs + architectural variants
RNNs
Schmidhuber, Sepp Hochreiter's Fundamental Deep Learning Problem (1991)
Pascanu, Mikolov, Bengio, On the difficulty of training RNNs
Molgedey et al., Suppressing Chaos in Neural Networks by Noise (1992)
Transformers
Dinan et al., Effective Theory of Transformers at Initialization (2023)
Cowsik et al., Geometric Dynamics of Signal Propagation Predict Trainability of Transformers (2024)
Part 2: LLMs
Scaling Laws
Empirical
**Kaplan et al., Scaling Laws for Neural Language Models (2020)**
Henighan et al., Scaling Laws for Autoregressive Generative Modeling (2020)
**Hoffmann et al., Training Compute-Optimal Large Language Models (Chinchilla) (2022) **
OpenAI, GPT-4 technical report
Hernandez et al. (Anthropic), Scaling Laws and Interpretability of Learning from Repeated Data (2022)
Theory
Bordelon et al. A Dynamical Model of Neural Scaling Laws (2024)
**Maloney et al., A Solvable Model of Neural Scaling Laws (2022)**
In-Context Learning
**Brown et al. (OpenAI), Language Models are Few-shot learners (2020)** (GPT-3 paper)
Lu et al., Asymptotic theory of in-context learning by linear attention (2024)
Emergence of Capabilities
Michaud et al., The quantization model of neural scaling (2024)
Nam et al., An exactly solvable model for emergence and scaling laws (2024)
Hidden Capabilities
Grokking
**Power et al., Grokking: Generalization Beyond Overfitting on Small Algorithmic Datasets (2022)**
**Varma et al, Explaining Grokking through circuit efficiency (2023)**
Reinforcement Learning from Human Feedback (RLHF)
Caspar, Davies et al., Open problems and fundamental limitations of RLHF
OpenAI, Training language models to follow instruction with human feedback
Mechanistic Interpretability