Paper presentations and projects: schedule & sign up sheet
Deadlines:
Oct 6 - project proposals due
TBA - final project submission due
Lecturer: Irina Rish
Topic: Intro and Overview: A brief history of AI at Scale (slides, video)
Papers: The Bitter Lesson, GPT-3 paper: Language Models are Few-Shot Learners
Topic: Intro and Overview: Continual Learning at Scale (slides, video)
Lecturer: Irina Rish
Topic: Overview of Papers to Present and Some Projects Topics (video-part1, video-part2 )
Class materials: some of the previous Topics & Papers (focus on: Continual Learning at Scale, Alignment and Safety, Emergence, Phase Transitions and Stat Physics of ML), Some previous large-scale projects, Towards Time_Series Foundation Models
Topic: Scaling Laws for Neural Language Models (slides, video)
Also covered: Training Compute-Optimal Large Language Models (Chinchilla Explained: video), Emergent Abilities of Large Language Models, Are emergent abilities of LLMs a Mirage? Additional materials: Neural Scaling Laws and GPT-3 (video); a nice overview of the history of scaling laws: Scaling Laws for LLMs: from GPT-3 to o3
Topic: An Empirical Model of Large-Batch Training (slides, video)