Confirmed Speakers
Stanford University
Title: The Art of Artificial Reasoning for Small Language Models
Abstract:
Large reasoning models such as Deepseek's R1 and OpenAI's O1/O3 have demonstrated the power of reinforcement learning to enable a new axis of scaling — test-time compute. This has catalyzed intensive research across the open-source community, generating rapid progress but also seemingly contradictory results. In this talk, I will present critical insights into the conditions under which reinforcement learning thrives or struggles, and how we can induce stronger reasoning capabilities from small language models, closing the gap against the larger counterparts in specific domains.
About the speaker:
Yejin Choi is the Dieter Schwartz Foundation HAI Professor, a Professor of Computer Science, and a Stanford HAI Senior Fellow. Prior, she was a professor of Paul G. Allen School of Computer Science & Engineering at the University of Washington, adjunct of the Linguistics department, and affiliate of the Center for Statistics and Social Sciences. Her research focuses on teaching AI systems commonsense reasoning, social intelligence, and moral understanding by grounding them in language, vision, and human norms.
University of Pennsylvania
Title: How Jailbreaking 1-Layer Transformers Taught us how to Steer LLMs
Abstract: Why are LLMs fundamentally so easily jailbroken? We will first analyze how to subvert and "jailbreak" simple, one-layer transformers, formalizing rule-breaking as a vulnerability in the model's attention mechanism. This theoretical work taught us a critical lesson: if attention is the key to breaking rules, it may also be the key to enforcing them. We will then introduce InstABoost, an incredibly simple yet highly effective steering method that directly applies this insight to boost the model's attention on user-provided instructions during generation. This technique, inspired by our theoretical understanding of how a small model is jailbroken, provides state-of-the-art control over large-scale LLMs with just five lines of code.
About the speaker:
Eric Wong is an assistant professor at the Department of Computer and Information Science at the University of Pennsylvania. His research spans machine learning, optimization, and robustness, in order to develop principled methods with an eye towards scalability and practicality in real-world settings such as cosmology and surgery.
Carnegie Mellon University
Title: Beyond benchmarks: the case for spherical cows in LLM research
Abstract:
Real-world benchmarks drive a lot of progress, but they cannot capture all key aspects of real-world deployment. How does one study notions like creativity and adaptability to new domains? These are messy, subjective, and cannot be captured easily in static benchmarks. I will argue that carefully constructed minimal examples and stylized settings---"spherical cows"---offer a powerful answer, helping surface important blind spots and limits in current paradigms.
For adaptation, I will walk through the story of how we discovered a surprising phenomenon of catastrophic overtraining, where pre-training on more tokens can hurt downstream fine-tuning. This challenges the core belief of machine learning that "more data is better". While initially conceptualized via stylized models, we put this to the test at scale on real-world datasets and observe that OLMo-1B model pre-trained on 3T tokens performs worse after fine-tuning than its less trained 2.3T token counterpart.
For creativity, we construct minimal examples that capture notions of creativity inspired by cognitive science. These examples show that next-token prediction is fundamentally myopic, underperforming multi-token approaches like teacherless training and diffusion models in generating diverse, original outputs. These settings also highlight that standard techniques such as temperature sampling may be suboptimal compared to injecting randomness through input prefixes.
About the speaker:
Aditi Raghunathan is an Assistant Professor in the Computer Science Department at Carnegie Mellon University. Her research focuses on actively developing new pre-training approaches for improved safety, privacy, and reasoning by design.
Princeton university
Title: Designing Efficient Attention: Insights from an Inference Perspective
Abstract:
Inference now drives the progress of AI thanks to test-time compute, necessitating efficient redesign of core architectural components such as the attention layer. We examine recent progress in addressing these efficiency challenges through two complementary directions. Starting from first principles of hardware efficiency with arithmetic intensity, we motivate the design of DeepSeek's multi-head latent attention (MLA) and recent variants such as Grouped-Tied Attention and Grouped Latent Attention. These variants reduce memory bandwidth requirements by performing more computation per byte loaded from memory, achieving up to 2× speedup in decoding scenarios. Second, we explore a complementary direction that reduces the FLOPS of attention from quadratic to linear or quasi-linear, bridging the gap between linear attention's efficiency and softmax attention's expressiveness. These new linear and quasi-linear attention methods enable sub-quadratic scaling while maintaining modeling capacity. Through systematic investigation at small scales, we demonstrate how inference-driven design principles can unlock new insights into attention mechanisms and provide practical pathways toward more efficient large-scale deployment.
About the speaker:
Tri Dao is an assistant professor at Princeton University and a chief scientist at Together AI. His research centers on Machine learning and systems, with a focus on efficient training and long-range context.
Panelists
University of California, San Diego
About the panelist:
Misha Belkin is a professor at the Halicioglu Data Science Institute, University of California San Diego. He is interested in questions concerning computation, statistics an optimization in Machine Learning, particularly for high dimensional data. His recent work has focused on the fundamental understanding of modern ML and deep learning, particularly on interpolation, overparemeterization and feature learning.
Haize Labs
About the panelist:
Nimit Kalra is a machine learning researcher at Haize Labs, where he is focused on aligning and robustifying real-world AI systems. His research develops methodologies across adversarial robustness, uncertainty calibration, and automated model-based evaluation for LLMs. He is the author of Verdict, an open-source framework for applying inference-time scaling to LLM judges.
EleutherAI
About the panelist:
Stella Biderman is a machine learning researcher and the executive director of the non-profit research center EleutherAI. She is a leading advocate for open source artificial intelligence and has trained many of the world's most powerful open source artificial intelligence algorithms.
Stanford University
About the panelist:
Yejin Choi is the incoming Dieter Schwarz Foundation HAI Professor, a Professor of Computer Science, and a Stanford HAI Senior Fellow.
Carnegie Mellon University
About the panelist:
Aditi Raghunathan is an Assistant Professor in the Computer Science Department at Carnegie Mellon University. Her research focuses on actively developing new pre-training approaches for improved safety, privacy, and reasoning by design.
University of British Columbia
About the panelist:
Christos Thrampoulidis is an Assistant Professor in the ECE Department at the University of British Columbia (UBC) in Vancouver, who works on machine learning and deep-learning theory, optimization and statistical signal processing.