Interests
Deep Learning
Transfer Learning
Dimension Reduction
Sequence Modeling
Dynamical Systems
I am a Research Scientist at the Lawrence Berkeley National Laboratory. I also lead the Deep Learning Group at the International Computer Science Institute (ICSI), an affiliated institute of UC Berkeley. Prior to this role, I was an Assistant Professor (Tenure-Track) for Data-driven Modeling and Artificial Intelligence in the Department of Mechanical Engineering and Materials Science at the University of Pittsburgh, from September 2021 to December 2022. Before joining Pitt, I was a postdoctoral researcher in the Department of Statistics at UC Berkeley, where I worked with Michael Mahoney. I was also part of the RISELab in the Department of Electrical Engineering and Computer Sciences (EECS) at UC Berkeley. Before that, I was a postdoc in the Department of Applied Mathematics at the University of Washington (UW), working with Nathan Kutz and Steven Brunton. I earned my PhD in Statistics at the University of St Andrews in December 2017. My MSc. in Applied Statistics is also from the University of St Andrews.
I am broadly interested in understanding what makes deep learning systems work, and how we can make them more robust, interpretable, and efficient. My work takes a scientific approach to these challenges. Viewing neural networks through the lens of dynamical systems theory helps explain issues like exploding and vanishing gradients, and points toward new ways of designing networks with better inductive biases, drawing on ideas from numerical integration and stochastic differential equations to make training more stable and inference more reliable.
Building on this foundation, I am currently working on large-scale generative diffusion models for spatio-temporal forecasting in areas such as earth science and fluid dynamics. I’m also exploring how foundation models can combine reasoning and multimodal information to make more informed predictions. A growing part of my research focuses on AI safety, aiming to understand and mitigate vulnerabilities in large language models, including jailbreaking and backdoor attacks.
One paper accepted in NeurIPS 2025
Block-Biased Mamba for Long-Range Sequence Processing (preprint)
One paper accepted in ICML 2025
Emoji Attack: Enhancing Jailbreak Attacks Against Judge LLM Detection (preprint)
I will serve as an area chair @ NeurIPS 2025.
I am co-organizing the Deep Learning for Science Summer School
Two papers accepted in ICLR 2025 (one as spotlight)
I will serve as an area chair @ ICML 2025.
One paper accepted in AISTATS 2025
Gated Recurrent Neural Networks with Weighted Time-Delay Feedback (preprint).
I am co-organizing The Berkeley Lab AI for Science Summit (BLASS 24)
I will serve as an area chair @ ICLR 2025.
Two papers accepted in ICLR 2024 (one as spotlight)
Robustifying State-space Models for Long Sequences via Approximate Diagonalization.
Generative Modeling of Regular and Irregular Time Series Data via Koopman VAEs.
One paper accepted in AISTATS 2024
Boosting model robustness to common corruptions.
One paper accepted in AISTATS 2023 (as oral presentation)
Error Estimation for Random Fourier Features.
Two papers accepted in ICLR 2022 (one as spotlight)
Noisy Feature Mixup.
Long Expressive Memory for Sequence Modeling.
Ananya Gupta, Undergrad Researcher
Aditi Gupta, Summer Intern
Pu Ren, joint Postdoc
with Michael Mahoney
Yihan Wang, Visiting PhD Student
Cici Wang (Undergraduate research intern 2023-25; now graduate student at U Chicago).
Garry Gao (Graduate researcher 2023-24; now Software Engineer at Amazon).
Junyi Guo (Graduate researcher 2022-24; now PhD student at University of Notre Dame).
Jialin Song (Graduate researcher 2021-24; now PhD student at Simon Fraser University).
Yixiao Kang (Graduate researcher 2023-24; now ML Engineer at Meta).
Daniel Barron (Graduate researcher 2023-24; now Software Engineer at Amazon).
Olina Mukherjee (High School student researcher, 2021-22; now undergraduate student at CMU).
Ziang Cao (Undergraduate research intern 2020-21; now graduate student at Stanford).
Francisco Utrera (Graduate researcher 2019-22; now Senior ML Engineer at Erithmitic) .
Evan Kravitz (Graduate researcher 2019-20; now Software Engineer at Amazon) .
Vanessa Lin (Undergraduate research intern 2018-19; now at Google).
Qixuan Wu (Undergraduate research intern 2018-19; now at Goldman Sachs).