The One World Seminar Series on the Mathematics of Machine Learning is an online platform for research seminars, workshops and seasonal schools in theoretical machine learning. The focus of the series lies on theoretical advances in machine learning and deep learning. The series was started during the Covid-19 epidemic in 2020 to bring together researchers from all over the world for presentations and discussions in a virtual environment. It follows in the footsteps of other community projects under the One World Umbrella which originated around the same time.
We welcome suggestions for speakers concerning new and exciting developments and are committed to providing a platform also for junior researchers. We recognize the advantages that online seminars provide in terms of flexibility. Any feedback on different events is welcome.
Zoom talks are held on Wednesdays at 12:00 pm New York time (9:00pm Pacific).
A list of past seminars can be found here and recordings can be viewed on our Youtube channel. The invitation to future seminars will be shared on this site before the talk and distributed via email.
Wed 28 Jan
Qing Gu
Understanding Generalization of Deep Generative Models Requires Rethinking Underlying Low-dimensional Structures
Diffusion models represent a remarkable new class of deep generative models, yet the mathematical principles underlying their generalization from finite training data are poorly understood. This talk offers novel theoretical insights into diffusion model generalization through the lens of "model reproducibility," revealing a surprising phase transition from memorization to generalization during training, notably occurring without the curse of dimensionality. Our theoretical framework hinges on two crucial observations: (i) the intrinsic low dimensionality of image datasets and (ii) the emergent low-rank property of the denoising autoencoder within trained neural networks. Under simplified settings, we rigorously establish that optimizing the training loss of diffusion models is mathematically equivalent to solving a canonical subspace clustering problem. This insight quantifies the minimal sample requirements for learning low-dimensional distributions, scaling linearly with the intrinsic dimension. Furthermore, by investigating this under a nonlinear two-layer network, we fully explain the memorization-to-generalization transition, highlighting inductive biases in learning dynamics and the models' strong representation learning ability. These theoretical insights have profound practical implications, enabling various applications for generation control and safety, including concept steering, watermarking, and memorization detection. This work not only advances theoretical understanding but also stimulates numerous directions for many applications in engineering and science.
Wed 11 Mar
Pierre-Alexandre Mattei
Ensembles in machine learning: (simple) theory and (simple) practice
Ensemble methods combine predictions from various statistical learning models. Their most famous representatives are random forests or neural network ensembles. This talk will center around the question: "How many models should I aggregate?" We will see that the answer depends on the chosen performance metric. Specifically, in the case of convex losses (such as cross-entropy in classification or mean squared error in regression), the error is a decreasing function of the number of models. In the case of non-convex losses (such as classification error in classification or the Fréchet Inception distance in generative modelling), things are more nuanced, and the error can sometimes be non-monotonic. These results will be illustrated with examples of neural network ensembles. This work is notably based on the paper JMLR paper "Are Ensembles Getting Better All the Time?" (http://jmlr.org/papers/v26/24-0408.html), joint work with Damien Garreau (Julius-Maximilians-Universität Würzburg). It will also feature recent work with Raphaël Razafindralambo, Rémy Sun, and Frédéric Precioso (Inria, Université Côte d'Azur).
Sign up here to join our mailing list and receive announcements. If your browser automatically signs you into a google account, it may be easiest to join on a university account by going through an incognito window. With other concerns, please reach out to one of the organizers.
Ricardo Baptista (University of Toronto)
Wuyang Chen (Simon Fraser University)
Bin Dong (Peking University)
Lyudmila Grigoryeva (University of St. Gallen)
Boumediene Hamzi (Caltech)
Yuka Hashimoto (NTT)
Qianxiao Li (National University of Singapore)
Lizao Li (Google)
George Stepaniants (Caltech)
Zhiqin Xu (Shanghai Jiao Tong University)
Simon Shaolei Du (University of Washington)
Franca Hoffmann (Caltech)
Surbhi Goel (Microsoft Research NY)
Issa Karambal (Quantum Leap Africa)
Tiffany Vlaar (University of Glasgow)
Chao Ma (Stanford University)
Song Mei (UC Berkeley)
Philipp Petersen (University of Vienna)
Matthew Thorpe (University of Warwick)
Stephan Wojtowytsch (University of Pittsburgh)