The 1W-MINDS Seminar was founded in the early days of the COVID-19 pandemic to mitigate the impossibility of travel. We have chosen to continue the seminar since to help form the basis of an inclusive community interested in mathematical data science, computational harmonic analysis, and related applications by providing free access to high quality talks without the need to travel. In the spirit of environmental and social sustainability, we welcome you to participate in both the seminar, and our slack channel community! Zoom talks are held on Thursdays either at 2:30 pm New York time or at 10:00 am Paris /4:00 pm summer Shanghai time/ 5:00 pm winter Shanghai time. To find and join the 1W-MINDS slack channel, please click here.
Current Organizers (September 2024 - May 2025): Axel Flinth (Principal Organizer for Europe/Asia, Umeå University), Christian Parkinson (Principal Organizer for The Americas, Michigan State University), Rima Alaifari (ETH Zürich), Alex Cloninger (UC San Diego), Longxiu Huang (Michigan State University), Mark Iwen (Michigan State University), Weilin Li (City College of New York), Siting Liu (UC Riverside), Kevin Miller (Brigham Young University), and Yong Sheng Soh (National University of Singapore).
Most previous talks are on the seminar YouTube channel. You can catch up there, or even subscribe if you like.
To sign up to receive email announcements about upcoming talks, click here.
To join MINDS slack channel, click here.
Passcode: the smallest prime > 100
Passcode: The integer part and first five decimals of e (Eulers number)
A central problem in data science is to use potentially noisy samples of an unknown function to predict values for unseen inputs. Classically, predictive error is understood as a trade-off between bias and variance that balances model simplicity with its ability to fit complex functions. However, over-parameterized models exhibit counterintuitive behaviors, such as “double descent” in which models of increasing complexity exhibit decreasing generalization error. Other models may exhibit more complicated patterns of predictive error with multiple peaks and valleys. Neither double descent nor multiple descent phenomena are well explained by the bias–variance decomposition. I present the generalized aliasing decomposition (GAD) to explain the relationship between predictive performance and model complexity. The GAD decomposes the predictive error into three parts: 1.) model insufficiency, which dominates when the number of parameters is much smaller than the number of data points, 2.) data insufficiency, which dominates when the number of parameters is much greater than the number of data points, and 3.) generalized aliasing, which dominates between these two extremes. I apply the GAD to linear regression problems from machine learning and materials discovery to explain salient features of the generalization curves in the context of the data and model class.