Machine Learning Seminar Series

Mar 16th

Speaker: Nicolas Gillis

  • Université de Mons, Belgium

  • Bio : Nicolas Gillis (nicolas.gillis@umons.ac.be) received his M.S. and Ph.D. degrees in applied mathematics from UCLouvain, Belgium, in 2007 and 2011, respectively. He is currently professor in the Department of Mathematics and Operational Research, University of Mons, Belgium. His research interests include optimization, numerical linear algebra, signal processing, machine learning, and data mining. He received the Householder Award in 2014 and a European Research Council starting grant in 2015. He currently serves as an associate editor of IEEE Transactions on Signal Processing and SIAM Journal on Matrix Analysis and Applications.


Talk information

  • Title: Identifiability and Computation of Nonnegative Matrix Factorizations

  • Time: Wednesday, Mar. 16th, 2022 12:001:00 pm

  • Location: Online via zoom (join) (video)

Abstract

Abstract: Given a nonnegative matrix X and a factorization rank r, nonnegative matrix factorization (NMF) approximates the matrix X as the product of a nonnegative matrix W with r columns and a nonnegative matrix H with r rows. NMF has become a standard linear dimensionality reduction technique in data mining and machine learning. In this talk, we will first briefly introduce NMF, and why it is useful in various applications. Then, we address the issue of non-uniqueness of NMF decompositions, also known as the identifiability issue, which is crucial in many applications. We discuss three key NMF models that allow us to obtain unique NMFs, namely, separable NMF, minimum-volume NMF, and sparse NMF. We also discuss how the factors (W,H) in such models can be computed. We illustrate these results on facial feature extraction, blind hyperspectral unmixing, and topic modeling.

Feb 23th

Speaker: Jingbo Liu

  • Department of Statistics, UIUC

  • Bio : Jingbo Liu received the B.S. in Electrical Engineering degree from Tsinghua University, Beijing, China in 2012, and the M.A. and Ph.D. degrees from Princeton University, Princeton, NJ, USA, in 2014 and 2017, all in electrical engineering. After two years of postdoc at MIT IDSS, he joined the Department of Statistics at the University of Illinois, Urbana-Champaign as an assistant professor. His research interests include signal processing, information theory, coding theory, high dimensional statistics, and the related fields. His undergraduate thesis received the best undergraduate thesis award at Tsinghua University (2012). He gave a semi-plenary presentation at the 2015 IEEE Int. Symposium on Information Theory, Hong-Kong, China. He was a recipient of the Princeton University Wallace Memorial Honorific Fellowship in 2016. His Ph.D. thesis received the Bede Liu Best Dissertation Award of Princeton and the Thomas M. Cover Dissertation Award of the IEEE Information Theory Society (2018).


Talk information

  • Title: A few interactions improve distributed nonparametric estimation

  • Time: Wednesday, Feb. 23th, 2022 12:001:00 pm

  • Location: Online via zoom (join)

Abstract

In recent years, the fundamental limits of distributed/federated learning have been studied under many statistical models, but often in the setting of horizontally partitioning, where data sets share the same feature space but differ in samples. Nevertheless, vertical federated learning, where data sets differ in features, have been in use in finance and medical care. In this talk, we consider a natural distributed nonparametric estimation problem with vertically partitioned datasets. Under a given budget of communication cost or information leakage constraint, we determine the minimax rates for estimating the density at a given point, which reveals that interactive protocols strictly improves over one-way protocols. Our novel estimation scheme in the interactive setting is constructed by carefully identifying a set of auxiliary random variables. The result also implies that interactive protocols strictly improve over one-way for biased binary sequences in the Gap-Hamming problem. (arXiv 2107.00211)

Feb 16th

Speaker: Hamed Hassani

  • Electrical and Systems Eng, U Penn

  • Bio : Hamed Hassani is currently an assistant professor of Electrical and Systems Engineering department as well as the Computer and Information Systems department, and the Statistics department at the University of Pennsylvania. Prior to that, he was a research fellow at Simons Institute for the Theory of Computing (UC Berkeley) affiliated with the program of Foundations of Machine Learning, and a post-doctoral researcher in the Institute of Machine Learning at ETH Zurich. He received a Ph.D. degree in Computer and Communication Sciences from EPFL, Lausanne. He is the recipient of the 2014 IEEE Information Theory Society Thomas M. Cover Dissertation Award, 2015 IEEE International Symposium on Information Theory Student Paper Award, 2017 Simons-Berkeley Fellowship, 2018 NSF-CRII Research Initiative Award, 2020 Air Force Office of Scientific Research (AFOSR) Young Investigator Award, 2020 National Science Foundation (NSF) CAREER Award, and 2020 Intel Rising Star award. He has recently been selected as the distinguished lecturer of the IEEE Information Theory Society in 2022-2023.


Talk information

  • Title: Learning in the Presence of Distribution Shifts: How does the Geometry of Perturbations Play a Role?

  • Time: Wednesday, Feb. 16th, 2022 12:001:00 pm

  • Location: Online via zoom (join) [slides] [video]

Abstract

In this talk, we will focus on the emerging field of (adversarially) robust machine learning. The talk will be self-contained and no particular background on robust learning will be needed. Recent progress in this field has been accelerated by the observation that despite unprecedented performance on clean data, modern learning models remain fragile to seemingly innocuous changes such as small, norm-bounded additive perturbations. Moreover, recent work in this field has looked beyond norm-bounded perturbations and has revealed that various other types of distributional shifts in the data can significantly degrade performance. However, in general our understanding of such shifts is in its infancy and several key questions remain unaddressed.

Feb 9th

Speaker: Qizhi He

  • CSGE, University of Minnesota

  • Bio : Qizhi He is an Assistant Professor in the Department of Civil, Environmental, and Geo- Engineering at the University of Minnesota. He received his M.A. in applied mathematics and Ph.D. in structural engineering and computational science from UC San Diego, in 2016 and 2018, respectively. Afterwards, he was a Postdoctoral Research Associate at Pacific Northwest National Laboratory (PNNL), where he developed scientific machine learning methods for modeling flow and transport processes in porous media. His current research interests lie at the intersection of computational mechanics, materials modeling, and data-driven computing, with a focus on advancing data-driven machine learning enabled computational tools to predict mechanics of complex multiphysical processes and improve our fundamental understanding of multiscale materials and structures in engineered and natural systems.

Talk information

  • Title: Machine Learning Enhanced Computational Mechanics for Materials Modeling

  • Time: Wednesday, Feb. 9th, 2022 12:001:00 pm

  • Location: Online via zoom (join)

Abstract

Qizhi He is an Assistant Professor in the Department of Civil, Environmental, and Geo- Engineering at the University of Minnesota. He received his M.A. in applied mathematics and Ph.D. in structural engineering and computational science from UC San Diego, in 2016 and 2018, respectively. Afterwards, he was a Postdoctoral Research Associate at Pacific Northwest National Laboratory (PNNL), where he developed scientific machine learning methods for modeling flow and transport processes in porous media. His current research interests lie at the intersection of computational mechanics, materials modeling, and data-driven computing, with a focus on advancing data-driven machine learning enabled computational tools to predict mechanics of complex multiphysical processes and improve our fundamental understanding of multiscale materials and structures in engineered and natural systems.

Jan. 26th

Speaker: Victor Zavala

  • Department of Chemical and Biological Engineering, University of Wisconsin-Madison

  • Bio : Victor M. Zavala is the Baldovin-DaPra Professor in the Department of Chemical and Biological Engineering at the University of Wisconsin-Madison and a computational mathematician in the Mathematics and Computer Science Division at Argonne National Laboratory. He holds a B.Sc. degree from Universidad Iberoamericana and a Ph.D. degree from Carnegie Mellon University, both in chemical engineering. He is on the editorial board of the Journal of Process Control, Mathematical Programming Computation, and Computers & Chemical Engineering. He is a recipient of NSF and DOE Early Career awards and of the Presidential Early Career Award for Scientists and Engineers. His research interests include computational modeling, statistics, control, and optimization.

Talk information

  • Title: The Euler Characteristic: A General Topological Descriptor for Complex Data

  • Time: Wednesday, Jan. 26th, 2022 12:001:00 pm

  • Location: Online via zoom (join) [slides]

Abstract

Datasets are mathematical objects (e.g., point clouds, matrices, graphs, images, fields/functions) that have shape. This shape encodes important knowledge about the system under study. Topology is an area of mathematics that provides diverse tools to characterize the shape of data objects. In this work, we study a specific tool known as the Euler characteristic (EC). The EC is a general, low-dimensional, and interpretable descriptor of topological spaces defined by data objects. We revise the mathematical foundations of the EC and highlight its connections with statistics, linear algebra, field theory, and graph theory. We discuss advantages offered by the use of the EC in the characterization of complex datasets; to do so, we illustrate its use in different applications of interest in chemical engineering such as process monitoring, flow cytometry, and microscopy. We show that the EC provides a descriptor that effectively reduces complex datasets and that this reduction facilitates tasks such as visualization, regression, classification, and clustering.

Jan. 19th

Speaker: Yihe Dong

  • Google Research

  • Bio : Yihe Dong is a machine learning researcher and engineer at Google, with interests in geometric deep learning and natural language processing.

Talk information

  • Title: Attention is not all you need

  • Time: Wednesday, Jan. 19th, 2021 12:001:00 pm

  • Location: Online via zoom (join) [slides] [video]

Abstract

I will be talking about our recent work on better understanding attention. Attention-based architectures have become ubiquitous in machine learning, yet our understanding of the reasons for their effectiveness remains limited. We show that self-attention possesses a strong inductive bias towards "token uniformity". Specifically, without skip connections or multi-layer perceptrons (MLPs), the output converges doubly exponentially to a rank-1 matrix. On the other hand, skip connections and MLPs stop the output from degeneration. Along the way, we develop a useful decomposition of attention architectures. This is joint work with Jean-Baptiste Cordonnier and Andreas Loukas.

Our paper and code are available online:

https://arxiv.org/abs/2103.03404;

https://github.com/twistedcubic/attention-rank-collapse.