Machine Learning Seminar Series

Dec. 16th

Speaker: Stefano Martiniani

  • Chemistry Engineering & Material Science, University of Minnesota

  • Bio : Stefano Martiniani is Assistant Professor in the Department of Chemical Engineering and Materials Science at the University of Minnesota, a member of the graduate faculty in the School of Physics & Astronomy, and in Data Science. Prior to joining UMN, he was a postdoc at the Center for Soft Matter Research at New York University, and a Gates Scholar at the University of Cambridge where he obtained an MPhil in Scientific Computing and a PhD in Theoretical Chemistry. Stefano’s research focuses on the design of novel theoretical and computational frameworks to address open problems in science and engineering. His work draws primarily from statistical and computational physics, dynamical systems, and machine learning. His theoretical interests span the energy landscapes of disordered systems, neuronal dynamics, bio/molecular design and simulation, and soft matter.

Talk information

  • Title: Design and Discovery in the Protein Fitness Landscape

  • Time: Wednesday, Dec. 16th, 2020 3:30–4:30 pm

  • Location: Online via zoom (join)

Abstract

Engineered proteins have emerged as novel diagnostics, therapeutics, and catalysts. Often, poor protein developability — quantified by expression, solubility, and stability — hinders commercialization. The ability to predict protein developability from amino acid sequence would reduce the experimental burden when selecting candidates. Recent advances in screening technologies enabled a high-throughput (HT) developability dataset for 105 of 1020 possible variants of protein scaffold Gp2. In this work, we evaluate the ability of neural networks to learn a developability representation from the HT dataset and transfer the knowledge to predict recombinant expression beyond the observed sequences. Our model predicts expression levels 42% closer to the experimental variance compared to a non-embedded control. We then seek to exploit this knowledge to design new sequences with high developability. To overcome the intractability of a brute force search of 1020 possible Gp2 variants, we descend through the protein fitness landscape by Nested Sampling, a Monte Carlo scheme for Bayesian parameter estimation and model selection, which is particularly suited for the analysis of multimodal distributions. In addition to identifying high-developability libraries, we obtain unprecedented insight into the structure of the protein fitness landscape through a “topographical” analysis and statistical mechanical interpretation of the results.

Dec. 9th

Speaker: Qi Zhang

  • Department of Chemical Engineering and Materials Science, University of Minnesota

  • Bio : Qi Zhang is an Assistant Professor in the Department of Chemical Engineering and Materials Science at the University of Minnesota. He received his Ph.D. in Chemical Engineering from Carnegie Mellon University and worked at BASF prior to joining the University of Minnesota. His research lies at the intersection of chemical engineering and operations research, particularly focusing on mixed-integer optimization, decision making under uncertainty, and data analytics, with applications in sustainable energy and process systems, advanced manufacturing, and supply chain management.

Talk information

  • Title: Data-driven inverse linear optimization with noisy observations

  • Time: Wednesday, Dec. 9th, 2020 3:30–4:30 pm

  • Location: Online via zoom (join)

Abstract

Inverse optimization refers to the inference of unknown optimization models given decision data that are assumed to be optimal or near-optimal solutions to the unknown optimization problem. In this work, we consider data-driven inverse linear optimization, where the underlying decision-making process can be modeled as a linear program whose cost vector is unknown. We propose a two-phase algorithm to determine the best estimate of the cost vector given noisy observations. Moreover, we propose an efficient decomposition algorithm to solve large instances of the problem. The algorithm extends naturally to an online learning environment where it can be used to provide quick updates of the cost estimate as new data becomes available over time. For the online setting, we further develop an effective adaptive sampling strategy that guides the selection of the next samples. The efficacy of the proposed methods is demonstrated in computational experiments involving two applications, customer preference learning and cost estimation for production planning. The results show significant reductions in computation and sampling efforts.

Dec. 2rd

Speaker: Vuk Mandic

  • School of Physics and Astronomy

  • Bio : Prof. Vuk Mandic is a Distinguished McKnight University Professor at the University of Minnesota Twin Cities. He obtained his Physics Ph.D. at UC Berkeley in 2004, and since then has been working in the field of gravitational wave astrophysics. He is a member of the LIGO Scientific Collaboration that discovered the first gravitational wave signal in 2015, generated in a merger of two black holes, which resulted in a Nobel Prize in Physics in 2017. Prof. Mandic is the founding director of the Center for Excellence in Sensing Technologies and Analytics (CESTA) at the University of Minnesota.

Talk information

  • Title: Sensing and Machine Learning

  • Time: Wednesday, Dec. 2rd, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

Center for Excellence in Sensing Technologies and Analytics (CESTA) at the University of Minnesota brings together researchers from diverse domain fields that use sensing in their work. While a number of CESTA members work on development of novel sensing technologies, there is also an increasing demand for using modern data analytic techniques to process and analyze diverse and large volumes of data produced by sensor arrays. I will survey the machine/deep learning needs of CESTA researches across multiple domain fields, highlighting the research projects that could benefit from machine/deep learning expertise.

Nov. 18th

Speaker: Jason Goodpaster

  • Department of Chemistry , University of Minnesota

  • Bio : Jason Goodpaster is an assistant professor of chemistry at the University of Minnesota Twin Cities. His research focuses on the development of new quantum chemistry methods and applying these methods to a wide variety of chemical systems including: metal organic frameworks, inorganic catalysis, surface enhanced ramen spectroscopy, and electrochemistry. Professor joined UMN in June 2016. Before that, he obtained his PhD at Caltech and performed his postdoctoral work at Lawrence Berkeley National Lab.

Talk information

  • Title: Transfer learning of neural network potentials for reactive chemistry

  • Time: Wednesday, Nov. 18th, 2020 3:30–4:30 pm

  • Location: Online via zoom (join)

Abstract

Large, condensed phase, and extended systems impose a challenge for theoretical studies due to the compromise between accuracy and computational cost in their calculations. Machine learning methods are an approach to solve this trade-off by leveraging large data sets to train on highly accurate calculations using small molecules and then apply them to larger systems. In this study, we are developing a method to train a neural network potential with high-level wavefunction theory on targeted systems of interest that are able to describe bond breaking. We combine density functional theory calculations and higher level ab initio wavefunction calculations, such as CASPT2, to train our neural network potentials. We first train our neural network at the DFT level of theory and using an adaptive active learning training scheme, we retrain the neural network potential to a CASPT2 level of accuracy. We demonstrate the process as well as report current progress and performance of this neural network potential for molecular dynamic simulations.

Nov. 11th

Speaker: Edward McFowland III

  • Information and Decision Sciences in the Carlson School of Management, University of Minnesota

  • Bio :Dr. Edward McFowland III is an Assistant Professor of Information and Decision Sciences in the Carlson School of Management, at the University of Minnesota; he received his Ph.D. in Information Systems and Management from Carnegie Mellon University. Dr. McFowland’s research interests—which lie at the intersection of Information Systems, Machine Learning, and Public Policy—include the development of computationally efficient algorithms for large-scale statistical machine learning and “big data” analytics. More specifically, his research seeks to demonstrate that many real-world problems faced by organizations, and society more broadly, can be reduced to the tasks of anomalous pattern detection and discovery. As a data and computational social scientist, Dr. McFowland’s broad research goal is bridging the gap between machine learning and the social sciences (e.g., economics, public policy, and management) both through the application of machine learning methods to social science problems and through the integration of machine learning and econometric methodologies.

Talk information

  • Title: Efficient Discovery of Heterogeneous Treatment Effects in Randomized Experiments via Anomalous Pattern Detection

  • Time: Wednesday, Nov. 11th, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

In the recent literature on estimating heterogeneous treatment effects, each proposed method makes its own set of restrictive assumptions about the intervention's effects and which subpopulations to explicitly estimate. Moreover, the majority of the literature provides no mechanism to identify which subpopulations are the most affected--beyond manual inspection--and provides little guarantee on the correctness of the identified subpopulations. Therefore, we propose Treatment Effect Subset Scan (TESS), a new method for discovering which subpopulation in a randomized experiment is most significantly affected by a treatment. We frame this challenge as a pattern detection problem where we efficiently maximize a nonparametric scan statistic over subpopulations. Furthermore, we identify the subpopulation which experiences the largest distributional change as a result of the intervention, while making minimal assumptions about the intervention's effects or the underlying data generating process. In addition to the algorithm, we demonstrate that the asymptotic Type I and II error can be controlled, and provide sufficient conditions for detection consistency--i.e., exact identification of the affected subpopulation. Finally, we validate the efficacy of the method by discovering heterogeneous treatment effects in simulations and in real-world data from a well-known program evaluation study.

Nov. 4th

Speaker: Raphael Stern

  • Department of Civil, Environmental, and Geo- Engineering, University of Minnesota

  • Bio: Raphael Stern is an Assistant Professor in the Department of Civil, Environmental, and Geo- Engineering at the University of Minnesota. Prior to joining UMN, Dr. Stern was a postdoctoral researcher in the Department of Informatics at the Technical University of Munich. Dr. Stern has also spent time as a visiting researcher at the Institute for Software Integrated Systems at Vanderbilt University. Dr. Stern received a bachelor of science degree (2013), master of science degree (2015), and Ph.D. (2018) all in Civil Engineering from the University of Illinois at Urbana-Champaign. Dr. Stern was a visiting researcher at the Institute for Pure and Applied Mathematics at UCLA, and a recipient of the Dwight David Eisenhower Graduate Fellowship from the Federal Highway Administration. Dr. Stern's research interests are in the area of traffic control and estimation with autonomous vehicles in the flow.

Talk information

  • Title: Impacts of automated vehicles: a traffic flow perspective

  • Time: Wednesday, Nov. 4th, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

This work is motivated by the possibility of a small number of automated vehicles (AVs) or partially automated vehicles that may soon be present on our roadways, and the impacts they will have on traffic flow. This automation may take the form of fully autonomous vehicles without human intervention (SAE Level 5) or, as is already the case in many modern vehicles, may take the form of driver assist features such as adaptive cruise control (ACC) or other SAE Level 1 features. Regardless of the extent of automation, the introduction of such vehicles has the potential to substantially alter emergent properties of the flow while also providing new opportunities for traffic state estimation. In this talk, I present some recent experimental and modeling work conducted to understand how AVs may be able to influence traffic flow.

Oct. 28th

Speaker: Sandra Safo

  • school of Public Health, University of Minnesota

  • Bio: Sandra Safo is an Assistant Professor of school of Public Health at University of Minnesota. Her primary research focuses on developing and applying statistical and machine learning methods and computational tools for big, biomedical data to advance clinical translational research and precision medicine. I have been developing multivariate statistical methods, statistical learning (including classification, discriminant analysis, association studies, biclustering), data integration, and feature selection methods for high dimensional data. Currently, I develop methods for integrative analysis of “omics” (including genomics, transcriptomics, and metabolomics) and clinical data to help elucidate complex interactions of these multifaceted data types.

Talk information

  • Title: Joint association and classification of multi-view structured data

  • Time: Wednesday, October 28, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

Classification methods that leverage the strengths of data from multiple sources (multi-view data) simultaneously have enormous potential to yield more powerful findings than two step methods: association followed by classification. We propose two methods, sparse integrative discriminant analysis (SIDA) and SIDA with incorporation of network information (SIDANet), for joint association and classification studies. The methods consider the overall association between multi-view data, and the separation within each view in choosing discriminant vectors that are associated and optimally separate subjects into different classes. SIDANet is among the first methods to incorporate prior structural information in joint association and classification studies. It uses the normalized Laplacian of a graph to smooth coefficients of predictor variables, thus encouraging selection of predictors that are connected. We demonstrate the effectiveness of our methods on a set of synthetic and real datasets. Our findings underscore the benefit of joint association and classification methods if the goal is to correlate multi-view data and to perform classification.

Oct. 21th

Speaker: Xiaoou Li

  • School of Statistics, University of Minnesota

  • Bio: Xiaoou Li is an assistant professor of statistics at the University of Minnesota. She holds a Ph.D. degree in Statistics from Columbia University in 2016. Her research interest includes latent variable models, sequential analysis, psychometrics, and applied probability.

Talk information

  • Title: Compound sequential change detection in multiple data streams

  • Time: Wednesday, October 21, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

We consider sequential change detection in multiple data streams, where each stream has its own change point. The goal is to maximize the normal operation of the pre-change streams while controlling the proportion of post-change streams among the active streams at all time points. Taking a Bayesian formulation, we develop a compound sequential decision theory framework for this problem and propose an oracle procedure under this framework. We also extend the problem to more general settings and apply the method to the monitoring of item pool quality in educational assessment.


Oct. 14th

Speaker: Yangzesheng Sun

  • Department of Chemistry and Department of Computer Science & Engineering, University of Minnesota

  • Bio: Yangzesheng (Andrew) Sun is a PhD student in the Department of Chemistry at University of Minnesota, advised by prof. J. Ilja Siepmann. His research interests include physics-informed machine learning, Monte Carlo simulations for molecular systems, and high-performance computing. He received BS in Chemistry with Honors from Wuhan University in 2017.

Talk information

  • Title: Physics-informed machine learning for molecular simulations in nanoporous materials discovery

  • Time: Wednesday, October 14, 2020 3:30–4:30 pm

  • Location: Online via zoom (join) (download slides)

Abstract

Nanoporous materials are promising candidates for clean-energy chemical storage and separation processes, and a combination of high-throughput molecular simulations with machine learning methods has dramatically accelerated the design of chemical systems involving nanoporous materials. Although machine learning has shown great success in various domains, physical inductive biases play a crucial role in the generalization and transferability of machine learning models for materials discovery. Based on the statistical thermodynamic nature of molecular simulations, the loss function of a machine learning model predicting molecular simulation results can be structured as minimizing the KL divergence between the statistical thermodynamics distribution and the approximating distribution parametrized by the model. When multiple types of guest molecules are separated by a nanoporous material, a strongly physics-informed and interpretable machine learning model based on the Transformer was developed by drawing an analogy between molecules in a chemical system and words in natural language. With almost trivial modifications to the original Transformer, the model dramatically outperformed a regular neural network on generalization in the state space for the separation of an 8-component benzene derivative mixture. Besides integrating physical principles into model architecture, meta-learning was also an effective approach to directly learning physical inductive biases from data. A metal-learning model was developed to jointly learn the hydrogen storage properties of multiple materials at multiple thermodynamic states using many small simulation datasets, improving extrapolation and few-shot performance. While machine learning was mainly employed as surrogate models for simulations, data-driven or differentiable simulations have also emerged as a new research direction.


Sep. 30th

Speaker: Jeff Calder

  • School of Mathematics, University of Minnesota

  • Bio: Jeff Calder is an assistant professor of mathematics at the University of Minnesota. He completed his PhD in 2014 at the University of Michigan advised by Selim Esedoglu (Math) and Alfred Hero (EECS), and was a Morrey assistant professor at UC Berkeley from 2014-2016. Calder was awarded an Alfred P. Sloan Research Fellowship in 2020. His research interests are focused on the intersection of partial differential equations, machine learning, and applied probability.

Talk information

  • Title: Poisson learning: Graph-based semi-supervised learning at very low label rates

  • Time: Wednesday, September 30, 2020 3:30–4:30 pm

  • Location: Online via zoom (download slides)

Abstract

Graph-based learning is a field within machine learning that uses similarities between datapoints to create efficient representations of high-dimensional data for tasks like semi-supervised classification, clustering and dimension reduction. For semi-supervised learning, the widely used Laplacian regularizer has recently been shown to be ill-posed at very low label rates, leading to poor classification similar to random guessing. In this talk, we will give a careful analysis of the ill-posedness of Laplacian regularization via random walks on graphs, and this will lead to a new algorithm for semi-supervised learning that we call Poisson learning. Poisson learning replaces the assignment of label values at training points with the placement of sources and sinks, and solves the resulting Poisson equation on the graph. The outcomes are provably more stable and informative than those of Laplacian learning. Poisson learning is efficient and simple to implement, and we will present numerical results on MNIST, FashionMNIST and Cifar-10 showing that the method is superior to other recent approaches to semi-supervised learning at low label rates.

This talk is joint work with Brendan Cook (UMN), Dejan Slepcev (CMU) and Matthew Thorpe (University Manchester).

Sep. 23th

Speaker: Xuan Bi

  • Information and Decision Sciences Carlson School of Management , University of Minnesota

  • Bio: Xuan Bi is an Assistant Professor of Information and Decision Sciences at the Carlson School of Management. He holds a PhD in Statistics from the University of Illinois at Urbana-Champaign and was a postdoc at Yale University before joining the University of Minnesota. Dr. Bi’s research mainly revolves around machine learning methodologies for personalization, such as recommender systems, and has several articles published on top journals in Statistics. Dr. Bi also holds broad interests in other machine learning and data science areas, such as differential privacy, product forecasting, and imaging data analysis, and is open to opportunities for collaboration

Talk information

  • Title: Improving Sales Forecasting Accuracy: A tensor factorization approach with demand awareness

  • Time: Wednesday, September 23, 2020 3:30–4:30 pm

  • Location: Online via zoom

Abstract

Due to accessible big data collections from consumers, products, and stores, advanced sales forecasting capabilities have drawn great attention from many companies especially in the retail business because of its importance in decision making. Improvement of the forecasting accuracy, even by a small percentage, may have a substantial impact on companies' production and financial planning, marketing strategies, inventory controls, supply chain management, and eventually stock prices. Specifically, our research goal is to forecast the sales of each product in each store in the near future. Motivated by tensor factorization methodologies for personalized context-aware recommender systems, we propose \iv{a novel approach} called the Advanced Temporal Latent-factor Approach to Sales forecasting (ATLAS), which achieves accurate and individualized prediction for sales by building a single tensor-factorization model across multiple stores and products. Our contribution is a combination of: tensor framework (to leverage information across stores and products), a new regularization function (to incorporate demand dynamics), and extrapolation of tensor into future time periods using state-of-the-art statistical (seasonal auto-regressive integrated moving-average models) and machine-learning (recurrent neural networks) models. The advantages of ATLAS are demonstrated on \iv{eight datasets} collected by the Information Resource, Inc., where a total of 165 million weekly sales transactions from more than 1,500 grocery stores over 15,560 products are analyzed.

Sep. 9th

Speaker: Dr Harini Veeraraghavan

  • Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York University

  • Bio: Harini Veeraraghavan is an assistant attending computer scientist at Memorial Sloan Kettering Cancer Center since 2012. Her research interests are in developing deep learning and medical image analysis methods for improving cancer treatments. Her group actively develops automated segmentation methods that are currently used for image-guided radiation therapy treatments at MSKCC. She also leads the development of integrated image-based biomarkers of cancer treatment response from radiological and surgical images for multiple cancers. Harini received her PhD in computer science from the University of Minnesota, Twin-Cities in 2006. Prior to joining MSKCC, she was a computer vision scientist at General Electric Research since 2008, and a postdoctoral fellow at Carnegie Mellon University from 2006 to 2008

Talk information

  • Title: Cross-domain deep distillation learning for image-guided therapies

  • Time: Wednesday, September 9th, 2020 3:30–4:30 pm

  • Location: Online via zoom

Abstract

Medical image analysis for oncology applications is plagued by problems of small labeled datasets and the difficulty in analyzing images that often have very low soft-tissue contrast to differentiate foreground and background structures. Deep learning techniques such as deep domain adaptation developed in computer vision, have limited success when applied to medical images. This is due to the wide variations in the images even in the same modalities and even larger variations of the different tissues appearing on multiple imaging modalities. In this talk, I will present some solutions developed by our group for solving these two problems. In brief, I will present an approach that uses joint distribution matching to learn from entirely different imaging modalities to segment on a target modality without any expert labels. Next, I will show how cross-domain adaptation can be combined with distillation learning to train models to segment centrally located lung tumors on CT and more challenging cone-beam CT images. I will also show how these methods are being translated from lab into clinical practice for image-guided radiation treatments in our hospital.