Keynote Talks

Towards Instance-Optimized Data Systems

[]

Abstract

Recently, there has been a lot of excitement around ML-enhanced (or learned) algorithms and data structures. For example, there has been work on applying machine learning to improve query optimization, indexing, storage layouts, scheduling, log-structured merge trees, sorting, compression, sketches, among many other data management tasks. Arguably, the ideas behind these techniques are similar: machine learning is used to model the data and/or workload in order to derive a more efficient algorithm or data structure. Ultimately, what these techniques will allow us to build are “instance-optimized” systems; systems that self-adjust to a given workload and data distribution to provide unprecedented performance and avoid the need for tuning by an administrator.

In this talk, I will first provide an overview of the opportunities and limitations of current ML-enhanced algorithms and data structures, present initial results of SageDB, a first instance-optimized system we are building as part of DSAIL@CSAIL at MIT, and finally outline remaining challenges and future directions.

Bio

Tim Kraska is an Associate Professor of Electrical Engineering and Computer Science in MIT's Computer Science and Artificial Intelligence Laboratory, co-director of the Data System and AI Lab at MIT (DSAIL@CSAIL), and co-founder of Einblick Analytics. Currently, his research focuses on building systems for machine learning, and using machine learning for systems. Before joining MIT, Tim was an Assistant Professor at Brown, spent time at Google Brain, and was a PostDoc in the AMPLab at UC Berkeley after he got his PhD from ETH Zurich. Tim is a 2017 Alfred P. Sloan Research Fellow in computer science and received several awards including the VLDB Early Career Research Contribution Award, the VMware Systems Research Award, the university-wide Early Career Research Achievement Award at Brown University, an NSF CAREER Award, as well as several best paper and demo awards at VLDB and ICDE.


Do Learned Patterns beat Smart Heuristics?

[]

Abstract

In this talk, I will tell 3 stories in learning patterns from ML in search and recommendations, to learned index structures and learned compilation.

Bio

Ed H. Chi is a Distinguished Scientist at Google, leading several machine learning research teams focusing on neural modeling, reinforcement learning, dialog modeling, reliable/robust machine learning, and recommendation systems in the Google Brain team. His team has delivered significant improvements for YouTube, News, Ads, Google Play Store at Google with >420 product improvements since 2013. With 39 patents and >150 research articles, he is also known for research on user behavior in web and social media.


Prior to Google, he was the Area Manager and a Principal Scientist at Palo Alto Research Center's Augmented Social Cognition Group, where he led the team in understanding how social systems help groups of people to remember, think and reason. Ed completed his three degrees (B.S., M.S., and Ph.D.) in 6.5 years from University of Minnesota. Recognized as an ACM Distinguished Scientist and elected into the CHI Academy, he recently received a 20-year Test of Time award for research in information visualization. He has been featured and quoted in the press, including the Economist, Time Magazine, LA Times, and the Associated Press. An avid swimmer, photographer and snowboarder in his spare time, he also has a blackbelt in Taekwondo.


Towards Automated Construction of Compiler Optimizations

[]

Abstract

Compilers are the workhorse that bridge the gap between human readable and machine executable code. The diversity of modern programs, along with the advent of new and complex hardware architectures, has strained the capabilities of current compilers, making development and maintenance of automatic program optimizations in compilers exceedingly challenging. In spite of this, modern compiler optimizations are still constructed manually, expending considerable human effort.

In this talk, I will show how to construct compiler optimizations in a more automated manner using compiler auto-vectorization as an example. More specifically, I will show how to construct the three main components of a compiler optimization pass, namely, the transformation space, cost model and the optimization algorithm using both formal reasoning and machine learning.

First, I will demonstrate how to automatically generate a richer transformation space for compiler auto-vectorization using just the formal ISA semantics of a given hardware platform with the vectorizer generator Vegen. Vegen automatically supports newer vector instructions compared to manually-written vectorizers and outperforms production compilers on a number of important kernels.

Second, I will show how to automatically learn the optimization algorithm that selects the best transformation choice among valid vectorization opportunities exposed through the transformation space in Vemal. Vemal formulates the optimization problem as a sequential decision making process and shows that it is feasible to learn how to imitate a costly oracle.

Third, I will briefly present how to use data-driven techniques to learn compiler cost models with Ithemal. Ithemal automatically learns how to predict the throughput of basic blocks represented in a given ISA and the results show that it is significantly more accurate than state-of-the-art hand-written cost models used in production compilers.

The synergy in design of these three components will pave the way towards developing more automated means of constructing compiler optimizations that achieve state-of-the-art results.

Bio

Charith Mendis is a visiting researcher at Google Brain and will be joining the University of Illinois at Urbana-Champaign as an Assistant Professor in Computer Science from Fall 2021. His research interests include Compilers, Machine Learning for Systems and Program Analysis. He completed his Master’s degree at MIT for which he received the William A. Martin Thesis Prize and his bachelor’s degree at University of Moratuwa, Sri Lanka for which he received the institute Gold Medal. Charith was the recipient of the best student paper award at IEEE Big Data conference and the best paper award at ML for Systems workshop at ISCA. He has published work at both top programming language venues such as PLDI and OOPSLA as well as at top machine learning venues such as ICML and NeurIPS.


Leveraging ML to Improve the Design of Large-Scale Cloud Systems

[]

Abstract

Cloud services are increasingly adopting new programming models, such as microservices and serverless compute. While these frameworks offer several advantages, such as better modularity, ease of maintenance and deployment, they also introduce new hardware and software challenges.

In this talk, I will briefly discuss the challenges that these new cloud models introduce in hardware and software, and present some of our work on accelerating critical microservice computation, and on employing ML to improve the cloud’s performance predictability and resource efficiency. I will first discuss Seer, a performance debugging system that identifies root causes of unpredictable performance in multi-tier interactive microservices, and Sage, which improves on Seer by taking a completely unsupervised learning approach to data-driven performance debugging, making it both practical and scalable.

Bio

Christina Delimitrou is an Assistant Professor and the John and Norma Balen Sesquicentennial Faculty Fellow at Cornell University, where she works on computer architecture and computer systems. She specifically focuses on improving the performance predictability and resource efficiency of large-scale cloud infrastructures by revisiting the way these systems are designed and managed. Christina is the recipient of the 2020 TCCA Young Computer Architect Award, an Intel Rising Star Award, a Microsoft Research Faculty Fellowship, an NSF CAREER Award, a Sloan Research Scholarship, two Google Research Award, and a Facebook Faculty Research Award. Her work has also received 4 IEEE Micro Top Picks awards and several best paper awards. Before joining Cornell, Christina received her PhD from Stanford University. She had previously earned an MS also from Stanford, and a diploma in Electrical and Computer Engineering from the National Technical University of Athens. More information can be found at: http://www.csl.cornell.edu/~delimitrou/