Zeroth-Order Machine Learning: Fundamental Principles and Emerging Applications in Foundation Models

AAAI 2024 Tutorial

February 20, 2024 (2:00 pm – 6:00 pm PST)

Room 118

Overview

With the swift progression of artificial intelligence, driven notably by the rise of foundation models (FMs) (e.g., LLMs), a plethora of fresh opportunities and challenges have arisen in the evolution of the next generation of ML algorithms.   While the auto-differentiation-based first-order (FO) optimizers, e.g., SGD and Adam, have been the predominant choices for model training and fine-tuning, increasing scenarios have emerged, where obtaining FO gradient information becomes infeasible or computationally prohibitive. For example, as LLMs continue to scale, they encounter significant memory overhead due to the back-propagation (BP), and advancements in addressing this challenge could also facilitate technological breakthroughs in related areas, such as on-device training. Similarly, a significant and recent challenge is the problem of prompt learning for foundation-model-as-a-service, exemplified by platforms like ChatGPT, where directly obtaining the FO gradients is impossible due to its black-box setting. Such challenges are also prevalent in numerous applications in AI for science (AI4S), where ML models might interact with non-differentiable or black-box simulators/experiments with analytically defined learning objectives. 

In stark contrast, the utilization of gradient-free zeroth-order (ZO) optimization techniques emerges as a viable approach for LLM fine-tuning, exhibiting an exceptional degree of memory efficiency and broader applicability for various black-box challenges. These instances underscore the imperative nature of exploring alternative avenues to FO-based machine learning algorithms in the current era of FMs. In this tutorial, the following aspects will be covered.

Table of Contents

Speakers

Michigan State University

IBM Research


DAMO Academy, Alibaba USA

Other Contributors

Graduate Student

Michigan State University

Graduate Student

UNC

Postdoc Fellow

UT Austin

Postdoc Fellow

UMN, Twin Cities

Graduate Student

Michigan State University

Graduate Student

UT Austin

Assistant Professor

UNC

Associate Professor

UT Austin

Associate Professor

UMN Twin Cities

Questions?

Contact Yihua Zhang [zhan1908@msu.edu] to get more information on the tutorial!