1st CVPR Workshop on
Dataset Distillation




Overview

In the past decade, deep learning has been mainly advanced by training increasingly large models on increasingly large datasets which comes with the price of massive computation and expensive devices for their training. As a result, research on designing state-of-the-art models gradually gets monopolized by large companies, while research groups with limited resources such as universities and small companies are unable to compete. Reducing the training dataset size while preserving model training effects is significant for reducing the training cost, enabling green AI, and encouraging the university research groups to engage in the latest research. 

This workshop focuses on the emerging research field of dataset distillation which aims to compress a large training dataset into a tiny informative one (e.g. 1% of the size of the original data) while maintaining the performance of models trained on this dataset. Besides general-purpose efficient model training, dataset distillation can also greatly facilitate downstream tasks such as neural architecture/hyperparameter search by speeding up model evaluation, continual learning by producing compact memory, federated learning by reducing data transmission, and privacy-preserving learning by removing data privacy. Dataset distillation is also closely related to research topics including core-set selection, prototype generation, active learning, few-shot learning, generative models, and a broad area of learning from synthetic data. 

Although DD has become an important paradigm in various machine-learning tasks, the potential of DD in computer vision (CV) applications, such as face recognition, person re-identification, and action recognition is far from being fully exploited. Moreover, DD has rarely been demonstrated effectively in advanced computer vision tasks such as object detection, image segmentation, and video understanding.

The purpose of this workshop is to unite researchers and professionals who share an interest in Dataset Distillation for computer vision for developing the next generation of dataset distillation methods for computer vision applications.

News

Feb. 14: We offer 3 free registration for the workshop for students

Feb. 12: The paper submission site will be opened soon.



Important Dates

Call for Papers

We invite papers related to Dataset Distillation and its related fields and applications. We ask for submissions along two tracks:

Accepted papers will be presented during a poster session and displayed on the workshop website.  A select few outstanding papers will also be offered an oral presentation

Topics

Potential topics may include, but are by no means limited to

For a comprehensive list of previous dataset distillation works, please see this Github repo: Guang000/Awesome-Dataset-Distillation 

Submission Instructions

Chairs

Saeed Vahidian

(Primary Contact)

 Duke University, USA

Yiran Chen

Duke University, USA

Bo Zhao, 

Beijing Academy of Artificial Intelligence, China

Ramin Hasani, 

MIT, USA

Alexander Amini, MIT, USA

Dongkuan (DK) Xu, North Carolina State University, USA

Ruochen Wang, UCLA, USA

Vyacheslav Kungurtsev, 

Czech Technical University, Czech Republic

Xinchao Wang, NUS, Singapore

Committee Members

George Cazenavette

MIT, USA

Justin Cui 

UCLA, USA

Jianyang Gu 

Zhejiang University, China

Noveen Sachdeva 

UC San Deigo,
Google DeepMind

Xindi Wu

Princeton University, USA

Invited Speakers

Olga Russakovsky, Princeton University

Cho-Jui Hsieh,

 University of California, Los Angeles


Furong Huang, University of Maryland, USA


Zhiwei Deng,
Google Research, USA


Hakan Bilen

University of Edinburgh, UK


Baharan Mirzasoleiman,

University of California, Los Angeles


Phillip Isola, 

MIT, USA


Program Schedule

TBD