East Ballroom A
Poster session in Exhibit hall #235-264
Transformers have recently emerged as promising and versatile deep neural architecture in various domains. Since the introduction of Vision Transformers (ViT) in 2020, the vision community has witnessed an explosion of transformer-based computer vision models with applications ranging from image classification to dense prediction (e.g., object detection, segmentation), video, self-supervised learning, 3D and multi-modal learning. This workshop presents a timely opportunity to bring together researchers across computer vision and machine learning communities to discuss the opportunities and open challenges in designing transformer models for vision tasks.
Microsoft
Columbia University
INRIA/Google
Meta AI
Meta AI
Columbia University
Google Research
NYU
ParisTech
We accept abstract submissions to our workshop. All submissions shall have maximally 4 pages (excluding references) following the CVPR 2023 author guidelines.
Abstract Submission Due: April 1st April 15th, 2023 (11:59pm PST)
Reviews Due: May 8th, 2023 (11:59pm PST)
Notification to Authors: May 15th 18th, 2023
Workshop: June 18th, 2023
Yilun Chen (CUHK)
De-An Huang (NVIDIA)
Chao Liu (NVIDIA)
Vaidehi Patil (IIT Bombay)
Zineng Tang (UNC)
Shoubin Yu (UNC)
Ziyang Wang (UNC)
Adyasha Maharana (UNC)
Yiming Li (NYU)
Haoxuan Wang (SJU)
Min-Hung Chen (NVIDIA)
Nishant Rai (Stanford)
Mingyu Ding (UC Berkeley)
Haotian Zhang (U. Washington)
Jim Fan (NVIDIA)
Xu Ma (Northeastern)
Shilong Liu (Tsinghua)
Feng Li (HKUST)
Hao Zhang (HKUST)
Xueyan Zou (UW Madison)
Subhashree Radhakrishnan (NVIDIA)
Andreas Steiner (Google)