Workshop location: 209 A-C
Poster session location: ExHall D
13:00 - 17:00 CDT
Transformers have recently emerged as promising and versatile deep neural architecture in various domains. Since the introduction of Vision Transformers (ViT) in 2020, the vision community has witnessed an explosion of transformer-based computer vision models with applications ranging from image classification to dense prediction (e.g., object detection, segmentation), video, self-supervised learning, 3D and multi-modal learning. This workshop presents a timely opportunity to bring together researchers across computer vision and machine learning communities to discuss the opportunities and open challenges in designing transformer models for vision tasks.
We accept abstract submissions to our workshop. All submissions shall have maximally 4 pages (excluding references) following the CVPR 2025 author guidelines.
CMT: https://cmt3.research.microsoft.com/T4V2025
Submission Deadline: April 18th, 2025 (11:59pm PST)
Paper Review Due: May 20th, 2025 (11:59pm PST)
Notification of Acceptance: May 25th, 2025
Camera-Ready Submission Deadline: June 8th, 2025
Workshop Date: June 12th, 2025
Shiyi Lan (NVIDIA) Xiaolong Li (NVIDIA) Subhashree Radhakrishnan (NVIDIA)
Ryo Hachiuma (NVIDIA) Hao Zhang (NVIDIA) Huck Yang (NVIDIA)
Yan-Bo Lin (UNC) Yi-Lin Sung (UNC) Jaemin Cho (UNC)
Ziyang Wang (UNC) Ce Zhang (UNC) Yue Yang (UNC)
Feng Cheng (ByteDance) Shilong Liu (Tsinghua Univ.) Fuxiao Liu (UMD)
Minkyoung Cho (UMich) Krishna Kanth Nakka (Huawei)