CVPR 2019

Workshop on Autonomous Driving

June 17th | Long Beach, USA

Note: this is the website of the Workshop on Autonomous Driving as listed in the CVPR program. It is different from the workshop on Autonomous Driving Beyond Single-Frame Perception which is hosted at wad.ai. We apologize for the confusion.

Introduction

The CVPR 2019 Workshop on Autonomous Driving (WAD) aims to gather researchers and engineers from academia and industry to discuss the latest advances in perception for autonomous driving. In this one-day workshop, we will have regular paper presentations, invited speakers, and technical benchmark challenges to present the current state of the art, as well as the limitations and future directions for computer vision in autonomous driving, arguably the most promising application of computer vision and AI in general. The previous chapters of the workshop at CVPR attracted hundreds of researchers to attend. This year, multiple industry sponsors also join our organizing efforts to push its success to a new level.

News and Updates

Participation

Paper Submission

We solicit paper submissions on novel methods and application scenarios of CV for Autonomous vehicles. We accept papers on a variety of topics, including autonomous navigation and exploration, ADAS, UAV, deep learning, calibration, SLAM, etc.. Papers will be peer reviewed under double-blind policy and the submission deadline is 20th March 2019. Accepted papers will be presented at the poster session, some as orals and one paper will be awarded as the best paper.

Challenge Track

We host a challenge to understand the current status of computer vision algorithms in solving the environmental perception problems for autonomous driving. We have prepared a number of large scale datasets with fine annotation, collected and annotated by Berkeley Deep Driving Consortium, nuTonomy and DiDi. Based on the datasets, we have define a set of four realistic problems and encourage new algorithms and pipelines to be invented for autonomous driving. More specifically, they are nuScenes challenge (nuScenes 3D detection challenge), Berkeley Deep Drive challenge (Object tracking and instance segmentation challenges), and DiDi (D²-City Detection Transfer Learning Challenge, D²-City Intelligent Annotation Teaser Challenge).

Invited Speakers

Trevor Darrell

UC Berkeley

Kurt Keutzer

UC Berkeley

Kilian Q. Weinberger

Cornell University

Raquel Urtasun

UofT & Uber

Edwin Olson

University of Michigan

Zhaoyin JIA

Didi Chuxing

Bo Li

UIUC

Manmohan Chandraker

UCSD & NEC Labs

Alex Kendall

Cambridge & Wayve

Diamond sponsors