The ROAD-R dataset

About the Dataset


ROAD-R is an extension of the ROAD dataset with a set of 243 manually annotated requirements over the 41 labels grouped into agents, actions and locations. The requirements are logical constraints provided in disjunctive normal form and express background knowledge applicable in autonomous driving scenarios, such as:


The dataset contains 22 videos, each about 8 minutes long, and is split into training (15 videos), validation (3 videos) and test (4 videos) partitions. Each video is annotated in terms of what we call road events (REs), as seen from the point of view of the autonomous vehicle capturing the video (Ego-vehicle). Road events are defined as a series of bounding boxes linked in time annotated with:

1. the label associated with the agent (e.g., “Pedestrian”),

2. the action(s) the agent is doing (e.g., “Pushing Object”, “Moving Away”), and

3. the location(s) where the agent is placed (e.g., “Right Pavement”, “Bus Stop”).



ROAD-R is part of a series of developments upon the ROAD dataset:


The dataset was designed according to the following principles:
  • A multi-label benchmark: each road event is composed of the label of the (moving) agent responsible, the label(s) of the type of action(s) being performed, and labels describing where the action is located. Each event can be assigned multiple instances of the same label type whenever relevant (e.g., an RE can be an instance of both moving away and turning left). 
  • The labelling is done from the point of view of the AV (ego-vehicle): the final goal is for the autonomous vehicle to use this information to make the appropriate decisions. The meta-data is intended to contain all the information required to fully describe a road scenario. After closing one’s eyes, the set of labels associated with the current video frame should be sufficient to recreate the road situation in one’s head (or, equivalently, sufficient for the AV to be able to make a decision). 
  • ROAD-R allows one to validate a manifold tasks associated with situation awareness for self-driving: (i) agent detection, (ii) action detection, and (iii) location detection.
  • The requirements of ROAD-R provide the ground for developing safer autonomous vehicles.

Demo

Main features


Download


To download ROAD-R dataset, please follow the instructions below:

a. requirements_dimacs.txt contains the requirements written in DIMACS format. Here, each label is represented as a number.

b. requirements_readable.txt contains the requirements written in a human-understandable format.

road-r

├── road_trainval_v1.0.json

├── requirements_dimacs.txt 

├── requirements_readable.txt 

├── videos

    ├── 2014-06-25-16-45-34_stereo_centre_02.mp4

    ├── 2014-06-26-09-53-12_stereo_centre_02.mp4

    ├── ...

├── rgb-images

    ├── 2014-06-25-16-45-34_stereo_centre_02

        ├── 00001.jpg

        ├── 00002.jpg

        ├── ...

    ├── 2014-06-26-09-53-12_stereo_centre_02

        ├── 00001.jpg

        ├── 00002.jpg

        ├── ...

    ├── ...

        


Annotation structure

The annotations for the train and validation split are saved in single json file named road_trainval_v1.0.json, which is located under root directory of the dataset as it can be seen above.

The first level of road_trainval_v1.0.json contains dataset level information like classes of each label type:


Finally, the db field contains all frame and tube level annotations for all the videos:

"annos": {"b19309": { "box": [0.34245960502692996,

                                0.423444976076555,

                                0.3631059245960503,

                                0.5179425837320574],

             "agent_ids": [0],

                      "loc_ids": [6],

                      "action_ids": [4],

                      "duplex_ids": [1],

                      "triplet_ids": [18],

                      "tube_uid": "bbef3659"

    }

"b433085": { "box": [0.5741350906095553,

                                0.44216691068814057,

                                0.58974519408777,

                                0.5230057739861901],

             "agent_ids": [1],

                      "loc_ids": [0],

                      "action_ids": [9, 12],

                      "duplex_ids": [26, 31],

                      "triplet_ids": [315, 329],

                      "tube_uid": "51522791"

    }

  }