Modality Agnostic Road Traffic Monitoring

NYU Center for Urban Science and Progress 2021 Capstone Project

Here We Provide...


  • Able to work with either and both audio and image.

  • Mitigates the limitations of poor visibility or noisy environments.


  • Provides different illumination and weather conditions.

  • Contains various car types, traffic volumes and camera angles.

The Problem We Want to Solve

State-of-the-art machine learning models of traffic monitoring focus on single-modal methods using only video, video-image or audio resources. However, audio analysis can only provide general classification of traffic conditions and is not reliable during dense traffic volumes and when certain events, such as helicopters passing by; video-image analysis faces challenges of poor illumination situations and inclement weather conditions. Besides the limitations of these methods, due to technical issues, one of the modalities could be missing or not informative for a specific period of time in a realistic scenario.

Thus, when the cameras or the microphones on the roads meet technical issues or bad environmental conditions, how can we estimate traffic conditions?

Our Team Members

Eve Shi

Software Engineer, Center for Urban Science & Progress, NYU

Linkedin Email

Yao Hou

Center for Urban Science & Progress, NYU

Email


Our Mentors

Magdalena Fuentes

PhD, Provost’s Postdoctoral Fellow

Music and Audio Research Laboratory, NYU


GitHub

Bea Steers

Research Scientist, Music and Audio Research Laboratory, NYU

GitHub

Our Sponsors