Distracted driving is a leading cause of accidents worldwide. The tasks of distraction detection and recognition have been traditionally addressed as computer vision problems. However, distracted behaviors are not always expressed in a visually observable way. In this project, we introduce a novel multimodal dataset of distracted driver behaviors, consisting of data collected using twelve information channels coming from visual, acoustic, near-infrared, thermal, physiological, and linguistic modalities. The data were collected from 45 subjects while being exposed to four different distractions (three cognitive and one physical). Data collection experiments took place during different times of the day to account for a greater drowsiness variability in our participants. Through the analysis of the data, we explore a variety of spatio-temporal machine learning techniques, ranging from traditional machine learning pipelines to more novel, state-of-the-art deep learning architectures. We identify modality-specific and multi-modal supervised and unsupervised features and we evaluate their ability to characterize the two different conditions within and across subjects. This project aims to explore how drowsy and distracted states overlap in drivers and how they can affect driver performance and safety. Our goal is to design personalized and adjustable AI models that learn through interaction and fit each individual's needs in order to increase road safety and offer just-in-time interventions. This project is in collaboration with the Toyota Research Institute.
Autonomous vehicles represent one of the most active technologies currently being developed, with research areas addressing, among others, the modeling of the states and behavioral elements of the occupants. This project contributes to this line of research by studying the circadian rhythm of individuals using a novel multimodal dataset of 36 subjects consisting of five information channels. These channels include visual, thermal, physiological, linguistic, and background data. Moreover, we propose a framework to explore whether the circadian rhythm can be modeled without continuous monitoring and investigate the hypothesis that multimodal features have a greater propensity for improved performance using data points specific to certain times during the day. We analyze and evaluate our approach using demographic and behavioral features produced via a series of surveys. Our goal is to provide a novel approach for future research and open the possibility of the integration of unrestrictive sensors in future automobiles. This project is in collaboration with the Ford Motor Company.
Much research is currently carried out with a focus on autonomous vehicles; research is starting to focus on areas such as the modeling of occupant states and behavioral elements. This project contributes to this line of research by developing a pipeline that extracts physiological signals from thermal imagery and models the occupant's physical and mental state at any given point in time using a fully non-contact based approach. In addition, this provides the opportunity to move towards an implementable technology in autonomous vehicles that does not rely on uncomfortable, restrictive contact-based sensors.