- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- Leveraging spatio-temporal features to forecast time-to-accident
Open Collections
UBC Theses and Dissertations
UBC Theses and Dissertations
Leveraging spatio-temporal features to forecast time-to-accident Anjum, Taif
Abstract
Globally, traffic accidents are the leading cause of death, according to reports published by the World Health Organization (WHO). Advanced Driver Assistance Systems (ADAS) effectively reduce the likelihood or severity of accidents. Time-To-Accident (TTA) forecasting is a critical component of ADAS; it can improve decision-making in traffic, dynamic path planning, and alert drivers to potential dangers. TTA is defined as the interval between the threat of an accident and the occurrence of the accident itself. Sensor or depth imaging-based TTA forecasting can be inaccurate and unaffordable to the mass. Existing vision-based traffic safety research in the literature focuses on accident anticipation with limited work on TTA prediction. We propose forecasting TTA by utilizing spatio-temporal features from dashboard camera videos and introducing a novel algorithm to compute ground truth TTA values. Additionally, we present two efficient architectures, a Light 3D Convolutional Neural Network (Li3D) and Hybrid CNN-Transformer (HyCT) network, for our application. Li3D uses significantly fewer parameters and larger kernels for efficient feature learning without sacrificing accuracy. We considered transfer learning to improve the model's performance further but faced limitations due to a lack of relevant large-scale video datasets and Li3D's input requirements. Thus we propose the HyCT network which is trained via transfer learning. HyCT outperformed a comprehensive list of networks, including Li3D on the Car Crash Dataset (CCD). Furthermore, with our approach, we can simultaneously forecast TTA and recognize accident and non-accident scenes. In real-world scenarios, it is vital to identify when a normal driving scene transitions into a potential accident scene. We assess our approach’s ongoing scene recognition ability using the TTAASiST@x (Time-to-Accident-based Accuracy at the xth frame Since Scene Transition) metric. With just 25% of the data, HyCT achieved over 85% accuracy in recognizing accident and non-accident scenes, further showcasing its performance on limited data. Our best-performing models can predict TTA values with an average prediction error of 0.25 and 0.79 seconds on the CCD and DoTA dataset, respectively.
Item Metadata
Title |
Leveraging spatio-temporal features to forecast time-to-accident
|
Creator | |
Supervisor | |
Publisher |
University of British Columbia
|
Date Issued |
2023
|
Description |
Globally, traffic accidents are the leading cause of death, according to reports published by the World Health Organization (WHO). Advanced Driver Assistance Systems (ADAS) effectively reduce the likelihood or severity of accidents. Time-To-Accident (TTA) forecasting is a critical component of ADAS; it can improve decision-making in traffic, dynamic path planning, and alert drivers to potential dangers. TTA is defined as the interval between the threat of an accident and the occurrence of the accident itself. Sensor or depth imaging-based TTA forecasting can be inaccurate and unaffordable to the mass. Existing vision-based traffic safety research in the literature focuses on accident anticipation with limited work on TTA prediction. We propose forecasting TTA by utilizing spatio-temporal features from dashboard camera videos and introducing a novel algorithm to compute ground truth TTA values.
Additionally, we present two efficient architectures, a Light 3D Convolutional Neural Network (Li3D) and Hybrid CNN-Transformer (HyCT) network, for our application. Li3D uses significantly fewer parameters and larger kernels for efficient feature learning without sacrificing accuracy. We considered transfer learning to improve the model's performance further but faced limitations due to a lack of relevant large-scale video datasets and Li3D's input requirements. Thus we propose the HyCT network which is trained via transfer learning. HyCT outperformed a comprehensive list of networks, including Li3D on the Car Crash Dataset (CCD). Furthermore, with our approach, we can simultaneously forecast TTA and recognize accident and non-accident scenes. In real-world scenarios, it is vital to identify when a normal driving scene transitions into a potential accident scene. We assess our approach’s ongoing scene recognition ability using the TTAASiST@x (Time-to-Accident-based Accuracy at the xth frame Since Scene Transition) metric. With just 25% of the data, HyCT achieved over 85% accuracy in recognizing accident and non-accident scenes, further showcasing its performance on limited data. Our best-performing models can predict TTA values with an average prediction error of 0.25 and 0.79 seconds on the CCD and DoTA dataset, respectively.
|
Genre | |
Type | |
Language |
eng
|
Date Available |
2023-07-12
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
Attribution-NonCommercial-NoDerivatives 4.0 International
|
DOI |
10.14288/1.0434193
|
URI | |
Degree | |
Program | |
Affiliation | |
Degree Grantor |
University of British Columbia
|
Graduation Date |
2023-09
|
Campus | |
Scholarly Level |
Graduate
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International