Please use this identifier to cite or link to this item: http://hdl.handle.net/1942/35365
Title: How to identify dangerous overtaking manoeuvres based on real-time sensor data?
Authors: Nguyen, Trang Thanh Kieu
Advisors: BRIJS, Tom
ADNAN, Muhammad
Issue Date: 2021
Publisher: UHasselt
Abstract: Overtaking manoeuvre which involves both lateral and longitudinal control is considered as one of the most dangerous and complex manoeuvres that a driver can perform. In all types of road accidents related to overtaking manoeuvres, there is a risk of rear-end accidents that the overtaking vehicle no longer maintains the safe distance from the car ahead in preparation for overtaking. The research is an effort to contribute to the development of an Advanced Driver Assistance System which helps predict dangerous overtaking manoeuvres with respect to rear-end accidents before the headway between the driven and preceding vehicles reaching its critical threshold and alert the driver about these possible dangers, giving him enough time to react. A sensory-fusion deep learning architecture based on Recurrent Neural Networks (RNNs) with Long Short-Term Memory (LSTM) units is proposed to monitor vehicle dynamics and driving context and signal predictions. The prediction performance of regular LSTM-RNN and bidirectional LSTM-RNN are also compared with other models based on Feedforward neural network (FFNN). In terms of experiment settings, the model is trained in simulation with driving scenarios on two-lane rural roads but is tested in natural freeway and city driving.
Notes: Master of Transportation Sciences-Traffic Safety
Document URI: http://hdl.handle.net/1942/35365
Category: T2
Type: Theses and Dissertations
Appears in Collections:Master theses

Files in This Item:
File Description SizeFormat 
af7cb238-255b-45fb-bcde-82c83727bb03.pdf3.95 MBAdobe PDFView/Open
Show full item record

Page view(s)

50
checked on Jul 2, 2022

Download(s)

28
checked on Jul 2, 2022

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.