Show simple item record

dc.contributor.authorMuzahid, Abu Jafar Md
dc.contributor.authorKamarulzaman, Syafiq Fauzi
dc.contributor.authorRahman, Md Arafatur
dc.contributor.authorAlenezi, Ali H.
dc.date.accessioned2022-06-30T08:38:10Z
dc.date.available2022-06-30T08:38:10Z
dc.date.issued2022-04-18
dc.identifier.citationMuzahid, A.J.M., Kamarulzaman, S.F., Rahman, M.A. and Alenezi, A.H. (2022) Deep reinforcement learning-based driving strategy for avoidance of chain collisions and its safety efficiency analysis in autonomous vehicles. IEEE Access, 10, pp. 43303-43319.en
dc.identifier.issn2169-3536en
dc.identifier.doi10.1109/ACCESS.2022.3167812en
dc.identifier.urihttp://hdl.handle.net/2436/624811
dc.description© 2022 The Authors. Published by IEEE. This is an open access article available under a Creative Commons licence. The published version can be accessed at the following link on the publisher’s website: https://ieeexplore.ieee.org/document/9758806en
dc.description.abstractVehicle control in autonomous traffic flow is often handled using the best decision-making reinforcement learning methods. However, unexpected critical situations make the collisions more severe and, consequently, the chain collisions. In this work, we first review the leading causes of chain collisions and their subsequent chain events, which might provide an indication of how to prevent and mitigate the crash severity of chain collisions. Then, we consider the problem of chain collision avoidance as a Markov Decision Process problem in order to propose a reinforcement learning-based decision-making strategy and analyse the safety efficiency of existing methods in driving security. To address this, A reward function is being developed to deal with the challenge of multiple vehicle collision avoidance. A perception network structure based on formation and on actor-critic methodologies is employed to enhance the decision-making process. Finally, in the safety efficiency analysis phase, we investigated the safety efficiency performance of the agent vehicle in both single-agent and multi-agent autonomous driving environments. Three state-of-the-art contemporary actor-critic algorithms are used to create an extensive simulation in Unity3D. Moreover, to demonstrate the accuracy of the safety efficiency analysis, multiple training runs of the neural networks in respect of training performance, speed of training, success rate, and stability of rewards with a trade-off between exploitation and exploration during training are presented. Two aspects (single-agent and multi-agent) have assessed the efficiency of algorithms. Every aspect has been analyzed regarding the traffic flows: (1) the controlling efficiency of unexpected traffic situations by the sudden slowdown, (2) abrupt lane change, and (3) smoothly reaching the destination. All the findings of the analysis are intended to shed insight on the benefits of a greater, more reliable autonomous traffic set-up for academics and policymakers, and also to pave the way for the actual carry-out of a driver-less traffic world.en
dc.description.sponsorshipThis work was supported in part by the Ministry of Higher Education of Malaysia through the Fundamental Research Grant Scheme under Grant FRGS/1/2018/TK08/UMP/02/2; and in part by the Deputyship for Research & Innovation, Ministry of Education in Saudi Arabia under Project IF-2020-NBU-418.en
dc.formatapplication/pdfen
dc.language.isoenen
dc.publisherIEEEen
dc.relation.urlhttps://ieeexplore.ieee.org/document/9758806en
dc.subjectautonomous vehiclesen
dc.subjectdeep reinforcement learning methoden
dc.subjectreward functionen
dc.subjectchain collision avoidanceen
dc.subjectautonomous traffic flowen
dc.subjectsafety efficiency analysisen
dc.subjectsafetyen
dc.subjectcollision avoidanceen
dc.subjectlong short term memoryen
dc.subjecttrainingen
dc.subjectfeature extractionen
dc.subjectdecision-makingen
dc.subjectuncertaintyen
dc.titleDeep reinforcement learning-based driving strategy for avoidance of chain collisions and its safety efficiency analysis in autonomous vehiclesen
dc.typeJournal articleen
dc.identifier.eissn2169-3536
dc.identifier.journalIEEE Accessen
dc.date.updated2022-06-29T10:30:25Z
dc.date.accepted2022-04-12
rioxxterms.funderMinistry of Higher Education of Malaysia, Deputyship for Research & Innovation, Ministry of Education in Saudi Arabiaen
rioxxterms.identifier.projectFRGS/1/2018/TK08/UMP/02/2, IF-2020-NBU-418en
rioxxterms.versionVoRen
rioxxterms.licenseref.urihttps://creativecommons.org/licenses/by/4.0/en
rioxxterms.licenseref.startdate2022-06-30en
dc.source.volume10
dc.source.beginpage43303
dc.source.endpage43319
dc.description.versionPublished version
refterms.dateFCD2022-06-30T08:37:46Z
refterms.versionFCDVoR
refterms.dateFOA2022-06-30T08:38:11Z


Files in this item

Thumbnail
Name:
Muzahid_et_al_Deep_Reinforceme ...
Size:
2.316Mb
Format:
PDF

This item appears in the following Collection(s)

Show simple item record

https://creativecommons.org/licenses/by/4.0/
Except where otherwise noted, this item's license is described as https://creativecommons.org/licenses/by/4.0/