Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/184552
Title: | Human-guided deep reinforcement learning for optimal decision making of autonomous vehicles | Authors: | Wu, Jingda Yang, Haohan Yang, Lie Huang, Yi He, Xiangkun Lv, Chen |
Keywords: | Engineering | Issue Date: | 2024 | Source: | Wu, J., Yang, H., Yang, L., Huang, Y., He, X. & Lv, C. (2024). Human-guided deep reinforcement learning for optimal decision making of autonomous vehicles. IEEE Transactions On Systems, Man, and Cybernetics: Systems, 54(11), 6595-6609. https://dx.doi.org/10.1109/TSMC.2024.3384992 | Project: | A2084c0156 M22K2c0079 NRF2021-NRF-ANR003 (HM Science) MOET2EP50222-0002 |
Journal: | IEEE Transactions on Systems, Man, and Cybernetics: Systems | Abstract: | Although deep reinforcement learning (DRL) methods are promising for making behavioral decisions in autonomous vehicles (AVs), their low training efficiency and difficulty to adapt to untrained cases hinder their applications. Introducing a human role in the DRL paradigm could improve training efficiency by using human prior knowledge and overcome untrained cases in deployment by online human takeover. In this study, a novel value-based DRL algorithm that leverages human guidance to improve its performance is proposed for addressing high-level decision-making problems in autonomous driving. We develop a new learning objective for DRL to increase the value of the human policy over the undertrained DRL policy so that the DRL agent can be encouraged to mimic human behaviors and thereby utilizing human guidance more efficiently. Our method can autonomously evaluate the importance of different human guidance, which makes it more robust for variation of human performance. The proposed DRL algorithm was used to address a challenging multiobjective lane-change decision-making problem. We collected human guidance from a human-in-the-loop driving experiment and evaluated our method in a high-fidelity simulator. Results validated the advantages of the proposed algorithm in terms of training efficiency and optimality in the decision-making problem compared to the baselines of state-of-the-art existing methods. Results also revealed the favorable fine-tuning ability of the proposed algorithm, which is promising for addressing the long-tail issue in DRL-based autonomous driving. Our methodology does not introduce additional domain knowledge so that it can be seamlessly applied to other similar issues. The supplementary video is available at https://youtu.be/Ec7WkqeLsB8. | URI: | https://hdl.handle.net/10356/184552 | ISSN: | 2168-2216 | DOI: | 10.1109/TSMC.2024.3384992 | Schools: | School of Mechanical and Aerospace Engineering | Rights: | © 2024 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/TSMC.2024.3384992. | Fulltext Permission: | open | Fulltext Availability: | With Fulltext |
Appears in Collections: | MAE Journal Articles |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Manuscript clear.pdf | 1.61 MB | Adobe PDF | ![]() View/Open |
Page view(s)
21
Updated on May 7, 2025
Download(s)
2
Updated on May 7, 2025
Google ScholarTM
Check
Altmetric
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.