International Journal of Control, Automation, and Systems 2023; 21(12): 4098-4110
https://doi.org/10.1007/s12555-023-0255-4
© The International Journal of Control, Automation, and Systems
Safe decision-making strategy of autonomous vehicles (AVs) plays a critical role in avoiding accidents. This study develops a safe reinforcement learning (safe-RL)-based driving policy for AVs on highways. The hierarchical framework is considered for the proposed safe-RL, where an upper layer executes a safe explorationexploitation by modifying the exploring process of the ε-greedy algorithm, and a lower layer utilizes a finite state machine (FSM) approach to establish the safe conditions for state transitions. The proposed safe-RL-based driving policy improves the vehicle’s safe driving ability using a Q-table that stores the values corresponding to each action state. Moreover, owing to the trade-off between the ε-greedy values and safe distance threshold, the simulation results demonstrate the superior performance of the proposed approach compared to other alternative RL approaches, such as the ε-greedy Q-learning (GQL) and decaying ε-greedy Q-learning (DGQL), in an uncertain traffic environment. This study’s contributions are twofold: it improves the autonomous vehicle’s exploration-exploitation and safe driving ability while utilizing the advantages of FSM when surrounding cars are inside safe-driving zones, and it analyzes the impact of safe-RL parameters in exploring the environment safely.
Keywords Autonomous vehicles, collision avoidance, decision-making, finite state machine, safe reinforcement learning.
International Journal of Control, Automation, and Systems 2023; 21(12): 4098-4110
Published online December 1, 2023 https://doi.org/10.1007/s12555-023-0255-4
Copyright © The International Journal of Control, Automation, and Systems.
Hung Duy Nguyen and Kyoungseok Han*
Kyungpook National University
Safe decision-making strategy of autonomous vehicles (AVs) plays a critical role in avoiding accidents. This study develops a safe reinforcement learning (safe-RL)-based driving policy for AVs on highways. The hierarchical framework is considered for the proposed safe-RL, where an upper layer executes a safe explorationexploitation by modifying the exploring process of the ε-greedy algorithm, and a lower layer utilizes a finite state machine (FSM) approach to establish the safe conditions for state transitions. The proposed safe-RL-based driving policy improves the vehicle’s safe driving ability using a Q-table that stores the values corresponding to each action state. Moreover, owing to the trade-off between the ε-greedy values and safe distance threshold, the simulation results demonstrate the superior performance of the proposed approach compared to other alternative RL approaches, such as the ε-greedy Q-learning (GQL) and decaying ε-greedy Q-learning (DGQL), in an uncertain traffic environment. This study’s contributions are twofold: it improves the autonomous vehicle’s exploration-exploitation and safe driving ability while utilizing the advantages of FSM when surrounding cars are inside safe-driving zones, and it analyzes the impact of safe-RL parameters in exploring the environment safely.
Keywords: Autonomous vehicles, collision avoidance, decision-making, finite state machine, safe reinforcement learning.
Vol. 22, No. 12, pp. 3545~3811
Jongsoo Lee, Jonghyeok Park, and Soohee Han*
International Journal of Control, Automation, and Systems 2024; 22(11): 3266-3274Joosun Lee, Taeyhang Lim, and Wansoo Kim*
International Journal of Control, Automation, and Systems 2024; 22(7): 2263-2272Guozhu Zhu, Hao Jie, Zekai Zheng, and Weirong Hong*
International Journal of Control, Automation, and Systems 2024; 22(5): 1666-1679