Department of Computer Science and Engineering, School of Data and Sciences, BRAC University, 66 Mohakhali, Dhaka 1212, Bangladesh.
Faculty of Computer Sciences, Østfold University College, 1783 Halden, Norway.
Sensors (Basel). 2021 Feb 20;21(4):1468. doi: 10.3390/s21041468.
Autonomous vehicle navigation in an unknown dynamic environment is crucial for both supervised- and Reinforcement Learning-based autonomous maneuvering. The cooperative fusion of these two learning approaches has the potential to be an effective mechanism to tackle indefinite environmental dynamics. Most of the state-of-the-art autonomous vehicle navigation systems are trained on a specific mapped model with familiar environmental dynamics. However, this research focuses on the cooperative fusion of supervised and Reinforcement Learning technologies for autonomous navigation of land vehicles in a dynamic and unknown environment. The Faster R-CNN, a supervised learning approach, identifies the ambient environmental obstacles for untroubled maneuver of the autonomous vehicle. Whereas, the training policies of Double Deep Q-Learning, a Reinforcement Learning approach, enable the autonomous agent to learn effective navigation decisions form the dynamic environment. The proposed model is primarily tested in a gaming environment similar to the real-world. It exhibits the overall efficiency and effectiveness in the maneuver of autonomous land vehicles.
自主车辆在未知动态环境中的导航对于基于监督学习和强化学习的自主机动都至关重要。这两种学习方法的协同融合有可能成为应对不确定环境动态的有效机制。大多数最先进的自主车辆导航系统都是在具有熟悉环境动态的特定映射模型上进行训练的。然而,本研究专注于监督学习和强化学习技术的协同融合,以实现陆地车辆在动态和未知环境中的自主导航。Faster R-CNN 是一种监督学习方法,用于识别周围环境障碍物,以便自主车辆能够畅通无阻地行驶。而 Double Deep Q-Learning 是一种强化学习方法,其训练策略使自主代理能够从动态环境中学习到有效的导航决策。所提出的模型主要在类似于现实世界的游戏环境中进行测试。它在自主陆地车辆的机动中表现出了整体效率和有效性。