School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China.
College of Mechanical and Electrical Engineering, Shihezi University, Shihezi 832003, China.
Sensors (Basel). 2022 Jun 7;22(12):4316. doi: 10.3390/s22124316.
Agricultural robots are one of the important means to promote agricultural modernization and improve agricultural efficiency. With the development of artificial intelligence technology and the maturity of Internet of Things (IoT) technology, people put forward higher requirements for the intelligence of robots. Agricultural robots must have intelligent control functions in agricultural scenarios and be able to autonomously decide paths to complete agricultural tasks. In response to this requirement, this paper proposes a Residual-like Soft Actor Critic (R-SAC) algorithm for agricultural scenarios to realize safe obstacle avoidance and intelligent path planning of robots. In addition, in order to alleviate the time-consuming problem of exploration process of reinforcement learning, this paper proposes an offline expert experience pre-training method, which improves the training efficiency of reinforcement learning. Moreover, this paper optimizes the reward mechanism of the algorithm by using multi-step TD-error, which solves the probable dilemma during training. Experiments verify that our proposed method has stable performance in both static and dynamic obstacle environments, and is superior to other reinforcement learning algorithms. It is a stable and efficient path planning method and has visible application potential in agricultural robots.
农业机器人是推动农业现代化、提高农业效率的重要手段之一。随着人工智能技术的发展和物联网(IoT)技术的成熟,人们对机器人的智能化提出了更高的要求。农业机器人在农业场景中必须具有智能控制功能,能够自主决定路径来完成农业任务。针对这一要求,本文提出了一种用于农业场景的残差式软动作控制器(R-SAC)算法,以实现机器人的安全避障和智能路径规划。此外,为了缓解强化学习探索过程中的耗时问题,本文提出了一种离线专家经验预训练方法,提高了强化学习的训练效率。并且,通过使用多步 TD 误差来优化算法的奖励机制,解决了训练过程中的可能困境。实验验证了我们提出的方法在静态和动态障碍物环境中都具有稳定的性能,并且优于其他强化学习算法。它是一种稳定且高效的路径规划方法,在农业机器人中有明显的应用潜力。