Park Jongcheon, Han Seungyong, Lee S M
Cyber Physical Systems & Control Laboratory, School of Electronic and Electrical Engineering, Kyungpook National University, Daehak-ro 80, Republic of Korea.
Cyber Physical Systems & Control Laboratory, School of Electronic and Electrical Engineering, Kyungpook National University, Daehak-ro 80, Republic of Korea.
ISA Trans. 2022 Oct;129(Pt B):684-690. doi: 10.1016/j.isatra.2022.02.041. Epub 2022 Mar 7.
In this paper, a new imitation learning algorithm is proposed based on the Restored Action Generative Adversarial Imitation Learning (RAGAIL) from observation. An action policy is trained to move a robot manipulator similar to a demonstrator's behavior by using the restored action from state-only demonstration. To imitate the demonstrator, the trajectory is generated by Recurrent Generative Adversarial Networks (RGAN), and the action is restored from the output of the tracking controller constructed by the state and the generated target trajectory. The proposed imitation learning algorithm is not required to access the demonstrator's action (internal control signal such as force/torque command) and provides better learning performances. The effectiveness of the proposed method is validated through the experimental results of the robot manipulator.
本文基于从观测中恢复的动作生成对抗模仿学习(RAGAIL)提出了一种新的模仿学习算法。通过使用仅基于状态的演示中恢复的动作,训练一个动作策略,使机器人操纵器的动作类似于演示者的行为。为了模仿演示者,由循环生成对抗网络(RGAN)生成轨迹,并根据状态和生成的目标轨迹构建的跟踪控制器的输出恢复动作。所提出的模仿学习算法不需要访问演示者的动作(如力/扭矩命令等内部控制信号),并具有更好的学习性能。通过机器人操纵器的实验结果验证了所提方法的有效性。