Restored Action Generative Adversarial Imitation Learning from observation for robot manipulator

ISA Trans. 2022 Oct;129(Pt B):684-690. doi: 10.1016/j.isatra.2022.02.041. Epub 2022 Mar 7.

Abstract

In this paper, a new imitation learning algorithm is proposed based on the Restored Action Generative Adversarial Imitation Learning (RAGAIL) from observation. An action policy is trained to move a robot manipulator similar to a demonstrator's behavior by using the restored action from state-only demonstration. To imitate the demonstrator, the trajectory is generated by Recurrent Generative Adversarial Networks (RGAN), and the action is restored from the output of the tracking controller constructed by the state and the generated target trajectory. The proposed imitation learning algorithm is not required to access the demonstrator's action (internal control signal such as force/torque command) and provides better learning performances. The effectiveness of the proposed method is validated through the experimental results of the robot manipulator.

Keywords: Imitation learning; Imitation learning from observation; Manipulator; Restored Action Generative Adversarial Imitation Learning.