Reinforcement learning based compensation methods for robot manipulators

Journal Article (2019)
Author(s)

Yudha P. Pane (Katholieke Universiteit Leuven)

Subramanya P. Nageshrao (Ford Motor Company)

Jens Kober (TU Delft - Learning & Autonomous Control)

Robert Babuska (TU Delft - Learning & Autonomous Control)

Research Group
Learning & Autonomous Control
Copyright
© 2019 Yudha P. Pane, Subramanya P. Nageshrao, J. Kober, R. Babuska
DOI related publication
https://doi.org/10.1016/j.engappai.2018.11.006
More Info
expand_more
Publication Year
2019
Language
English
Copyright
© 2019 Yudha P. Pane, Subramanya P. Nageshrao, J. Kober, R. Babuska
Research Group
Learning & Autonomous Control
Volume number
78
Pages (from-to)
236-247
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Smart robotics will be a core feature while migrating from Industry 3.0 (i.e., mass manufacturing) to Industry 4.0 (i.e., customized or social manufacturing). A key characteristic of a smart system is its ability to learn. For smart manufacturing, this means incorporating learning capabilities into the current fixed, repetitive, task-oriented industrial manipulators, thus rendering them ‘smart’. In this paper we introduce two reinforcement learning (RL) based compensation methods. The learned correction signal, which compensates for unmodeled aberrations, is added to the existing nominal input with an objective to enhance the control performance. The proposed learning algorithms are evaluated on a 6-DoF industrial robotic manipulator arm to follow different kinds of reference paths, such as square or a circular path, or to track a trajectory on a three dimensional surface. In an extensive experimental study we compare the performance of our learning-based methods with well-known tracking controllers, namely, proportional-derivative (PD), model predictive control (MPC), and iterative learning control (ILC). The experimental results show a considerable performance improvement thanks to our RL-based methods when compared to PD, MPC, and ILC.

Files

1_s2.0_S0952197618302446_main.... (pdf)
(pdf | 1.74 Mb)
- Embargo expired in 13-06-2019
License info not available