TY - JOUR
T1 - Shared Control of Robot Manipulators With Obstacle Avoidance
T2 - A Deep Reinforcement Learning Approach
AU - Rubagotti, Matteo
AU - Sangiovanni, Bianca
AU - Nurbayeva, Aigerim
AU - Incremona, Gian Paolo
AU - Ferrara, Antonella
AU - Shintemirov, Almas
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2023/2/1
Y1 - 2023/2/1
N2 - The recent surge of interest in the application of learning-based methods to control systems motivates this work
to investigate how a purely model-free and a purely model-based method compare when applied to a shared control
problem for a robot manipulator. Specifically, we propose a
method based on model-free deep reinforcement learning
(DRL) for tracking the position of an operator’s hand with
the end effector of a manipulator while automatically avoiding obstacles in the workspace with the whole robot frame.
The obtained control strategy generates joint reference velocities via a deep neural network trained using Q-learning.
The method is tested in simulation and experimentally on a
UR5 manipulator, and it is compared with a model predictive control (MPC) approach for solving the same problem.
It is observed that DRL presents better performance than
MPC but only if the provided reference falls within the distribution of the DRL algorithm policy. As expected, the model-based nature of MPC allows the approach to address
unforeseen situations as long as these are compatible with its
process model. This is not the case for DRL, for which an
unexpected (not seen during the training process) human
hand reference would lead to extremely poor performance.
AB - The recent surge of interest in the application of learning-based methods to control systems motivates this work
to investigate how a purely model-free and a purely model-based method compare when applied to a shared control
problem for a robot manipulator. Specifically, we propose a
method based on model-free deep reinforcement learning
(DRL) for tracking the position of an operator’s hand with
the end effector of a manipulator while automatically avoiding obstacles in the workspace with the whole robot frame.
The obtained control strategy generates joint reference velocities via a deep neural network trained using Q-learning.
The method is tested in simulation and experimentally on a
UR5 manipulator, and it is compared with a model predictive control (MPC) approach for solving the same problem.
It is observed that DRL presents better performance than
MPC but only if the provided reference falls within the distribution of the DRL algorithm policy. As expected, the model-based nature of MPC allows the approach to address
unforeseen situations as long as these are compatible with its
process model. This is not the case for DRL, for which an
unexpected (not seen during the training process) human
hand reference would lead to extremely poor performance.
UR - http://www.scopus.com/inward/record.url?scp=85147534627&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147534627&partnerID=8YFLogxK
U2 - 10.1109/MCS.2022.3216653
DO - 10.1109/MCS.2022.3216653
M3 - Article
AN - SCOPUS:85147534627
SN - 1066-033X
VL - 43
SP - 44
EP - 63
JO - IEEE Control Systems
JF - IEEE Control Systems
IS - 1
ER -