Back to Search Start Over

Dexterous robotic manipulation using deep reinforcement learning and knowledge transfer for complex sparse reward‐based tasks.

Authors :
Wang, Qiang
Sanchez, Francisco Roldan
McCarthy, Robert
Bulens, David Cordova
McGuinness, Kevin
O'Connor, Noel
Wüthrich, Manuel
Widmaier, Felix
Bauer, Stefan
Redmond, Stephen J.
Source :
Expert Systems; Jul2023, Vol. 40 Issue 6, p1-19, 19p
Publication Year :
2023

Abstract

This paper describes a deep reinforcement learning (DRL) approach that won Phase 1 of the Real Robot Challenge (RRC) 2021, and then extends this method to a more difficult manipulation task. The RRC consisted of using a TriFinger robot to manipulate a cube along a specified positional trajectory, but with no requirement for the cube to have any specific orientation. We used a relatively simple reward function, a combination of a goal‐based sparse reward and a distance reward, in conjunction with Hindsight Experience Replay (HER) to guide the learning of the DRL agent (Deep Deterministic Policy Gradient [DDPG]). Our approach allowed our agents to acquire dexterous robotic manipulation strategies in simulation. These strategies were then deployed on the real robot and outperformed all other competition submissions, including those using more traditional robotic control techniques, in the final evaluation stage of the RRC. Here we extend this method, by modifying the task of Phase 1 of the RRC to require the robot to maintain the cube in a particular orientation, while the cube is moved along the required positional trajectory. The requirement to also orient the cube makes the agent less able to learn the task through blind exploration due to increased problem complexity. To circumvent this issue, we make novel use of a Knowledge Transfer (KT) technique that allows the strategies learned by the agent in the original task (which was agnostic to cube orientation) to be transferred to this task (where orientation matters). KT allowed the agent to learn and perform the extended task in the simulator, which improved the average positional deviation from 0.134 to 0.02 m, and average orientation deviation from 142° to 76° during evaluation. This KT concept shows good generalization properties and could be applied to any actor‐critic learning algorithm. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
02664720
Volume :
40
Issue :
6
Database :
Complementary Index
Journal :
Expert Systems
Publication Type :
Academic Journal
Accession number :
164116238
Full Text :
https://doi.org/10.1111/exsy.13205