Back to Search Start Over

Neural inverse reinforcement learning in autonomous navigation.

Authors :
Xia, Chen
El Kamel, Abdelkader
Source :
Robotics & Autonomous Systems. Oct2016, Vol. 84, p1-14. 14p.
Publication Year :
2016

Abstract

Designing intelligent and robust autonomous navigation systems remains a great challenge in mobile robotics. Inverse reinforcement learning (IRL) offers an efficient learning technique from expert demonstrations to teach robots how to perform specific tasks without manually specifying the reward function. Most of existing IRL algorithms assume the expert policy to be optimal and deterministic, and are applied to experiments with relatively small-size state spaces. However, in autonomous navigation tasks, the state spaces are frequently large and demonstrations can hardly visit all the states. Meanwhile the expert policy may be non-optimal and stochastic. In this paper, we focus on IRL with large-scale and high-dimensional state spaces by introducing the neural network to generalize the expert’s behaviors to unvisited regions of the state space and an explicit policy representation is easily expressed by neural network, even for the stochastic expert policy. An efficient and convenient algorithm, Neural Inverse Reinforcement Learning (NIRL), is proposed. Experimental results on simulated autonomous navigation tasks show that a mobile robot using our approach can successfully navigate to the target position without colliding with unpredicted obstacles, largely reduce the learning time, and has a good generalization performance on undemonstrated states. Hence prove the robot intelligence of autonomous navigation transplanted from limited demonstrations to completely unknown tasks. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09218890
Volume :
84
Database :
Academic Search Index
Journal :
Robotics & Autonomous Systems
Publication Type :
Academic Journal
Accession number :
117737390
Full Text :
https://doi.org/10.1016/j.robot.2016.06.003