Back to Search Start Over

Integrated Ray-Tracing and Coverage Planning Control using Reinforcement Learning

Authors :
Papaioannou, Savvas
Kolios, Panayiotis
Theocharides, Theocharis
Panayiotou, Christos G.
Polycarpou, Marios M.
Source :
2022 IEEE 61st Conference on Decision and Control (CDC), 06-09 December 2022, Cancun, Mexico
Publication Year :
2023

Abstract

In this work we propose a coverage planning control approach which allows a mobile agent, equipped with a controllable sensor (i.e., a camera) with limited sensing domain (i.e., finite sensing range and angle of view), to cover the surface area of an object of interest. The proposed approach integrates ray-tracing into the coverage planning process, thus allowing the agent to identify which parts of the scene are visible at any point in time. The problem of integrated ray-tracing and coverage planning control is first formulated as a constrained optimal control problem (OCP), which aims at determining the agent's optimal control inputs over a finite planning horizon, that minimize the coverage time. Efficiently solving the resulting OCP is however very challenging due to non-convex and non-linear visibility constraints. To overcome this limitation, the problem is converted into a Markov decision process (MDP) which is then solved using reinforcement learning. In particular, we show that a controller which follows an optimal control law can be learned using off-policy temporal-difference control (i.e., Q-learning). Extensive numerical experiments demonstrate the effectiveness of the proposed approach for various configurations of the agent and the object of interest.<br />Comment: 2022 IEEE 61st Conference on Decision and Control (CDC), 06-09 December 2022, Cancun, Mexico

Details

Database :
arXiv
Journal :
2022 IEEE 61st Conference on Decision and Control (CDC), 06-09 December 2022, Cancun, Mexico
Publication Type :
Report
Accession number :
edsarx.2304.09631
Document Type :
Working Paper
Full Text :
https://doi.org/10.1109/CDC51059.2022.9992360