Treffer: Quadrotor path following and reactive obstacle avoidance with deep reinforcement learning

Title:
Quadrotor path following and reactive obstacle avoidance with deep reinforcement learning
Contributors:
Universitat Politècnica de Catalunya. Departament d'Enginyeria de Sistemes, Automàtica i Informàtica Industrial
Publication Year:
2021
Collection:
Universitat Politècnica de Catalunya, BarcelonaTech: UPCommons - Global access to UPC knowledge
Document Type:
Fachzeitschrift article in journal/newspaper
File Description:
application/pdf
Language:
English
Relation:
https://link.springer.com/article/10.1007/s10846-021-01491-2; info:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2013-2016/DPI2017-88403-R/ES/SEGURIDAD Y CONTROL EN VEHICULOS AUTONOMOS/; http://hdl.handle.net/2117/359100
DOI:
10.1007/s10846-021-01491-2
Rights:
Open Access
Accession Number:
edsbas.60FC5B83
Database:
BASE

Weitere Informationen

A deep reinforcement learning approach for solving the quadrotor path following and obstacle avoidance problem is proposed in this paper. The problem is solved with two agents: one for the path following task and another one for the obstacle avoidance task. A novel structure is proposed, where the action computed by the obstacle avoidance agent becomes the state of the path following agent. Compared to traditional deep reinforcement learning approaches, the proposed method allows to interpret the training process outcomes, is faster and can be safely trained on the real quadrotor. Both agents implement the Deep Deterministic Policy Gradient algorithm. The path following agent was developed in a previous work. The obstacle avoidance agent uses the information provided by a low-cost LIDAR to detect obstacles around the vehicle. Since LIDAR has a narrow field-of-view, an approach for providing the agent with a memory of the previously seen obstacles is developed. A detailed description of the process of defining the state vector, the reward function and the action of this agent is given. The agents are programmed in python/tensorflow and are trained and tested in the RotorS/gazebo platform. Simulations results prove the validity of the proposed approach. ; This work has been partially funded by the Spanish Government (MINECO) through the project CICYT (ref. DPI2017-88403-R). ; Peer Reviewed ; Postprint (published version)