Go to:
Logótipo
Comuta visibilidade da coluna esquerda
Você está em: Start > Publications > View > Deep Reinforcement Learning for Zero-Shot Coverage Path Planning With Mobile Robots
Publication

Deep Reinforcement Learning for Zero-Shot Coverage Path Planning With Mobile Robots

Title
Deep Reinforcement Learning for Zero-Shot Coverage Path Planning With Mobile Robots
Type
Article in International Scientific Journal
Year
2025
Authors
Carvalho, JP
(Author)
Other
The person does not belong to the institution. The person does not belong to the institution. The person does not belong to the institution. Without AUTHENTICUS Without ORCID
Journal
The Journal is awaiting validation by the Administrative Services.
Vol. 12
Pages: 1594-1609
ISSN: 2329-9266
Indexing
Other information
Authenticus ID: P-018-3AT
Abstract (EN): The ability of mobile robots to plan and execute a path is foundational to various path-planning challenges, particularly Coverage Path Planning. While this task has been typically tackled with classical algorithms, these often struggle with flexibility and adaptability in unknown environments. On the other hand, recent advances in Reinforcement Learning offer promising approaches, yet a significant gap in the literature remains when it comes to generalization over a large number of parameters. This paper presents a unified, generalized framework for coverage path planning that leverages value-based deep reinforcement learning techniques. The novelty of the framework comes from the design of an observation space that accommodates different map sizes, an action masking scheme that guarantees safety and robustness while also serving as a learning-from-demonstration technique during training, and a unique reward function that yields value functions that are size-invariant. These are coupled with a curriculum learning-based training strategy and parametric environment randomization, enabling the agent to tackle complete or partial coverage path planning with perfect or incomplete knowledge while generalizing to different map sizes, configurations, sensor payloads, and sub-tasks. Our empirical results show that the algorithm can perform zero-shot learning scenarios at a near-optimal level in environments that follow a similar distribution as during training, outperforming a greedy heuristic by sixfold. Furthermore, in out-of-distribution environments, our method surpasses existing state-of-the-art algorithms in most zero-shot and all few-shot scenarios, paving the way for generalizable and adaptable path-planning algorithms.
Language: English
Type (Professor's evaluation): Scientific
No. of pages: 16
Documents
We could not find any documents associated to the publication.
Related Publications

Of the same authors

A Unified Stability Analysis of Safety-Critical Control using Multiple Control Barrier Functions (2025)
Article in International Scientific Journal
Reis, MF; Carvalho, JP; Aguiar, AP
Model Predictive Control for B-Spline Trajectory Tracking in Omnidirectional Robots (2024)
Article in International Conference Proceedings Book
Carvalho, JP; António Paulo Moreira; Aguiar, AP
A Reinforcement Learning Based Online Coverage Path Planning Algorithm (2023)
Article in International Conference Proceedings Book
Carvalho, JP; Aguiar, AP

Of the same journal

Two-Stage Robust Optimization Under Decision Dependent Uncertainty (2022)
Article in International Scientific Journal
Zhang, YF; Liu, F; Su, YF; Chen, Y; Wang, ZJ; Catalao, JPS
Recommend this page Top
Copyright 1996-2025 © Centro de Desporto da Universidade do Porto I Terms and Conditions I Acessibility I Index A-Z
Page created on: 2025-10-30 01:45:36 | Privacy Policy | Personal Data Protection Policy | Whistleblowing | Electronic Yellow Book