Abstract
We propose a deep reinforcement learning approach for solving a mapless navigation problem in warehouse scenarios. In our approach, an automatic guided vehicle is equipped with two LiDAR sensors and one frontal RGB camera and learns to perform a targeted navigation task. The challenges reside in the sparseness of positive samples for learning, multi-modal sensor perception with partial observability, the demand for accurate steering maneuvers together with long training cycles. To address these points, we propose NavACL-Q as an automatic curriculum learning method in combination with a distributed version of the soft actor-critic algorithm. The performance of the learning algorithm is evaluated exhaustively in a different warehouse environment to validate both robustness and generalizability of the learned policy. Results in NVIDIA Isaac Sim demonstrates that our trained agent significantly outperforms the map-based navigation pipeline provided by NVIDIA Isaac Sim with an increased agent-goal distance of 3 m and a wider initial relative agent-goal rotation of approximately 45∘. The ablation studies also suggest that NavACL-Q greatly facilitates the whole learning process with a performance gain of roughly 40% compared to training with random starts and a pre-trained feature extractor manifestly boosts the performance by approximately 60%.
Originalsprache | Englisch |
---|---|
Aufsatznummer | 3153 |
Seitenumfang | 30 |
Fachzeitschrift | Applied Sciences : open access journal |
Jahrgang | 12.2022 |
Ausgabenummer | 6 |
DOIs | |
Publikationsstatus | Veröffentlicht - 19 März 2022 |
Bibliographische Notiz
Funding Information:Funding: This research is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation)–No 430054590 (TRAIN, to E.R.). The work is done in KION Group AG, Technology and Innovation.
Publisher Copyright:
© 2022 by the authors. Licensee MDPI, Basel, Switzerland.
Schlagwörter
- Deep Learning
- Machine learning
- navigation
- Warehouse