We present a semantically rich graph representation for indoor robotic navigation. Our graph representation encodes: semantic locations such as offices or corridors as nodes, and navigational behaviors such as enter office or cross a corridor as edges. In particular, our navigational behaviors operate directly from visual inputs to produce motor controls and are implemented with deep learning architectures. This enables the robot to avoid explicit computation of its precise location or the geometry of the environment, and enables navigation at a higher level of semantic abstraction. We evaluate the effectiveness of our representation by simulating navigation tasks in a large number of virtual environments. Our results show that using a simple sets of perceptual and navigational behaviors, the proposed approach can successfully guide the way of the robot as it completes navigational missions such as going to a specific office. Furthermore, our implementation shows to be effective to control the selection and switching of behaviors.
Registro Sencillo
Registro Completo
Autor | Sepulveda Villalobos, Gabriel Andres Niebles, Juan Carlos Soto Arriaza, Álvaro Marcelo |
Título | A Deep Learning Based Behavioral Approach to Indoor Autonomous Navigation |
ISSN | 2577-087X |
ISBN | 9781538630822 |
ISBN electrónico | 9781538630815 |
Página inicio | 4646 |
Página final | 4653 |
Fecha de publicación | 2018 |
Resumen | We present a semantically rich graph representation for indoor robotic navigation. Our graph representation encodes: semantic locations such as offices or corridors as nodes, and navigational behaviors such as enter office or cross a corridor as edges. In particular, our navigational behaviors operate directly from visual inputs to produce motor controls and are implemented with deep learning architectures. This enables the robot to avoid explicit computation of its precise location or the geometry of the environment, and enables navigation at a higher level of semantic abstraction. We evaluate the effectiveness of our representation by simulating navigation tasks in a large number of virtual environments. Our results show that using a simple sets of perceptual and navigational behaviors, the proposed approach can successfully guide the way of the robot as it completes navigational missions such as going to a specific office. Furthermore, our implementation shows to be effective to control the selection and switching of behaviors. |
Derechos | acceso restringido |
DOI | 10.1109/ICRA.2018.8460646 |
Editorial | IEEE |
Enlace | https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=8460646 |
Palabra clave | Navigation Semantics Visualization Simultaneous localization and mapping Robustness Measurement |
Publicado en / Colección | IEEE International Conference on Robotics and Automation (ICRA) (2018 : Brisbane, Australia) |
Tipo de documento | comunicación de congreso |