DOI QR코드

DOI QR Code

도시환경 매핑 시 SLAM 불확실성 최소화를 위한 강화 학습 기반 경로 계획법

RL-based Path Planning for SLAM Uncertainty Minimization in Urban Mapping

  • Cho, Younghun (Dept. of Civil and Environmental Engineering, KAIST) ;
  • Kim, Ayoung (Dept. of Civil and Environmental Engineering, KAIST)
  • 투고 : 2021.02.26
  • 심사 : 2021.04.05
  • 발행 : 2021.05.31

초록

For the Simultaneous Localization and Mapping (SLAM) problem, a different path results in different SLAM results. Usually, SLAM follows a trail of input data. Active SLAM, which determines where to sense for the next step, can suggest a better path for a better SLAM result during the data acquisition step. In this paper, we will use reinforcement learning to find where to perceive. By assigning entire target area coverage to a goal and uncertainty as a negative reward, the reinforcement learning network finds an optimal path to minimize trajectory uncertainty and maximize map coverage. However, most active SLAM researches are performed in indoor or aerial environments where robots can move in every direction. In the urban environment, vehicles only can move following road structure and traffic rules. Graph structure can efficiently express road environment, considering crossroads and streets as nodes and edges, respectively. In this paper, we propose a novel method to find optimal SLAM path using graph structure and reinforcement learning technique.

키워드

과제정보

This research was supported by a grant (21TSRD-B151228-03) from Urban Declining Area Regenerative Capacity-Enhancing Technology Research Program funded by Ministry of Land, Infrastructure and Transport of Korean government

참고문헌

  1. C. Leung, S. Huang, and G. Dissanayake, "Active SLAM using Model Predictive Control and Attractor based Exploration," 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 2006, DOI: 10.1109/IROS.2006.282530.
  2. E. W. Dijkstra, "A note on two problems in connexion with graphs," Numerische Mathematik, vol. 1, pp. 269-271, 1959, DOI: 10.1007/BF01386390.
  3. P. E. Hart, N. J. Nilsson, and B. Raphael, " A Formal Basis for the Heuristic Determination of Minimum Cost Paths," IEEE Transactions on Systems Science and Cybernetics, vol. 4, no. 2, pp. 100-107, Jul., 1968, DOI: 10.1109/TSSC.1968.300136.
  4. S. Singh, A. Barto, R. Grupen, and C. Connolly, "Robust Reinforcement Learning in Motion Planning," Advances in Neural Information Processing Systems 6 (NIPS 1993), 1993, [Online], https://proceedings.neurips.cc/paper/1993/hash/3d8e28caf901313a554cebc7d32e67e5-Abstract.html.
  5. A. W. Moore and C. G. Atkeson, "The parti-game algorithm for variable resolution reinforcement learning in multidimensional state-spaces," Machine Learning, vol. 21, pp. 199-233, 1995, DOI: 10.1007/bf00993591.
  6. J. J. Park, J. H. Kim, and J. B. Song, "Path planning for a robot manipulator based on probabilistic roadmap and reinforcement learning," International Journal of Control, Automation, and Systems, vol. 5, no. 6, pp. 674-680, 2007, [Online], https://koreauniv.pure.elsevier.com/en/publications/path-planning-fora-robot-manipulator-based-on-probabilistic-road.
  7. B. Zhang, Z. Mao, W. Liu, and J. Liu, "Geometric Reinforcement Learning for Path Planning of UAVs," Journal of Intelligent & Robotic Systems, vol. 77, no. 2, pp. 391-409, 2015, DOI: 10.1007/s10846-013-9901-z.
  8. T. Kollar and N. Roy, "Trajectory Optimization using Reinforcement Learning for M ap Exploration," The International Journal of Robotics Research, vol. 27, no. 2, pp. 175-196, 2008. https://doi.org/10.1177/0278364907087426
  9. L. Murphy and P. Newman, "Risky Planning on Probabilistic Costmaps for Path Planning in Outdoor Environments," IEEE Transactions on Robotics, vol. 29, no. 2, pp. 445-457, 2012, DOI: 10.1109/TRO.2012.2227216.
  10. L. Nardi and C. Stachniss, "Actively Improving Robot Navigation On Different Terrains Using Gaussian Process Mixture Models," 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 2019, DOI: 10.1109/ICRA.2019.8794079.
  11. T. Shan and B. Englot, "LeGO-LOAM: Lightweight and Ground-Optimized Lidar Odometry and Mapping on Variable Terrain," 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 2018, DOI: 10.1109/IROS.2018.8594299.
  12. A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, "CARLA: An Open Urban Driving Simulator," 1st Annual Conference on Robot Learning, PMLR, vol. 78, pp. 1-16, 2017, [Online], http://proceedings.mlr.press/v78/dosovitskiy17a.html.