posted on 2024-01-03, 14:48authored byRuizhen Gao, Ziheng Li, Junfu Li, Baihua LiBaihua Li, Jingjun Zhang, Jun Liu
Slam (simultaneous localization and mapping) play an important role in the field of artificial and driverless intelligence. A real-time dynamic visual SLAM algorithm based on an object detection network is proposed to address the robustness and camera localization accuracy issues caused by dynamic objects in indoor dynamic scenes. The YOLOv5s model, which has the smallest depth and feature map width in the YOLOv5 series, is chosen as the object detection network. The backbone network is replaced with the lightweight ShuffleNetv2 network. Experimental results on the VOC2007 dataset show that the YOLOv5-LITE model reduces the network parameters by 41.89% and speeds up the runtime by 39.00% compared to the YOLOv5s model. A motion level division strategy is adopted to provide prior information to the object detection network. In the tracking thread of the visual SLAM system, a parallel thread combining the improved object detection network and multi-view geometry is introduced to eliminate dynamic feature points. The experimental results demonstrate that in dynamic scenes, the proposed algorithm improves the camera localization accuracy by an average of 85.38% compared to ORB-SLAM2. Finally, experiments in a real environment are conducted to validate the effectiveness of the algorithm.
Funding
Science and Technology Project of the Hebei Education Department, Science and Technology Research Projects of Colleges and Universities in Hebei
Handan Science and Technology Bureau Project
School Level Project of Handan University under Grant QN2021405, Grant ZD2018207, Grant 21422021173, and Grant XZ2021202
History
School
Science
Department
Computer Science
Published in
IEEE Access
Volume
11
Pages
113952 - 113964
Publisher
Institute of Electrical and Electronics Engineers (IEEE)
This is an Open Access Article. It is published by the Institute of Electrical and Electronics Engineers (IEEE) under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Licence (CC BY-NC-ND). Full details of this licence are available at: https://creativecommons.org/licenses/by-nc-nd/4.0/