The presence of moving objects in real-world scenarios can lead to mismatches in visual odometry feature points, thereby affecting the accuracy of positioning and mapping by the SLAM system and reducing its robustness in practical applications. This paper introduces a visual SLAM algorithm that leverages the ORB-SLAM3 framework and deep learning techniques. Enhancements to the SLAM system’s tracking thread enable identifying and removing dynamic feature points, thus increasing its adaptability to dynamic environments. Concurrently, YOLOv8s, known for its minimal depth and feature map width in the YOLOv5 series, is selected as the object detection network, with VanillaNet, a lightweight network, replacing its backbone network. This combination effectively determines the mobility of objects within the environment. Consequently, we propose an enhanced algorithm based on YOLOv8s capable of performing both object detection and semantic segmentation to eliminate dynamic feature points precisely. Ultimately, the algorithm’s accuracy and real-time performance were assessed using indoor dynamic scene data from the TUM RGB datasets. In comparison to models lacking any strategic approach, test results on the TUM datasets reveal that the experimental outcomes are more favorable in dynamic environments.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.