2023
DOI: 10.1088/1361-6501/acfb2d
|View full text |Cite
|
Sign up to set email alerts
|

DIG-SLAM: an accurate RGB-D SLAM based on instance segmentation and geometric clustering for dynamic indoor scenes

Rongguang Liang,
Jie Yuan,
Benfa Kuang
et al.

Abstract: Simultaneous localization and mapping (SLAM) has emerged as a critical technology enabling robots to navigate in unknown environments, drawing extensive attention within the robotics research community. However, traditional visual SLAM ignores the presence of dynamic objects in indoor scenes, and dynamic point features of dynamic objects can lead to incorrect data correlation, making the traditional visual SLAM is difficult to accurately estimate the camera’s pose when the objects in the scenes are moving. Usi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 37 publications
0
6
0
Order By: Relevance
“…Ran et al [48] proposed a novel RGB-D-inertial dynamic SLAM method that enables accurate localization even when a significant portion of the camera's view is obstructed by multiple dynamic objects for an extended period.SVD-SLAM [49] has incorporated an improved YOLACT++ lightweight instance segmentation network and calculates the motion probability of each potential moving object based on the camera pose and polar constraints.DIG-SLAM [50] utilizes YOLOv7 instance segmentation to extract object contours, while optimizing line features through line segment detection and K-means clustering. It employs motion consistency checks to identify dynamic areas and removes points and line features within them.DYS-SLAM [51] Current literature mainly leans towards combining geometric constraints and deep learning for Visual SLAM in dynamic settings.…”
Section: ) Deep Learning-based Methodsmentioning
confidence: 99%
“…Ran et al [48] proposed a novel RGB-D-inertial dynamic SLAM method that enables accurate localization even when a significant portion of the camera's view is obstructed by multiple dynamic objects for an extended period.SVD-SLAM [49] has incorporated an improved YOLACT++ lightweight instance segmentation network and calculates the motion probability of each potential moving object based on the camera pose and polar constraints.DIG-SLAM [50] utilizes YOLOv7 instance segmentation to extract object contours, while optimizing line features through line segment detection and K-means clustering. It employs motion consistency checks to identify dynamic areas and removes points and line features within them.DYS-SLAM [51] Current literature mainly leans towards combining geometric constraints and deep learning for Visual SLAM in dynamic settings.…”
Section: ) Deep Learning-based Methodsmentioning
confidence: 99%
“…Based on case segmentation, Li et al [15] introduces a confidence-based matching algorithm to correlate dynamic objects and differentiate moving objects from static objects. To eliminate the interference of dynamic objects in the room, Liang et al [16] determined the dynamic area by moving one-time inspection combined with case segmentation, thus removing points and features within the dynamic area. Although the above methods can filter dynamic objects, they are not real-time.…”
Section: Related Workmentioning
confidence: 99%
“…For example, generalized-ICP (GICP) [6] attempts to match local planar patches from both scans and integrates pointto-plane distances to ICP. Alternatively, feature-based matching methods [4,5] minimize computational efforts by only matching feature points extracted from raw point clouds.…”
Section: Related Workmentioning
confidence: 99%
“…The conventional LiDAR SLAM framework [3][4][5][6][7][8][9][10][11][12] operates by sequentially registering LiDAR scans, thereby estimating the position of the present frame for each iteration. However, it is worth noting that the information from the current frame can also enhance the estimation of historical frames, consequently improving the accuracy of the current frame itself.…”
Section: Introductionmentioning
confidence: 99%