高级检索

基于语义分割网络的动态场景视觉SLAM算法

Dynamic Scene Vision SLAM Optimization Based on Semantic Segmentation Network

  • 摘要: 针对传统即时定位与建图(SLAM)算法在动态场景中位姿估计精度不高的问题,提出一种动态场景下基于语义分割网络的视觉SLAM算法。通过RGB-D相机采集彩色图与深度图,将彩色图输入轻量级语义分割网络LR−ASPP,剔除先验动态物体,同时获得语义图;采用多视角几何算法剔除非先验动态物体,分离动静态特征点后得到优化后的位姿,并结合语义图和深度图构建纯静态语义八叉树地图,提高对动态场景的适应能力并直接用于导航系统。公开数据集TUM的实验测试结果表明:本文算法的最小绝对定位误差仅0.007 6 m,相比于ORB−SLAM3算法,在高动态场景中定位精度提升了80%以上,并能获取精确的动态区域及准确的语义地图,本文算法在复杂动态场景中具有良好的定位精度和鲁棒性。

     

    Abstract: Aiming at the problem that the traditional real-time location and mapping (SLAM) algorithm does not have high accuracy in pose estimation in dynamic scenes, a visual SLAM algorithm based on semantic segmentation network in dynamic scenes was proposed. Firstly, the color map and depth map were collected by RGB-D camera, and the color map was input to the lightweight semantic segmentation network LR-ASPP to eliminate the a priori dynamic objects and obtain the semantic map at the same time. Then, the multi-view geometry algorithm was used to reject the non-prior dynamic objects, after separating the dynamic and static feature points, the optimized position and pose was obtained, and the pure static semantic octree map was constructed by combining the semantic map and depth map to improve the adaptability to dynamic scenes and directly used in the navigation system.The test results of the public dataset TUM show that the minimum absolute localization error of the algorithm in this paper is only 0.007 6 m. Compared with ORB-SLAM3, the localization accuracy in highly dynamic scenes is improved by more than 80%, and precise dynamic regions and accurate semantic maps can be obtained, which verifies that the algorithm has good localization accuracy and robustness in complex dynamic scenes.

     

/

返回文章
返回