检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:曹雏清 罗海南 马玉洁 CAO Chuqing;LUO Hainan;MA Yujie(School of Computer and Information,Anhui Polytechnic University,Wuhu 241000,Anhui,China;Yangtze River Delta Hart Robotics Industrial Technology Research Institute,Wuhu 241000,Anhui,China)
机构地区:[1]安徽工程大学计算机与信息学院,安徽芜湖241000 [2]长三角哈特机器人产业技术研究院,安徽芜湖241000
出 处:《计算机工程》2024年第11期130-141,共12页Computer Engineering
基 金:安徽省教育厅科学研究重点项目(KJ2020A0364);国家自然科学基金面上项目(62073101)。
摘 要:视觉重定位技术是室内服务机器人关键技术之一,其主要目的是精确确定机器人在场景中的六自由度位姿。在室内环境中,稀疏纹理区域的普遍存在对视觉重定位的精度构成了挑战,因为这些区域中的相似图像块会大幅干扰定位准确性。此外,现有的视觉重定位网络往往忽视图像中角点的重要性,限制了卷积神经网络在编码场景信息时的能力,因为角点中蕴含着丰富的几何特征。为此,提出一种结合元素级注意力机制和角点特征的视觉重定位网络。为了解决相似图像块的问题,提出元素级注意力机制,通过预测元素级注意力加权系数来评估特征图中每个元素的重要性。该方法可有效融合多级特征图,利用低级特征图中的几何结构信息与高级特征图中的语义信息来提升相似图像块的区分度。针对角点特征被忽视的问题,提出一种角点特征整合模块,利用角点提取网络SuperPoint提取大量角点进行聚类,并选择距离聚类中心最近的角点来保证其均匀地分布在图像中。该网络将提取的角点特征整合进高维特征图中,从而保证网络充分地提取角点中所包含的图像几何特征,进而提升网络的场景解析能力。在7-Scenes数据集上的实验结果表明,在包含大量稀疏纹理的室内场景中,提升相似图像块间的区分度并整合角点特征可有效提升视觉重定位精度,使该方法实现了0.025 m的中值平移误差、0.83°的中值旋转误差以及87.43%的重定位准确率。Visual relocalization,an essential technique for indoor service robots,aims to recover the six Degree-Of-Freedom(6-DOF)pose of a robot.However,numerous textureless regions in indoor environments pose a challenge in achieving accurate visual relocalization because similar image patches in these regions significantly disturb the relocalization accuracy.In addition,current visual relocalization networks ignore the importance of corners,which means that the abundant geometric features are not fully leveraged.This limits the encoding capability of a network for scene information.To resolve these issues,this paper proposes a novel visual localization network containing an element-wise attention mechanism and corner features.First,to solve the problems caused by similar image patches,the network introduces an element-wise weighting mechanism that predicts element-wise weighting factors to measure the importance of each element in the feature maps.Thereafter,multi-level features are fused effectively,and the structural information in low-level features and semantic information in high-level features are leveraged to distinguish similar image patches.To ignore corner features,the network introduces a corner-feature integration module that clusters numerous corners extracted using SuperPoint and selects the corners closest to the clustering centers to ensure a uniform corner distribution.Subsequently,the network integrates the corner features into high-dimensional features,which ensures an adequate extraction of the geometric features contained in the corners.Thus,the scene parsing capability of the network is boosted.Experimental results using the 7-Scenes dataset demonstrate that in indoor scenes with large sparse textures,distinguishing similar image patches and integrating corner information is conducive for boosting the relocalization accuracy.The method achieves median positional error,median rotational error,and accuracy of 87.43%,0.025 m,and 0.83°,respectively.
关 键 词:视觉重定位 室内服务机器人 卷积神经网络 多级特征融合 角点特征
分 类 号:TP391[自动化与计算机技术—计算机应用技术]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.222