检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Jing Li Chenhui Shi Jun Chen Ruisheng Wang Zhiyuan Yang Fan Zhang Jianhua Gong
机构地区:[1]National Engineering Research Center for Geoinformatics,Aerospace Information Research Institute,Chinese Academy of Sciences,Beijing,People’s Republic of China [2]National Geomatics Center of China,Beijing,People’s Republic of China [3]Department of Geomatics Engineering,University of Calgary,Calgary,Canada [4]Department of Urban Studies and Planning,Massachusetts Institute of Technology,Cambridge,MA,USA
出 处:《International Journal of Digital Earth》2022年第1期1929-1946,共18页国际数字地球学报(英文)
基 金:supported by the National Key Research and Development Program of China under[Grant number 2019YFC1511304];supported by the Pilot Fund of Frontier Science and Disruptive Technology of Aerospace Information Research Institute,Chinese Academy of Sciences under[Grant number E0Z21101].
摘 要:Simultaneous Localization and Mapping(SLAM)has been widely used in emergency response,self-driving and city-scale 3D mapping and navigation.Recent deep-learning based feature point extractors have demonstrated superior performance in dealing with the complex environmental challenges(e.g.extreme lighting)while the traditional extractors are struggling.In this paper,we have successfully improved the robustness and accuracy of a monocular visual SLAM system under various complex scenes by adding a deep learning based visual localization thread as an augmentation to the visual SLAM framework.In this thread,our feature extractor with an efficient lightweight deep neural network is used for absolute pose and scale estimation in real time using the highly accurate georeferenced prior map database at 20cm geometric accuracy created by our in-house and low-cost LiDAR and camera integrated device.The closed-loop error provided by our SLAM system with and without this enhancement is 1.03m and 18.28m respectively.The scale estimation of the monocular visual SLAM is also significantly improved(0.01 versus 0.98).In addition,a novel camera-LiDAR calibration workflow is also provided for large-scale 3D mapping.This paper demonstrates the application and research potential of deep-learning based vision SLAM with image and LiDAR sensors.
关 键 词:Deep local features lightweight network visual localization SLAM LIDAR
分 类 号:TN96[电子电信—信号与信息处理]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.249