检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:解宇敏 张浪文[1,2] 余孝源 谢巍 XIE Yu-min;ZHANG Lang-wen;YU Xiao-yuan;XIE Wei(College of Automation Science and Technology,South China University of Technology,Guangzhou Guangdong 510640,China;Yueyang Goaland Energy Conservation Equipment Manufacturing Co.,Ltd,Yueyang Hunan 414000,China;College of Physics and Telecommunication Engineering,South China Normal University,Guangzhou Guangdong 510006,China)
机构地区:[1]华南理工大学自动化科学与工程学院,广东广州510640 [2]岳阳高澜节能装备制造有限公司,湖南岳阳414000 [3]华南师范大学物理与电信工程学院,广东广州510006
出 处:《控制理论与应用》2024年第5期914-922,共9页Control Theory & Applications
基 金:国家自然科学基金项目(61803161);广东省自然科学基金项目(2022A1515011887,2023A1515030119);清远市科技计划项目(2023DZX006);佛山市重点领域科技攻关项目(2020001006812);顺德区核心攻关项目(2030218000174);广州市科技计划项目(202102020379);江门市基础与应用基础研究项目(2020030103080008999)资助。
摘 要:目标检测是自动驾驶系统的关键技术,普通RGB目标检测算法在夜间和恶劣气候等场景往往表现一般,融合可见光和红外信息的目标检测算法因而受到诸多研究关注.现有方法通常融合结构复杂,且忽视了模态间信息交流的重要性.对此,本文以YOLOv5为基本框架,提出一种可见光–红外特征交互与融合的目标检测算法,使用一种新的主干网络跨阶段局部(CSPDarknet53-F),采用双分支结构分别提取可见光和红外特征.然后,通过特征交互模块重构各模态的信息成分和比例,提升模态间信息交流,使可见光和红外特征进行更充分的融合.在FLIR-aligned和M3FD数据集上的大量实验证明,本文算法使用的CSPDarknet53-F在协同利用可见光和红外信息方面更加出色,提升了模型精度,同时,拥有对抗光照强度骤变的鲁棒性.Object detection is the key technology of the autonomous driving system,but object detection algorithms based on RGB often perform poorly in scenarios such as nighttime and severe weather.Therefore,the object detection algorithms fusing visible and infrared information have begun to receive a lot of research attention.However,the existing methods usually have complex fusion structures and ignore the importance of information exchange between modalities.In this paper,we take YOLOv5 as the basic framework,and propose an object detection algorithm with visible-infrared feature interaction and fusion.It uses a new backbone network,CSPDarknet53-F,which uses a dual branch structure to extract visible and infrared features,respectively,and then reconstructs the information components and proportions of each mode through feature interaction modules to improve the information exchange between modalities so that visible and infrared features can be more fully integrated.Extensive experiments on the FLIR-aligned dataset and the M3FD dataset show that the CSPDarknet53-F used in our algorithm is more excellent in terms of synergistically utilizing visible and infrared information,which improves the detection accuracy of the model and has robustness against sudden changes in light intensity.
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:3.145.197.164