可见光–红外特征交互与融合的YOLOv5目标检测算法  被引量:3

YOLOv5 object detection algorithm with visible-infrared feature interaction and fusion

在线阅读下载全文

作  者:解宇敏 张浪文[1,2] 余孝源 谢巍 XIE Yu-min;ZHANG Lang-wen;YU Xiao-yuan;XIE Wei(College of Automation Science and Technology,South China University of Technology,Guangzhou Guangdong 510640,China;Yueyang Goaland Energy Conservation Equipment Manufacturing Co.,Ltd,Yueyang Hunan 414000,China;College of Physics and Telecommunication Engineering,South China Normal University,Guangzhou Guangdong 510006,China)

机构地区:[1]华南理工大学自动化科学与工程学院,广东广州510640 [2]岳阳高澜节能装备制造有限公司,湖南岳阳414000 [3]华南师范大学物理与电信工程学院,广东广州510006

出  处:《控制理论与应用》2024年第5期914-922,共9页Control Theory & Applications

基  金:国家自然科学基金项目(61803161);广东省自然科学基金项目(2022A1515011887,2023A1515030119);清远市科技计划项目(2023DZX006);佛山市重点领域科技攻关项目(2020001006812);顺德区核心攻关项目(2030218000174);广州市科技计划项目(202102020379);江门市基础与应用基础研究项目(2020030103080008999)资助。

摘  要:目标检测是自动驾驶系统的关键技术,普通RGB目标检测算法在夜间和恶劣气候等场景往往表现一般,融合可见光和红外信息的目标检测算法因而受到诸多研究关注.现有方法通常融合结构复杂,且忽视了模态间信息交流的重要性.对此,本文以YOLOv5为基本框架,提出一种可见光–红外特征交互与融合的目标检测算法,使用一种新的主干网络跨阶段局部(CSPDarknet53-F),采用双分支结构分别提取可见光和红外特征.然后,通过特征交互模块重构各模态的信息成分和比例,提升模态间信息交流,使可见光和红外特征进行更充分的融合.在FLIR-aligned和M3FD数据集上的大量实验证明,本文算法使用的CSPDarknet53-F在协同利用可见光和红外信息方面更加出色,提升了模型精度,同时,拥有对抗光照强度骤变的鲁棒性.Object detection is the key technology of the autonomous driving system,but object detection algorithms based on RGB often perform poorly in scenarios such as nighttime and severe weather.Therefore,the object detection algorithms fusing visible and infrared information have begun to receive a lot of research attention.However,the existing methods usually have complex fusion structures and ignore the importance of information exchange between modalities.In this paper,we take YOLOv5 as the basic framework,and propose an object detection algorithm with visible-infrared feature interaction and fusion.It uses a new backbone network,CSPDarknet53-F,which uses a dual branch structure to extract visible and infrared features,respectively,and then reconstructs the information components and proportions of each mode through feature interaction modules to improve the information exchange between modalities so that visible and infrared features can be more fully integrated.Extensive experiments on the FLIR-aligned dataset and the M3FD dataset show that the CSPDarknet53-F used in our algorithm is more excellent in terms of synergistically utilizing visible and infrared information,which improves the detection accuracy of the model and has robustness against sudden changes in light intensity.

关 键 词:可见光图像 红外图像 特征融合 交互 YOLOv5 

分 类 号:TP391.41[自动化与计算机技术—计算机应用技术] U495[自动化与计算机技术—计算机科学与技术]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象