基于ViT-改进YOLOv7的稻田杂草识别  被引量:1

Recognizing weed in rice field using ViT-improved YOLOv7

在线阅读下载全文

作  者:陈学深[1] 吴昌鹏 党佩娜 张恩造 陈彦学 汤存耀 齐龙[1] CHEN Xueshen;WU Changpeng;DANG Peina;ZHANG Enzao;CHEN Yanxue;TANG Cunyao;QI Long(College of Engineering,South China Agricultural University,Guangzhou 510642,China)

机构地区:[1]华南农业大学工程学院,广州510642

出  处:《农业工程学报》2024年第10期185-193,共9页Transactions of the Chinese Society of Agricultural Engineering

基  金:广东省自然科学基金项目(2021A1515010831);广州市科技计划项目(202206010125);广东省杰出青年基金项目(2019B151502056);国家自然科学基金项目(51575195)。

摘  要:为解决光线遮蔽、藻萍干扰以及稻叶尖形状相似等复杂环境导致稻田杂草识别效果不理想问题,该研究提出一种基于组合深度学习的杂草识别方法。引入MSRCP(multi-scale retinex with color preservation)对图像进行增强,以提高图像亮度及对比度;加入ViT分类网络去除干扰背景,以提高模型在复杂环境下对小目标杂草的识别性能。在YOLOv7模型中主干特征提取网络替换为GhostNet网络,并引入CA注意力机制,以增强主干特征提取网络对杂草特征提取能力及简化模型参数计算量。消融试验表明:改进后的YOLOv7模型平均精度均值为88.2%,较原YOLOv7模型提高了3.3个百分点,参数量减少10.43 M,计算量减少66.54×109次/s。识别前先经过MSRCP图像增强后,与原模型相比,改进YOLOv7模型的平均精度均值提高了2.6个百分点,光线遮蔽、藻萍干扰以及稻叶尖形状相似的复杂环境下平均精度均值分别提高5.3、3.6、3.1个百分点,加入ViT分类网络后,较原模型平均精度均值整体提升了4.4个百分点,光线遮蔽、藻萍干扰一级稻叶尖形状相似的复杂环境下的平均精度均值较原模型整体提升了6.2、6.1、5.7个百分点。ViT-改进YOLOv7模型的平均精度均值为92.6%,相比于YOLOv5s、YOLOXs、MobilenetV3-YOLOv7、YOLOv8和改进YOLOv7分别提高了11.6、10.1、5.0、4.2、4.4个百分点。研究结果可为稻田复杂环境的杂草精准识别提供支撑。A weed recognition was proposed using combinatorial deep learning,in order to reduce the influencing factors,such as the light shading of rice plants,interference of rice field algae and weeds with small targets.Data enhancement of weed sample was used to improve the model training and generalization for less overfitting.The MSRCP was introduced to enhance the image quality in complex environments.Weed target recognition was realized in the low contrast and clarity of rice field images,due to the light blockage of the rice plant.Front-end slicing and ViT(vision of transformer)classification were performed on the HD images.Information loss was avoided to detect the images in the process of network compression input.Small targets were retained in the high-definition images,in order to improve the effectiveness of the model in complex environments.The YOLOv7 model was replaced by the lightweight network GhostNet,and then embedded by the CA attention mechanism.The number of parameters and computations was reduced to enhance the feature extraction,particularly for the high accuracy and real-time performance of weed recognition.After the classification of target recognition,the image compression was attributed to the small target blurring and loss of effective information that was caused by only the image recognition model.The experiment showed that the weed dataset was expanded to improve the recognition of the model.The ablation test showed that the average mean accuracy of the test set after data enhancement was 84.9%,which was 10.8 percentage points better than the model trained on the original dataset.The ViT classification network outperformed Resnet 50 and Vgg,in terms of accuracy,recall and detection speed.Among them,the accuracy rate increased by 7.9 and 7.5 percentage points,respectively,and the recall rate increased by 7.1 and 5.3 percentage points,respectively.Comparative tests showed that the ViT network also achieved high classification accuracy and speed.The ablation test showed that the mean average acc

关 键 词:机器视觉 深度学习 YOLOv7 VIT 稻田杂草 识别 

分 类 号:S985.2[农业科学—捕捞与储运]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象