基于车载成像与深度卷积神经网络的地表残膜识别方法
CSTR:
作者:
作者单位:

作者简介:

通讯作者:

中图分类号:

基金项目:

国家重点研发计划项目(2022YFD2002400)、国家棉花产业技术体系岗位科学家项目(CARS-15-17)和石河子大学高层次人才科研启动项目(RCZK202442)


Surface Residual Film Recognition Method Based on Vehicle-mounted Imaging and Deep Convolutional Neural Networks
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    针对残膜回收机实际作业过程中存在多种相似非目标场景干扰,目标场景图像背景复杂且地表残膜尺寸小、破碎度大、无固定轮廓导致残膜覆盖率难以准确评估的问题,提出基于车载成像和深度卷积神经网络的地表残膜识别方法。构建了一种基于多重特征增强的SE-DenseNet-DC分类模型,在DenseNet121模型每个稠密块的非线性组合函数前后引入通道注意力机制增强有效特征信息通道的权重,然后引入多尺度串联空洞卷积替换原始模型第1层卷积提升感受野并保持细节敏感度,实现目标场景图像的有效提取;构建了一种基于细节信息增强和多尺度特征融合的CDC-TransUnet分割模型,在TransUnet模型的编码器部分引入CBAM模块提取更加细微和精确的全局特征,在跳跃连接部分引入DAB模块融合多尺度语义信息并弥补编码和解码阶段特征之间的语义差距,然后在解码器部分引入CCAF模块减少上采样丢失的细节信息,实现目标场景图像复杂背景中地表残膜的精准分割。试验结果表明,SE-DenseNet-DC分类模型对目标场景图像的分类准确率、查准率、查全率和F1值分别达到96.26%、91.54%、94.49%和92.83%,CDC-TransUnet分割模型对目标场景图像中地表残膜分割平均交并比(MIOU)达到77.17%,模型预测残膜覆盖率与人工标注残膜覆盖率决定系数(R2)为0.92,均方根误差(RMSE)为0.23%,平均相对误差为2.95%,单幅图像评估时间平均为0.54s。本文方法在残膜回收机回收后地表残膜覆盖率监测评估中具有较高的准确率和较快的推理速度,为残膜回收机回收质量实时准确评估提供技术支撑。

    Abstract:

    Aiming to address the challenges in accurately assessing residual film coverage due to interference from multiple similar non-target scenarios, complex background textures in target scene images, and the small size, high fragmentation, and irregular contours of residual films during the operational process of residual film recovery machinery, a residual film recognition method was proposed based on vehicle-mounted imaging and deep convolutional neural networks. A multi-feature-enhanced SE-DenseNet-DC classification model was developed by integrating channel attention mechanisms before and after the nonlinear combination functions in each dense block of the DenseNet121 architecture, the model enhanced the weighting of effective feature channels. Additionally, the first-layer convolution of the original model was replaced with multi-scale cascaded dilated convolutions to expand the receptive field while preserving sensitivity to fine details, enabling effective extraction of target scene images. Furthermore, a CDC-TransUnet segmentation model was constructed with enhanced detail information and multi-scale feature fusion. In the encoder of the TransUnet framework, CBAM modules were introduced to capture finer and more precise global features. DAB modules were embedded in the skip connections to fuse multi-scale semantic information and bridge the semantic gap between encoder and decoder features. CCAF modules were then incorporated into the decoder to mitigate detail loss during upsampling, achieving precise segmentation of residual films against complex backgrounds in target scenes. Experimental results demonstrated that the SE-DenseNet-DC classification model achieved classification accuracy, precision, recall, and F1 score of 96.26%, 91.54%, 94.49%, and 92.83%, respectively, for target scene image classification. The CDC-TransUnet segmentation model achieved an average intersection over union (MIOU) of 77.17% for surface residual film segmentation. The coefficient of determination (R2) between the predicted and manually annotated film coverage was 0.92, with root mean square error (RMSE) of 0.23%, and average relative error of 2.95%. The average evaluation time was 0.54 s per image. This method demonstrated high accuracy and rapid processing capabilities for real-time monitoring and evaluation of residual film coverage post-recovery, providing robust technical support for quality assessment in residual film recovery operations.

    参考文献
    相似文献
    引证文献
引用本文

吕继东,翟志强,孟庆建,苗璐鹏,陈悦,张若宇.基于车载成像与深度卷积神经网络的地表残膜识别方法[J].农业机械学报,2025,56(5):26-37,70. Lü Jidong, ZHAI Zhiqiang, MENG Qingjian, MIAO Lupeng, CHEN Yue, ZHANG Ruoyu. Surface Residual Film Recognition Method Based on Vehicle-mounted Imaging and Deep Convolutional Neural Networks[J]. Transactions of the Chinese Society for Agricultural Machinery,2025,56(5):26-37,70.

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2024-11-04
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2025-05-10
  • 出版日期:
文章二维码