基于改进DeepLab V3+的果园场景多类别分割方法
作者:
作者单位:

作者简介:

通讯作者:

中图分类号:

基金项目:

国家自然科学基金项目(32171908)、江苏省现代农机装备与技术示范推广项目(NJ2021-14)、宁夏回族自治区重点研发计划重大项目(2018BBF02020)、江苏省重点研发计划项目(BE2018372)和江苏高校优势学科建设工程项目(PAPD)


Multi-category Segmentation of Orchard Scene Based on Improved DeepLab V3+
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    果园环境实时检测是保证果园喷雾机器人精准作业的重要前提。本文提出了一种基于改进DeepLab V3+语义分割模型的果园场景多类别分割方法。为了在果园喷雾机器人上部署,使用轻量化MobileNet V2网络替代原有的Xception网络以减少网络参数,并在空洞空间金字塔池化(Atrous spatial pyramid pooling,ASPP)模块中运用ReLU6激活函数减少部署在移动设备的精度损失,此外结合混合扩张卷积(Hybrid dilated convolution,HDC),以混合扩张卷积替代原有网络中的空洞卷积,将ASPP中的扩张率设为互质以减少空洞卷积的网格效应。使用视觉传感器采集果园场景RGB图像,选取果树、人、天空等8类常见的目标制作了数据集,并在该数据集上基于Pytorch对改进前后的DeepLab V3+进行训练、验证和测试。结果表明,改进后DeepLab V3+模型的平均像素精度、平均交并比分别达到62.81%和56.64%,比改进前分别提升5.52、8.75个百分点。模型参数量较改进前压缩88.67%,单幅图像分割时间为0.08s,与原模型相比减少0.09s。尤其是对树的分割精度达到95.61%,比改进前提高1.31个百分点。该方法可为喷雾机器人精准施药和安全作业提供有效决策,具有实用性。

    Abstract:

    Real-time detection of orchard environment is an important prerequisite to ensure the accurate operation of orchard spray robot. An improved DeepLab V3+ semantic segmentation model was proposed for multi-category segmentation in orchard scene. For deployment on the orchard spray robot, the lightweight MobileNet V2 network was used to replace the original Xception network to reduce the network parameters, and ReLU6 activation function was applied in atrous spatial pyramid pooling (ASPP) module to reduce the loss of accuracy when deployed in mobile devices. In addition, hybrid dilated convolution (HDC) was combined to replace the void convolution in the original network. The dilated rates in ASPP were prime to each other to reduce the grid effect of dilated convolution. The RGB images of orchard scene were collected by using visual sensor, and eight common targets were selected to make the dataset, such as fruit trees, pedestrians and sky. On this dataset, DeepLab V3+ before and after improvement was trained, verified and tested based on Pytorch. The results showed that the mean pixel accuracy and mean intersection over union of the improved Deeplab V3+ model were 62.81% and 56.64%, respectively, which were 5.52 percentage points and 8.75 percentage points higher than before improvement. Compared with the original model, the parameters were reduced by 88.67%. The segmentation time of a single image was 0.08s, which was 0.09s less than the original model. In particular, the accuracy of tree segmentation reached 95.61%, which was 1.31 percentage points higher than before improvement. This method can provide an effective decision for precision spraying and safe operation of the spraying robot, and it was practical.

    参考文献
    相似文献
    引证文献
引用本文

刘慧,姜建滨,沈跃,贾卫东,曾潇,庄珍珍.基于改进DeepLab V3+的果园场景多类别分割方法[J].农业机械学报,2022,53(11):255-261.

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2022-01-13
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2022-11-10
  • 出版日期: