基于实例分割的番茄串视觉定位与采摘姿态估算方法
作者:
作者单位:

作者简介:

通讯作者:

中图分类号:

基金项目:

广东省现代农业产业共性关键技术研发创新团队建设项目(2019KJ129)


Visual Positioning and Picking Pose Estimation of Tomato Clusters Based on Instance Segmentation
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    准确识别定位采摘点,根据果梗方向,确定合适的采摘姿态,是机器人实现高效、无损采摘的关键。由于番茄串的采摘背景复杂,果实颜色、形状各异,果梗姿态多样,叶子藤枝干扰等因素,降低了采摘点识别准确率和采摘成功率。针对这个问题,考虑番茄串生长特性,提出基于实例分割的番茄串视觉定位与采摘姿态估算方法。首先基于YOLACT实例分割算法的实例特征标准化和掩膜评分机制,保证番茄串和果梗感兴趣区域 (Region of interest, ROI)、掩膜质量和可靠性,实现果梗粗分割;通过果梗掩膜信息和ROI位置关系匹配可采摘果梗,基于细化算法、膨胀操作和果梗形态特征实现果梗精细分割;再通过果梗深度信息填补法与深度信息融合,精确定位采摘点坐标。然后利用果梗几何特征、八邻域端点检测算法识别果梗关键点预测果梗姿态,并根据果梗姿态确定适合采摘的末端执行器姿态,引导机械臂完成采摘。研究和大量现场试验结果表明,提出的方法在复杂采摘环境中具有较高的定位精度和稳定性,对4个品种的番茄串采摘点平均识别成功率为98.07%,图像分辨率为1280像素×720像素时算法处理速率达到21f/s,采摘点图像坐标最大定位误差为3像素,深度误差±4mm,成功定位采摘点后采摘成功率为98.15%。与现有的同类方法相比,采摘点图像坐标定位精度提高76.80个百分点,采摘成功率提高15.17个百分点,采摘效率提高31.18个百分点,满足非结构化种植环境中番茄串采摘需求。

    Abstract:

    Recognizing and positioning the picking points (spatial position and coordinate points), and determining the appropriate picking pose according to the direction of fruit stem, are the keys for the robot to achieve efficient and lossless picking during harvesting. However, the harvesting environment is complex and changeable, the color of fruit stem is similar to the branches and leaves, and the tomato clusters are always with different colors and shapes. Furthermore, tomato clusters grow in different directions, and the end effector frequently interferes with the leaves and vine during picking, there are often situations of “not picking when robot see it”, which reduces the recognition accuracy of picking points and picking rate. Aiming at this problem, considering the growth characteristics of tomato clusters, a method for visual positioning and picking pose estimation of tomato clusters based on instance segmentation was proposed. Firstly, based on the instance feature standardization, and the mask scoring mechanism of the YOLACT algorithm, the high quality and reliable region of interest (ROIs) and masks of tomato clusters were collected. Specifically, in order to efficiently achieve the coarse segmentation of fruit stems via the YOLACT. Then, according to the stem mask information and the neighbor relationship between tomato ROIs and stem ROIs, the ROIs of pickable stems were determined. Meanwhile, the pickable stem edges were finely extracted from the stem ROI by using the thinning algorithm, together with expansion operation and shape characteristics of stem. Secondly, the picking point in image coordinate system was obtained, which was set as the center point of stem skeleton along the X (or Y) axis. Subsequently, the depth map of pickable stem ROI was used for obtaining the original depth value of picking point. Specifically, due to the large depth value errors, or even a lack of depth values when capturing small objects by the economical RGB-D depth camera. By using only the depth map corresponding to the stem mask area, the average depth value of picking point was calculated. The accurate depth value of picking point was obtained by comparing the average with the original depth value. Thirdly, according to the geometric features of fruit stem, the tangent slope of fruit stem at the picking point was calculated, and the search algorithm was used for finding the endpoints of fruit stem. Correspondingly, fruit stem direction was estimated by the vector composed of two endpoints of fruit stem. Finally, the picking point was converted to the robot coordinate system. Simultaneously, according to the tangent slope and the direction of fruit stem, the picking pose of the end effector was determined. Eventually, the robot was guided to complete the picking task with an appropriate pose. A large number of field test verified that the average recognition rate of pickings point was 98.07%, while the image resolution was 1280 pixel×720 pixel, the processing rate of the algorithm was 21f/s, the maximum positioning error of the image coordinates of picking points was 3 pixels, and the depth value error was ±4mm. After the picking points were successfully positioned, the picking rate was 98.15%. Compared with the existing similar methods, the positioning accuracy of the picking point was increased by 76.80 percentage points, the picking rate was increased by 15.17 percentage points, and picking efficiency was increased by 31.18 percentage points. Therefore, the proposed method fully met the requirements for robots in unstructured environment during harvesting.

    参考文献
    相似文献
    引证文献
引用本文

张勤,庞月生,李彬.基于实例分割的番茄串视觉定位与采摘姿态估算方法[J].农业机械学报,2023,54(10):205-215. ZHANG Qin, PANG Yuesheng, LI Bin. Visual Positioning and Picking Pose Estimation of Tomato Clusters Based on Instance Segmentation[J]. Transactions of the Chinese Society for Agricultural Machinery,2023,54(10):205-215.

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2023-04-08
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2023-05-03
  • 出版日期: