冯青春, 成伟, 李亚军, 王博文, 陈立平. 基于Mask R-CNN的番茄植株整枝操作点定位方法[J]. 农业工程学报, 2022, 38(3): 128-135. DOI: 10.11975/j.issn.1002-6819.2022.03.015
    引用本文: 冯青春, 成伟, 李亚军, 王博文, 陈立平. 基于Mask R-CNN的番茄植株整枝操作点定位方法[J]. 农业工程学报, 2022, 38(3): 128-135. DOI: 10.11975/j.issn.1002-6819.2022.03.015
    Feng Qingchun, Cheng Wei, Li Yajun, Wang Bowen, Chen Liping. Method for identifying tomato plants pruning point using Mask R-CNN[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2022, 38(3): 128-135. DOI: 10.11975/j.issn.1002-6819.2022.03.015
    Citation: Feng Qingchun, Cheng Wei, Li Yajun, Wang Bowen, Chen Liping. Method for identifying tomato plants pruning point using Mask R-CNN[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2022, 38(3): 128-135. DOI: 10.11975/j.issn.1002-6819.2022.03.015

    基于Mask R-CNN的番茄植株整枝操作点定位方法

    Method for identifying tomato plants pruning point using Mask R-CNN

    • 摘要: 针对工厂化番茄智能化整枝打叶作业需要,研究了基于Mask R-CNN模型的整枝操作点识别定位方法,以期为整枝机器人的精准操作提供依据。鉴于丛生植株中主茎和侧枝茎秆目标随机生长、形态各异,结合植株在不同生长阶段、远近视场尺度和观测视角等条件下的成像特征,构建了温室番茄植株图像样本数据集。采用学习率微调训练方法,对Mask R-CNN预训练模型进行迁移训练,建立了主茎和侧枝像素区域的识别分割模型。在对视场内同株相邻主茎和侧枝目标进行判别基础上,提出基于图像矩特征的茎秆中心线拟合方法。以中心线交点为参考,沿侧枝进行定向偏移,实现对整枝操作点图像坐标的定位。最后,通过测试试验评估该方法对目标识别和定位的效果。试验结果表明,模型对番茄主茎和侧枝目标识别的错误率、精确率和召回率分别为0.12、0.93和0.94,对整枝操作点平均定位偏差与对应主茎像素宽度的比值为0.34,模型对于近景仰视图像中目标的识别和定位效果优于其他视场的图像。该研究可为整枝机器人视觉系统的研发提供技术依据。

       

      Abstract: An accurate identification of the pruning point was proposed for the stem segmentation among the complex background with clustered plants using the improved Mask RCNN. As such, the robot was guided to precisely operate at the pruning point for tomatoes in the greenhouse. The standard requirement was then introduced on the pruning lateral branch of the tomato plant in the greenhouse. The pruning point was determined, where the intersection of center lines between the main stem and lateral branch was taken as a reference. The Mask R-CNN with feature extraction network of ResNet50 was adopted as the target area segmentation model. An images dataset of tomato plant was constructed to consider the growth stage of tomato plant (the growing and productive period of the plant), the view field scale (the close and distant view, upward and front view), and imaging posture. After that, 3 000 images were collected, among which 2 400 ones were set as a training set, A fine-tuning method was adopted to transfer the pre-training model from the MMDetection algorithms library. The various learning rates of 0.02 and 0.002 were used for the head net and backbone net, respectively. A segmentation model was established for the main stem and the lateral branch using the Mask R-CNN, with the loss was tended to convergence with training iteration. Two types of target areas were identified and located to form the plants image. The main stem and lateral branch adjacent to each other were identified as the same plant, according to the distance of the center point. The centerline of the stem was fitted, according to the second central moment feature. The pruning point was located as the point, with an offset distance of the main stem’s radius along the centerline of the lateral branch. The performance of the model was verified to identify and locate the pruning point on the test set with 80 images, excluding both in the train set and validation set. The experimental results showed that the error rate, precision rate, and recall rate on the tomato stem target’s recognition were 0.12, 0.93, and 0.94, respectively. In particular, the error rate on the main stem was less than that on the lateral branch, where the occlusion of leaves was the main reason for the false detection. Besides, the fruit stem in the front view was occasionally identified as the lateral branch. In addition, the upward view images of the plants in the productive period presented a higher accuracy rate, in which fewer leaves and fruit stems existed. In terms of location, the average error of locating pruning points was 0.34 of the diameter of the main stem, and the upward view images from the plant in the production period presented less locating error. As the empirical value of main-stem diameter was set as 15 mm, the average locating error was 5.12 mm, indicating that was easily tolerated by an additional movement of the stem grasper. Furthermore, the location error was also lower for the image of the close-upwards view. The finding can provide promising technical support to the tomato pruning robots.

       

    /

    返回文章
    返回