蔡舒平, 孙仲鸣, 刘慧, 吴翃轩, 庄珍珍. 基于改进型YOLOv4的果园障碍物实时检测方法[J]. 农业工程学报, 2021, 37(2): 36-43. DOI: 10.11975/j.issn.1002-6819.2021.2.005
    引用本文: 蔡舒平, 孙仲鸣, 刘慧, 吴翃轩, 庄珍珍. 基于改进型YOLOv4的果园障碍物实时检测方法[J]. 农业工程学报, 2021, 37(2): 36-43. DOI: 10.11975/j.issn.1002-6819.2021.2.005
    Cai Shuping, Sun Zhongming, Liu Hui, Wu Hongxuan, Zhang Zhenzhen. Real-time detection methodology for obstacles in orchards using improved YOLOv4[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2021, 37(2): 36-43. DOI: 10.11975/j.issn.1002-6819.2021.2.005
    Citation: Cai Shuping, Sun Zhongming, Liu Hui, Wu Hongxuan, Zhang Zhenzhen. Real-time detection methodology for obstacles in orchards using improved YOLOv4[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2021, 37(2): 36-43. DOI: 10.11975/j.issn.1002-6819.2021.2.005

    基于改进型YOLOv4的果园障碍物实时检测方法

    Real-time detection methodology for obstacles in orchards using improved YOLOv4

    • 摘要: 针对农业机器人在复杂的果园环境中作业时需要精确快速识别障碍物的问题,该研究提出了一种改进型的YOLOv4目标检测模型对果园障碍物进行分类和识别。为了减少改进后模型的参数数量并提升检测速度,该研究使用了深度可分离卷积代替模型中原有的标准卷积,并将主干网络CSP-Darknet中的残差组件(Residual Unit)改进为逆残差组件(Inverted Residual Unit)。此外,为了进一步增强模型对目标密集区域的检测能力,使用了软性非极大值抑制(Soft DIoU-Non-Maximum Suppression,Soft-DIoU-NMS)算法。为了验证该研究所提方法的有效性,选取果园中常见的3种障碍物作为检测对象制作图像数据集,在Tensorflow深度学习框架上训练模型。然后将测试图片输入训练好的模型中检测不同距离下的目标障碍物,并在同一评价指标下,将该模型的测试结果与改进前YOLOv4模型的测试结果进行评价对比。试验结果表明,改进后的YOLOv4果园障碍物检测模型的平均准确率和召回率分别为96.92%和91.43%,视频流检测速度为58.5帧/s,相比于原模型,改进后的模型在不损失精度的情况下,将模型大小压缩了75%,检测速度提高了29.4%。且改进后的模型具有鲁棒性强、实时性更好、轻量化的优点,能够更好地实现果园环境下障碍物的检测,为果园智能机器人的避障提供了有力的保障。

       

      Abstract: Abstract: China is one of the countries with the largest cultivation area of the orchards in the world. The traditional orchard planting is quite time-consuming and laborious. An orchard robot can be expected as an important artificial intelligence (AI) tool to replace the manual labor in orchard management. However, the robot can encounter various obstacles in the actual operation, due to the complex and changeable environment of an orchard. It is necessary for the agricultural robots to real-time detect obstacles during operation. In recent years, various target detection systems have been widely used for agricultural robot avoidance, such as YOLOv4, YOLOv3, and Faster-RCNN, particularly with the rise of intelligent deep learning. Generally, there were some problems, including the unsatisfactory detection accuracy, a large number of parameters required in the models, low real-time performance, and difficulty in detecting densely overlapping target areas. In this study, an improved YOLOv4-based model of target detection was proposed with the help of the latest vision sensor technology to realize that the agricultural robots can quickly and accurately identify and classify the obstacles in the orchard. A deep separable convolution was utilized to reduce the number of parameters, and further improve the detection speed. An Inverted Residual Unit was selected to replace the Residual Unit in the core network CSP-Darknet in the previous model. In addition, a Soft DIoU-Non-Maximum Suppression (Soft-DIoU-NMS) algorithm was employed to detect the dense areas. Three common obstacles, including pedestrians, fruit trees, and telegraph poles, in the orchards were selected as the detection objects to generate an image dataset. The improved model was trained on the Tensorflow deep learning framework, and then the test images were input into the trained model to detect target obstacles at different distances. Under the same evaluation index, an evaluation was made on the improved YOLOv4, the original YOLOv4, YOLOv3, and Faster-RCNN. The results showed that the improved YOLOv4-based detection model for orchard obstacles achieved an average accuracy rate of 96.92%, 0.61 percent point higher than that of the original YOLOv4 model, 4.18 percent point higher than that of the YOLOv3 model, and 0.04 percent point higher than that of Faster-RCNN model. The recall rate of the proposed model reached 96.31%, 0.68 percent point higher than that of the original YOLOv4, 6.37 percent point higher than that of YOLOv3, and 0.18 percent point higher than that of Faster-RCNN. The detection speed in the improved YOLOv4-based video stream was 58.5 frames/s, 29.4% faster than that in the original YOLOv4, 22.1% faster than that in YOLOv3, and 346% faster than that in Faster-RCNN. The number of parameters in the improved YOLOv4-based model was reduced by 75%, compared with the original YOLOv4 model, 68.7% less than that of the YOLOv3 model, and 81% less than that of the Fasters-RCNN model. In general, the proposed model can greatly reduce its size without losing accuracy, and thereby enhance the real-time performance and robustness in the actual orchard environment. The improved YOLOv4-based model achieved ideal effects in different distance tests, indicating better performance for the obstacle detection in the orchard environment. The findings can provide a strong guarantee for the obstacle avoidance of intelligent robots in orchards.

       

    /

    返回文章
    返回