吕石磊, 卢思华, 李震, 洪添胜, 薛月菊, 吴奔雷. 基于改进YOLOv3-LITE轻量级神经网络的柑橘识别方法[J]. 农业工程学报, 2019, 35(17): 205-214. DOI: 10.11975/j.issn.1002-6819.2019.17.025
    引用本文: 吕石磊, 卢思华, 李震, 洪添胜, 薛月菊, 吴奔雷. 基于改进YOLOv3-LITE轻量级神经网络的柑橘识别方法[J]. 农业工程学报, 2019, 35(17): 205-214. DOI: 10.11975/j.issn.1002-6819.2019.17.025
    Lü Shilei, Lu Sihua, Li Zhen, Hong Tiansheng, Xue Yueju, Wu Benlei. Orange recognition method using improved YOLOv3-LITE lightweight neural network[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2019, 35(17): 205-214. DOI: 10.11975/j.issn.1002-6819.2019.17.025
    Citation: Lü Shilei, Lu Sihua, Li Zhen, Hong Tiansheng, Xue Yueju, Wu Benlei. Orange recognition method using improved YOLOv3-LITE lightweight neural network[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2019, 35(17): 205-214. DOI: 10.11975/j.issn.1002-6819.2019.17.025

    基于改进YOLOv3-LITE轻量级神经网络的柑橘识别方法

    Orange recognition method using improved YOLOv3-LITE lightweight neural network

    • 摘要: 柑橘识别是实现柑橘园果实自动采摘、果树精细化管理以及实现果园产量预测的关键技术环节。为实现自然环境下柑橘果实的快速精准识别,该文提出一种基于改进YOLOv3-LITE轻量级神经网络的柑橘识别方法。在采摘机器人领域,果实识别回归框的准确率直接决定了机器手的采摘成功率,该方法通过引入GIoU边框回归损失函数来提高果实识别回归框准确率;为便于迁移到移动终端,提出一种YOLOv3-LITE轻量级网络模型,使用MobileNet-v2作为模型的骨干网络;使用混合训练与迁移学习结合的预训练方式来提高模型的泛化能力。通过与Faster-RCNN以及SSD模型对比在不同遮挡程度的测试样本下模型的识别效果,用F1值与AP值评估各模型的差异,试验结果表明:该文提出的模型识别效果提升显著,对于果实轻度遮挡的数据集,该文提出的柑橘识别模型的F1值和AP值分别为95.27%和92.75%,Average IoU为90.65%;在全部测试集上,F1值和AP值分别为93.69%和91.13%,Average IoU为87.32%,在GPU上对柑橘目标检测速度可达246帧/s,对单张416×416的图片推断速度为16.9ms,在CPU上检测速度可达22帧/s,推断速度为80.9 ms,模型占用内存为28 MB。因此,该文提出的柑橘识别方法具有模型占用内存低、识别准确率高及识别速度快等优点,可为柑橘采摘机器人以及柑橘产业产量预测提出新的解决方案,为柑橘产业智能化提供新的思路。

       

      Abstract: Abstract: Orange recognition is one of the key technologies for automatical fruits picking, delicacy management and orchard yield forecast etc. In order to realize the rapid and accurate identification of orange fruits in natural environment, this paper presents an orange recognition method using improved YOLOv3-LITE lightweight neural network. In the field of fruit-picking robots, the accuracy of bounding box regression for fruit recognition directly determines the success rate of robotic hand-picking. In this paper, the proposed method, which introduces GIoU bounding box regression loss function (GIoU reflects the relationship between the target frame and forecast box) to replace traditional loss function of MSE (mean square error) part of the bounding box regression, effectively improves the accuracy of bounding boxregression. In order to facilitate the migration of the model to mobile terminals, theYOLOv3-LITE lightweight neural network model is proposed. Mobilenet-v2, a lightweight network proposed based on the mobile terminal, can effectively reduce the complexity of the model, and is thus used in this paper to replace the backbone network Darknet-53 in the original network model. The mixup-training method combines 2 pictures into 1 with a certain weight to achieve the effect of target occlusion, reduce the impact of differences between images, improve the generalization ability of the model, and reduce over fitting. In this paper, the mixup-training is combined with transfer learning for the purpose of pre-training. The deep learning framework built in the proposed model is TensorFlow and DarkNet. In order to verify the superiority and feasibility of the proposed model, its recognition effect is tested on test samples with different obscured degrees, comparing with the Faster-RCNN, SSD models and original YOLOv3 network model. The values of F1 and AP are used to assess the differences among models. Test results show that, the model proposed in this paper can get significantly good recognition results. In the data sets of lightly obscured fruit, the F1 value and AP value of the orange recognition model proposed in this paper can reach 95.27% and 92.75%, and the Average IoU is as high as 90.65%. In the data sets of severely obscured fruit, the F1 value and AP value of the orange recognition model proposed in this paper are 91.43% and 89.10%, respectively, and the Average IoU is 83.73%. In all the test sets, the F1 and AP values are 93.69% and 91.13%, respectively, and Average IoU is 87.32%. Compared with the original YOLOv3, the values of F1and AP increase by 1.77% and 2.43%, respectively, and Average IoU increases by 4.72%. The detection speed of orange target can reach 246 frames per second; the inferred speed of a single 416×416 picture on GPU is 16.9ms. The detection speed on the CPU can reach 22 frames per second, and the inferred speed can reach 80.9ms. The memory occupied by the proposed model is 28 MB. Moreover, the pre-training method, combined with the transfer learning and mixup training, reduces the training time and memory consumption of the proposed model. Therefore, the orange recognition method proposed in this paper has the advantages of low model memory footprint, high recognition accuracy and fast recognition speed, etc. It can provide new solutions for orange picking robot and orchard yield forecast, and new ideas for intelligent orange industry as well.

       

    /

    返回文章
    返回