张志博, 赵西宁, 姜海晨, 袁浩, 杨龙, 高晓东, 石亮亮, 牛雨婷. 基于无人机影像和深度学习的渭北旱塬区土地利用精准分类[J]. 农业工程学报, 2022, 38(22): 199-209. DOI: 10.11975/j.issn.1002-6819.2022.22.022
    引用本文: 张志博, 赵西宁, 姜海晨, 袁浩, 杨龙, 高晓东, 石亮亮, 牛雨婷. 基于无人机影像和深度学习的渭北旱塬区土地利用精准分类[J]. 农业工程学报, 2022, 38(22): 199-209. DOI: 10.11975/j.issn.1002-6819.2022.22.022
    Zhang Zhibo, Zhao Xining, Jiang Haichen, Yuan Hao, Yang Long, Gao Xiaodong, Shi Liangliang, Niu Yuting. Precise classification of land use in Weibei Dryland using UAV images and deep learning[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2022, 38(22): 199-209. DOI: 10.11975/j.issn.1002-6819.2022.22.022
    Citation: Zhang Zhibo, Zhao Xining, Jiang Haichen, Yuan Hao, Yang Long, Gao Xiaodong, Shi Liangliang, Niu Yuting. Precise classification of land use in Weibei Dryland using UAV images and deep learning[J]. Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 2022, 38(22): 199-209. DOI: 10.11975/j.issn.1002-6819.2022.22.022

    基于无人机影像和深度学习的渭北旱塬区土地利用精准分类

    Precise classification of land use in Weibei Dryland using UAV images and deep learning

    • 摘要: 为明确基于无人机影像的旱塬区土地利用精准分类方法,尤其是算法的选择,该研究通过获取渭北旱塬区白水县通积村不同航拍高度无人机正射遥感影像,利用多种深度学习算法和机器学习算法对土地利用分类进行研究。首先,采用大疆御2Pro获取研究区80和160 m不同高度航拍影像;然后对不同航拍高度目视解译结果和多种深度学习、机器学习模型预测结果进行对比分析;最后,基于表现最佳算法对其进行创新和改进。结果表明:深度学习算法的表现远远优于传统机器学习算法,其中深度学习算法中表现最好的DeepLabv3+像素精度为90.06%,比随机森林(Random Forest,RF)和支持向量机(Support Vector Machine,SVM)分别高出24.65和21.32个百分点。基于DeepLabv3+改进的DeepLabv3+_BA模型整体分类效果最好,其像素精度为91.37%,比FCN、SegNet、UNet和DeepLabv3+分别高出7.43、10.12、2.27和1.31个百分点。各种算法在160 m数据集上各指标精度高于80 m。改进模型DeepLabv3+_BA具有较高的地物分类精度及较强的鲁棒性,该研究可为基于无人机影像和深度学习的土地利用信息普查提供技术参考。

       

      Abstract: Abstract: Accurate land use classification is highly required using Unmanned Aerial Vehicle (UAV) images, especially the data selection. In this study, the UAV orthographic remote sensing images were acquired at different aerial heights in Tongji Village, Baishui County, Weibei dry land, China. The land use was then classified using a variety of deep learning and machine learning. The DJI Mavic 2Pro was used to obtain 80 and 160m aerial images in the study area. There were 96 routes, the total length of routes was 42.43 km, the heading overlap degree was 75%, the side overlap degree was 60%, and a total of 2 248 original aerial photos were taken at a flight height of 80 m. At 160 m flight height, there were 20 routes with a total length of 17.90 km, the heading overlap degree was 70%, the side overlap degree was 55%, and a total of 502 original aerial images were taken in this case. The geo-positioning of the photo control points was performed on the Zhuolin A8 handheld Beidou GPS locator. Agisoft PhotoScan 1.4.5 software was used to splice and process the original single-image data. A comparison was made on the visual interpretation of different aerial photography heights and the prediction of various deep learning and machine learning models. Labelme4.5.6 software was used for the visual interpretation. As such, the best performance was achieved during this time. The results show that the performance of deep learning was far better than that of traditional machine learning. The best-performing of deep learning (DeepLabv3+) presented a pixel accuracy of 90.06%, which was 24.65, and 21.32 percentage points higher than that of random forest (RF) and support vector machine (SVM), respectively. The improved DeepLabv3+_BA model performed the best overall classification. The improvement of deep learning was attributed to two aspects. Firstly, the BN layer was removed after the first two separate convolution layers in the Entry flow in the encoder Xception part of the original DeepLabv3+ model. The BN layer was removed in ASPP after the last three separate convolutional layers in the Exit flow. The BN layer was removed after each dilated convolutional layer. Secondly, the ASPP atrous rate combination design was re-optimized, according to the characteristics of the data set. The pixel accuracy of the improved model was 91.37%, which was 7.43, 10.12, 2.27, and 1.31 percentage points higher than those of FCN, SegNet, UNet, and DeepLabv3+, respectively. The number of iterations required for the best accuracy was reduced by about 50%, compared with the other four deep-learning models. Taking the extraction of apple orchard as an example, the F1 value of DeepLabv3+_BA was 89.10%, which was 19.94, 23.68, 2.04, 2.97, 2.4, and 0.78 percentage points higher than those of SVM, RF, FCN, SegNet, UNet, and DeepLabv3+, respectively. The accuracy of various algorithms was higher than 80 m on 160 m datasets. The performance of various deep learning on the test set demonstrated that the accuracy of DeepLabv3+_BA reached more than 90% for the apple orchard, bare field, stubble field, and road ground object classification. The improved model DeepLabv3+_BA presented higher accuracy and robustness of ground object classification. This finding can also provide a strong reference for the land use information census using UAV images and deep learning.

       

    /

    返回文章
    返回