Multi-Temporal Unmanned Aerial Vehicle Remote Sensing for Vegetable Mapping Using an Attention-Based Recurrent Convolutional Neural Network
Open Access
- 21 May 2020
- journal article
- research article
- Published by MDPI AG in Remote Sensing
- Vol. 12 (10), 1668
- https://doi.org/10.3390/rs12101668
Abstract
Vegetable mapping from remote sensing imagery is important for precision agricultural activities such as automated pesticide spraying. Multi-temporal unmanned aerial vehicle (UAV) data has the merits of both very high spatial resolution and useful phenological information, which shows great potential for accurate vegetable classification, especially under complex and fragmented agricultural landscapes. In this study, an attention-based recurrent convolutional neural network (ARCNN) has been proposed for accurate vegetable mapping from multi-temporal UAV red-green-blue (RGB) imagery. The proposed model firstly utilizes a multi-scale deformable CNN to learn and extract rich spatial features from UAV data. Afterwards, the extracted features are fed into an attention-based recurrent neural network (RNN), from which the sequential dependency between multi-temporal features could be established. Finally, the aggregated spatial-temporal features are used to predict the vegetable category. Experimental results show that the proposed ARCNN yields a high performance with an overall accuracy of 92.80%. When compared with mono-temporal classification, the incorporation of multi-temporal UAV imagery could significantly boost the accuracy by 24.49% on average, which justifies the hypothesis that the low spectral resolution of RGB imagery could be compensated by the inclusion of multi-temporal observations. In addition, the attention-based RNN in this study outperforms other feature fusion methods such as feature-stacking. The deformable convolution operation also yields higher classification accuracy than that of a standard convolution unit. Results demonstrate that the ARCNN could provide an effective way for extracting and aggregating discriminative spatial-temporal features for vegetable mapping from multi-temporal UAV RGB imagery.Funding Information
- China Postdoctoral Science Foundation (2018M641529, 2019T120155)
- National Key Research and Development Program of China (2018YFE0122700)
This publication has 54 references indexed in Scilit:
- Fully Convolutional Networks for Semantic SegmentationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2016
- Bag-of-Visual-Words Scene Classifier With Local and Global Features for High Spatial Resolution Remote Sensing ImageryIEEE Geoscience and Remote Sensing Letters, 2016
- Classification of riparian forest species and health condition using multi-temporal and hyperspatial imagery from unmanned aerial systemEnvironmental Monitoring and Assessment, 2016
- Monitoring Cropland Dynamics of the Yellow River Delta based on Multi-Temporal Landsat Imagery over 1986 to 2015Sustainability, 2015
- Rule-based land use/land cover classification in coastal areas using seasonal remote sensing imagery: a case study from Lianyungang City, ChinaEnvironmental Monitoring and Assessment, 2015
- Deep learningNature, 2015
- Urban Flood Mapping Based on Unmanned Aerial Vehicle Remote Sensing and Random Forest Classifier—A Case of Yuyao, ChinaWater, 2015
- UAV Remote Sensing for Urban Vegetation Mapping Using Random Forest and Texture AnalysisRemote Sensing, 2015
- Multi-temporal mapping of the vegetation fraction in early-season wheat fields using images from UAVComputers and Electronics in Agriculture, 2014
- Long Short-Term MemoryNeural Computation, 1997