Back to Search
Start Over
Full Convolution Neural Network Combined with Contextual Feature Representation for Cropland Extraction from High-Resolution Remote Sensing Images.
- Source :
-
Remote Sensing . May2022, Vol. 14 Issue 9, p2157-2157. 24p. - Publication Year :
- 2022
-
Abstract
- The quantity and quality of cropland are the key to ensuring the sustainable development of national agriculture. Remote sensing technology can accurately and timely detect the surface information, and objectively reflect the state and changes of the ground objects. Using high-resolution remote sensing images to accurately extract cropland is the basic task of precision agriculture. The traditional model of cropland semantic segmentation based on the deep learning network is to down-sample high-resolution feature maps to low resolution, and then restore from low-resolution feature maps to high-resolution ideas; that is, obtain low-resolution feature maps through a network, and then recover to high resolution by up-sampling or deconvolution. This will bring about the loss of features, and the segmented image will be more fragmented, without very clear and smooth boundaries. A new methodology for the effective and accurate semantic segmentation cropland of high spatial resolution remote sensing images is presented in this paper. First, a multi-temporal sub-meter cropland sample dataset is automatically constructed based on the prior result data. Then, a fully convolutional neural network combined with contextual feature representation (HRNet-CFR) is improved to complete the extraction of cropland. Finally, the initial semantic segmentation results are optimized by the morphological post-processing approach, and the broken spots are ablated to obtain the internal homogeneous cropland. The proposed method has been validated on the Jilin-1 data and Gaofen Image Dataset (GID) public datasets, and the experimental results demonstrate that it outperforms the state-of-the-art method in cropland extraction accuracy. We selected the comparison of Deeplabv3+ and UPerNet methods in GID. The overall accuracy of our approach is 92.03%, which is 3.4% higher than Deeplabv3+ and 5.12% higher than UperNet. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 20724292
- Volume :
- 14
- Issue :
- 9
- Database :
- Academic Search Index
- Journal :
- Remote Sensing
- Publication Type :
- Academic Journal
- Accession number :
- 156874495
- Full Text :
- https://doi.org/10.3390/rs14092157