Processing math: 100%
BSNet: Dynamic Hybrid Gradient Convolution Based Boundary-Sensitive Network for Remote Sensing Image Segmentation | IEEE Journals & Magazine | IEEE Xplore

BSNet: Dynamic Hybrid Gradient Convolution Based Boundary-Sensitive Network for Remote Sensing Image Segmentation


Abstract:

Boundary information is essential for the semantic segmentation of remote sensing images. However, most existing methods were designed to establish strong contextual info...Show More

Abstract:

Boundary information is essential for the semantic segmentation of remote sensing images. However, most existing methods were designed to establish strong contextual information while losing detailed information, making it challenging to extract and recover boundaries accurately. In this article, a boundary-sensitive network (BSNet) is proposed to address this problem via dynamic hybrid gradient convolution (DHGC) and coordinate sensitive attention (CSA). Specifically, in the feature extraction stage, we propose DHGC to replace vanilla convolution (VC), which adaptively aggregates one VC kernel and two gradient convolution kernels (GCKs) into a new operator to enhance boundary information extraction. The GCKs are proposed to explicitly encode boundary information, which is inspired by traditional Sobel operators. In the feature recovery stage, the CSA is introduced. This module is used to reconstruct the sharp and detailed segmentation results by adaptively modeling the boundary information and long-range dependencies in the low-level features as the assistance of high-level features. Note that DHGC and CSA are plug-and-play modules. We evaluate the proposed BSNet on three public datasets: the ISPRS 2-D semantic labeling Vaihingen, the Potsdam benchmark, and the iSAID dataset. The experimental results indicate that BSNet is a highly effective architecture that produces sharper predictions around object boundaries and significantly improves the segmentation accuracy. Our method demonstrates superior performance on the Vaihingen, the Potsdam benchmark, and the iSAID dataset in terms of the mean F_{1} , with improvements of 4.6%, 2.3%, and 2.4% over strong baselines, respectively. The code and models will be made publicly available.
Article Sequence Number: 5624022
Date of Publication: 18 May 2022

ISSN Information:

References is not available for this document.

I. Introduction

With the rapid development of remote sensing technology, an increasing number of remote sensing images with high resolution (e.g., 5–10 cm) can be obtained. In these images, small objects, such as cars and trees, can be clearly observed, which makes pixel-level semantic segmentation possible. Image semantic segmentation allocates pixelwise semantic labels in an image, segmenting the image into several regions based on semantic units. Semantic segmentation of remote sensing images has a wide range of applications, including environmental monitoring [1], postdisaster reconstruction [2], agriculture [3], forestry [4], and urban planning [5], [6]. Traditional image semantic segmentation methods [7]–[10] manually design feature classifiers but do not achieve satisfactory performance due to their limited feature extraction and data fitting abilities. These shortcomings have prompted researchers to search for more stable and efficient methods.

Select All
1.
L. Lopez-Fuentes, C. Rossi, H. Skinnemoen et al., "River segmentation for flood monitoring", Proc. IEEE Int. Conf. Big Data (Big Data), pp. 3746-3749, Dec. 2017.
2.
K. Nogueira et al., "Exploiting convnet diversity for flooding identification", IEEE Geosci. Remote Sens. Lett., vol. 15, no. 9, pp. 1446-1450, Sep. 2018.
3.
H. Sheng, X. Chen, J. Su, R. Rajagopal and A. Ng, "Effective data fusion with generalized vegetation index: Evidence from land cover segmentation in agriculture", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 267-276, Jun. 2020.
4.
L. Mi and Z. Chen, "Superpixel-enhanced deep neural forest for remote sensing image semantic segmentation", ISPRS J. Photogramm. Remote Sens., vol. 159, pp. 140-152, 2020.
5.
M. M. Nielsen, "Remote sensing for urban planning and management: The use of window-independent context segmentation to extract urban features in Stockholm", Comput. Environ. Urban Syst., vol. 52, pp. 1-9, Jul. 2015.
6.
Y. Liu, L. Gross, Z. Li, X. Li, X. Fan and W. Qi, "Automatic building extraction on high-resolution remote sensing imagery using deep convolutional encoder-decoder with spatial pyramid pooling", IEEE Access, vol. 7, pp. 128774-128786, 2019.
7.
M. Fauvel, Y. Tarabalka, J. A. Benediktsson, J. Chanussot and J. C. Tilton, "Advances in spectral-spatial classification of hyperspectral images", Proc. IEEE, vol. 101, no. 3, pp. 652-675, Mar. 2013.
8.
L. Gueguen, "Classifying compound structures in satellite images: A compressed representation for fast queries", IEEE Trans. Geosci. Remote Sens., vol. 53, no. 4, pp. 1803-1818, Apr. 2015.
9.
F. Li, X. Jia, D. Fraser and A. Lambert, "Super resolution for remote sensing images based on a universal hidden Markov tree model", IEEE Trans. Geosci. Remote Sens., vol. 48, no. 3, pp. 1270-1278, Mar. 2010.
10.
F. Tupin and M. Roux, "Markov random field on region adjacency graph for the fusion of SAR and optical data in radargrammetric applications", IEEE Trans. Geosci. Remote Sens., vol. 43, no. 8, pp. 1920-1928, Aug. 2005.
11.
J. Long, E. Shelhamer and T. Darrell, "Fully convolutional networks for semantic segmentation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 3431-3440, Jun. 2015.
12.
H. Noh, S. Hong and B. Han, "Learning deconvolution network for semantic segmentation", Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 1520-1528, Dec. 2015.
13.
V. Badrinarayanan, A. Kendall and R. Cipolla, "SegNet: A deep convolutional encoder-decoder architecture for image segmentation", IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 12, pp. 2481-2495, Jan. 2017.
14.
A. Li, L. Jiao, H. Zhu, L. Li and F. Liu, "Multitask semantic boundary awareness network for remote sensing image segmentation", IEEE Trans. Geosci. Remote Sens., vol. 60, pp. 1-14, 2022.
15.
S. Liu, W. Ding, C. Liu, Y. Liu, Y. Wang and H. Li, "ERN: Edge loss reinforced semantic segmentation network for remote sensing images", Remote Sens., vol. 10, no. 9, pp. 1339, Aug. 2018.
16.
G. Cheng, X. Xie, J. Han, L. Guo and G.-S. Xia, "Remote sensing image scene classification meets deep learning: Challenges methods benchmarks and opportunities", IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 13, no. 99, pp. 3735-3756, Jun. 2020.
17.
G. Cheng, C. Yang, X. Yao, L. Guo and J. Han, "When deep learning meets metric learning: Remote sensing image scene classification via learning discriminative CNNs", IEEE Trans. Geosci. Remote Sens., vol. 56, no. 5, pp. 2811-2821, May 2018.
18.
Y. Li, T. Shi, Y. Zhang, W. Chen, Z. Wang and H. Li, "Learning deep semantic segmentation network under multiple weakly-supervised constraints for cross-domain remote sensing image semantic segmentation", ISPRS J. Photogramm. Remote Sens., vol. 175, pp. 20-33, May 2021.
19.
Y. Li, W. Chen, Y. Zhang, C. Tao, R. Xiao and Y. Tan, "Accurate cloud detection in high-resolution remote sensing imagery by weakly supervised deep learning", Remote Sens. Environ., vol. 250, Dec. 2020.
20.
T. Takikawa, D. Acuna, V. Jampani and S. Fidler, "Gated-SCNN: Gated shape CNNs for semantic segmentation", Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 5228-5237, Nov. 2019.
21.
L.-C. Chen, J. T. Barron, G. Papandreou, K. Murphy and A. L. Yuille, "Semantic image segmentation with task-specific edge detection using CNNs and a discriminatively trained domain transform", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 4545-4554, Jun. 2016.
22.
K. Gong, X. Liang, Y. Li, Y. Chen, M. Yang and L. Lin, "Instance-level human parsing via part grouping network", CoRR, vol. abs/1808.00157, 2018, [online] Available: http://arxiv.org/abs/1808.00157.
23.
X. Li et al., "Semantic flow for fast and accurate scene parsing", CoRR, vol. abs/2002.10120, 2020, [online] Available: https://arxiv.org/abs/2002.10120.
24.
H. Zhao, J. Shi, X. Qi, X. Wang and J. Jia, "Pyramid scene parsing network", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 6230-6239, Jul. 2017.
25.
L.-C. Chen, G. Papandreou, F. Schroff and H. Adam, "Rethinking atrous convolution for semantic image segmentation" in arXiv:1706.05587, 2017.
26.
L. Chen, Y. Zhu, G. Papandreou, F. Schroff and H. Adam, "Encoder-decoder with atrous separable convolution for semantic image segmentation", CoRR, vol. abs/1802.02611, 2018, [online] Available: http://arxiv.org/abs/1802.02611.
27.
Y. Wang, W. Ding, R. Zhang and H. Li, "Boundaryaware multitask learning for remote sensing imagery", IEEE J. Sel. Topics. Appl. Earth Observ. Remote. Sens., vol. 14, pp. 951-963, 2021.
28.
G. Yang, Q. Zhang and G. Zhang, "EANet: Edge-aware network for the extraction of buildings from aerial images", Remote Sens., vol. 12, no. 13, pp. 2161, Jul. 2020.
29.
X. Sun, A. Shi, H. Huang and H. Mayer, "BAS⁴Net: Boundary-aware semi-supervised semantic segmentation network for very high resolution remote sensing images", IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 13, pp. 5398-5413, 2020.
30.
Z. Su et al., "Pixel difference networks for efficient edge detection" in arXiv:2108.07009, 2021.
Contact IEEE to Subscribe

References

References is not available for this document.