Loading web-font TeX/Math/Italic
Semantic Category Balance-Aware Involved Anti-Interference Network for Remote Sensing Semantic Segmentation | IEEE Journals & Magazine | IEEE Xplore

Semantic Category Balance-Aware Involved Anti-Interference Network for Remote Sensing Semantic Segmentation


Abstract:

In recent years, semantic segmentation technology plays an important role in land resource management tasks. However, many classic semantic segmentation methods often fai...Show More

Abstract:

In recent years, semantic segmentation technology plays an important role in land resource management tasks. However, many classic semantic segmentation methods often fail to obtain satisfactory results for remote sensing images with a large amount of interference information. To improve this situation, we propose semantic category balance-aware involved anti-interference network (SCBANet). SCBANet has an encoder–decoder structure similar to DeeplabV3+. On this basis, we propose clustering-guided semantic decoupling module (CGSDM), consistency-based anti-interference feature extraction module (CAFEM), relevance-based anti-interference feature extraction module (RAFEM), and optional decoder module based on semantic category balance (ODMSCB) to improve the accuracy of semantic segmentation. CGSDM aims to obtain the information of different semantic categories through K - means clustering algorithm. CAFEM performs an average operation on the feature vectors in each semantic category to obtain semantic consistency information. RAFEM deeply excavates the information contained in each semantic category through the modeling method with self-attention mechanism as the core, making the relationship between pixels within each semantic category to be better understood by the model. ODMSCB classifies the feature map according to the balance of different semantic categories, so that different decoders can be applied to feature maps with different semantic category balance. These four parts complement each other, greatly improving the model’s anti-interference ability while also enhancing the ability to handle category imbalance issue. We compared our method with several of the most advanced deep learning methods on the Vaihingen and Potsdam datasets. The final results demonstrate the superiority of our method.
Article Sequence Number: 4409712
Date of Publication: 18 October 2023

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

The data required by traditional geography, environmental science and earth science research are obtained by ground monitoring stations or field visits, which have the defects of long cycle, long time-consuming, and high cost. However, remote sensing technology is a technology to obtain various information about ground objects from a long distance through sensors, which overcomes the shortcomings of traditional methods. Its main research object is ground object information, and its main content is to obtain ground object information for various applications, such as land change surveys, forest cover surveys, marine pollution surveys, and urban land surveys. It is of great significance to provide more useful information for geographical science research and development. In recent years, remote sensing technology has developed rapidly. Semantic segmentation based on remote sensing images has become a very important task in land resource management and road detection [2], [3], [4], [5].

Select All
1.
L. Chen, Y. Zhu, G. Papandreou, F. Schroff and H. Adam, "Encoder–decoder with atrous separable convolution for semantic image segmentation", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 801-818, 2018.
2.
B. Huang, B. Zhao and Y. Song, "Urban land-use mapping using a deep convolutional neural network with high spatial resolution multispectral remote sensing imagery", Remote Sens. Environ., vol. 214, pp. 73-86, Sep. 2018.
3.
C. Zhang, P. A. Harrison, X. Pan, H. Li, I. Sargent and P. M. Atkinson, "Scale sequence joint deep learning (SS-JDL) for land use and land cover classification", Remote Sens. Environ., vol. 237, Feb. 2020.
4.
W. Zhang, W. Li, C. Zhang, D. M. Hanink, X. Li and W. Wang, "Parcel-based urban land use classification in megacity using airborne LiDAR high resolution orthoimagery and Google street view", Comput. Environ. Urban Syst., vol. 64, pp. 215-228, Jul. 2017.
5.
W. Sun and R. Wang, "Fully convolutional networks for semantic segmentation of very high resolution remotely sensed images combined with DSM", IEEE Geosci. Remote Sens. Lett., vol. 15, no. 3, pp. 474-478, Mar. 2018.
6.
J. Long, E. Shelhamer and T. Darrell, "Fully convolutional networks for semantic segmentation", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 3431-3440, Jun. 2015.
7.
O. Ronneberger, P. Fischer and T. Brox, "U-Net: Convolutional networks for biomedical image segmentation", Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Intervent., pp. 234-241, 2015.
8.
V. Badrinarayanan, A. Kendall and R. Cipolla, "SegNet: A deep convolutional encoder–decoder architecture for image segmentation", IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 12, pp. 2481-2495, Dec. 2017.
9.
Q. Liu, M. Kampffmeyer, R. Jenssen and A.-B. Salberg, "Multi-view self-constructing graph convolutional networks with adaptive class weighting loss for semantic segmentation", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 44-45, Jun. 2020.
10.
Y. Chen, M. Rohrbach, Z. Yan, Y. Shuicheng, J. Feng and Y. Kalantidis, "Graph-based global reasoning networks", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 433-442, Jun. 2019.
11.
H. Zhao, J. Shi, X. Qi, X. Wang and J. Jia, "Pyramid scene parsing network", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2881-2890, Jul. 2017.
12.
J. Fu et al., "Dual attention network for scene segmentation", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 3146-3154, Jun. 2019.
13.
J. Hu, L. Shen and G. Sun, "Squeeze-and-excitation networks", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 7132-7141, Jun. 2018.
14.
S. Woo, J. Park, J.-Y. Lee and I. S. Kweon, "CBAM: Convolutional block attention module", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 3-19, 2018.
15.
Z. Liu et al., "Swin transformer: Hierarchical vision transformer using shifted windows", Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 10012-10022, Oct. 2021.
16.
Y. Zhang, B. Pang and C. Lu, "Semantic segmentation by early region proxy", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1258-1268, Jun. 2022.
17.
H. Cao et al., "Swin-UNet: UNet-like pure transformer for medical image segmentation" in Computer Vision—ECCV 2022 Workshops, Tel Aviv, Israel:Springer, vol. 13803, pp. 205-218, 2022.
18.
L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy and A. L. Yuille, "Semantic image segmentation with deep convolutional nets and fully connected CRFs", arXiv:1412.7062, 2014.
19.
L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy and A. L. Yuille, "DeepLab: Semantic image segmentation with deep convolutional nets atrous convolution and fully connected CRFs", IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, no. 4, pp. 834-848, Apr. 2018.
20.
L. Chen, G. Papandreou, F. Schroff and H. Adam, "Rethinking atrous convolution for semantic image segmentation", arXiv:1706.05587, 2017.
21.
M. Kampffmeyer, A.-B. Salberg and R. Jenssen, "Semantic segmentation of small objects and modeling of uncertainty in urban remote sensing images using deep convolutional neural networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops (CVPRW), pp. 1-9, Jun. 2016.
22.
Z. Guo, Q. Chen, G. Wu, Y. Xu, R. Shibasaki and X. Shao, "Village building identification based on ensemble convolutional neural networks", Sensors, vol. 17, no. 11, pp. 2487, Oct. 2017.
23.
A. Krizhevsky, I. Sutskever and G. E. Hinton, "ImageNet classification with deep convolutional neural networks", Proc. Adv. Neural Inf. Process. Syst., vol. 25, pp. 1-9, 2012.
24.
K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition", arXiv:1409.1556, 2014.
25.
C. Szegedy et al., "Going deeper with convolutions", Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1-9, Jun. 2015.
26.
F. N. Iandola, S. Han, M. W. Moskewicz, K. Ashraf, W. J. Dally and K. Keutzer, "SqueezeNet: AlexNet-level accuracy with 50× fewer parameters and < 0.5 MB model size", arXiv:1602.07360, 2016.
27.
X. Pan et al., "Building extraction from high-resolution aerial imagery using a generative adversarial network with spatial and channel attention mechanisms", Remote Sens., vol. 11, no. 8, pp. 917, Apr. 2019.
28.
P. He et al., "MANet: Multi-scale aware-relation network for semantic segmentation in aerial scenes", IEEE Trans. Geosci. Remote Sens., vol. 60, 2022.
29.
L. Ding et al., "Looking outside the window: Wide-context transformer for the semantic segmentation of high-resolution remote sensing images", IEEE Trans. Geosci. Remote Sens., vol. 60, 2022.
30.
K. Yue, L. Yang, R. Li, W. Hu, F. Zhang and W. Li, "TreeUNet: Adaptive tree convolutional neural networks for subdecimeter aerial image segmentation", ISPRS J. Photogramm. Remote Sens., vol. 156, pp. 1-13, Oct. 2019.

Contact IEEE to Subscribe

References

References is not available for this document.