Loading [MathJax]/extensions/MathMenu.js
Small-Object Sensitive Segmentation Using Across Feature Map Attention | IEEE Journals & Magazine | IEEE Xplore

Small-Object Sensitive Segmentation Using Across Feature Map Attention


Abstract:

Semantic segmentation is an important step in understanding the scene for many practical applications such as autonomous driving. Although Deep Convolutional Neural Netwo...Show More

Abstract:

Semantic segmentation is an important step in understanding the scene for many practical applications such as autonomous driving. Although Deep Convolutional Neural Networks-based methods have significantly improved segmentation accuracy, small/thin objects remain challenging to segment due to convolutional and pooling operations that result in information loss, especially for small objects. This article presents a novel attention-based method called Across Feature Map Attention (AFMA) to address this challenge. It quantifies the inner-relationship between small and large objects belonging to the same category by utilizing the different feature levels of the original image. The AFMA could compensate for the loss of high-level feature information of small objects and improve the small/thin object segmentation. Our method can be used as an efficient plug-in for a wide range of existing architectures and produces much more interpretable feature representation than former studies. Extensive experiments on eight widely used segmentation methods and other existing small-object segmentation models on CamVid and Cityscapes demonstrate that our method substantially and consistently improves the segmentation of small/thin objects.
Page(s): 6289 - 6306
Date of Publication: 30 September 2022

ISSN Information:

PubMed ID: 36178991

Funding Agency:

Citations are not available for this document.

1 Introduction

Semantic segmentation is an important processing step in natural or medical image analysis for the detection of distinct types of objects in images [1]. In this process, a semantic label is assigned to each pixel of a given image. The breakthrough of semantic segmentation methods came when fully convolutional neural networks (FCN) were first used by [2] to perform end-to-end segmentation of images. While semantic segmentation has achieved significant improvement based on the conception of fully convolutional networks, small and thin items in the scene remain difficult to segment because the information of small objects is lost throughout the convolutional and pooling processes [3], [4], [5], [6]. For example, Fig. 1a is an image of size 800 by 1200 pixels, which contains two cars: the larger car is 160 by 220 pixels (Fig. 1b), and the smaller one is 30 by 40 (Fig. 1c). After a convolution operation with a convolution kernel of 10×10, the length and width of the image are compressed to one-tenth of the original size (as shown in Fig. 1d). Accordingly, the dimensions of the large and small cars become 16 by 22 and 3 by 4 pixels, respectively. As seen from the example, we can still see the car's features from Fig. 1e (feature map of the large car), but we can hardly see the features of the small car from the 12-pixel size Fig. 1c (feature map of the small car). This is because the high-level representation from convolutional and pooling operations generated along lowers the resolution, which often leads to the loss of the detailed information of small/thin objects [3] — as a result, recovering the car information from the coarse feature maps is difficult for segmentation models [7]. However, accurately segmenting small objects is critical in many applications, such as autonomous driving, where the segmentation and recognition of small-sized cars and pedestrians in the distance is critical [8], [9], [10], [11].

Cites in Papers - |

Cites in Papers - IEEE (8)

Select All
1.
Yu-Hsien Chung, Chi-Hsuan Lu, Jung-Hui Cho, Chih-Chang Yu, "Utilizing Cross Layer Attentions for Semantic Segmentation of Small Objects", 2024 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), pp.1-6, 2024.
2.
Zhenhao Yang, Fukun Bi, Xinghai Hou, Dehao Zhou, Yanping Wang, "DDRNet: Dual-Domain Refinement Network for Remote Sensing Image Semantic Segmentation", IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol.17, pp.20177-20189, 2024.
3.
Mengjie Gao, Fang Chen, Lei Wang, Huichen Zhao, Bo Yu, "Swin Transformer-Based Multiscale Attention Model for Landslide Extraction From Large-Scale Area", IEEE Transactions on Geoscience and Remote Sensing, vol.62, pp.1-14, 2024.
4.
Zheng Wang, Ryojun Ikeura, Soichiro Hayakawa, Zhiliang Zhang, "Road Marking Damage Degree Detection Based on Adaptive Feature Fusion", 2024 IEEE International Conference on Mechatronics and Automation (ICMA), pp.1819-1823, 2024.
5.
Jinhuan Shan, Wei Jiang, Yue Huang, Dongdong Yuan, Yaohan Liu, "Unmanned Aerial Vehicle (UAV)-Based Pavement Image Stitching Without Occlusion, Crack Semantic Segmentation, and Quantification", IEEE Transactions on Intelligent Transportation Systems, vol.25, no.11, pp.17038-17053, 2024.
6.
Yuxuan Wang, Shuailin Chen, Ruixiang Zhang, Fang Xu, Pingping Huang, Wen Yang, "Cross-Resolution Distillation for Building Extraction in Medium-Resolution Satellite Imagery", IGARSS 2024 - 2024 IEEE International Geoscience and Remote Sensing Symposium, pp.8122-8126, 2024.
7.
Yasin Yari, Ingun Næve, Per Helge Bergtun, Asle Hammerdal, Svein-Erik Måsøy, Marco Marien Voormolen, Lasse Løvstakken, "Deep Learning for Automated Egg Maturation Prediction of Atlantic Salmon Using Ultrasound Imaging", IEEE Access, vol.12, pp.80233-80243, 2024.
8.
Fanzhao Lin, Kexin Bao, Yong Li, Dan Zeng, Shiming Ge, "Learning Contrast-Enhanced Shape-Biased Representations for Infrared Small Target Detection", IEEE Transactions on Image Processing, vol.33, pp.3047-3058, 2024.

Cites in Papers - Other Publishers (8)

1.
Jun Wu, Jinshu Gao, Jingjie Lin, Zeyu Huang, Yang Liu, Zhengyu Chen, Qin Long, Jianchun Zhao, Dayong Ding, "A Dual-Stream Network for\\xa0Langerhans’ Cells Segmentation in\\xa0CCM Images", Ophthalmic Medical Image Analysis, vol.15188, pp.126, 2025.
2.
Ziang Zhou, Wensong Zhao, Kechen Song, Yanyan Wang, Jun Li, "EAFNet: Extraction-amplification-fusion network for tiny cracks detection", Engineering Applications of Artificial Intelligence, vol.134, pp.108691, 2024.
3.
Mingxin Yuan, Jie Li, Borui Cao, Shihao Bao, Li Sun, Xiangbin Li, "Research on Target Hybrid Recognition and Localization Methods Based on an Industrial Camera and a Depth Camera in Complex Scenes", Electronics, vol.13, no.22, pp.4381, 2024.
4.
Ronny Velastegui, Maxim Tatarchenko, Sezer Karaoglu, Theo Gevers, "Image semantic segmentation of indoor scenes: A survey", Computer Vision and Image Understanding, pp.104102, 2024.
5.
Apoorva Safai, Colin Froines, Robert Slater, Rachel E. Linderman, Jacob Bogost, Caleb Pacheco, Rickie Voland, Jeong Pak, Pallavi Tiwari, Roomasa Channa, Amitha Domalpally, "Quantifying Geographic Atrophy in Age-Related Macular Degeneration: A Comparative Analysis Across 12 Deep Learning Models", Investigative Ophthalmology & Visual Science, vol.65, no.8, pp.42, 2024.
6.
Hongtao Huang, Xiaofeng Tian, Wei Tian, "Research on Semantic Segmentation of Fish-Eye Images for Autonomous Driving", International Journal of Robotics and Automation Technology, vol.10, pp.138, 2023.
7.
Chunyu Zhang, Fang Xu, Chengdong Wu, Jinzhao Li, "Lightweight semantic segmentation network with configurable context and small object attention", Frontiers in Computational Neuroscience, vol.17, 2023.
8.
Zhiqiang Li, Jie Jiang, Xi Chen, Robert Laganiere, Qingli Li, Min Liu, Honggang Qi, Yong Wang, Min Zhang, "Dense-scale dynamic network with filter-varying atrous convolution for semantic segmentation", Applied Intelligence, 2023.
Contact IEEE to Subscribe

References

References is not available for this document.