Loading [MathJax]/extensions/MathMenu.js
Parameter-Free Similarity-Aware Attention Module for Medical Image Classification and Segmentation | IEEE Journals & Magazine | IEEE Xplore

Parameter-Free Similarity-Aware Attention Module for Medical Image Classification and Segmentation


Abstract:

Automatic classification and segmentation of medical images play essential roles in computer-aided diagnosis. Deep convolutional neural networks (DCNNs) have shown their ...Show More

Abstract:

Automatic classification and segmentation of medical images play essential roles in computer-aided diagnosis. Deep convolutional neural networks (DCNNs) have shown their advantages on image classification and segmentation. However, they have not achieved the same success on medical images as they have done on natural images. In this paper, two challenges are exploited for DCNNs on medical images, including 1) lack of feature diversity; 2) neglect of small lesions. These two issues heavily influence the classification and segmentation performances. To improve the performance of DCNN on medical images, similarity-aware attention (simi-attention) module is proposed, including a Feature-similarity-aware Channel Attention (FCA) and a Region-similarity-aware Spatial Attention (RSA). Our simi-attention provides three advantages: 1) higher accuracy can be achieved since it extracts both diverse and discriminant features from medical images via our FCA and RSA; 2) the lesions can be exactly focused and located by it even for the data with low intensity contrast and small lesions; 3) it does not increase the complexity of backbone models due to NO trainable parameters in its module. The experimental results are conducted on both classification and segmentation tasks under four public medical classification datasets and two public medical segmentation datasets. The visualization results show that our simi-attention can accurately focuses on the lesions for classification and generate fine segmentation results even for small objects. The overall performances show that our simi-attention can significantly improve the performances of backbone models and outperforms compared attention models on most of datasets for both classification and segmentation.
Page(s): 845 - 857
Date of Publication: 30 August 2022
Electronic ISSN: 2471-285X

Funding Agency:

References is not available for this document.

I. Introduction

Medical image classification aims to distinguish medical images based on clinical pathologies or imaging modalities [1], and medical image segmentation can locate the position of lesions and assist clinicians to quantitatively evaluate the effect in preoperative and postoperative examination [2]. In recent years, deep learning methods, especially deep convolutional neural networks (DCNNs), have gained significant achievements in medical image classification and segmentation tasks. However, they have not achieved the same success on medical images as they have done on natural images [3], [4]. In this work, two challenges in medical image classification and segmentation are exploited and considered.

Lack of feature diversity: Different from natural color images, most medical images are gray-scale images and the lesion area usually has low intensity contrast with the neighboring normal tissues [5], [6]. As shown in Fig. 1(a) of one malignant ultrasound image for breast, the lesion area is blocked with yellow dotted line, which is shown to have very low intensity contrast with the surrounding normal tissues. Hence, it is difficult to extract as diverse features as that extracted from natural color images through convolution operation only [7];

Neglect of small lesions: The key information such as the lesion area in medical images often occupies much fewer pixels than normal tissue and the lesion area is different from normal tissue on particulars [8], [9]. In other words, the particulars in medical images are not so unimportant as in natural images. As shown in Fig. 1(b) of one AMD (Age-related Macular Degeneration) fundus images. The lesion areas are blocked with yellow dotted line, which are too small and inconspicuous when compared to normal tissues. Hence, it is difficult to learn discriminant features from the small and inconspicuous lesions through convolution operation only due to the neglect of small lesions in high level feature maps [10].

Illustration of medical image examples: (a) a malignant ultrasound image; and (b) an AMD fundus image.

Select All
1.
M. Goyal, N. D. Reeves, A. K. Davison, S. Rajbhandari, J. Spragg and M. H. Yap, "DFUNet: Convolutional neural networks for diabetic foot ulcer classification", IEEE Trans. Emerg. Topics Comput. Intell., vol. 4, no. 5, pp. 728-739, Oct. 2020.
2.
Q. Chen, W. Xie, P. Zhou, C. Zheng and D. Wu, "Multi-crop convolutional neural networks for fast lung nodule segmentation", IEEE Trans. Emerg. Topics Comput. Intell..
3.
J. Zhang, Y. Xie, Q. Wu and Y. Xia, "Medical image classification using synergic deep learning", Med. Image Anal., vol. 54, pp. 10-19, 2019.
4.
L. Zhang et al., "Generalizing deep learning for medical image segmentation to unseen domains via deep stacked transformation", IEEE Trans. Med. Imag., vol. 39, no. 7, pp. 2531-2540, Jul. 2020.
5.
Q. Huang, Y. Huang, Y. Luo, F. Yuan and X. Li, "Segmentation of breast ultrasound image with semantic classification of superpixels", Med. Image Anal., vol. 61, 2020.
6.
X. Yao, Z. Zhu, C. Kang, S.-H. Wang, J. M. Gorriz and Y.-D. Zhang, "AdaD-FNN for chest CT-based COVID-19 diagnosis", IEEE Trans. Emerg. Topics Comput. Intell..
7.
W. S. Evans et al., "Effects of semantic feature type diversity and quantity on semantic feature analysis treatment outcomes in aphasia", Amer. J. Speech- Lang. Pathol., vol. 30, no. 1S, pp. 344-358, 2021.
8.
M. H. Hesamian, W. Jia, X. He and P. Kennedy, "Deep learning techniques for medical image segmentation: Achievements and challenges", J. Digit. Imag., vol. 32, no. 4, pp. 582-596, 2019.
9.
A. Bria, C. Marrocco and F. Tortorella, "Addressing class imbalance in deep learning for small lesion detection on medical images", Comput. Biol. Med., vol. 120, 2020.
10.
L. Bo, T. Hai and F. Qiang, "A small object detection method based on local maxima and SSD", Proc. AI Opt. Photon., 2019.
11.
W. Huang et al., "Feature pyramid network with level-aware attention for meningioma segmentation", IEEE Trans. Emerg. Topics Comput. Intell..
12.
Z. Li et al., "A novel deep learning framework based mask-guided attention mechanism for distant metastasis prediction of lung cancer", IEEE Trans. Emerg. Topics Comput. Intell..
13.
S. Woo, J. Park, J.-Y. Lee and I. S. Kweon, "CBAM: Convolutional block attention module", Proc. Eur. Conf. Comput. Vis., pp. 3-19, 2018.
14.
H. Zhang, I. Goodfellow, D. Metaxas and A. Odena, "Self-attention generative adversarial networks", Proc. Int. Conf. Mach. Learn., pp. 7354-7363, 2019.
15.
T. Qu et al., "M3Net: A multi-scale multi-view framework for multi-phase pancreas segmentation based on cross-phase non-local attention", Med. Image Anal., vol. 75, 2022.
16.
X. Wang et al., "SSA-Net: Spatial self-attention network for COVID-19 pneumonia infection segmentation with semi-supervised few-shot learning", Med. Image Anal., vol. 79, 2022.
17.
M. Tan and Q. Le, "EfficientNet: Rethinking model scaling for convolutional neural networks", Proc. Int. Conf. Mach. Learn., pp. 6105-6114, 2019.
18.
M. Canayaz, "C+EffxNet: A novel hybrid approach for COVID-19 diagnosis on CT images based on CBAM and EfficientNet", Chaos Solitons Fractals, vol. 151, 2021.
19.
Y. Li, H. Huang, L. Zhang, G. Wang, H. Zhang and W. Zhou, "Super-resolution and self-attention with generative adversarial network for improving malignancy characterization of hepatocellular carcinoma", Proc. IEEE 17th Int. Symp. Biomed. Imag., pp. 1556-1560, 2020.
20.
P. Wang and A. C. Chung, "Relax and focus on brain tumor segmentation", Med. Image Anal., vol. 75, 2022.
21.
Y. Li, K. Guo, Y. Lu and L. Liu, "Cropping and attention based approach for masked face recognition", Appl. Intell., vol. 51, no. 5, pp. 3012-3025, 2021.
22.
M. Toğaçar, K. B. Özkurt, B. Ergen and Z. Cömert, "Breastnet: A novel convolutional neural network model through histopathological images for the diagnosis of breast cancer", Physica A: Stat. Mechanics Appl., vol. 545, 2020.
23.
S. Wang, B. Z. Li, M. Khabsa, H. Fang and H. Ma, "Linformer: Self-attention with linear complexity", 2020.
24.
B. Yang, L. Wang, D. F. Wong, S. Shi and Z. Tu, "Context-aware self-attention networks for natural language processing", Neurocomputing, vol. 458, pp. 157-169, 2021.
25.
K. He, X. Zhang, S. Ren and J. Sun, "Deep residual learning for image recognition", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 770-778, 2016.
26.
F. Chollet, "Xception: Deep learning with depthwise separable convolutions", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1251-1258, 2017.
27.
N. Tajbakhsh, L. Jeyaseelan, Q. Li, J. N. Chiang, Z. Wu and X. Ding, "Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation", Med. Image Anal., vol. 63, 2020.
28.
J. Hu, L. Shen and G. Sun, "Squeeze-and-excitation networks", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 7132-7141, 2018.
29.
A. S. Lundervold and A. Lundervold, "An overview of deep learning in medical imaging focusing on MRI", Zeitschrift für Medizinische Physik, vol. 29, no. 2, pp. 102-127, 2019.
30.
Q. Hou, D. Zhou and J. Feng, "Coordinate attention for efficient mobile network design", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 13713-13722, 2021.

Contact IEEE to Subscribe

References

References is not available for this document.