Loading [MathJax]/extensions/MathMenu.js
OADB-Net: An Occlusion-Aware Dual-Branch Network for Pedestrian Detection | IEEE Journals & Magazine | IEEE Xplore

OADB-Net: An Occlusion-Aware Dual-Branch Network for Pedestrian Detection


Abstract:

With the advancements in deep learning, detecting occluded pedestrians has become a focal point of research. Extracting pedestrian parts has proven to be an effective sol...Show More

Abstract:

With the advancements in deep learning, detecting occluded pedestrians has become a focal point of research. Extracting pedestrian parts has proven to be an effective solution for handling occlusion. However, existing methods mainly rely on Region Proposal Networks (RPNs) for the part feature extraction. These RPNs-based methods suffer from limitations such as complex structures and limited receptive fields, which hinder their ability to capture global dependency information. To overcome these challenges, we propose a simple but effective Occlusion-Aware Dual-Branch Network (OADB-Net) based on an anchor-free framework for pedestrian detection in crowded scenes. Specifically, we design a dual-branch occlusion-aware detection head, consisting of a full-body detection branch and a head-shoulder detection branch, to address the occlusion issue in crowded scenes. The head-shoulder detection branch to handle heavily occluded instances and the full-body branch to focus on non-heavily occluded instances. Furthermore, we propose a Cross-Layer Non-Local Module (CLNL-Module), which captures long-range dependencies across feature layers, to effectively differentiate the relationship between pedestrian body and body parts while integrating essential features for more accurate and discriminative pedestrian representation. This strengthens the connections between the dual detection branches and enhances the responses of their respective center heatmaps. Our OADB-Net leverages part and full-body features to handle pedestrians with varying degrees of occlusion, while avoiding the limitations of RPNs-based methods. In heavy occlusion settings, OADB-Net achieves the average miss rates of 39.9%, 26.8%, and 43.1% on the Citypersons, Caltech, and CrowdHuman datasets, respectively, and demonstrates superior performance in traffic scenes.
Published in: IEEE Transactions on Intelligent Transportation Systems ( Volume: 26, Issue: 2, February 2025)
Page(s): 1617 - 1630
Date of Publication: 20 November 2024

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Pedestrian detection is a specialized form of object detection that has gained increasing interest in computer vision and multimedia analysis [1], [2], [3], [4], [5], [6], [7], [8], [9]. It is crucial for applications like autonomous driving systems [1], [2], [3], [4], human-robot interaction [5], [6], and intelligent video surveillance [7], [8], [9]. However, pedestrian detection in crowded scenes is more challenging than general object detection due to diverse appearances, scale variations, and occlusions. Especially, the detection performance for heavily occluded samples still remains unsatisfactory, even many Convolutional Neural Networks (CNNs) have been applied to handle the issue.

Select All
1.
L. Chen, S. Lin, X. Lu, D. Cao and F. Y. Wang, "Deep neural network based vehicle and pedestrian detection for autonomous driving: A survey", IEEE Trans. Intell. Transp. Syst., vol. 22, no. 6, pp. 3234-3246, Jun. 2021.
2.
M. M. Islam, v. A. R. Newaz and A. Karimoddini, "Pedestrian detection for autonomous cars: Inference fusion of deep neural networks", IEEE Trans. Intell. Transp. Syst., vol. 23, no. 12, pp. 23358-23368, Dec. 2022.
3.
F. Camara et al., "Pedestrian models for autonomous driving—Part I: Low-level models from sensing to tracking", IEEE Trans. Intell. Transp. Syst., vol. 22, no. 10, pp. 6131-6151, Oct. 2021.
4.
D. Feng et al., "Deep multi-modal object detection and semantic segmentation for autonomous driving: Datasets methods and challenges", IEEE Trans. Intell. Transp. Syst., vol. 22, no. 3, pp. 1341-1360, Mar. 2021.
5.
F. Bu, T. Le, X. Du, R. Vasudevan and M. Johnson-Roberson, "Pedestrian planar LiDAR pose (PPLP) network for oriented pedestrian detection based on planar LiDAR and monocular images", IEEE Robot. Autom. Lett., vol. 5, no. 2, pp. 1626-1633, Apr. 2020.
6.
S. Paisitkriangkrai, C. Shen and A. V. D. Hengel, "Pedestrian detection with spatially pooled features and structured ensemble learning", IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 6, pp. 1243-1257, Jun. 2016.
7.
S. Wu, H.-S. Wong and S. Wang, "Variant SemiBoost for improving human detection in application scenes", IEEE Trans. Circuits Syst. Video Technol., vol. 28, no. 7, pp. 1595-1608, Jul. 2018.
8.
K. Chen and Z. Zhang, "Pedestrian counting with back-propagated information and target drift remedy", IEEE Trans. Syst. Man Cybern. Syst., vol. 47, no. 4, pp. 639-647, Apr. 2017.
9.
X. Wang et al., "S3D: Scalable pedestrian detection via score scale surface discrimination", IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 10, pp. 3332-3344, Oct. 2020.
10.
S. Zhang, R. Benenson and B. Schiele, "CityPersons: A diverse dataset for pedestrian detection", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 3213-3221, Jul. 2017.
11.
P. Dollar, C. Wojek, B. Schiele and P. Perona, "Pedestrian detection: An evaluation of the state of the art", IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no. 4, pp. 743-761, Apr. 2012.
12.
S. Shao et al., "CrowdHuman: A benchmark for detecting human in a crowd", arXiv:1805.00123, 2018.
13.
Q. Li, Y. Su, Y. Gao, F. Xie and J. Li, "OAF-net: An occlusion-aware anchor-free network for pedestrian detection in a crowd", IEEE Trans. Intell. Transp. Syst., vol. 23, no. 11, pp. 21291-21300, Nov. 2022.
14.
X. Wang, T. Xiao, Y. Jiang, S. Shao, J. Sun and C. Shen, "Repulsion loss: Detecting pedestrians in a crowd", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 7774-7783, Jun. 2018.
15.
S. Zhang, L. Wen, X. Bian, Z. Lei and S. Z. Li, "Occlusion-aware R-CNN: Detecting pedestrians in a crowd", Proc. Eur. Conf. Comput. Vis., pp. 637-653, 2018.
16.
Q. Li, Y. Bi, R. Cai and J. Li, "Occluded pedestrian detection through bi-center prediction in anchor-free network", Neurocomputing, vol. 507, pp. 199-207, Oct. 2022.
17.
Y. Pang, J. Xie, M. H. Khan, R. M. Anwer, F. S. Khan and L. Shao, "Mask-guided attention network for occluded pedestrian detection", Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 4967-4975, Oct. 2019.
18.
S. Zhang, J. Yang and B. Schiele, "Occluded pedestrian detection through guided attention in CNNs", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 6995-7003, Jun. 2018.
19.
C. Zhou and J. Yuan, "Bi-box regression for pedestrian detection and occlusion estimation", Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 135-151, 2018.
20.
M. Xu, Y. Bai, S. S. Qu and B. Ghanem, "Semantic part RCNN for real-world pedestrian detection", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 45-54, Jun. 2019.
21.
L. Zhang, L. Lin, X. Liang and K. He, "Is faster R-CNN doing well for pedestrian detection?", Proc. Eur. Conf. Comput. Vis., pp. 443-457, Oct. 2016.
22.
J. Li, X. Liang, S. Shen, T. Xu, J. Feng and S. Yan, "Scale-aware fast R-CNN for pedestrian detection", IEEE Trans. Multimedia, vol. 20, no. 4, pp. 985-996, Apr. 2018.
23.
C. Chi, S. Zhang, J. Xing, Z. Lei, S. Z. Li and X. Zou, "PedHunter: Occlusion robust pedestrian detector in crowded scenes", Proc. AAAI Conf. Artif. Intell., vol. 34, no. 7, pp. 10639-10646, Apr. 2020.
24.
P. Yang, G. Zhang, L. Wang, L. Xu, Q. Deng and M. Yang, "A part-aware multi-scale fully convolutional network for pedestrian detection", IEEE Trans. Intell. Transp. Syst., vol. 22, no. 2, pp. 1125-1137, Feb. 2021.
25.
N. Bodla, B. Singh, R. Chellappa and L. S. Davis, "Soft-NMS—Improving object detection with one line of code", Proc. IEEE Int. Conf. Comput. Vis. (ICCV), pp. 5562-5570, Oct. 2017.
26.
X. Huang, Z. Ge, Z. Jie and O. Yoshie, "NMS by representative region: Towards crowded pedestrian detection by proposal pairing", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 10750-10759, Jun. 2020.
27.
P. Zhou et al., "NOH-NMS: Improving pedestrian detection by nearby objects hallucination", Proc. 28th ACM Int. Conf. Multimedia, pp. 1967-1975, Oct. 2020.
28.
S. Liu, D. Huang and Y. Wang, "Adaptive NMS: Refining pedestrian detection in a crowd", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 6459-6468, Jun. 2019.
29.
A. Abdelmutalab and C. Wang, "Pedestrian detection using MB-CSP model and boosted identity aware non-maximum suppression", IEEE Trans. Intell. Transp. Syst., vol. 23, no. 12, pp. 24454-24463, Dec. 2022.
30.
Y. Tang, M. Liu, B. Li, Y. Wang and W. Ouyang, "OTP-NMS: Toward optimal threshold prediction of NMS for crowded pedestrian detection", IEEE Trans. Image Process., vol. 32, pp. 3176-3187, 2023.
Contact IEEE to Subscribe

References

References is not available for this document.