Loading [MathJax]/extensions/MathMenu.js
Enhancing Traffic Object Detection in Variable Illumination With RGB-Event Fusion | IEEE Journals & Magazine | IEEE Xplore

Enhancing Traffic Object Detection in Variable Illumination With RGB-Event Fusion


Abstract:

Traffic object detection under variable illumination is challenging due to the information loss caused by the limited dynamic range of conventional frame-based cameras. T...Show More

Abstract:

Traffic object detection under variable illumination is challenging due to the information loss caused by the limited dynamic range of conventional frame-based cameras. To address this issue, we introduce bio-inspired event cameras and propose a novel Structure-aware Fusion Network (SFNet) that extracts sharp and complete object structures from the event stream to compensate for the lost information in images through cross-modality fusion, enabling the network to obtain illumination-robust representations for traffic object detection. Specifically, to mitigate the sparsity or blurriness issues arising from diverse motion states of traffic objects in fixed-interval event sampling methods, we propose the Reliable Structure Generation Network (RSGNet) to generate Speed Invariant Frames (SIF), ensuring the integrity and sharpness of object structures. Next, we design a novel Adaptive Feature Complement Module (AFCM) which guides the adaptive fusion of two modality features to compensate for the information loss in the images by perceiving the global lightness distribution of the images, thereby generating illumination-robust representations. Finally, considering the lack of large-scale and high-quality annotations in the existing event-based object detection datasets, we build a DSEC-Det dataset, which consists of 53 sequences with 63,931 images and more than 208,000 labels for 8 classes. Extensive experimental results demonstrate that our proposed SFNet can overcome the perceptual boundaries of conventional cameras and outperform the frame-based methods, e.g., YOLOX by 7.9% in mAP50 and 3.8% in mAP50:95. Our code and dataset will be available at https://github.com/YN-Yang/SFNet.
Published in: IEEE Transactions on Intelligent Transportation Systems ( Volume: 25, Issue: 12, December 2024)
Page(s): 20335 - 20350
Date of Publication: 17 September 2024

ISSN Information:

Funding Agency:


I. Introduction

Traffic object detection aims to accurately recognize and locate traffic objects, making it a critical component of autonomous driving systems [1], [2], [3], [4], [5], [6], [7] and serving as the foundation for downstream tasks such as object tracking and trajectory prediction [8], [9], [10], [11], [12]. Notably, the success of existing traffic object detection methods is heavily contingent upon image quality [13], [14], [15], [16]. However, the limited capacitance capacity in the integral imaging circuit of frame-based cameras restricts their dynamic range [17], making it difficult to achieve stable imaging in poor lighting conditions (e.g., low-light and over-exposure), which will result in decreased image contrast and information loss and hinder the extraction of representative features for traffic object detection [18].

Contact IEEE to Subscribe

References

References is not available for this document.