Loading web-font TeX/Math/Italic
Self-Supervised Adversarial Training of Monocular Depth Estimation Against Physical-World Attacks | IEEE Journals & Magazine | IEEE Xplore

Self-Supervised Adversarial Training of Monocular Depth Estimation Against Physical-World Attacks


Abstract:

Monocular Depth Estimation (MDE) plays a vital role in applications such as autonomous driving. However, various attacks target MDE models, with physical attacks posing s...Show More

Abstract:

Monocular Depth Estimation (MDE) plays a vital role in applications such as autonomous driving. However, various attacks target MDE models, with physical attacks posing significant threats to system security. Traditional adversarial training methods, which require ground-truth labels, are not directly applicable to MDE models that lack ground-truth depth. Some self-supervised model hardening techniques (e.g., contrastive learning) overlook the domain knowledge of MDE, resulting in suboptimal performance. In this work, we introduce a novel self-supervised adversarial training approach for MDE models, leveraging view synthesis without the need for ground-truth depth. We enhance adversarial robustness against real-world attacks by incorporating L_{0}-norm-bounded perturbation during training. We evaluate our method against supervised learning-based and contrastive learning-based approaches specifically designed for MDE. Our experiments with two representative MDE networks demonstrate improved robustness against various adversarial attacks, with minimal impact on benign performance.
Page(s): 9084 - 9101
Date of Publication: 17 June 2024

ISSN Information:

PubMed ID: 38885107

Funding Agency:


I. Introduction

Monocular Depth Estimation (MDE) is a deep neural network (DNN)-based task that estimates depth from a single image, allowing for 2D-to-3D projection by predicting the distance for each pixel in a 2D image [2]. This makes it a cost-effective alternative to pricey Lidar sensors. Applications of MDE are vast, including autonomous driving [3], visual SLAM [4], and visual relocalization [5], etc. Specifically, self-supervised MDE has become increasingly popular in the industry (e.g., Tesla Autopilot [3]) due to its ability to achieve comparable accuracy without requiring ground-truth depth data from Lidar during training. However, due to the known vulnerabilities in DNNs [6], [7], several digital-world [8], [9] and physical-world adversarial attacks [10] has been developed against MDE. These attacks primarily use optimization-based methods to create adversarial input that deceive the MDE models. Given the significance and widespread use of MDE models, these adversarial attacks pose a substantial risk to the security of applications such as autonomous driving. As a result, there is an urgent need to develop and strengthen MDE models against these threats.

Contact IEEE to Subscribe

References

References is not available for this document.