Loading [MathJax]/extensions/MathMenu.js
Mixture of Teacher Experts for Source-Free Domain Adaptive Object Detection | IEEE Conference Publication | IEEE Xplore

Mixture of Teacher Experts for Source-Free Domain Adaptive Object Detection


Abstract:

Unsupervised domain adaptive object detection methods transfer knowledge from the labelled source domain to a visually distinct and unlabeled target domain. Most methods ...Show More

Abstract:

Unsupervised domain adaptive object detection methods transfer knowledge from the labelled source domain to a visually distinct and unlabeled target domain. Most methods achieve this by training the detector model with the help of both labeled source and unlabeled target data. However, in real-world scenarios, gaining access to source data is not practical due to privacy concerns, legal issues and inefficient data transmission. To this end, we tackle the problem of Source-Free Domain Adaptive Object Detection, where during adaptation, we do not have access to the source data but only the source trained model. Specifically, we introduce Mixture of Teacher Experts (MoTE) method, where our key idea is to exploit the prediction uncertainty through a mixture of teacher models and progressively train the student model. We evaluate the proposed method by conducting extensive experiments on several object detection benchmark datasets to demonstrate the effectiveness of the proposed mixture of teacher expert based student-teacher training, specifically for source-free adaptation.
Date of Conference: 16-19 October 2022
Date Added to IEEE Xplore: 18 October 2022
ISBN Information:

ISSN Information:

Conference Location: Bordeaux, France
References is not available for this document.

1. INTRODUCTION

In the past decade, Convolutional Neural Network (CNN) based methods have achieved significant improvements for the task of object detection [1], [2], [3], [4], [5], [6]. These advancements have led to several successful applications in the real-world like video surveillance, autonomous navigation, image analysis, etc. However, existing object detectors are highly data-driven and susceptible to data distribution shift/domain shift. Consider a real-world application such as an autonomous car, where the object detectors are likely trained on data from one particular city, e.g. Tokyo City (TC), also referred to as source domain. While deploying the same model in a different city, e.g., New York City (NYC), also referred to as the target domain, the detector experiences severe performance degradation. This is due to the fact that object appearance, scene type, illumination, background, or weather condition in NYC are visually distinct from the TC. This problem is widely studied under the Unsupervised Domain Adaption (UDA) setting [7, 8, 9, 10, 11, 12, 13, 14], where the detection model is adopted to transfer the knowledge from the labelled source (e.g. TC) to unlabeled target (e.g. NYC) domain to overcome this poor generalization problem.

Select All
1.
Paul Viola, Michael Jones et al., "Rapid object detection using a boosted cascade of simple features", CVPR (1), vol. 1, pp. 511-518, 2001.
2.
Pedro F Felzenszwalb, Ross B Girshick, David McAllester and Deva Ramanan, "Object detection with discriminatively trained part-based models", IEEE transactions on pattern analysis and machine intelligence, vol. 32, no. 9, pp. 1627-1645, 2010.
3.
Ross Girshick, Jeff Donahue, Trevor Darrell and Jitendra Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation", Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 580-587, 2014.
4.
Ross Girshick, "Fast r-cnn", Proceedings of the IEEE international conference on computer vision, pp. 1440-1448, 2015.
5.
Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, et al., "Ssd: Single shot multi-box detector", European conference on computer vision, pp. 21-37, 2016.
6.
Shaoqing Ren, Kaiming He, Ross Girshick and Jian Sun, "Faster r-cnn: Towards real-time object detection with region proposal networks", Advances in neural information processing systems, pp. 91-99, 2015.
7.
Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira and Jennifer Wortman Vaughan, "A theory of learning from different domains", Machine learning, vol. 79, no. 1-2, pp. 151-175, 2010.
8.
Yuhua Chen, Wen Li, Christos Sakaridis, Dengxin Dai and Luc Van Gool, "Domain adaptive faster r-cnn for object detection in the wild", 2018 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3339-3348, 2018.
9.
Yuhu Shan, Wen Feng Lu and Chee Meng Chew, "Pixel and feature level based domain adaption for object detection in autonomous driving", 09 2018.
10.
Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada and Kate Saenko, "Strong-weak distribution alignment for adaptive object detection", CoRR, vol. abs/1812.04798, 2018.
11.
Naoto Inoue, Ryosuke Furuta, Toshihiko Yamasaki and Kiyoharu Aizawa, "Cross-domain weakly-supervised object detection through progressive domain adaptation", Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5001-5009, 2018.
12.
Aruni RoyChowdhury, Prithvijit Chakrabarty, Ashish Singh, SouY-oung Jin, Huaizu Jiang, Liangliang Cao, et al., "Automatic adaptation of object detectors to new domains using self-training", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 780-790, 2019.
13.
Han-Kai Hsu, Chun-Han Yao, Yi-Hsuan Tsai, Wei-Chih Hung, Hung-Yu Tseng, Maneesh Singh, et al., "Progressive do-main adaptation for object detection", Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 749-757, 2020.
14.
Taekyung Kim, Minki Jeong, Seunghyeon Kim, Seokeon Choi and Changick Kim, "Diversify and match: A domain adaptive representation learning paradigm for object detection", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 12456-12465, 2019.
15.
Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku and Tatsuya Harada, "Maximum classifier discrepancy for unsupervised domain adaptation", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3723-3732, 2018.
16.
Eric Tzeng, Judy Hoffman, Kate Saenko and Trevor Darrell, "Adversarial discriminative domain adaptation", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7167-7176, 2017.
17.
Mehran Khodabandeh, Arash Vahdat, Mani Ranjbar and William G Macready, "A robust learning approach to domain adaptive object detection", Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 480-490, 2019.
18.
Swami Sankaranarayanan, Yogesh Balaji, Carlos D Castillo and Rama Chellappa, "Generate to adapt: Aligning domains using generative adversarial networks", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8503-8512, 2018.
19.
Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, et al., "Cycada: Cycle-consistent adversarial domain adaptation", 2017.
20.
Zak Murez, Soheil Kolouri, David Kriegman, Ravi Ramamoorthi and Kyungnam Kim, "Image to image translation for domain adaptation", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4500-4509, 2018.
21.
Lanqing Hu, Meina Kan, Shiguang Shan and Xilin Chen, "Duplex generative adversarial network for unsupervised domain adaptation", Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1498-1507, 2018.
22.
Xianfeng Li, Weijie Chen, Di Xie, Shicai Yang, Peng Yuan, Shiliang Pu, et al., "A free lunch for unsupervised domain adaptive object detection without source data", 2020.
23.
Geoffrey Hinton, Oriol Vinyals and Jeff Dean, "Distilling the knowledge in a neural network", 2015.
24.
Antti Tarvainen and Harri Valpola, "Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results", 2017.
25.
Duc Tam Nguyen, Chaithanya Kumar Mummadi, Thi Phuong Nhung Ngo, Thi Hoai Phuong Nguyen, Laura Beggel and Thomas Brox, "Self: Learning to filter noisy labels with self-ensembling", 2019.
26.
Yen-Cheng Liu, Chih-Yao Ma, Zijian He, Chia-Wen Kuo, Kan Chen, Peizhao Zhang, et al., "Unbiased teacher for semi-supervised object detection", International Conference on Learning Representations, 2021.
27.
Balaji Lakshminarayanan, Alexander Pritzel and Charles Blundell, "Simple and scalable predictive uncertainty estimation using deep ensembles", 2016.
28.
Xu Lan, Xiatian Zhu and Shaogang Gong, "Knowledge distillation by on-the-fly native ensemble", 2018.
29.
Yarin Gal and Zoubin Ghahramani, "Dropout as a bayesian approximation: Representing model uncertainty in deep learning", international conference on machine learning, pp. 1050-1059, 2016.
30.
Chang-Dong Xu, Xing-Ran Zhao, Xin Jin and Xiu-Shen Wei, "Exploring categorical regularization for domain adaptive object detection", Proceedings of the IEEE/CVF Conference on Computer Vision and Pat-tern Recognition, pp. 11724-11733, 2020.

Contact IEEE to Subscribe

References

References is not available for this document.