Loading [MathJax]/extensions/MathMenu.js
Boosting Contrastive Self-Supervised Learning with False Negative Cancellation | IEEE Conference Publication | IEEE Xplore

Boosting Contrastive Self-Supervised Learning with False Negative Cancellation


Abstract:

Self-supervised representation learning has made significant leaps fueled by progress in contrastive learning, which seeks to learn transformations that embed positive in...Show More

Abstract:

Self-supervised representation learning has made significant leaps fueled by progress in contrastive learning, which seeks to learn transformations that embed positive input pairs nearby, while pushing negative pairs far apart. While positive pairs can be generated reliably (e.g., as different views of the same image), it is difficult to accurately establish negative pairs, defined as samples from different images regardless of their semantic content or visual features. A fundamental problem in contrastive learning is mitigating the effects of false negatives. Contrasting false negatives induces two critical issues in representation learning: discarding semantic information and slow convergence. In this paper, we propose novel approaches to identify false negatives, as well as two strategies to mitigate their effect, i.e. false negative elimination and attraction, while systematically performing rigorous evaluations to study this problem in detail. Our method exhibits consistent improvements over existing contrastive learning-based methods. Without labels, we identify false negatives with ~40% accuracy among 1000 semantic classes on ImageNet, and achieve 5.8% absolute improvement in top-1 accuracy over the previous state-of-the-art when finetuning with 1% labels. Our code is available at https://github.com/google-research/fnc
Date of Conference: 03-08 January 2022
Date Added to IEEE Xplore: 15 February 2022
ISBN Information:

ISSN Information:

Conference Location: Waikoloa, HI, USA
References is not available for this document.

1. Introduction

Representation learning has become the backbone of most modern AI agents. High quality pretrained representations are essential to improving performance on downstream tasks [16], [22], [52], [29]. While conventional approaches rely on labeled data, there has been a recent surge in self-supervised representation learning [20], [15], [37], [46], [39], [8], [53], [32]. In fact, self-supervised representation learning has been closing the gap with and, in some cases, even surpassing its supervised counterpart [9], [24], [11], [10]. Notably, most state-of-the-art methods are converging around and fueled by the central concept of contrastive learning [45], [25], [26], [43], [35], [24], [9].

Select All
1.
Y. M. Asano, C. Rupprecht and A. Vedaldi, "Self-labelling via simultaneous clustering and representation learning", Proc. Int’l Conf. on Learning Representations (ICLR), 2020.
2.
P. Bachman, R. D. Hjelm and W. Buchwalter, "Learning representations by maximizing mutual information across views", Advances in Neural Information Processing Systems (NeurIPS), pp. 15535-15545, 2019.
3.
M. A. Bautista, A. Sanakoyeu, E. Tikhoncheva and B. Ommer, "Cliquecnn: Deep unsupervised exemplar learning", Advances in Neural Information Processing Systems (NeurIPS), pp. 3846-3854, 2016.
4.
T. Berg, J. Liu, S. W. Lee, M. L. Alexander, D. W. Jacobs and P. N. Belhumeur, "Large-scale fine-grained visual categorization of birds", Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2014.
5.
L. Bossard, M. Guillaumin and L. V. Gool, "Food-101 Mining discriminative components with random forests", Proc. European Conf. on Computer Vision (ECCV), 2014.
6.
M. Caron, P. Bojanowski, A. Joulin and M. Douze, "Deep clustering for unsupervised learning of visual features", Proc. European Conf. on Computer Vision (ECCV), 2018.
7.
M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski and A. Joulin, "Unsupervised learning of visual features by contrasting cluster assignments", Advances in Neural Information Processing Systems (NeurIPS), 2020.
8.
M. Chen, A. Radford, R. Child, J. Wu, H. Jun, P. Dhariwal, et al., "Generative pretraining from pixels", Proc. Int’l Conf. on Machine Learning (ICML), 2020.
9.
T. Chen, S. Kornblith, M. Norouzi and G. Hinton, "A simple framework for contrastive learning of visual representations", Proc. Int’l Conf. on Machine Learning (ICML), 2020.
10.
T. Chen, S. Kornblith, K. Swersky, M. Norouzi and G. Hinton, "Big self-supervised models are strong semi-supervised learners", Advances in Neural Information Processing Systems (NeurIPS), 2020.
11.
X. Chen, H. Fan, R. Girshick and K. He, Improved baselines with momentum contrastive learning, 2020.
12.
C.-Y. Chuang, J. Robinson, L. Yen-Chen, A. Torralba and S. Jegelka, "Debiased contrastive learning", Advances in Neural Information Processing Systems (NeurIPS), 2020.
13.
M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed and A. Vedaldi, "Describing textures in the wild", Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2014.
14.
E. D. Cubuk, B. Zoph, J. Shlens and Q. V. Le, "Randaugment: Practical data augmentation with no separate search", Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2020.
15.
C. Doersch, A. Gupta and A. A. Efros, "Unsupervised visual representation learning by context prediction", Proc. Int’l. Conf. on Computer Vision (ICCV), 2015.
16.
J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, et al., "DeCAF: A deep convolutional activation feature for generic visual recognition", Proc. Int’l Conf. on Machine Learning (ICML), pp. 647-655, 2014.
17.
M. Everingham, L. V. Gool, C. K. Williams, J. Winn and A. Zisserman, "The Pascal visual object classes (VOC) challenge", Int’l J. on Computer Vision, vol. 88, no. 2, pp. 303-338, 2010.
18.
L. Fei-Fei, R. Fergus and P. Perona, "Learning generative visual models from few training examples: An incremental Bayesian approach tested on 101 object categories", Proceedings of the CVPR Workshop on Generative-Model Based Vision, 2004.
19.
R. Geirhos, J.-H. Jacobsen, C. Michaelis, R. Zemel, W. Brendel, M. Bethge, et al., Shortcut learning in deep neural networks, 2020.
20.
S. Gidaris, P. Singh and N. Komodakis, "Unsupervised representation learning by predicting image rotations", Proc. Int’l Conf. on Learning Representations (ICLR), 2018.
21.
B. Ginsburg, I. Gitman and Y. You, "Large batch training of convolutional networks with layer-wise adaptive rate scaling", Proc. Int’l Conf. on Learning Representations (ICLR), 2018.
22.
R. Girshick, J. Donahue, T. Darrell and J. Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation", Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), pp. 580-587, 2014.
23.
J.-B. Grill, F. Strub, F. Altché, C. Tallec, P. H. Richemond, E. Buchatskaya, et al., Bootstrap your own latent: A new approach to self-supervised learning, 2020.
24.
K. He, H. Fan, Y. Wu, S. Xie and R. Girshick, "Momentum contrast for unsupervised visual representation learning", Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019.
25.
O. J. Henaff, A. Srinivas, J. D. Fauw, A. Razavi, C. Doersch, S. M. A. Eslami, et al., "Data-efficient image recognition with contrastive predictive coding", Proc. Int’l Conf. on Learning Representations (ICLR), 2020.
26.
R. D. Hjelm, A. Fedorov, S. Lavoie-Marchildon, K. Grewal, P. Bachman, A. Trischler, et al., "Learning deep representations by mutual information estimation and maximization", Proc. Int’l Conf. on Learning Representations (ICLR), 2019.
27.
Y. Kalantidis, M. B. Sariyildiz, N. Pion, P. Weinzaepfel and D. Larlus, "Hard negative mixing for contrastive learning", Advances in Neural Information Processing Systems (NeurIPS), 2020.
28.
P. Khosla, P. Teterwak, C. Wang, A. Sarna, Y. Tian, P. Isola, et al., Supervised contrastive learning, 2020.
29.
D.-K. Kim and M. R. Walter, "Satellite image-based localization via learned embeddings", Proc. IEEE Int’l Conf. on Robotics and Automation (ICRA), May 2017.
30.
J. Krause, M. S. J. Dengand and L. Fei-Fei, "3D object representations for fine-grained categorization", Proc. Int’l. Conf. on Computer Vision (ICCV), 2013.

Contact IEEE to Subscribe

References

References is not available for this document.