Loading [MathJax]/extensions/MathMenu.js
Robust Commonsense Reasoning Against Noisy Labels Using Adaptive Correction | IEEE Journals & Magazine | IEEE Xplore

Robust Commonsense Reasoning Against Noisy Labels Using Adaptive Correction


Abstract:

Commonsense reasoning based on knowledge graphs (KGs) is a challenging task that requires predicting complex questions over the described textual contexts and relevant kn...Show More

Abstract:

Commonsense reasoning based on knowledge graphs (KGs) is a challenging task that requires predicting complex questions over the described textual contexts and relevant knowledge about the world. However, current methods typically assume clean training scenarios with accurately labeled samples, which are often unrealistic. The training set can include mislabeled samples, and the robustness to label noises is essential for commonsense reasoning methods to be practical, but this problem remains largely unexplored. This work focuses on commonsense reasoning with mislabeled training samples and makes several technical contributions: 1) we first construct diverse augmentations from knowledge and model, and offer a simple yet effective multiple-choice alignment method to divide the training samples into clean, semi-clean, and unclean parts; 2) we design adaptive label correction methods for the semi-clean and unclean samples to exploit the supervised potential of noisy information; and 3) finally, we extensively test these methods on noisy versions of commonsense reasoning benchmarks (CommonsenseQA and OpenbookQA). Experimental results show that the proposed method can significantly enhance robustness and improve overall performance. Furthermore, the proposed method is generally applicable to multiple existing commonsense reasoning frameworks to boost their robustness. The code is available at https://github.com/xdxuyang/CR_Noisy_Labels.
Published in: IEEE Transactions on Cybernetics ( Volume: 54, Issue: 7, July 2024)
Page(s): 4138 - 4149
Date of Publication: 27 December 2023

ISSN Information:

PubMed ID: 38150342

Funding Agency:

References is not available for this document.

I. Introduction

Commonsense reasoning [1], [2] is a challenging task, requiring complex question predictions using the described textual contexts and unstated association knowledge about the world, which has made significant progress with the development of language models (LMs) [3]. Language encoder models [4], such as BERT [5], [6] and its variants [7], [8], have outperformed on various commonsense reasoning tasks, like question answering (QA) [9], natural language inference (NLI) [10], and text generation. Recently, massive studies have demonstrated the significant role of knowledge graphs (KGs) [11], which obtain external knowledge explicitly using relationships between entities and can play in structured reasoning and query answering [12]. However, recent commonsense reasoning advances are restricted to absolutely right supervised scenarios. Consequently, they are not scalable to real-world applications [13], [14], [15] where the training labels may be incorrect, which would mislead the model to learn or memorize wrong correlations.

Select All
1.
D. Khashabi et al., "UnifiedQA: Crossing format boundaries with a single QA system", arXiv:2005.00700, 2020.
2.
S. Ji, S. Pan, E. Cambria, P. Marttinen and S. Y. Philip, "A survey on knowledge graphs: Representation acquisition and applications", IEEE Trans. Neural Netw. Learn. Syst., vol. 33, no. 2, pp. 494-514, Feb. 2022.
3.
Y. Zhu et al., "Configurable graph reasoning for visual relationship detection", IEEE Trans. Neural Netw. Learn. Syst., vol. 33, no. 1, pp. 117-129, Jan. 2022.
4.
W. Liu, X. Zhao, Z. Zhao, Q. Ju, X. Yang and W. Lu, "An empirical study on adaptive inference for pretrained language model", IEEE Trans. Neural Netw. Learn. Syst., vol. 34, no. 8, pp. 4321-4331, Aug. 2023.
5.
J. Devlin, M.-W. Chang, K. Lee and K. Toutanova, "Bert: Pre-training of deep bidirectional transformers for language understanding", arXiv:1810.04805, 2018.
6.
X. Yang, H. Wang, D. Xie, C. Deng and D. Tao, "Object-agnostic transformers for video referring segmentation", IEEE Trans. Image Process., vol. 31, pp. 2839-2849, 2022.
7.
Y. Liu et al., "ROBERTa: A robustly optimized BERT pretraining approach", arXiv:1907.11692, 2019.
8.
Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma and R. Soricut, "ALBERT: A lite BERT for self-supervised learning of language representations", arXiv:1909.11942, 2019.
9.
B. Y. Lin, X. Chen, J. Chen and X. Ren, "KagNet: Knowledge-aware graph networks for commonsense reasoning", arXiv:1909.02151, 2019.
10.
Q. Chen, X. Zhu, Z.-H. Ling, D. Inkpen and S. Wei, "Neural natural language inference models enhanced with external knowledge", arXiv:1711.04289, 2017.
11.
X. Zhao, X. Feng and H. Chen, "A background knowledge revising and incorporating dialogue model", IEEE Trans. Neural Netw. Learn. Syst., vol. 34, no. 8, pp. 3874-3884, Aug. 2023.
12.
Y.-B. Kang, S. Krishnaswamy and A. Zaslavsky, "A retrieval strategy for case-based reasoning using similarity and association knowledge", IEEE Trans. Cybern., vol. 44, no. 4, pp. 473-487, Apr. 2014.
13.
X. Huang, Y. Peng and Z. Wen, "Visual-textual hybrid sequence matching for joint reasoning", IEEE Trans. Cybern., vol. 51, no. 12, pp. 5692-5705, Dec. 2021.
14.
K. Wei, C. Deng, X. Yang and D. Tao, "Incremental zero-shot learning", IEEE Trans. Cybern., vol. 52, no. 12, pp. 13788-13799, Dec. 2022.
15.
X. Yang, C. Deng, T. Liu and D. Tao, "Heterogeneous graph attention network for unsupervised multiple-target domain adaptation", IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 4, pp. 1992-2003, Apr. 2022.
16.
A. Talmor, J. Herzig, N. Lourie and J. Berant, "CommonsenseQA: A question answering challenge targeting commonsense knowledge", arXiv:1811.00937, 2018.
17.
X. Zhang et al., "GreaseLM: Graph reasoning enhanced language models for question answering", arXiv:2201.08860, 2022.
18.
S. Wu et al., "Class2Simi: A noise reduction perspective on learning with noisy labels", Proc. ICML, pp. 11285-11295, 2021.
19.
T. Liu and D. Tao, "Classification with noisy labels by importance reweighting", IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 3, pp. 447-461, Mar. 2016.
20.
X. Xia et al., "Sample selection with uncertainty of losses for learning with noisy labels", arXiv:2106.00445, 2021.
21.
S. Yang et al., "Estimating instance-dependent label-noise transition matrix using DNNs", arXiv:2105.13001, 2021.
22.
S. Wu, M. Gong, B. Han, Y. Liu and T. Liu, "Fair classification with instance-dependent label noise", Proc. CLeaR, pp. 927-943, 2022.
23.
B. Han et al., "Co-teaching: Robust training of deep neural networks with extremely noisy labels", Proc. NeurIPS, vol. 31, pp. 8536-8546, 2018.
24.
Y. Wang et al., "Iterative learning with open-set noisy labels", Proc. CVPR, pp. 8688-8696, 2018.
25.
H. Wei, L. Feng, X. Chen and B. An, "Combating noisy labels by agreement: A joint training method with co-regularization", Proc. CVPR, pp. 13726-13735, 2020.
26.
T. Wei, J.-X. Shi, W.-W. Tu and Y.-F. Li, "Robust long-tailed learning under label noise", arXiv:2108.11569, 2021.
27.
B. Du, T. Xinyao, Z. Wang, L. Zhang and D. Tao, "Robust graph-based semisupervised learning for noisy labeled data via maximum correntropy criterion", IEEE Trans. Cybern., vol. 49, no. 4, pp. 1440-1453, Apr. 2019.
28.
Y. Yao et al., "Learning latent stable patterns for image understanding with weak and noisy labels", IEEE Trans. Cybern., vol. 49, no. 12, pp. 4243-4252, Dec. 2019.
29.
Y. Feng, H. Zhu, D. Peng, X. Peng and P. Hu, "RONO: Robust discriminative learning with noisy labels for 2D-3D cross-modal retrieval", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 11610-11619, 2023.
30.
T. Xiao, T. Xia, Y. Yang, C. Huang and X. Wang, "Learning from massive noisy labeled data for image classification", Proc. CVPR, pp. 2691-2699, 2015.
Contact IEEE to Subscribe

References

References is not available for this document.