I. Introduction
Commonsense reasoning [1], [2] is a challenging task, requiring complex question predictions using the described textual contexts and unstated association knowledge about the world, which has made significant progress with the development of language models (LMs) [3]. Language encoder models [4], such as BERT [5], [6] and its variants [7], [8], have outperformed on various commonsense reasoning tasks, like question answering (QA) [9], natural language inference (NLI) [10], and text generation. Recently, massive studies have demonstrated the significant role of knowledge graphs (KGs) [11], which obtain external knowledge explicitly using relationships between entities and can play in structured reasoning and query answering [12]. However, recent commonsense reasoning advances are restricted to absolutely right supervised scenarios. Consequently, they are not scalable to real-world applications [13], [14], [15] where the training labels may be incorrect, which would mislead the model to learn or memorize wrong correlations.