Loading [MathJax]/extensions/MathMenu.js
ITSMatch: Improved Safe Semi-Supervised Text Classification Under Class Distribution Mismatch | IEEE Journals & Magazine | IEEE Xplore

ITSMatch: Improved Safe Semi-Supervised Text Classification Under Class Distribution Mismatch


Abstract:

Deep semi-supervised learning (SSL) brings deep learning from lab with expensive label data costs to real-world commercial application. Today, deep SSL is being universal...Show More

Abstract:

Deep semi-supervised learning (SSL) brings deep learning from lab with expensive label data costs to real-world commercial application. Today, deep SSL is being universally applied in various artificial intelligence commercial technologies. However, there may be a distribution mismatch between labeled and unlabeled datasets in practical application, which is a key issue that degrades deep SSL performance. Some recent studies deal with out-of-distribution (OOD) data by directly removing or uniformly reducing weights, which ignore potential value of OOD data. To address the issue, we propose ITSMatch, a simple, safe and effective SSL method to process text classification by recycling OOD data near labeled domain to fully utilize data information. Specifically, a weighted adversarial domain adaptation is applied to OOD data to project it into the space of labeled and in-distribution (ID) data, and its recover ability is accurately quantified by the transferable score. ITSMatch unifies mainstream methods, including pseudo-labels generation and consistency regularization on unlabeled data and its augmented data. Besides, we also perform metric learning on labeled data and ID data with pseudo-labels to fully acquire sample space features. Experiment results on the AG News and Yelp datasets demonstrate that our ITSMatch method performs better than the baseline methods including TextSMatch, MixText, UDA, and BERT. This method of semi-supervised text classification can be applied to the analysis of product reviews on e-commerce platforms to improve customers’ online shopping experience.
Published in: IEEE Transactions on Consumer Electronics ( Volume: 70, Issue: 2, May 2024)
Page(s): 4729 - 4740
Date of Publication: 13 October 2023

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Text classification technology, powered by artificial intelligence, has been widely used in intelligent electronic systems. Sentiment analysis is a fundamental application of text classification to analyze users’ emotional tendencies towards specific products, services, or events by recognizing and classifying emotions. Afzaal et al. [1] proposed a framework for sentiment classification of tourism reviews. Kim et al. [2] introduced a system that classifies music based on users’ emotional responses. Similarly, Chatterjee et al. [3] noted the widespread application of speech emotion recognition (SER) in the consumer field and introduced a comprehensive method for analyzing sentiment in human speech. Prabhakar et al. [4] introduced a multi-channel CNN-BLSTM architecture to enhance traditional speech emotion recognition techniques. Moreover, text classification can also be employed for user intention recognition, analyzing user comments, feedback, or social media posts to discern their intentions. Pal et al. [5] analyzed factors influencing users’ decisions to purchase IoT (Internet of Things) consumer electronic devices by classifying text from dialogues. The growing popularity and development of consumer electronics underscore the increasing demand for intelligent systems to effectively process and classify text.

Select All
1.
M. Afzaal, M. Usman and A. Fong, "Tourism mobile app with aspect-based sentiment classification framework for tourist reviews", IEEE Trans. Consum. Electron., vol. 65, no. 2, pp. 233-242, May 2019.
2.
H.-G. Kim, G. Y. Lee and M.-S. Kim, "Dual-function integrated emotion-based music classification system using features from physiological signals", IEEE Trans. Consum. Electron., vol. 67, no. 4, pp. 341-349, Nov. 2021.
3.
R. Chatterjee, S. Mazumdar, R. S. Sherratt, R. Halder, T. Maitra and D. Giri, "Real-time speech emotion analysis for smart home assistants", IEEE Trans. Consum. Electron., vol. 67, no. 1, pp. 68-76, Feb. 2021.
4.
G. A. Prabhakar, B. Basel, A. Dutta and C. V. Rama Rao, "Multichannel CNN-BLSTM architecture for speech emotion recognition system by fusion of magnitude and phase spectral features using DCCA for consumer applications", IEEE Trans. Consum. Electron., vol. 69, no. 2, pp. 226-235, May 2023.
5.
D. Pal, V. Vanijja, X. Zhang and H. Thapliyal, "Exploring the antecedents of consumer electronics IoT devices purchase decision: A mixed methods study", IEEE Trans. Consum. Electron., vol. 67, no. 4, pp. 305-318, Nov. 2021.
6.
K. Wang, C. Han, S. Long and J. Poon, "ME-GCN: Multi-dimensional edge-embedded graph convolutional networks for semi-supervised text classification", Proc. Workshop Deep Learn. Graphs Nat. Lang. Process., pp. 1-13, 2022.
7.
J. Chen, Z. Yang and D. Yang, "MixText: Linguistically-informed interpolation of hidden space for semi-supervised text classification", Proc. ACL, pp. 2147-2157, 2020.
8.
P. Karisani and N. Karisani, "Semi-supervised text classification via self-pretraining", Proc. 14th ACM Int. Conf. Web Search Data Min., pp. 40-48, 2021.
9.
T. Miyato, A. M. Dai and I. Goodfellow, "Adversarial training methods for semi-supervised text classification", arXiv:1605.07725, 2016.
10.
D. S. Sachan, M. Zaheer and R. Salakhutdinov, "Revisiting LSTM networks for semi-supervised text classification via mixed objective function", Proc. AAAI Conf. Artif. Intell., vol. 33, pp. 6940-6948, 2019.
11.
P. Kirichenko, P. Izmailov and A. G. Wilson, "Why normalizing flows fail to detect out-of-distribution data", Proc. Adv. Neural Inf. Process. Syst., vol. 33, pp. 20578-20589, 2020.
12.
C. Liu, Z. Mengchao, F. Zhibing, P. Hou and Y. Li, "FLiText: A faster and lighter semi-supervised text classification with convolution networks", Proc. Conf. Empiric. Methods Nat. Lang. Process., pp. 2481-2491, 2021.
13.
A. Tarvainen and H. Valpola, "Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results", Proc. Adv. Neural Inf. Process. Syst., vol. 30, pp. 1195-1204, 2017.
14.
H. Jo and C. Cinarel, "Delta-training: Simple semi-supervised text classification using Pretrained word Embeddings", Proc. Conf. Empiric. Methods Nat. Lang. Process. 9th Int. Joint Conf. Nat. Lang. Process. (EMNLP-IJCNLP), pp. 3458-3463, 2019.
15.
S. Gururangan, T. Dang, D. Card and N. A. Smith, "Variational pretraining for semi-supervised text classification", arXiv:1906.02242, 2019.
16.
H.-M. Xu, L. Liu and E. Abbasnejad, "Progressive class semantic matching for semi-supervised text classification", arXiv:2205.10189, 2022.
17.
C. Wei, K. Sohn, C. Mellina, A. Yuille and F. Yang, "CReST: A class-rebalancing self-training framework for imbalanced semi-supervised learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 10857-10866, 2021.
18.
J. Kim, Y. Hur, S. Park, E. Yang, S. J. Hwang and J. Shin, "Distribution aligning refinery of pseudo-label for imbalanced semi-supervised learning", Proc. Adv. Neural Inf. Process. Syst., vol. 33, pp. 14567-14579, 2020.
19.
Z. Wang, X. Wang, R. Sun and T.-H. Chang, "Federated semi-supervised learning with class distribution mismatch", arXiv:2111.00010, 2021.
20.
A. Oliver, A. Odena, C. A. Raffel, E. D. Cubuk and I. Goodfellow, "Realistic evaluation of deep semi-supervised learning algorithms", Proc. Adv. Neural Inf. Process. Syst., vol. 31, pp. 3239-3250, 2018.
21.
Y. Chen, X. Zhu, W. Li and S. Gong, "Semi-supervised learning under class distribution mismatch", Proc. AAAI Conf. Artif. Intell., vol. 34, pp. 3569-3576, 2020.
22.
L.-Z. Guo, Z.-Y. Zhang, Y. Jiang, Y.-F. Li and Z.-H. Zhou, "Safe deep semi-supervised learning for unseen-class unlabeled data", Proc. Int. Conf. Mach. Learn., pp. 3897-3906, 2020.
23.
S. Calderon-Ramirez et al., "MixMOOD: A systematic approach to class distribution mismatch in semi-supervised learning using deep dataset dissimilarity measures", arXiv:2006.07767, 2020.
24.
A. Nguyen, J. Yosinski and J. Clune, "Deep neural networks are easily fooled: High confidence predictions for unrecognizable images", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 427-436, 2015.
25.
D.-H. Lee et al., "Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks", Proc. Workshop Challenges Represent. Learn. (ICML), vol. 3, pp. 896, 2013.
26.
W. D.-D. Chen and Z.-H. WeiGao, "Tri-net for semi-supervised deep learning", Proc. IJCAI, pp. 2014-2020, 2018.
27.
W. Sun, T.-L. B. Tseng, J. Zhang and W. Qian, "Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data", Comput. Med. Imag. Graph., vol. 57, pp. 4-9, Apr. 2017.
28.
K. Chen, L. Yao, D. Zhang, X. Wang, X. Chang and F. Nie, "A semisupervised recurrent convolutional attention model for human activity recognition", IEEE Trans. Neural Netw. Learn. Syst., vol. 31, no. 5, pp. 1747-1756, May 2020.
29.
E. Yu, J. Sun, J. Li, X. Chang, X.-H. Han and A. G. Hauptmann, "Adaptive semi-supervised feature selection for cross-modal retrieval", IEEE Trans. Multimedia, vol. 21, no. 5, pp. 1276-1288, May 2019.
30.
M. Luo, X. Chang, L. Nie, Y. Yang, A. G. Hauptmann and Q. Zheng, "An adaptive semisupervised feature analysis for video semantic recognition", IEEE Trans. Cybern., vol. 48, no. 2, pp. 648-660, Feb. 2018.
Contact IEEE to Subscribe

References

References is not available for this document.