With the renaissance of deep neural networks (DNNs), humans have made a huge breakthrough in computer vision (CV) tasks, especially due to the emergence of convolutional neural networks (CNNs)-based and transformer-based variants. Nowadays, the ability of deep models on various tasks has far exceeded human standards. However, they still have many fatal defects. One of the most serious things is that they need numerous labeled instances to optimize parameters during training. For some species without enough or no training samples, they cannot be recognized as flexibly as humans. In order to resolve this dilemma, researchers attempt to classify certain species when training samples do not exist, which introduces a new enlightenment to image recognition called zero-shot learning (ZSL). Nonetheless, researchers gradually discover that the conventional ZSL task does not fit the realistic scenarios. In real scenarios, the conditions faced are more harsh, that is, when performing ZSL task, not only new unseen classes must be introduced, but also the model’s ability to classify the original seen classes should be also maintained. Based on the abovementioned, Chao et al. proposed another experimental setup and named it generalized ZSL (GZSL).
Abstract:
Due to the prosperous development of generative models, research works have achieved great success on the generalized zero-shot learning (GZSL) task. In most generative m...Show MoreMetadata
Abstract:
Due to the prosperous development of generative models, research works have achieved great success on the generalized zero-shot learning (GZSL) task. In most generative methods of GZSL, researchers try to utilize attributes and normally distributed noise to generate visual features, which ignores whether the normal distribution can perfectly represent all categories. Therefore, in this article, we exploit variational auto-encoders (VAE) and visual features to generate image-level noise that can preserve class-level characteristics in more detail and propose a mechanism called more factual generative network (MFGN) to achieve more authentic generative process. In other words, it is to transfer the seen feature distribution to the unseen domains and regulate the knowledge to correct the generation of unseen samples. Extensive experiments are conducted on four popular datasets and the results demonstrate the effectiveness of the proposed work.
Published in: IEEE MultiMedia ( Volume: 29, Issue: 3, 01 July-Sept. 2022)
Funding Agency:
References is not available for this document.
Select All
1.
W.-L. Chao, S. Changpinyo, B. Gong and F. Sha, "An empirical study and analysis of generalized zero-shot learning for object recognition in the wild" in Proc. Eur. Conf. Comput. Vis., pp. 52-68, 2016.
2.
J. Li, M. Jing, K. Lu, Z. Ding, L. Zhu and Z. Huang, "Leveraging the invariant side of generative zero-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 7394-7403, 2019.
3.
E. Schonfeld, S. Ebrahimi, S. Sinha, T. Darrell and Z. Akata, "Generalized zero-and few-shot learning via aligned variational autoencoders", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 8239-8247, 2019.
4.
R. Keshari, R. Singh and M. Vatsa, "Generalized zero-shot learning via over-complete distribution", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 13297-13305, 2020.
5.
Z. Ding and H. Liu, "Marginalized latent semantic encoder for zero-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 6184-6192, 2019.
6.
Y. Xian, T. Lorenz, B. Schiele and Z. Akata, "Feature generating networks for zero-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 5542-5551, 2018.
7.
M. B. Sariyildiz and R. G. Cinbis, "Gradient matching generative networks for zero-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 2163-2173, 2019.
8.
Z. Han, Z. Fu and J. Yang, "Learning the redundancy-free features for generalized zero-shot object recognition", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 12862-12871, 2020.
9.
C. H. Lampert, H. Nickisch and S. Harmeling, "Learning to detect unseen object classes by between-class attribute transfer", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 951-958, 2009.
10.
Y. Xian, C. H. Lampert, B. Schiele and Z. Akata, "Zero-shot learning—A comprehensive evaluation of the good the bad and the ugly" in IEEE Trans. Pattern Anal. Mach. Intell., vol. 41, no. 9, pp. 2251-2265, Sep. 2019.
11.
A. Farhadi, I. Endres, D. Hoiem and D. Forsyth, "Describing objects by their attributes", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 1778-1785, 2009.
12.
P. Welinder et al., "The caltech-UCSD birds-200-2011 dataset", 2010.
13.
G. Patterson and J. Hays, "Sun attribute database: Discovering annotating and recognizing scene attributes", Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 2751-2758, 2012.
14.
F. Sung, Y. Yang, L. Zhang, T. Xiang, P. H. Torr and T. M. Hospedales, "Learning to compare: Relation network for few-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 1199-1208, 2018.
15.
H. Jiang, R. Wang, S. Shan and X. Chen, "Transferable contrastive network for generalized zero-shot learning", Proc. IEEE/CVF Int. Conf. Comput. Vis., pp. 9764-9773, 2019.
16.
R. Felix, M. Sasdelli, I. Reid and G. Carneiro, "Multi-modal ensemble classification for generalized zero shot learning", 2019.
17.
K. Li, M. R. Min and Y. Fu, "Rethinking zero-shot learning: A conditional visual classification perspective", Proc. IEEE/CVF Int. Conf. Comput. Vis., pp. 3582-3591, 2019.
18.
S. Liu, M. Long, J. Wang and M. I. Jordan, "Generalized zero-shot learning with deep calibration network", Proc. Adv. Neural Inf. Process. Syst., pp. 2005-2015, 2018.
19.
H. Huang, C. Wang, P. S. Yu and C.-D. Wang, "Generative dual adversarial network for generalized zero-shot learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 801-810, 2019.
20.
J. Liu, H. Bai, H. Zhang and L. Liu, "Near-real feature generative network for generalized zero-shot learning", Proc. IEEE Int. Conf. Multimedia Expo, pp. 1-6, 2021.