Dual VAEGAN : A generative model for generalized zero-shot learning

Yuxuan LUO, Xizhao WANG*, Farhad POURPANAH

*Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

30 Citations (Scopus)

Abstract

Generalized zero-shot learning (GZSL) aims to recognize samples from all classes based on training samples of seen classes by bridging the gap between the seen and unseen classes through the semantic descriptions (attributes). Recently, generative-based methods have been used to convert the GZSL task into a supervised learning problem by generating visual features for unseen classes. In this paper, we propose a dual framework based on variational auto-encoder (VAE) and generative adversarial network (GAN), known as dual VAEGAN, to produce more clear visual features than VAE and alleviate the model collapse problem of GAN. To avoid generating unconstraint visual features, the generated visual features are forced to map back into their respective common space using a cycle consistency loss. Meanwhile, the cycle-consistency loss promotes the diversity and preserves the semantic consistency of the generated visual features. The experimental results on the six standard datasets indicate that dual VAEGAN can produce promising results as compared with other methods reported in the literature.

Original languageEnglish
Article number107352
Number of pages9
JournalApplied Soft Computing
Volume107
Early online date7 Apr 2021
DOIs
Publication statusPublished - Aug 2021
Externally publishedYes

Bibliographical note

This work was supported in part by the National Natural Science Fund of China (Grant 61976141) and in part by Key Project of Natural Science Foundation of China (Grant 61732011).

Keywords

  • Deep learning
  • Generalized zero-shot learning
  • Generative adversarial network
  • Variational auto-encoder

Fingerprint

Dive into the research topics of 'Dual VAEGAN : A generative model for generalized zero-shot learning'. Together they form a unique fingerprint.

Cite this