Abstract
Generalized zero-shot learning (GZSL) aims to recognize samples from all classes based on training samples of seen classes by bridging the gap between the seen and unseen classes through the semantic descriptions (attributes). Recently, generative-based methods have been used to convert the GZSL task into a supervised learning problem by generating visual features for unseen classes. In this paper, we propose a dual framework based on variational auto-encoder (VAE) and generative adversarial network (GAN), known as dual VAEGAN, to produce more clear visual features than VAE and alleviate the model collapse problem of GAN. To avoid generating unconstraint visual features, the generated visual features are forced to map back into their respective common space using a cycle consistency loss. Meanwhile, the cycle-consistency loss promotes the diversity and preserves the semantic consistency of the generated visual features. The experimental results on the six standard datasets indicate that dual VAEGAN can produce promising results as compared with other methods reported in the literature.
Original language | English |
---|---|
Article number | 107352 |
Number of pages | 9 |
Journal | Applied Soft Computing |
Volume | 107 |
Early online date | 7 Apr 2021 |
DOIs | |
Publication status | Published - Aug 2021 |
Externally published | Yes |
Bibliographical note
This work was supported in part by the National Natural Science Fund of China (Grant 61976141) and in part by Key Project of Natural Science Foundation of China (Grant 61732011).Keywords
- Deep learning
- Generalized zero-shot learning
- Generative adversarial network
- Variational auto-encoder