Generation with Nuanced Changes: Continuous Image-to-Image Translation with Adversarial Preferences

Yinghua YAO, Yuangang PAN, Ivor W. TSANG*, Xin YAO*

*Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

Abstract

Most previous methods for continuous image-to-image translation resorted to binary attributes with restrictive description ability and thus cannot achieve satisfactory performance. Some works proposed to use fine-grained semantic information, relative attributes (RAs), preferences over pairs of images on the strength of a specified attribute. However, they still failed to reconcile both goals for smooth translation and for high-quality generation simultaneously. In this work, we propose a new model continuous translation via adversarial preferences (CTAP) to coordinate these two goals for high-quality continuous translation based on RAs. In CTAP, we simultaneously train two modules: a generator that translates an input image to the desired image with smooth nuanced changes w.r.t. the interested attributes; and a ranker that executes adversarial preferences consisting of the input image and the desired image. Particularly, adversarial preferences involve an adversarial ranking process: 1) the ranker thinks no difference between the desired image and the input image in terms of the interested attributes; 2) the generator fools the ranker to believe the attributes of its output image changes as expect compared with the input image. RAs over pairs of real images are introduced to guide the ranker to rank image pairs regarding the interested attributes only. With an effective ranker, the generator would "win"the adversarial game by producing high-quality images that present smooth changes. The experiments on two face datasets and one shoe dataset demonstrate that our CTAP achieves state-of-art results in generating high-fidelity images which exhibit smooth changes over the interested attributes.

Original languageEnglish
Pages (from-to)816-828
Number of pages13
JournalIEEE Transactions on Artificial Intelligence
Volume6
Issue number4
Early online date13 Nov 2024
DOIs
Publication statusPublished - 2025

Bibliographical note

Publisher Copyright:
© 2024 IEEE.

Funding

This work was supported in part by the Agency for Science, Technology and Research (A*STAR) Centre for Frontier AI Research, the A*STAR GAP project under Grant I23D1AG079, in part by the NSFC under Grant 62250710682, and in part by the Program for Guangdong Provincial Key Laboratory under Grant 2020B121201001. This work was supported by the Agency for Science, Technology and Research (A*STAR) Centre for Frontier AI Research, the A*STAR GAP project (Grant No. I23D1AG079), NSFC (Grant No. 62250710682), and the Program for Guangdong Provincial Key Laboratory (Grant No. 2020B121201001). Yinghua Yao, Yuangang Pan and Ivor W. Tsang are with the Centre for Frontier AI Research, Agency for Science, Technology and Research (A*STAR), 138632, Singapore and also with the Institute of High Performance Computing, Agency for Science, Technology and Research (A*STAR), 138632, Singapore. Xin Yao is with the School of Data Science, Lingnan University, Hong Kong and also with the School of Computer Science, University of Birmingham, B15 2TT Birmingham, UK. Email: [email protected], [email protected], [email protected], [email protected]. (Cor-resonding author: Ivor W. Tsang and Xin Yao)

Keywords

  • Adversarial preferences
  • continuous image-to-image translation
  • generative adversarial network (GAN)
  • relative attributes (RAs)

Fingerprint

Dive into the research topics of 'Generation with Nuanced Changes: Continuous Image-to-Image Translation with Adversarial Preferences'. Together they form a unique fingerprint.

Cite this