Abstract
Sketch-to-image (S2I) translation plays an important role in image synthesis and manipulation tasks, such as photo editing and colorization. Some specific S2I translation including sketch-to-photo and sketch-to-painting can be used as powerful tools in the art design industry. However, previous methods only support S2I translation with a single level of density, which gives less flexibility to users for controlling the input sketches. In this work, we propose the first multi-level density sketch-to-image translation framework, which allows the input sketch to cover a wide range from rough object outlines to micro structures. Moreover, to tackle the problem of noncontinuous representation of multi-level density input sketches, we project the density level into a continuous latent space, which can then be linearly controlled by a parameter. This allows users to conveniently control the densities of input sketches and generation of images. Moreover, our method has been successfully verified on various datasets for different applications including face editing, multi-modal sketchto- photo translation, and anime colorization, providing coarse-tofine levels of controls to these applications.
Original language | English |
---|---|
Pages (from-to) | 4002-4015 |
Journal | IEEE Transactions on Multimedia |
Volume | 24 |
Early online date | 14 Sept 2021 |
DOIs | |
Publication status | Published - 2022 |
Externally published | Yes |
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 61672443, in part by Hong Kong GRF-RGC General Research Fund under Grants 9042322 (CityU 11200116), 9042489 (CityU 11206317), and 9042816 (CityU 11209819), in part by Hong Kong Research Grants Council (RGC) Early Career Scheme under Grant 9048148 (CityU 21209119), and in part by the CityU of Hong Kong under APRC Grant 9610488.
Keywords
- Deep image synthesis
- GAN
- interactive editing
- multi-scale disentangle