Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/172645
Title: AgileGAN: stylizing portraits by inversion-consistent transfer learning
Authors: Song, Guoxian
Luo, Linjie
Liu, Jing
Ma, Wan-Chun
Lai, Chunpong
Zheng, Chuanxia
Cham, Tat-Jen
Keywords: Engineering::Computer science and engineering::Computing methodologies::Computer graphics
Issue Date: 2021
Source: Song, G., Luo, L., Liu, J., Ma, W., Lai, C., Zheng, C. & Cham, T. (2021). AgileGAN: stylizing portraits by inversion-consistent transfer learning. ACM Transactions On Graphics, 40(4), 117-. https://dx.doi.org/10.1145/3450626.3459771
Journal: ACM Transactions on Graphics
Abstract: Portraiture as an art form has evolved from realistic depiction into a plethora of creative styles. While substantial progress has been made in automated stylization, generating high quality stylistic portraits is still a challenge, and even the recent popular Toonify suffers from several artifacts when used on real input images. Such StyleGAN-based methods have focused on finding the best latent inversion mapping for reconstructing input images; however, our key insight is that this does not lead to good generalization to different portrait styles. Hence we propose AgileGAN, a framework that can generate high quality stylistic portraits via inversion-consistent transfer learning. We introduce a novel hierarchical variational autoencoder to ensure the inverse mapped distribution conforms to the original latent Gaussian distribution, while augmenting the original space to a multi-resolution latent space so as to better encode different levels of detail. To better capture attribute-dependent stylization of facial features, we also present an attribute-aware generator and adopt an early stopping strategy to avoid overfitting small training datasets. Our approach provides greater agility in creating high quality and high resolution (1024×1024) portrait stylization models, requiring only a limited number of style exemplars (∼100) and short training time (∼1 hour). We collected several style datasets for evaluation including 3D cartoons, comics, oil paintings and celebrities. We show that we can achieve superior portrait stylization quality to previous state-of-the-art methods, with comparisons done qualitatively, quantitatively and through a perceptual user study. We also demonstrate two applications of our method, image editing and motion retargeting.
URI: https://hdl.handle.net/10356/172645
ISSN: 0730-0301
DOI: 10.1145/3450626.3459771
Schools: School of Computer Science and Engineering 
Rights: © 2021 Copyright held by the owner/author(s). All rights reserved.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:SCSE Journal Articles

SCOPUSTM   
Citations 5

70
Updated on May 5, 2025

Page view(s)

117
Updated on May 5, 2025

Google ScholarTM

Check

Altmetric


Plumx

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.