STALP: Style Transfer with Auxiliary Limited Pairing
Date
2021Metadata
Show full item recordAbstract
We present an approach to example-based stylization of images that uses a single pair of a source image and its stylized counterpart. We demonstrate how to train an image translation network that can perform real-time semantically meaningful style transfer to a set of target images with similar content as the source image. A key added value of our approach is that it considers also consistency of target images during training. Although those have no stylized counterparts, we constrain the translation to keep the statistics of neural responses compatible with those extracted from the stylized source. In contrast to concurrent techniques that use a similar input, our approach better preserves important visual characteristics of the source style and can deliver temporally stable results without the need to explicitly handle temporal consistency. We demonstrate its practical utility on various applications including video stylization, style transfer to panoramas, faces, and 3D models.
BibTeX
@article {10.1111:cgf.142655,
journal = {Computer Graphics Forum},
title = {{STALP: Style Transfer with Auxiliary Limited Pairing}},
author = {Futschik, David and Kucera, Michal and Lukác, Mike and Wang, Zhaowen and Shechtman, Eli and Sýkora, Daniel},
year = {2021},
publisher = {The Eurographics Association and John Wiley & Sons Ltd.},
ISSN = {1467-8659},
DOI = {10.1111/cgf.142655}
}
journal = {Computer Graphics Forum},
title = {{STALP: Style Transfer with Auxiliary Limited Pairing}},
author = {Futschik, David and Kucera, Michal and Lukác, Mike and Wang, Zhaowen and Shechtman, Eli and Sýkora, Daniel},
year = {2021},
publisher = {The Eurographics Association and John Wiley & Sons Ltd.},
ISSN = {1467-8659},
DOI = {10.1111/cgf.142655}
}