Back to Search Start Over

A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive Learning.

Authors :
YUXIN ZHANG
FAN TANG
WEIMING DONG
HAIBIN HUANG
CHONGYANG MA
TONG-YEE LEE
CHANGSHENG XU
Source :
ACM Transactions on Graphics; Oct2023, Vol. 42 Issue 5, p1-16, 16p
Publication Year :
2023

Abstract

This work presents Unified Contrastive Arbitrary Style Transfer (UCAST), a novel style representation learning and transfer framework, that can fit in most existing arbitrary image style transfer models, such as CNN-based, ViT-based, and flow-based methods. As the key component in image style transfer tasks, a suitable style representation is essential to achieve satisfactory results. Existing approaches based on deep neural networks typically use second-order statistics to generate the output. However, these hand-crafted features computed from a single image cannot leverage style information sufficiently, which leads to artifacts such as local distortions and style inconsistency. To address these issues, we learn style representation directly from a large number of images based on contrastive learning by considering the relationships between specific styles and the holistic style distribution. Specifically, we present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature. Our framework consists of three key components: a parallel contrastive learning scheme for style representation and transfer, a domain enhancement (DE) module for effective learning of style distribution, and a generative network for style transfer. Qualitative and quantitative evaluations show the results of our approach are superior to those obtained via state-of-the-art methods. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
07300301
Volume :
42
Issue :
5
Database :
Complementary Index
Journal :
ACM Transactions on Graphics
Publication Type :
Academic Journal
Accession number :
172854412
Full Text :
https://doi.org/10.1145/3605548