Back to Search
Start Over
Taming diffusion model for exemplar-based image translation.
- Source :
- Computational Visual Media; Dec2024, Vol. 10 Issue 6, p1031-1043, 13p
- Publication Year :
- 2024
-
Abstract
- Exemplar-based image translation involves converting semantic masks into photorealistic images that adopt the style of a given exemplar. However, most existing GAN-based translation methods fail to produce photorealistic results. In this study, we propose a new diffusion model-based approach for generating high-quality images that are semantically aligned with the input mask and resemble an exemplar in style. The proposed method trains a conditional denoising diffusion probabilistic model (DDPM) with a SPADE module to integrate the semantic map. We then used a novel contextual loss and auxiliary color loss to guide the optimization process, resulting in images that were visually pleasing and semantically accurate. Experiments demonstrate that our method outperforms state-of-the-art approaches in terms of both visual quality and quantitative metrics. [ABSTRACT FROM AUTHOR]
- Subjects :
- TRANSLATING & interpreting
Subjects
Details
- Language :
- English
- ISSN :
- 20960433
- Volume :
- 10
- Issue :
- 6
- Database :
- Complementary Index
- Journal :
- Computational Visual Media
- Publication Type :
- Academic Journal
- Accession number :
- 180589649
- Full Text :
- https://doi.org/10.1007/s41095-023-0371-3