Back to Search
Start Over
Embedding an Ethical Mind: Aligning Text-to-Image Synthesis via Lightweight Value Optimization
- Publication Year :
- 2024
-
Abstract
- Recent advancements in diffusion models trained on large-scale data have enabled the generation of indistinguishable human-level images, yet they often produce harmful content misaligned with human values, e.g., social bias, and offensive content. Despite extensive research on Large Language Models (LLMs), the challenge of Text-to-Image (T2I) model alignment remains largely unexplored. Addressing this problem, we propose LiVO (Lightweight Value Optimization), a novel lightweight method for aligning T2I models with human values. LiVO only optimizes a plug-and-play value encoder to integrate a specified value principle with the input prompt, allowing the control of generated images over both semantics and values. Specifically, we design a diffusion model-tailored preference optimization loss, which theoretically approximates the Bradley-Terry model used in LLM alignment but provides a more flexible trade-off between image quality and value conformity. To optimize the value encoder, we also develop a framework to automatically construct a text-image preference dataset of 86k (prompt, aligned image, violating image, value principle) samples. Without updating most model parameters and through adaptive value selection from the input prompt, LiVO significantly reduces harmful outputs and achieves faster convergence, surpassing several strong baselines and taking an initial step towards ethically aligned T2I models.<br />Comment: Accepted by ACM Multimedia 2024. The dataset and code can be found at https://github.com/achernarwang/LiVO
Details
- Database :
- arXiv
- Publication Type :
- Report
- Accession number :
- edsarx.2410.12700
- Document Type :
- Working Paper
- Full Text :
- https://doi.org/10.1145/3664647.3681652