Back to Search Start Over

Unified Generation, Reconstruction, and Representation: Generalized Diffusion with Adaptive Latent Encoding-Decoding

Authors :
Liu, Guangyi
Wang, Yu
Feng, Zeyu
Wu, Qiyu
Tang, Liping
Gao, Yuan
Li, Zhen
Cui, Shuguang
McAuley, Julian
Yang, Zichao
Xing, Eric P.
Hu, Zhiting
Publication Year :
2024

Abstract

The vast applications of deep generative models are anchored in three core capabilities -- generating new instances, reconstructing inputs, and learning compact representations -- across various data types, such as discrete text/protein sequences and continuous images. Existing model families, like variational autoencoders (VAEs), generative adversarial networks (GANs), autoregressive models, and (latent) diffusion models, generally excel in specific capabilities and data types but fall short in others. We introduce Generalized Encoding-Decoding Diffusion Probabilistic Models (EDDPMs) which integrate the core capabilities for broad applicability and enhanced performance. EDDPMs generalize the Gaussian noising-denoising in standard diffusion by introducing parameterized encoding-decoding. Crucially, EDDPMs are compatible with the well-established diffusion model objective and training recipes, allowing effective learning of the encoder-decoder parameters jointly with diffusion. By choosing appropriate encoder/decoder (e.g., large language models), EDDPMs naturally apply to different data types. Extensive experiments on text, proteins, and images demonstrate the flexibility to handle diverse data and tasks and the strong improvement over various existing models.<br />Comment: ICML 2024 camera-ready. Code is available at https://github.com/guangyliu/EDDPM

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2402.19009
Document Type :
Working Paper