Back to Search Start Over

Exploring Adversarial Attacks against Latent Diffusion Model from the Perspective of Adversarial Transferability

Authors :
Chen, Junxi
Dong, Junhao
Xie, Xiaohua
Chen, Junxi
Dong, Junhao
Xie, Xiaohua
Publication Year :
2024

Abstract

Recently, many studies utilized adversarial examples (AEs) to raise the cost of malicious image editing and copyright violation powered by latent diffusion models (LDMs). Despite their successes, a few have studied the surrogate model they used to generate AEs. In this paper, from the perspective of adversarial transferability, we investigate how the surrogate model's property influences the performance of AEs for LDMs. Specifically, we view the time-step sampling in the Monte-Carlo-based (MC-based) adversarial attack as selecting surrogate models. We find that the smoothness of surrogate models at different time steps differs, and we substantially improve the performance of the MC-based AEs by selecting smoother surrogate models. In the light of the theoretical framework on adversarial transferability in image classification, we also conduct a theoretical analysis to explain why smooth surrogate models can also boost AEs for LDMs.<br />Comment: 24 pages, 13 figures

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1438515619
Document Type :
Electronic Resource