TY - GEN
T1 - Latent Diffusion Unlearning
T2 - 33rd ACM International Conference on Multimedia, MM 2025
AU - Devulapally, Naresh Kumar
AU - Agarwal, Shruti
AU - Gokhale, Tejas
AU - Lokhande, Vishnu Suresh
N1 - Publisher Copyright: © 2025 ACM.
PY - 2025/10/27
Y1 - 2025/10/27
N2 - Text-to-image diffusion models have demonstrated remarkable effectiveness in rapid and high-fidelity personalization, even when provided with only a few user images. However, the effectiveness of personalization techniques has lead to concerns regarding data privacy, intellectual property protection, and unauthorized usage. To mitigate such unauthorized usage and model replication, the idea of generating ''unlearnable'' training samples utilizing image poisoning techniques has emerged. Existing methods for this have limited imperceptibility as they operate in the pixel space which results in images with noise and artifacts. In this work, we propose a novel model-based perturbation strategy that operates within the latent space of diffusion models. Our method alternates between denoising and inversion while modifying the starting point of the denoising trajectory: of diffusion models. This trajectory-shifted sampling ensures that the perturbed images maintain high visual fidelity to the original inputs while being resistant to inversion and personalization by downstream generative models. This approach integrates unlearnability into the framework of Latent Diffusion Models (LDMs), enabling a practical and imperceptible defense against unauthorized model adaptation. We validate our approach on four benchmark datasets to demonstrate robustness against state-of-the-art inversion attacks. Results demonstrate that our method achieves significant improvements in imperceptibility (∼8% - 10% on perceptual metrics including PSNR, SSIM, and FID) and robustness (∼10% on average across five adversarial settings), highlighting its effectiveness in safeguarding sensitive data. https://github.com/naresh-ub/unlearnable-samples.
AB - Text-to-image diffusion models have demonstrated remarkable effectiveness in rapid and high-fidelity personalization, even when provided with only a few user images. However, the effectiveness of personalization techniques has lead to concerns regarding data privacy, intellectual property protection, and unauthorized usage. To mitigate such unauthorized usage and model replication, the idea of generating ''unlearnable'' training samples utilizing image poisoning techniques has emerged. Existing methods for this have limited imperceptibility as they operate in the pixel space which results in images with noise and artifacts. In this work, we propose a novel model-based perturbation strategy that operates within the latent space of diffusion models. Our method alternates between denoising and inversion while modifying the starting point of the denoising trajectory: of diffusion models. This trajectory-shifted sampling ensures that the perturbed images maintain high visual fidelity to the original inputs while being resistant to inversion and personalization by downstream generative models. This approach integrates unlearnability into the framework of Latent Diffusion Models (LDMs), enabling a practical and imperceptible defense against unauthorized model adaptation. We validate our approach on four benchmark datasets to demonstrate robustness against state-of-the-art inversion attacks. Results demonstrate that our method achieves significant improvements in imperceptibility (∼8% - 10% on perceptual metrics including PSNR, SSIM, and FID) and robustness (∼10% on average across five adversarial settings), highlighting its effectiveness in safeguarding sensitive data. https://github.com/naresh-ub/unlearnable-samples.
KW - diffusion models
KW - identity protection
KW - personalization
KW - unlearning
UR - https://www.scopus.com/pages/publications/105024076265
U2 - 10.1145/3746027.3755112
DO - 10.1145/3746027.3755112
M3 - Conference contribution
T3 - MM 2025 - Proceedings of the 33rd ACM International Conference on Multimedia, Co-Located with MM 2025
SP - 11376
EP - 11384
BT - MM 2025 - Proceedings of the 33rd ACM International Conference on Multimedia, Co-Located with MM 2025
PB - Association for Computing Machinery, Inc
Y2 - 27 October 2025 through 31 October 2025
ER -