Fine-Grained Erasure in Text-To-Image Diffusion-Based Foundation Models

Kartik Thakral, Tamar Glaser, Tal Hassner, Mayank Vatsa, Richa Singh

Research output: Contribution to journalConference articlepeer-review

Abstract

Existing unlearning algorithms in text-to-image generative models often fail to preserve the knowledge of semantically related concepts when removing specific target concepts - a challenge known as adjacency. To address this, we propose FADE (Fine-Grained Attenuation for Diffusion Erasure), introducing adjacency-aware unlearning in diffusion models. FADE comprises two components: (1) the Concept Neighborhood, which identifies an adjacency set of related concepts, and (2) Mesh Modules, employing a structured combination of Expungement, Adjacency, and Guidance loss components. These enable precise erasure of target concepts while preserving fidelity across related and unrelated concepts. Evaluated on datasets like Stanford Dogs, Oxford Flowers, CUB, I2P, Imagenette, and ImageNet-1k, FADE effectively removes target concepts with minimal impact on correlated concepts, achieving at least a 12% improvement in retention performance over state-of-the-art methods. Our code and models are available on the project page: iab-rubric/unlearning/FG-Un.

Original languageEnglish
Pages (from-to)9121-9130
Number of pages10
JournalProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
DOIs
StatePublished - 2025
Externally publishedYes
Event2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2025 - Nashville, United States
Duration: 11 Jun 202515 Jun 2025

Bibliographical note

Publisher Copyright:
© 2025 IEEE.

Keywords

  • diffusion
  • fine-grained
  • text-to-image
  • unlearning

Fingerprint

Dive into the research topics of 'Fine-Grained Erasure in Text-To-Image Diffusion-Based Foundation Models'. Together they form a unique fingerprint.

Cite this