Chasing Illumination
智者说,太阳下山就到有太阳的地方去

前言

本文为个人学习笔记整理,行文较为随性,未尽完善之处,敬请谅解。

Abstract

Task

Copyright issues related to Diffusion Models (DMs)

Technical challenge for previous methods

There don’t seem to be any established methods for adding visible watermarks on the generated images. Instead, most current approaches focus on adding perturbations, such as adversarial examples, to the original images to prevent diffusion models from imitating unauthorized images.

Key Insight/Motivation

  1. They adds visible watermarks on the generated images
    • which is a more straightforward way to indicate copyright violations? —— I agree

Technical contributions

  1. Using a generator instead of the original diffusion model
    • Benefit: High computation and memory requirements, lower to 0.2s per image after the generator is trained (Diff-protect needs 8G30S per image on ? GPU)
      • NVIDIA A100 80GB GPU, Training the generator using 100 samples with 200 epochs takes about 20 minutes? —— maybe, more like a better trade-off
  2. Good transferability across unknown generative models
    • Benefit: A more effective approach to copyright protection
      • Is it difficult for attackers to remove the perturbations? —— Didn’t OpenSource

Introdution

Task and application

potential copyright violations when unauthorized images are used to train DMs or to generate new creations by DMs.

For example, an entity could use copyrighted paintings shared on the Internet to generate images with a similar style and appearance, potentially earning illegal revenue.

Technical challenge for previous methods

  1. Technical challenge 1
    1. Previous method
      1. Watermark ------ From: https://www.mihuashi.com/artworks/5827524
      2. Failure cases (Limitation)
        1. visiable & DMs may ignore the waterrmark
      3. Technical reason
        1. Costly & Need special design for each paintings
  2. Technical challenge 2
    • Previous method
    • Failure cases (Limitation)
      1. they are primarily designed to protect the copyright of DMs or to distinguish generated images from natural ones, which differs from our goal of protecting the copyright of human creations.
      2. Succeed in their task (Totally white-box attack)
    • Technical reason
      • Costly and Not practical
  3. Technical challenge 3
    • Previous method
    • Failure cases (Limitation)
      1. Difficult to comprehend (add chaotic textures to the generated images)
      2. The original image’s copyright is not traceable
    • Technical reason
      • Easy to remove (Not included in the original paper)
      • Time-consuming (Each AdvE need to be optimized separately & iterative optimization)

Introdution of the challenges and pipelines

  1. key innovation/insight/contribution
    1. Force DMs to generate images with visible watermarks as well as chaotic textures.
  2. contribution 1
    1. Creat a novel framework that embeds personal watermarks into the generation of adversarial examples to prevent copyright violations caused by DM-based imitation.
    2. advantage/insight
      1. Force DMs to generate images with visible watermarks for tracing copyright.
  3. contribution 2
    1. Design three losses: adversarial loss, GAN loss, and weighted perturbation loss
    2. advantage/insight: ---
  4. contribution 3
    1. Experiments shows that the method is robustness and have the transferability to other models
    2. advantage/insight: ---

Demos/applications


Method

overview

The original figure wasn't very clear, especially regarding the inputs for each part, so I added some annotation to make it clearer. I hope it looks better now.😂

Task Overview

  1. Input(What we have): Original x ( 100+ samples from artist ), diffusion model
  2. Output: Adversarial x'

Modules Overview

Pipeline module 1

  1. Motivation
  2. How to do
  3. Why it works?
  4. technical advantage

Pipeline module 2

  1. Motivation
  2. How to do
  3. Why it works?
  4. technical advantage

Pipeline module 3

  1. Motivation
    1. For transferability
    2. LDMs are more common in practice (efficiency & high-quality)
  2. How to do
  3. Why it works?
    Use a loss to constrain the x' to resemble the watermark m, simple but work
  4. technical advantage
    1. No need to attack the U-Net to produce satisfactory results, faster

Experiments

Comparision experiment

Ablation studies

  1. Core contribution to performance
  2. Main components to performance

Reference

Papers are arranged according to their similarity (this is not a citation tree); Node size is the number of citations; Node color is the publishing year; Similar papers have strong connecting lines and cluster together;
TitleAuthorsYearCitationsReferencesSimilarityto origin
GLAZE: Protecting Artists from Style Mimicry by Text-to-Image ModelsShawn Shan, Jenna Cryan, Emily Wenger, Haitao Zheng, Rana Hanocka, Ben Y. Zhao202311613118.3
The Stable Signature: Rooting Watermarks in Latent Diffusion ModelsPierre Fernandez, Guillaume Couairon, Herv'e J'egou, Matthijs Douze, T. Furon20238710617.5
A Recipe for Watermarking Diffusion ModelsYunqing Zhao, Tianyu Pang, Chao Du, Xiao Yang, Ngai-Man Cheung, Min Lin2023668116.6
Adversarial Example Does Good: Preventing Painting Imitation from Diffusion Models via Adversarial ExamplesChumeng Liang, Xiaoyu Wu, Yang Hua, Jiaru Zhang, Yiming Xue, Tao Song, Zhengui Xue, Ruhui Ma, Haibing Guan2023656624.6
Raising the Cost of Malicious AI-Powered Image EditingHadi Salman, Alaa Khaddaj, Guillaume Leclerc, Andrew Ilyas, A. Madry2023655619.1
End-to-End Diffusion Latent Optimization Improves Classifier GuidanceBram Wallace, Akash Gokul, Stefano Ermon, N. Naik2023535516.8
Anti-DreamBooth: Protecting users from personalized text-to-image synthesisT. Le, Hao Phung, Thuan Hoang Nguyen, Quan Dao, Ngoc N. Tran, A. Tran2023476718.8
SyncDiffusion: Coherent Montage via Synchronized Joint DiffusionsYuseung Lee, Kunho Kim, Hyunjin Kim, Minhyuk Sung2023385515.9
DiffusionShield: A Watermark for Copyright Protection against Generative Diffusion ModelsYingqian Cui, J. Ren, Han Xu, Pengfei He, Hui Liu, Lichao Sun, Jiliang Tang2023354719.2
Mist: Towards Improved Adversarial Examples for Diffusion ModelsChumeng Liang, Xiaoyu Wu2023241319.7
FreeControl: Training-Free Spatial Control of Any Text-to-Image Diffusion Model with Any ConditionSicheng Mo, Fangzhou Mu, Kuan Heng Lin, Yanli Liu, Bochen Guan, Yin Li, Bolei Zhou2023205916.8
Unlearnable Examples for Diffusion Models: Protect Data from Unauthorized ExploitationZhengyue Zhao, Jinhao Duan, Xingui Hu, Kaidi Xu, Chenan Wang, Rui Zhang, Zidong Du, Qi Guo, Yunji Chen2023184517.5
The Blessing of Randomness: SDE Beats ODE in General Diffusion-based Image EditingShen Nie, Hanzhong Guo, Cheng Lu, Yuhao Zhou, Chenyu Zheng, Chongxuan Li2023188016.6
Generative Watermarking Against Unauthorized Subject-Driven Image SynthesisY. Ma, Zhengyu Zhao, Xinlei He, Zheng Li, M. Backes, Yang Zhang2023176919.5
Copyright Protection in Generative AI: A Technical PerspectiveJie Ren, Han Xu, Pengfei He, Yingqian Cui, Shenglai Zeng, Jiankun Zhang, Hongzhi Wen, Jiayuan Ding, Hui Liu, Yi Chang, Jiliang Tang20241517817.7
Toward effective protection against diffusion based mimicry through score distillationHaotian Xue, Chumeng Liang, Xiaoyu Wu, Yongxin Chen2023144018.3
Receler: Reliable Concept Erasing of Text-to-Image Diffusion Models via Lightweight ErasersChi-Pin Huang, Kai-Po Chang, Chung-Ting Tsai, Yung-Hsuan Lai, Yu-Chiang Frank Wang2023135718.6
Improving Adversarial Attacks on Latent Diffusion ModelBoyang Zheng, Chumeng Liang, Xiaoyu Wu, Yan Liu2023134016.7
FT-Shield: A Watermark Against Unauthorized Fine-tuning in Text-to-Image Diffusion ModelsYingqian Cui, Jie Ren, Yuping Lin, Han Xu, Pengfei He, Yue Xing, Wenqi Fan, Hui Liu, Jiliang Tang202393519.6
Enhanced Controllability of Diffusion Models via Feature Disentanglement and Realism-Enhanced Sampling MethodsWonwoong Cho, Hareesh Ravi, Midhun Harikumar, V. Khuc, Krishna Kumar Singh, Jingwan Lu, David I. Inouye, Ajinkya Kale202365216.8
Toward Robust Imperceptible Perturbation against Unauthorized Text-to-image Diffusion-based SynthesisYixin Liu, Chenrui Fan, Yutong Dai, Xun Chen, Pan Zhou, Lichao Sun202365316.5
Towards Prompt-robust Face Privacy Protection via Adversarial Decoupling Augmentation FrameworkRuijia Wu, Yuhang Wang, Huafeng Shi, Zhipeng Yu, Yichao Wu, Ding Liang202364716.4
IMMA: Immunizing text-to-image Models against Malicious AdaptationYijia Zheng, Raymond A. Yeh202355218.7
Watermark-embedded Adversarial Examples for Copyright Protection against Diffusion ModelsPeifei Zhu, Tsubasa Takahashi, Hirokatsu Kataoka2024466100
Catch You Everything Everywhere: Guarding Textual Inversion via Concept WatermarkingWeitao Feng, Jiyan He, Jie Zhang, Tianwei Zhang, Wenbo Zhou, Weiming Zhang, Neng H. Yu202345422
Perturbing Attention Gives You More Bang for the Buck: Subtle Imaging Perturbations That Efficiently Fool Customized Diffusion ModelsJingyao Xu, Yuetong Lu, Yandong Li, Siyang Lu, Dongdong Wang, Xiang Wei202433621.4
Imperceptible Protection against Style Imitation from Diffusion ModelsNamhyuk Ahn, Wonhyuk Ahn, Kiyoon Yoo, Daesik Kim, Seung-Hun Nam202437020.4
R.A.C.E.: Robust Adversarial Concept Erasure for Secure Text-to-Image Diffusion ModelC. Kim, Kyle Min, Yezhou Yang202437418.2
Pick-and-Draw: Training-free Semantic Guidance for Text-to-Image PersonalizationHenglei Lv, Jiayu Xiao, Liang Li, Qingming Huang202433316
Pruning for Robust Concept Erasing in Diffusion ModelsTianyun Yang, Juan Cao, Chang Xu202424022.9
A Somewhat Robust Image Watermark against Diffusion-based Editing ModelsMingtian Tan, Tianhao Wang, Somesh Jha202326622.5
Pixel is a Barrier: Diffusion Models Are More Adversarially Robust Than We ThinkHaotian Xue, Yongxin Chen202424921.3
Towards Test-Time Refusals via Concept NegationPeiran Dong, Song Guo, Junxiao Wang, Bingjie Wang, Jiewei Zhang, Ziming Liu202323518.8
Towards Memorization-Free Diffusion ModelsChen Chen, Daochang Liu, Chang Xu202423917.1
VA3: Virtually Assured Amplification Attack on Probabilistic Copyright Protection for Text-to-Image Generative ModelsXiang Li, Qianli Shen, Kenji Kawaguchi202325916.6
PiGW: A Plug-in Generative Watermarking FrameworkRui Ma, Mengxi Guo, Yuming Li, Hengyuan Zhang, Cong Ma, Yuan Li, Xiaodong Xie, Shanghang Zhang202415318.9
Privacy-Preserving Low-Rank Adaptation for Latent Diffusion ModelsZihao Luo, Xilie Xu, Feng Liu, Yun Sing Koh, Di Wang, Jingfeng Zhang202414318
SimAC: A Simple Anti-Customization Method for Protecting Face Privacy against Text-to-Image Synthesis of Diffusion ModelsFeifei Wang, Zhentao Tan, Tianyi Wei, Yue Wu, Qidong Huang202313917.1
Prompt-Agnostic Adversarial Perturbation for Customized Diffusion ModelsCong Wan, Yuhang He, Xiang Song, Yihong Gong202406320
PaRa: Personalizing Text-to-Image Diffusion via Parameter Rank ReductionShangyu Chen, Zizheng Pan, Jianfei Cai, Dinh Q. Phung202404216.6
PID: Prompt-Independent Data Protection Against Latent Diffusion ModelsAng Li, Yichuan Mo, Mingjie Li, Yisen Wang202405116.6

Appendix