Recently, text-to-image generative models have been misused to create unauthorized malicious images of individuals, posing a growing social problem. Previous solutions, such as Anti-DreamBooth, add adversarial noise to images to protect them from being used as training data for malicious generation. However, we found that the adversarial noise can be removed by adversarial purification methods such as DiffPure. Therefore, we propose a new adversarial attack method that adds strong perturbation on the high-frequency areas of images to make it more robust to adversarial purification. Our experiment showed that the adversarial images retained noise even after adversarial purification, hindering malicious image generation.
- Create virtual environment with conda
conda env create -n {env_name} --file env/env-{env_name}.txt
necessary
env/env-anti-dreambooth.yaml
optional (for evaluation)
env/env-bilateral.yaml
env/env-brisque.yaml
env/env-diffpure.yaml
env/env-fdfr.yaml
env/env-ism.yaml
env/env-l_p.yaml
env/env-ser-fiq.yaml
- Install PyTorch manually
conda activate anti-dreambooth
conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.6 -c pytorch -c nvidia
- Download Stable Diffusion model from following url, and put it in
model/
↓ One of the easiest way of downloading Stable Diffusion using python.
import huggingface_hub
model_id = "stabilityai/stable-diffusion-2-1-base"
local_dir = "models/stable-diffusion/stable-diffusion-2-1-base"
huggingface_hub.snapshot_download(model_id, local_dir=local_dir, local_dir_use_symlinks=False)
We used NVIDIA A100 SXM.
- Set experiment configurations on
job_scripts/hfadb.sh
- ALPHA: The amount of noise added in each step
- NOISE_LEVEL_IN_MASK: The maximum amount of noise added on high-frequency area
- NOISE_LEVEL_OUT_MASK: The maximum amount of noise added on low-frequency area
- MASK_AREA: The ratio of high-frequency area
- STEP: The number of steps to add noise
- EXPERIMENT_NAME: Arbitrary experiment name
- CLASS_DIR: Direntory where you put class image for Stable Diffusion
- INSTANCE_DIR: Directory of sample images
- HOME_DIR: Directory of outputs
- PROMPT: Prompt for generation
- PROMPT_DIR: Replace
' '
ofPROMPT
with'_'
- SRC_DIR: Root directory of the source code
- Run this command on terminal
bash ./job_scripts/hfadb.sh
- arxiv
- bibtex
@inproceedings{onikubo2024high,
title={High-Frequency Anti-DreamBooth: Robust Defense against Personalized Image Synthesis},
author={Onikubo, Takuto and Matsui, Yusuke},
booktitle={ECCV 2024 Workshop The Dark Side of Generative AIs and Beyond}
}
The code of our method is based on that of Anti-DreamBooth. For evaluation, we used insightface repo and retinaface repo. We would like to express our sincere gratitude to the contributors of the Anti-DreamBooth, insightface, and retinaface repositories.