PaperHub
7.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
5
4
5
4.0
置信度
创新性3.0
质量3.3
清晰度2.8
重要性3.5
NeurIPS 2025

BlurGuard: A Simple Approach for Robustifying Image Protection Against AI-Powered Editing

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29
TL;DR

A surprisingly simple plug-and-play method to strengthen adversarial image protection against diverse purification techniques.

摘要

关键词
copyright protectionadversarial examplestext-to-image modelscomputer vision

评审与讨论

审稿意见
5

This paper introduces a simple but effective method to enhance the robustness of image protection methods against noise reversal techniques, employing an adaptive Gaussian blur on the noise to adjust the overall frequency spectrum. Unlike previous efforts primarily focused on imperceptibility, this approach emphasizes the irreversibility of adversarial noise, making it difficult to detect as noise.

优缺点分析

Strengths:

  1. The paper offers a novel perspective by focusing on "irreversibility" rather than solely "imperceptibility" for image protection. And the proposed approach is surprisingly simple but effective.
  2. The method's effectiveness is thoroughly demonstrated through comprehensive experiments across various AI-generated scenarios and diverse noise purification techniques.

Weakness:

  1. The ablation study indicates that "Per-region adaptation" might contribute less significantly compared to other modules. It would be beneficial to qualitatively show the different blur intensities across different regions for specific cases. Additionally, since the blur intensity depends on the semantic regions, a statistical correlation between specific semantic categories and their corresponding blur intensities may reveal interesting insights into how this module functions.
  2. Most inpainting methods mask edited regions during processing. A concern arises if the masked area is a region with strong adversarial noise; it's unclear if the proposed method would remain effective. The inpainting experiments (line 278) in this paper are limited to facial images and focus only on modifying areas around the face. However, real-world inpainting scenarios are far more diverse, with unpredictable manipulation regions. This limited experiment might not fully cover all types of inpainting tasks.
  3. More comprehensive details are needed for the ImageNet-Edit dataset. This includes clearer information about the prompts used and a statistic of image categories.

问题

Please refer to the weaknesses mentioned above.

  1. The ablation study indicates that "Per-region adaptation" might contribute less significantly compared to other modules. It would be beneficial to qualitatively show the different blur intensities across different regions for specific cases. Additionally, since the blur intensity depends on the semantic regions, a statistical correlation between specific semantic categories and their corresponding blur intensities may reveal interesting insights into how this module functions.
  2. Most inpainting methods mask edited regions during processing. A concern arises if the masked area is a region with strong adversarial noise; it's unclear if the proposed method would remain effective. The inpainting experiments (line 278) in this paper are limited to facial images and focus only on modifying areas around the face. However, real-world inpainting scenarios are far more diverse, with unpredictable manipulation regions. This limited experiment might not fully cover all types of inpainting tasks.
  3. More comprehensive details are needed for the ImageNet-Edit dataset. This includes clearer information about the prompts used and a statistic of image categories.

局限性

Yes

最终评判理由

Based on the author's detailed rebuttal, which effectively addressed all of my concerns, and after considering the comments and discussions from other reviewers, I have decided to maintain my original score. I recommend the paper for acceptance.

格式问题

No paper formatting concerns.

作者回复

We sincerely appreciate your thoughtful and constructive feedback. We respond to each of your questions one-by-one in what follows. Should you have any additional comments or concerns, we would be happy to discuss or clarify them.


Q1. Further analysis on per-region adaptation

Thank you for your thoughtful suggestion. We agree that a further analysis of the per-region adaptation module would enhance clarity. For your information, our extended ablation study in Appendix G provides a qualitative comparison between BlurGuard and the vanilla PhotoGuard, showing how BlurGuard adaptively allocates noise frequencies and produces perceptually natural perturbations. In the final draft, we will incorporate additional qualitative examples (e.g., more BlurGuard perturbations across different region types) and quantitative results (e.g., correlation between the learned σ\sigma values and semantic regions, as you suggested) to further improve clarity of the module.


Q2. Inpainting under mask variability

As you mentioned, our inpainting experiments primarily focus on “region-targeted” protection, where adversarial perturbations are selectively applied to the region intended for editing. We emphasize that this setup is practically motivated, as it reflects realistic misuse scenarios where a sensitive region, e.g., a person’s face, is kept intact while the surrounding areas are manipulated.

That said, we agree that real-world inpainting scenarios may involve unpredictable and diverse masking patterns. Indeed, this challenge has recently attracted attention in the literature; for example, DiffusionGuard [1], which we also include as a baseline, proposes the InpaintGuardBench dataset to benchmark robustness of image protection methods under mask variability.

To further address your concern, we conducted additional experiments on InpaintGuardBench, evaluating BlurGuard across six different masks per image, including both “seen” (used for protection) and “unseen” (novel at test time) masks. As shown in the table below, BlurGuard maintains strong protection performance across these mask variations, suggesting that BlurGuard not only provides robustness on the region-targeted setup but also generalizes well to more diverse inpainting scenarios. We will incorporate this discussion and the additional results into the final draft.

InpaintGuardBenchNaturalnessWorst-caseeffect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑FID↑ (Seen)FID↑ (Unseen)│ PSNR↓ (Seen)PSNR↓ (Unseen)
PhotoGuard0.03±0.02\underline{\text{0.03±0.02}}0.96±0.02103.8367.05│ 37.2±14.235.5±8.2
AdvDM0.03±0.02\underline{\text{0.03±0.02}}0.97±0.02125.6671.0635.3±10.3\underline{\text{35.3±10.3}}35.4±7.0\underline{\text{35.4±7.0}}
Mist0.02±0.010.98±0.01\underline{\text{0.98±0.01}}112.2358.48│37.1±14.236.4±9.8
SDS0.05±0.030.96±0.02113.3083.91\underline{\text{83.91}}│37.1±14.235.6±7.2
DiffusionGuard0.03±0.02\underline{\text{0.03±0.02}}0.97±0.02134.25\underline{\text{134.25}}61.32│37.1±14.235.9±9.9
BlurGuard0.03±0.02\underline{\text{0.03±0.02}}0.99±0.01140.8290.3234.9±17.035.3±8.9

Q3. ImageNet-Edit: Additional details needed

Thank you for the suggestion. We agree that providing additional clarification on the ImageNet-Edit dataset we constructed would enhance the reproducibility of our work. For your information, we note that we have included the full dataset as part of our supplementary material submission, along with the complete set of prompts used. We also plan to publicly release the dataset upon acceptance to further support reproducibility.


[1] Choi et al., DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing. ICLR 2025.


评论

Thank you for the detailed rebuttal. I have no further questions and will take these updates into account.

评论

Dear Reviewer Zo4E,

Thank you for your response and for considering our updates.

We are glad to hear that our rebuttal addressed your questions.

Should any additional input be helpful, please let us know.

Best regards,
Authors

审稿意见
5

The paper tackles the problem that existing adversarial image-protection schemes are easily neutralised by simple post-processing such as JPEG compression or diffusion-based “purification”. The authors observe that many prior perturbations concentrate energy in high-frequency bands and are therefore easy to detect and strip away in the frequency domain. BlurGuard addresses this with two ingredients: (1) Per-region adaptive Gaussian blurring: using SAM masks, different σ values are learned for each semantic region so the perturbation’s spectrum better follows the underlying content. (2) Power-spectrum regularisation: an ℓ∞ penalty on the log-ratio between the RAPSD of the original and protected images keeps the overall spectrum “natural”.

优缺点分析

Strengths

  1. Re-framing protection as spectral naturalness instead of pure imperceptibility is simple yet insightful.
  2. The paper is well written with clear figures.

Weaknesses

  1. The testing primarily relies on outdated UNet-based diffusion models. Experiments on modern models with MMDIT architecture and flow-matching training paradigms, such as SD3.5 [1], FLUX.Fill [2], or Step1X-Edit [3], would strengthen the evaluation.

  2. The authors should consider including a section on additional related work, as some studies, such as VINE [4] and Robust-Wide [5], focus on enhancing image watermark robustness to protect images against editing. These works are relevant to the authors’ research, particularly those analyzing image editing from a frequency perspective.

  3. The transferability analysis lacks persuasiveness, as SD 1.4 shares the same architecture and some training data with SD 2.1. Demonstrating that image protections optimized for SD-v1.4 are effective against models like FLUX.Fill [2] or Step1X-Edit [3] would provide more robust evidence of transferability.

[1] SD3.5

[2] Flux.Fill

[3] Step1X-Edit

[4] Robust Watermarking Using Generative Priors Against Image Editing: From Benchmarking to Advances

[5] Robust-Wide: Robust Watermarking against Instruction-driven Image Editing

问题

Please review the Weaknesses section. If all of my concerns are fully resolved, I’ll gladly raise my rating; if the assessment is still lacking, I may lower it instead.

局限性

yes

最终评判理由

Thank you very much for your effort and thorough response.

I have carefully read all the review comments as well as the authors’ responses. I believe the authors have adequately addressed my concerns, and I have raised my score.

格式问题

NA

作者回复

We sincerely appreciate your thoughtful and constructive feedback. We respond to each of your questions one-by-one in what follows. Should you have any additional comments or concerns, we would be happy to discuss or clarify them.


Q1. BlurGuard with MMDiT-based models

Thank you for the suggestion. We agree that including evaluations on more recent models, e.g., the MMDiT-based architectures with flow-matching you mentioned, would further strengthen our work. Following your suggestion, we have extended our experiments to two additional models, viz., SD-v3.5 [1] and FLUX.1-dev [2], as shown in the following two tables. Overall, the results confirm again that BlurGuard offers the most effective protection against worst-case purification attempts, while also demonstrating the high adaptability of our framework. We will incorporate these additional results into the final draft, and continue to extend our evaluation to other models, e.g., Step1X-Edit [3] you also mentioned.

Table 1: Comparison of image protection on ImageNet-Edit, tested on SD-v3.5 [1].

SD-v3.5NaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑PSNR↑FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
PhotoGuard0.24±0.090.79±0.0931.5±1.30121.58\underline{\text{121.58}}0.45±0.080.48±0.1928.6±0.640.88±0.05
AdvDM0.28±0.100.77±0.0829.5±0.26116.360.42±0.080.48±0.1929.0±0.940.89±0.05\underline{\text{0.89±0.05}}
Mist0.25±0.09\underline{\text{0.25±0.09}}0.80±0.0930.9±1.14\underline{\text{30.9±1.14}}118.910.45±0.080.48±0.1928.6±0.640.88±0.05
SDS0.25±0.11\underline{\text{0.25±0.11}}0.86±0.08\underline{0.86±0.08}29.5±0.32119.200.46±0.08\underline{\text{0.46±0.08}}0.47±0.19\underline{\text{0.47±0.19}}28.5±0.56\underline{\text{28.5±0.56}}0.88±0.05
BlurGuard0.25±0.10\underline{\text{0.25±0.10}}0.88±0.1330.4±1.97125.320.48±0.070.46±0.1828.4±0.460.88±0.05

Table 2: Comparison of image protection on ImageNet-Edit, tested on FLUX.1-dev [2].

FLUX.1-devNaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑PSNR↑FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
PhotoGuard0.21±0.100.86±0.0732.1±0.3380.220.28±0.060.68±0.11\underline{\text{0.68±0.11}}29.5±0.310.91±0.05
AdvDM0.27±0.100.85±0.0732.6±1.10\underline{\text{32.6±1.10}}83.420.28±0.060.68±0.12\underline{\text{0.68±0.12}}28.8±0.720.90±0.05\underline{\text{0.90±0.05}}
Mist0.26±0.090.87±0.0832.3±1.4084.10\underline{\text{84.10}}0.29±0.07\underline{\text{0.29±0.07}}0.67±0.1229.7±0.600.90±0.04\underline{\text{0.90±0.04}}
SDS0.24±0.100.88±0.08\underline{\text{0.88±0.08}}32.2±0.8081.650.28±0.060.68±0.11\underline{\text{0.68±0.11}}28.6±0.48\underline{\text{28.6±0.48}}0.90±0.05\underline{\text{0.90±0.05}}
BlurGuard0.23±0.11\underline{\text{0.23±0.11}}0.89±0.0733.4±2.9085.210.30±0.100.67±0.1328.4±0.330.89±0.05

Q2. Additional related work

Thank you for suggesting additional related works. For your information, we note that Appendix A.1 already includes a brief discussion on watermarking-based approaches, including the VINE work [4] you referred to. Nevertheless, we agree that providing a dedicated discussion on frequency-based methods explored in other lines of work, e.g., Robust-Wide [5] you mentioned, would help better contextualize our contributions. We will revise our manuscript to include an additional discussion of these related works in the main text of the final draft.


Q3. Transferability analysis to different architectures, e.g., MMDiT?

Following your suggestion, we have additionally conducted black-box transfer experiments from SD-v1.4 to SD-v3.5 and FLUX.1-dev; both of which are based on MMDiT with flow-matching training, i.e., significantly diverging from SD-v1.4 in both architecture and training scheme. From the results summarized in the table below, we observe two key points: (a) BlurGuard consistently improves performance over the vanilla PhotoGuard across all metrics, showing the clear effectiveness of BlurGuard to enhance transferability; and (b) BlurGuard remains either the most effective or competitive across the board, further supporting general applicability of our approach. These observations confirm that our proposed framework offers a promising direction for improving the transferability of adversarial image protection methods. We will incorporate these additional results and the respective discussion in the final draft.

ImageNet-EditSD‑v1.4     \ \ \ \ \ \rightarrowSD‑v3.5SD‑v1.4     \ \ \ \ \ \rightarrowFLUX1‑dev
(ε=16/255\varepsilon=16/255)FID↑LPIPS↑SSIM↓PSNR↓I.A.↓FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
PhotoGuard112.580.40±0.090.48±0.1829.1±0.850.90±0.0472.60\underline{\text{72.60}}0.26±0.080.80±0.0730.7±0.950.95±0.03
AdvDM98.250.34±0.100.71±0.0929.9±1.040.92±0.0471.740.26±0.080.78±0.0630.6±1.060.94±0.04\underline{\text{0.94±0.04}}
Mist120.170.43±0.080.47±0.1928.8±0.830.89±0.05\underline{\text{0.89±0.05}}63.590.21±0.06\underline{\text{0.21±0.06}}0.80±0.0630.3±0.87\underline{\text{30.3±0.87}}0.95±0.03
SDS102.120.41±0.090.52±0.1429.4±0.880.89±0.05\underline{\text{0.89±0.05}}58.590.19±0.060.81±0.0731.0±1.370.96±0.03
BlurGuard117.71\underline{\text{117.71}}0.46±0.080.46±0.1928.4±0.540.88±0.0574.340.26±0.090.79±0.09\underline{\text{0.79±0.09}}29.0±0.690.93±0.04

[1] stabilityai/stable-diffusion-3.5-medium
[2] black-forest-labs/FLUX.1-dev
[3] Liu et al., Step1X-Edit: A Practical Framework for General Image Editing, 2025.
[4] Lu et al., Robust Watermarking Using Generative Priors Against Image Editing: From Benchmarking to Advances, ICLR 2025.
[5] Hu et al., Robust-Wide: Robust Watermarking against Instruction-driven Image Editing, ECCV 2024.


评论

Thank you very much for your effort and thorough response. I can easily imagine that conducting additional experiments must have been quite challenging, and I truly appreciate your dedication.

I have carefully read all the review comments as well as the authors’ responses. I believe the authors have adequately addressed my concerns, and I am inclined to raise my score. I would also like to observe the ongoing discussions between the authors and the other reviewers before making a final decision.

Thank you again.

评论

Dear Reviewer jjT1,

Thank you very much for your prompt response and encouraging words!

We truly value your thoughtful consideration, and we would be happy to address any further clarification you may need as the discussion continues.

Best regards,
Authors

审稿意见
4

This paper asserts that existing data poisons for diffusion models are weak to purification-based defenses because they fail to conform to the data prior expected by the diffusion models. They propose that this nonconformity can be identified as a deviation from real (unpoisoned) images in the frequency spectrum, as measured by RAPSD curves. They propose an adaptive per-region blurring operation applied to adversarial noise in order to encourage the adversarial images to align closer to real images. They present experiments, comparing BlurGuard to multiple current poisons and validating their method across multiple defenses. They measure the naturalness (i.e., imperceptibility) and effectiveness (i.e., worst-case performance decrease after purification) of adversarial images.

优缺点分析

  • Significance: Protection for personal data or copyright against copying or theft is an important topic. Moreover, existing protections for images are easily circumvented by purification-based (or other) defenses. Research into effective data poisons is necessary if artists and copyright holders are to continue publishing their work in a public setting.
  • Originality: BlurGuard has moderate originality. Most poisons in diffusion literature utilize an adversarial component and a perceptibility constraint (e.g., LPIPS in Glaze/Nightshade or a simple norm-ball on adversarial signal). Prior works on poisoning in other domains have explored blurring and sharpening filters as poisons (e.g., https://arxiv.org/abs/2303.04278, https://link.springer.com/chapter/10.1007/978-3-031-73464-9_5). In the diffusion domain, Gaussian blur has been used to model transitions between probability distributions (i.e., heat dissipation vs diffusion) (e.g., https://arxiv.org/pdf/2206.13397). Papers in that area have also utilized spectral density plots for analysis. Regardless, I have not seen these specific components combined for the purpose of poisoning. I think that more exploration into poisons that focus on the frequency-spectrum analysis is necessary, and so BlurGuard is a step in the right direction. The existing ablations in the paper and Appendix G are insightful and valuable. It would be even better if the paper could conduct further ablations to serve as a foundation for further research. Some that come to mind include testing other adversarial objectives instead of just EncoderAttack, analyzing sharpening in addition to blur, and applying blur to the adversarial image (x + delta) instead of just the adversarial noise.
  • Clarity: Clarity is a particular weakness of this paper. My main complaint is the organization of sections 3 and 4. Equation 1 has a notational error, as the argmin of L_enc is equal to the MSE objective, but its argmin is not. Equation 4 is introduced and then never used. I think "Per-Region Adaptation" should be introduced before "Learnable Gaussian Blurring" in section 4.2 since masking and blurring are strongly linked in BlurGuard. Notation for distance constraints seems inconsistent, using d(-,-), ball B, or ||-||. Also it seems epsilon and eta are used interchangeably. Perhaps a simplified (shorter) Algorithm 1 should be shown in the main paper because the multi-stage optimization procedure is unintuitive, especially with using dummy deltas during stage 1. Elsewhere, some figures have near-unreadable labels (e.g., Fig. 2, Fig. 4). Fig. 4 and its discussion are confusing: I don't understand what is meant by the "Worst-Case Gap", does the low value of BlurGuard's "Worst-case Gap" indicate that it performs strongly even with a 1/255 adversarial perturbation?
  • Quality: The paper is generally well-written, with straightforward introduction and experiments sections, as well as a nicely organized Appendix. I cite the extensive studies in the appendix as a strength. The authors have also benchmarked performance against a large set of modern adversarial methods and purification methods.

问题

I have given multiple suggestions in Strengths and Weaknesses. In particular, I emphasize additional ablation studies and improved clarity in section 4. I think these improvements are necessary to accept the paper.

局限性

Yes

最终评判理由

The authors have answered my questions and addressed my concerns.

格式问题

Some near unreadable figure labels

作者回复

We sincerely appreciate your thoughtful and constructive feedback. We respond to each of your questions one-by-one in what follows. Should you have any additional comments or concerns, we would be happy to discuss or clarify them.


Q1. BlurGuard with other adversarial objectives

As you mentioned, BlurGuard is compatible with a variety of adversarial image protection objectives beyond the encoder attack primarily considered in our experiments. Following your suggestion, we have additionally tested BlurGuard in combination with other adversarial objectives, viz., AdvDM [1] and Mist [2]. Specifically, AdvDM aims to directly maximize the denoising loss of the diffusion process, and Mist considers a combined form of encoder and denoising attacks. As shown in the table below, we observe that BlurGuard preserves its effectiveness in improving both naturalness and worst-case effectiveness of both attacks, confirming the broad applicability of our framework across diverse protection objectives. We will incorporate these additional results in the final draft.

ImageNet-EditNaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑PSNR↑FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
AdvDM0.36±0.120.74±0.0928.9±0.2898.270.30±0.080.73±0.0930.2±1.170.93±0.04
AdvDM + BlurGuard0.25±0.070.81±0.0930.1±2.19101.840.35±0.090.70±0.0928.2±2.030.93±0.05
Mist0.35±0.120.72±0.1028.6±0.2590.960.30±0.070.71±0.0929.7±0.860.93±0.04
Mist + BlurGuard0.33±0.110.86±0.0832.3±2.21137.910.42±0.100.63±0.0728.8±0.470.89±0.06

Q2. Additional ablation study

We appreciate your insightful suggestions regarding further ablation studies. In response, we have conducted additional experiments based on the following two ablations of BlurGuard: (a) applying blurring directly to the protected image (i.e., x+δ\mathbf{x} + \boldsymbol{\delta}) rather than only to the perturbation (i.e., δ\boldsymbol{\delta}), and (b) replacing the blurring operation with sharpening, which effectively acts as a high-pass filter. From the results shown in the table below, we observe the following:

  • (a) Blurring x+δ\mathbf{x} + \boldsymbol{\delta} has a moderate effect in enhancing robustness, e.g., over the vanilla PhotoGuard. However, it also results in a notable drop in naturalness metrics compared to BlurGuard, as blurring the entire image erases high-frequency details from the image itself and thus compromises perceptual quality. In contrast, BlurGuard can avoid the issue by applying blur only to δ\boldsymbol{\delta}, preserving both robustness and visual fidelity.
  • (b) The adaptive sharpening scheme, implemented via unsharp masking [3], did not provide effective protection after purification. This is because sharpening rather amplifies high-frequency patterns, which can be easily suppressed by known purification techniques, e.g., JPEG compression.

Overall, these results highlight our proposed adaptive blurring scheme as a more effective strategy, achieving both robustness and perceptual naturalness. We will incorporate these results in our final manuscript.

ImageNet-EditNaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑PSNR↑FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
PhotoGuard0.34±0.110.70±0.1128.2±0.3292.210.27±0.070.73±0.1029.9±1.070.94±0.04
BlurGuard0.21±0.100.93±0.0931.1±2.29107.880.32±0.090.70±0.0928.8±0.420.92±0.05
└─blur(x+δ\mathbf{x}+\boldsymbol{\delta})0.32±0.110.79±0.0832.3±1.1198.450.30±0.070.71±0.1030.1±1.210.94±0.04
└─sharpen(δ\boldsymbol{\delta})0.35±0.120.65±0.1030.9±1.0882.720.28±0.080.72±0.1030.4±1.380.94±0.04

Q3. Editorial comments

Thank you very much for your detailed reading. We agree that your editorial suggestions will help strengthen the clarity of our manuscript. We will thoroughly review all of your comments and incorporate them into the final draft, e.g., including the following points:
Equation (1) will be rewritten to separately define L_encL\_{\mathrm{enc}} apart from the argmin\mathrm{argmin} formulation. The distance notations will be clarified, e.g., by standardizing the use of ϵ\epsilon in place of η\eta for consistency. A simplified version of Algorithm 1 will be incorporated into the main text. The font sizes of Figure 2 and 4 will be increased to improve readability.


Q4. Figure 4: “Worst-case gap”?

We use the term “worst-case gap” in Figure 4 to refer to the drop in protection effectiveness (e.g., in LPIPS) between a protected image before and after the strongest purification attempt among those considered. In other words, it measures how much the protection effect is weakened by the strongest purification. Therefore, the small worst-case gaps of BlurGuard in Figure 4 show that its protections remain robust even under aggressive purification attacks. This trend holds consistently across a wide range of protection budgets; e.g., including those from ϵ=1255\epsilon=\tfrac{1}{255}, as you mentioned, up to ϵ=16255\epsilon=\tfrac{16}{255}. We will clarify this point in the final draft.


[1] Liang et al., Adversarial Example Does Good: Preventing Painting Imitation from Diffusion Models via Adversarial Examples. ICML 2023.
[2] Liang & Wu, Mist: Towards improved adversarial examples for diffusion models. 2023.
[3] Polesel et al., Image enhancement via adaptive unsharp masking. IEEE TIP 2000.


评论

The authors apply BlurGuard to additional base adversarial objectives (AdvDM and Mist) and BlurGuard shows improvement for each. They have also conducted additional studies on sharpening and blur applied to the adversarial image; the original formulation outperforms both ablations, but I think the ablations are useful nonetheless. The authors intend to incorporate some of the formatting/clarity suggestions.

I have additionally reviewed responses to other reviewer comments. Some notable highlights:

  • fDXr: Additional experiments with EditShield
  • jjT1: Additional validation on SD3.5 and FLUX1
  • Z04E: Additional comparison to DiffusionGuard for inpainting

I am pleased with these improvements and raise my score to 4.

评论

Dear Reviewer Y1wo,

Thank you very much for your thoughtful feedback, and for raising your score!

We are grateful that our additional experiments and clarifications, including those for other reviewers, were carefully considered and helped address your concerns. We will make sure to incorporate all these discussions into the final draft.

Thank you once again for your time and support.

Best regards,
Authors

评论

Dear Reviewer Y1wo,

Thank you again for your time and effort in reviewing our manuscript.

As we are now midway through the discussion period, we would like to kindly check if you have any remaining questions or concerns we could help clarify.

We believe we have made a sincere effort to address your earlier comments, and would be happy to receive any additional feedback you may have.

Your insights are invaluable and would greatly help us further strengthen our manuscript.

Best regards,
Authors

审稿意见
5

BlurGuard is a effective image protection method designed to improve robustness against AI-driven image editing. It addresses the vulnerability that prior adversarial noise-based protections can be easily reversed by simple post-processing. BlurGuard’s approach is to apply an adaptive local Gaussian blur to the embedded perturbation to adjust its frequency spectrum, making the protective noise difficult to extract while mitigating image quality loss. Experiments show that this method significantly improves worst-case protection performance under diverse image editing scenarios and noise-removal attacks, and achieves superior image quality compared to existing methods.

优缺点分析

Strengths:

  1. Innovativeness of the Work.This study demonstrates a notable degree of innovation. The proposed BlurGuard method adjusts the frequency bandwidth of adversarial noise via adaptive Gaussian blurring, presenting a simple yet effective strategy to enhance the robustness of adversarial image protection. Compared to existing approaches, this method emphasizes alignment with the natural frequency spectrum of images, thereby improving the perceptual naturalness of the protected outputs.
  2. Comprehensiveness of the Experimental Evaluation. The experimental section of the paper is well-developed and comprehensive. It provides strong empirical evidence supporting the effectiveness and robustness of BlurGuard. The experiments are rigorously designed and the results are presented in detail, highlighting the advantages of the method under various attack scenarios.
  3. Construction of a Dataset. The study acknowledges the lack of publicly available standardized datasets for malicious image editing and accordingly constructs a tailored evaluation dataset. This dataset encompasses a diverse set of images and editing intentions, facilitating consistent and meaningful comparisons across different protection methods.
  4. Clarity and Logical Organization of the Paper. The paper is clearly written and logically organized. The arguments are well-structured and coherent, enabling readers to follow the motivation, methodology, and contributions of the work in a systematic and persuasive manner. Weakness:
  5. Lack of Comparison with the Recent Methods.The comparative methods used in the experimental section are primarily from 2023 and do not include the most up-to-date techniques. For example, more recent approaches such as Distraction is All You Need and EditShield could be incorporated to provide a more comprehensive evaluation.
  6. Insufficient Explanation in Certain Experimental Results.Some parts of the experimental analysis lack detailed interpretation. For instance, in Table 9, the BlurGuard method exhibits relatively lower performance in terms of PSNR under the "Naturalness" metric, yet no corresponding explanation is provided in the text.
  7. Lack of Discussion on Alternative Tampering Methods.BlurGuard is based on adversarial perturbation. If an attacker employs alternative editing techniques, such as methods beyond diffusion models, the protective effectiveness of BlurGuard may be limited. This potential limitation is practically relevant and deserves further discussion.
  8. Omission of Acronym Explanations.Certain acronyms are not clearly defined in the main text. For example, the “I.A.” metric in Table 1 is not explicitly explained in the body of the paper. Although it is addressed in the appendix, its meaning should be clarified in the main text or table caption to avoid confusion, especially for first-time readers.

问题

  1. Lack of Comparison with the Most Recent Methods. The comparative methods used in the experimental section are primarily from 2023 and do not include the most up-to-date techniques. For example, more recent approaches such as Distraction is All You Need and EditShield could be incorporated to provide a more comprehensive evaluation.
  2. Insufficient Explanation in Certain Experimental Results. Some parts of the experimental analysis lack detailed interpretation. For instance, in Table 9, the BlurGuard method exhibits relatively lower performance in terms of PSNR under the "Naturalness" metric, yet no corresponding explanation is provided in the text.
  3. Lack of Discussion on Alternative Tampering Methods. BlurGuard is based on adversarial perturbation. If an attacker employs alternative editing techniques, such as methods beyond diffusion models, the protective effectiveness of BlurGuard may be limited. This potential limitation is practically relevant and deserves further discussion.
  4. Omission of Acronym Explanations. Certain acronyms are not clearly defined in the main text. For example, the “I.A.” metric in Table 1 is not explicitly explained in the body of the paper. Although it is addressed in the appendix, its meaning should be clarified in the main text or table caption to avoid confusion, especially for first-time readers.

局限性

Lack of Discussion on Alternative Tampering Methods. BlurGuard is based on adversarial perturbation. If an attacker employs alternative editing techniques, such as methods beyond diffusion models, the protective effectiveness of BlurGuard may be limited. This potential limitation is practically relevant and deserves further discussion.

最终评判理由

The authors have addressed most of our concerns in the rebuttal process.

格式问题

no

作者回复

We sincerely appreciate your thoughtful and constructive feedback. We respond to each of your questions one-by-one in what follows. Should you have any additional comments or concerns, we would be happy to discuss or clarify them.


Q1. Additional baselines

Thank you for suggesting additional baselines. In our experiments, we selected our baselines that are most commonly considered in recent literature on adversarial image protection. We note that we also made an effort to include more recent techniques where possible, e.g., DiffusionGuard (2025) [1] and High-Frequency Anti-DreamBooth (2024) [2], as well as state-of-the-art purification-based defenses like GrID‑Pure (2024) [3] and PDM‑Pure (2024) [4]. In response to your suggestion, we have additionally compared BlurGuard with EditShield [5] you mentioned, although we were unfortunately unable to include “Distraction is All You Need” [6] due to the lack of publicly available code. As summarized in the table below, the results further confirm the strong robustness of BlurGuard as a natural and effective image protection method. We will incorporate these and more results in the final draft.

MagicBrushNaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑PSNR↑FID↑LPIPS↑SSIM↓PSNR↓I.A.↓
PhotoGuard0.19±0.100.74±0.0931.1±0.2296.230.21±0.090.73±0.1130.7±1.500.96±0.03
AdvDM0.27±0.080.76±0.07\underline{0.76\pm 0.07}28.9±0.24123.390.30±0.090.67±0.1330.2±1.250.93±0.04\underline{0.93\pm 0.04}
Mist0.26±0.090.74±0.0828.6±0.17123.54\underline{123.54}0.30±0.090.66±0.1229.8±0.94\underline{29.8\pm 0.94}0.93±0.04\underline{0.93\pm 0.04}
SDS0.23±0.070.74±0.0729.1±0.37115.230.27±0.090.68±0.1330.3±1.380.94±0.04
EditShield\textcolor{blue}{\text{EditShield}}0.36±0.130.75±0.0931.0±0.23\underline{31.0\pm 0.23}120.130.33±0.09\underline{0.33\pm 0.09}0.65±0.13\underline{0.65\pm 0.13}29.9±1.260.93±0.04\underline{0.93\pm 0.04}
BlurGuard0.20±0.08\underline{0.20\pm 0.08}0.89±0.0730.1±1.76138.220.36±0.100.64±0.1228.8±0.550.90±0.06

Q2. Table 9: Clarification on PSNR results

We remark that PSNR primarily captures pixel-wise differences to measure naturalness, and accounts less for perceptual relevance; often penalizing even imperceptible changes. BlurGuard intentionally applies stronger perturbations in perceptually insensitive regions to improve robustness while maintaining visual quality. These imperceptible changes can slightly lower PSNR, despite not affecting human perception. This aspect is better reflected in stronger performance on perceptual metrics like SSIM and LPIPS, and is further supported by qualitative results in Figure 11 in Appendix. We will clarify this point in the final version.


Q3. Discussion on non-diffusion editing methods

We acknowledge that the transferability of adversarial image protection methods, e.g., to non-diffusion editing methods, is an important step toward increasing their practical applicability. In particular, the question of whether adversarial perturbations crafted for one type of editor (e.g., a diffusion model) can effectively transfer to other types of (black-box) editing tools, e.g., GAN-based models, remains a technically challenging and, to our knowledge, under-explored problem. In this work, we focus on diffusion-based pipelines, as they currently represent the most widely used class of image editing models in both academic and commercial settings. We think extending BlurGuard to cover non-diffusion editors is a promising future direction, and we will reflect this point in the final draft by revising the discussion of limitations (e.g., in Appendix A.2).


Q4. “I.A.”: Should be defined in the main text

Thank you for the incisive comment. We will ensure that the I.A. metric is clearly defined and explained in the main text of the final draft.


[1] Choi et al., DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing. ICLR 2025.
[2] Onikubo et al., High-Frequency Anti-DreamBooth: Robust Defense against Personalized Image Synthesis. ECCV Workshop 2024.
[3] Zhao et al., Can Protective Perturbation Safeguard Personal Data from Being Exploited by Stable Diffusion? CVPR 2024.
[4] Xue et al., Pixel is a Barrier: Diffusion Models Are More Adversarially Robust Than We Think. 2024.
[5] Chen et al., EditShield: Protecting Unauthorized Image Editing by Instruction-guided Diffusion Models. ECCV 2024.
[6] Lo et al., Distraction is all you need: Memory-Efficient Image Immunization against Diffusion-Based Image Editing. CVPR 2024.


评论

Dear Reviewer fDxR,

Thank you again for your time and effort in reviewing our manuscript. We are also delighted and encouraged by your positive review, and we truly appreciate your support for our manuscript.

As we are now midway through the discussion period, we would like to kindly check if you have any remaining questions or concerns we could help clarify.

We believe we have made a sincere effort to address your earlier comments, and would be happy to receive any additional feedback you may have.

Your insights are invaluable and would greatly help us further strengthen our manuscript.

Best regards,
Authors

评论

While BlurGuard shows strong robustness against diffusion-based models, the practical relevance of this work could be limited if the defense cannot be easily adapted to these other techniques. The increasing use of GAN-based editors (such as in commercial applications like deepfake generation) raises important questions regarding the applicability of BlurGuard outside of diffusion models. It would be helpful to address how these alternative methods could potentially circumvent BlurGuard’s protections and what specific challenges this creates.

评论

Dear Reviewer fDxR,

Thank you for your follow-up and for raising this point before the end of the discussion period, allowing us to address it more thoroughly.

Q3-2. BlurGuard on other editing methods, e.g., GANs

In response to your concern, we have additionally conducted a black-box transfer experiment from SD-v1.4 to SimSwap [7], a popular GAN-based method for deepfake generation, using the human portrait subset of InpaintGuardBench [8] as the test set. As shown in the table below, BlurGuard continues to demonstrate strong performance in both naturalness and worst-case effectiveness. We find this to be a particularly interesting observation, as it confirms the strong transferability of our approach not only across architectures, but also across fundamentally different generative paradigms, i.e., from diffusion models to GANs. We will incorporate these additional results and the corresponding discussion into the final draft.

SD-v1.4 \rightarrow SimSwapNaturalnessWorstcase effect
(ε=16/255\varepsilon=16/255)LPIPS↓SSIM↑FID↑PSNR↓
PhotoGuard0.04±0.020.96±0.0218.3440.0±2.67
AdvDM0.04±0.030.97±0.0117.1640.4±2.43
Mist0.02±0.010.98±0.01\underline{\text{0.98±0.01}}14.1041.0±2.43
SDS0.06±0.030.96±0.0225.11\underline{\text{25.11}}39.3±2.44\underline{\text{39.3±2.44}}
DiffusionGuard0.04±0.020.97±0.0117.3940.2±2.33
BlurGuard0.03±0.02\underline{\text{0.03±0.02}}0.99±0.0131.5438.4±2.37

Furthermore, we clarify that our BlurGuard framework is not limited to diffusion-based models even in white-box setups, i.e., it can be readily applied to protect images directly against other editing models. This is because our approach does not rely on any prior specific to diffusion models; rather, it introduces a simple regularization term that can be added to arbitrary loss objectives, exploiting frequency characteristics of adversarial noise shared across models. In other words, we expect that the protective performance of BlurGuard could be further enhanced if it were directly applied to disrupt GAN outputs (e.g., for deepfakes, where GANs remain widely used), although we consider this to be beyond the current scope.

Thank you once again for your engagement and valuable feedback!

Best regards,
Authors


[7] Chen et al., SimSwap: An Efficient Framework For High Fidelity Face Swapping. ACM MM 2020.
[8] Choi et al., DiffusionGuard: A Robust Defense Against Malicious Diffusion-based Image Editing. ICLR 2025.


最终决定

This paper presents a simple method for protecting an image against AI-based image editing. The idea extends existing perturbation-based protection methods to local regions by utilizing segmentation masks. Experiments demonstrate more effective protection over baselines. Overall, the paper is easy to understand and well motivated. The reviewers requested several additional experiments, which the authors have addressed; hence, the reviewers are supportive of this paper. The AC agrees with this evaluation.