PaperHub
3.0
/10
withdrawn4 位审稿人
最低3最高3标准差0.0
3
3
3
3
4.0
置信度
ICLR 2024

Diffusion Denoising as a Certified Defense Against Clean-Label Poisoning Attacks

OpenReviewPDF
提交: 2023-09-16更新: 2024-03-26
TL;DR

We show that our certified defense against data poisoning that leverages the diffusion denoising approach renders existing clean-label poisoning ineffective while preserving a model utility.

摘要

We present a certified defense to clean-label poisoning attacks. These attacks work by injecting poisoning samples that contain $p$-norm bounded adversarial perturbations into the training data to induce a targeted misclassification of a test-time input. Inspired by the adversarial robustness achieved by $denoised$ $smoothing$, we show how a pre-trained diffusion model can sanitize the training data before a model training. We extensively test our defense against seven clean-label poisoning attacks and reduce their attack success to 0-16% with only a small drop in the test time accuracy. We compare our defense with existing countermeasures against clean-label poisoning, showing that the defense reduces the attack success the most and offers the best model utility. Our results highlight the need for future work on developing stronger clean-label attacks and using our certified yet practical defense as a strong baseline to evaluate these attacks.
关键词
Certified defenseData poisoningDiffusion denoising

评审与讨论

审稿意见
3

In this paper, the authors propose a certified defense against clean-label data poisoning attacks using diffusion denoising. The authors sanitize the training dataset by using an off-the-shelf diffusion model to remove adversarial perturbations on poisoned data. They also propose a warm-starting initialization to improve the model utility under the defense. Experiments are carried out on CIFAR-10 to evaluate the defense against multiple attacks and comparisons are performed with existing heuristic and certified defenses.

优点

  1. The proposed approach is simple. The authors leverage the diffusion model to remove adversarial noise added on the poisoned data.

  2. The authors proposed the strategy of warm-starting to improve the model utility.

  3. The authors evaluate the proposed defense over difference settings such as training from scratch and transfer-learning.

缺点

  1. One of the most important claims of this work is that the proposed approach is a certified defense and it offers a provable guarantee. However, there is missing theoretical proof for the certified defense in the paper.

  2. I favor the simplicity of the proposed approach (i.e., the authors use an off-the-shelf diffusion model to remove the noise imposed on the poisoned samples), but the motivation of this idea is not well explained. Why pick diffusion model as a denoising approach over other techniques? Why does diffusion model work well in removing the adversarial perturbations?

  3. Experiments and evaluations are preformed only on CIFAR-10 dataset, which is insufficient to showcase the effectiveness and efficiency of the proposed attack. The authors mentioned that the proposed defense requires training of many models and is computationally expensive. Thus it is necessary to show the performance of the defense on more and large-scale datasets.

  4. The comparison to existing defenses looks insufficient. Only three attacks and one certified defense are evaluated in Table 5. It would be desired to have a more comprehensive evaluation that includes more attacks and certified defenses.

  5. The contribution of the proposed defense could be limited because it only considers the clean-label integrity attacks and backdoor attacks. Would it still be effective against clean-label availability attacks such as CLPA[1] or dirty-label backdoor attacks such as BadNets[2], LIRA[3], WaNet[4], etc? I suggest the authors better specify the scope of the work and clarify why other attacks are not considered.

  6. This paper is not well polished. For example, the caption of Algorithm 1 is Figure 1 and there is no function called Certify in the algorithm. There are also typos and grammatical issues that makes the paper hard to follow.

[1]. CLPA: Clean-Label Poisoning Availability Attacks Using Generative Adversarial Nets, AAAI 2022

[2]. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain

[3]. LIRA: Learnable, Imperceptible and Robust Backdoor Attacks, ICCV 2021

[4]. WaNet -- Imperceptible Warping-based Backdoor Attack, ICLR 2021

问题

Please see weaknesses for details.

审稿意见
3

This paper studies the adversarial robustness against the poisoning attack. The core idea of this paper is based on using diffusion denoising to improve the performance of randomized smoothing. In the randomized smoothing mechanism, inputs are injected by noise drawn from a noise distribution, both in the training and the inference. This training process can degrade the task performance due to the distortion of the injected noise. The diffusion model can be used as a denoiser to clean this noise and recover the semantic representation of the inputs. It starts with a noise variable and then denoises the variable continuously toward a sample from the normal data distribution. Therefore, this paper uses the diffusion model to denoise the noisy input in randomized smoothing, by considering the noisy input as the middle variable in the reverse diffusion process. Different from other works that improve the randomized smoothing using the diffusion model, this paper focuses on the certified robustness against poisoning attacks.

优点

  1. To the best of my knowledge, this is the first work that applies diffusion denoising to randomized smoothing against poisoning attack, which is a novel application.

  2. The writing is easy to understand.

缺点

  1. Some core algorithms/methodologies are missing. It is not clear how the randomized smoothing is applied to the poisoning attack. In Section 3.2, this paper states the certificate follows the standard randomized smoothing, which is confusing since the randomized smoothing is proposed to certify the robustness of evasion attacks. Since the poisoning attack shares a different setting with evasion attacks, the authors may want to clarify how the randomized smoothing guarantees the robustness against the poisoning attack. In addition, how the time step in the reverse diffusion process is computed according to σ\sigma is not clear. The algorithm GetTStep is not well clarified in the main text. This is actually an important algorithm that is non-trivial.

  2. No ablation study supports the main contribution. The main contribution is using the diffusion denoiser to improve the randomized smoothing, while this paper lacks the ablation study to compare the randomized smoothing performance w/ and w/o the diffusion denoising. It is not clear whether the diffusion denoising improves the randomized smoothing.

  3. This paper claims that one of the contributions is to provide a provable defense that also works in realistic settings. This is actually controversial. When using the certified defense to defend against real attacks, the settings are very different and the results may be meaningless. For example, the randomized smoothing used in this paper provides the certified robustness in the criteria of 2\ell_2 norm radius, but the real attacks used in this paper adopt the \ell_\infty constrain, so it is not fair to evaluate the 2\ell_2-norm defense against the \ell_\infty-norm attacks. In addition, the threat model could be very different in the empirical/certified defense. The model owner does not need to know the perturbation of the inputs in empirical defense, while the model owner needs to know the perturbation of the inputs when comparing the perturbation size to the radius, which is not realistic.

  4. The application of randomized smoothing in poisoning attacks will raise a severe problem in running time. In this paper, to certify the perturbation to the training set on the prediction w.r.t. a specific sample, it requires training 10,000 models, plus the computation on the multi-step diffusion process, even for the fine-tuning setting, the running time overhead will be extremely large.

问题

  1. Can you provide some quantitative results on the overhead of the robust training and the diffusion process?

  2. How do you compute the step T in the diffusion process?

  3. How does randomized smoothing guarantee robustness against poisoning attacks?

  4. Can you provide the ablation study on randomized smoothing w/ and w/o denoising?

审稿意见
3

This paper introduces a certified defense mechanism designed to counter clean-label poisoning attacks. These attacks involve injecting malicious samples into the training data, which carry adversarial perturbations bounded by a specified p-norm. These perturbations aim to manipulate the model's predictions for certain test inputs intentionally. Drawing inspiration from the concept of adversarial robustness achieved through denoised smoothing, the authors demonstrate how a pre-trained diffusion model can effectively cleanse the training data before model training.

The defense undergoes extensive testing against seven distinct clean-label poisoning attacks, significantly reducing their success rates to a range of 0–16%. This defense accomplishes this without a significant decrease in test-time accuracy. In comparative evaluations against existing countermeasures for clean-label poisoning, the method stands out by delivering the most substantial reduction in attack success while maintaining high model utility.

These results underscore the importance of future research efforts in developing more potent clean-label attacks. Moreover,the certified yet practical defense serves as a robust baseline for evaluating these attacks and enhancing the security of machine learning models.

优点

  1. The paper introduces a defense method based on DDPM to counter clean-label poisoning attacks, and it demonstrates strong performance against various attack methods.

缺点

  1. The primary results presented in Table 1 are obtained through empirical experiments rather than relying on certified defense methods that guarantee a specific radius where the model is proven to provide the correct predictions. I would suggest a mathematical formulation to provide a certified bound.

  2. Randomized smoothing has already been established as an effective technique for ensuring l2 or l_infinite robustness against various types of perturbations, including backdoor perturbations. Moreover, it offers a clear and established framework for calculating certified l2 or l_infinite robustness bounds. The novelty of this paper is somewhat limited, as its core idea revolves around utilizing randomized smoothing to provide certified defense specifically against backdoor perturbations.

  3. Experiments on the indiscriminate data poisoning [1] should also be included.

[1] Yiwei Lu, Gautam Kamath, and Yaoliang Yu. 2023. Exploring the limits of model-targeted indiscriminate data poisoning attacks. In Proceedings of the 40th International Conference on Machine Learning (ICML'23), Vol. 202. JMLR.org, Article 949, 22856–22879.

问题

See weakness above.

审稿意见
3

The paper proposes a certified defense against clean label poisoning attacks, where an attacker is allowed to modify each sample in the training dataset by ϵ\epsilon in an p\ell_p norm in order to obtain a misclassification on a target test point xx. The defense is a direct application of Randomized Smoothing to this problem, and proceeds by first adding gaussian noise to each image in the dataset, then denoising each image by a diffusion model, and finally training a classifier on this denoised dataset. The majority vote on xx over multiple classifiers obtained from such noise-denoise steps gives rise to a certifiably robust prediction.

优点

  1. Denoising has been popular recently as a defense against adversarial examples, and this paper extends these works to the poisoning setting, which is interesting. 


  2. The paper contains a nice overview of the field related to poisoning attacks, which is very helpful in contextualizing this work.

缺点

The paper consists of two parts: a robustness certificate (on the amount of poisoning adversarial corruption allowed per image in the dataset), and an empirical evaluation of the method with respect to known attacks in the literature.

  1. The presentation of the first part is unclear, missing key technical details


  2. The implementation of the second part raises several questions, pertinent to proper evaluation of defenses with random components in general, and defense using diffusion models in particular.

问题

  1. The authors should clarify the following technical details:

    1. Is it right to understand that just as in traditional RS, we have g(x)=argmaxkPrv[f(x+v)=k]g(x) = {\rm argmax}_k Pr_v[f(x + v) = k], in this paper we have g(x)=argmaxkPrv[Algx(D+v)=k]g(x) = {\rm argmax}_k \Pr_v[{\rm Alg}_x(D + v) = k], where Algx{\rm Alg}_x is an algorithm where the test point xx is fixed, and it produces a classification for xx by examining the dataset DD. In this paper, Alg corresponds to (Denoise D) -> Train classifier on D -> output prediction on xx. 

However, this procedure should produce an 2\ell_2 certificate on the entirety of DD, which is unstructured, i.e., there is no restriction that the perturbation made to each image is 2\ell_2 bounded. This is different from an element-wise certificate on each image in DD, which was the goal presumably. 

    2. Now, in the previous point, if the norm used is \ell_\infty, then it doesn’t matter because \ell_\infty decomposes into individual pixels. However, now traditional RS certificates do not hold for \ell_\infty, and other noise distributions are more suitable [1].

    3. In light of the above, it is extremely unclear what does “and from this we can derive a radius…” mean, and for that matter what is Figure 2 plotting. A derivation of the certificate (even if it is a reproduction of Cohen et. al.) obtained would make the statements clearer.

  2. Empirical Evaluation: One needs to be very careful while backpropagating gradients through diffusion models to attack them. Accordingly, the optimization steps in many of the tested poisoning attacks which explicitly use gradients of the networks are doomed to fail from the start. In particular:

    1. See Section 3.2 in [1] for a detailed evaluation of how diffusion denoising can be broken as an adversarial defense, and what can be used to improve its real robust performance. Specifically, how does incorporating the techniques used in [2] (gradient estimation) to the optimization steps of all the tested attacks affect the evaluation?

    2. More generally, adversarial evaluation of defenses with random components is very tricky [3], and typically the gradients need to be stabilized with EoT attacks as a first step towards proper evaluation. Accordingly, it is important to understand how do all the evaluated attacks perform when the gradients of the defense are stabilized. 

    3. The empirical evaluation seems to be on \ell_\infty attacks, but the certificates were presumably on 2\ell_2 attacks. Gaussian smoothing is used to obtain certificates against 2\ell_2 attacks. For \ell_\infty attacks, gaussian smoothing is not the best noise model, perhaps the results can be improved further by using the correct noise model.

    4. The defense performance seems to shoot up even at very small magnitude of the injected noise σ\sigma. A reasonable conjecture would be that σ>0\sigma > 0 is not even needed for the empirical performances observed currently. In particular, the σ=0\sigma = 0 column should have the title “no defense”, and a separate σ=0\sigma = 0 column should be added which measures performance where no noise is added to the image, and it is simply passed through the forward and the backward process of a diffusion model. If the performance is indeed bad, then there needs to be an experiment finding exactly at what σ\sigma does the defense start performing well, in other words, where is the jump from bad defense performance to good defense performance?

    5. Related to the above, this shooting up of empirical defense performance is a tell-tale sign of gradient masking [3], and the empirical evaluation should include replacement of the optimization steps in all the poisoning defenses studied with a black-box gradient estimation oracle.

[1]: Randomized Smoothing of All Shapes and Sizes, Greg Yang, Tony Duan, J. Edward Hu, Hadi Salman, Ilya Razenshteyn, Jerry Li

[2]: Robust Evaluation of Diffusion-Based Adversarial Purification. Minjong Lee, Dongwoo Kim

[3]: On Evaluating Adversarial Robustness. Nicholas Carlini, Anish Athalye, Nicolas Papernot, Wieland Brendel, Jonas Rauber, Dimitris Tsipras, Ian Goodfellow, Aleksander Madry, Alexey Kurakin.

评论

Dear Reviewers,

We thank the reviewers for taking the time to read and evaluate our paper. We also thank the reviewers for their valuable comments and suggestions. Reading the reviews carefully, we agree that the current manuscript lacks clarity in problem formulation and the evaluation of our claims. We thus decided to withdraw our submission from this batch for the major revisions.

Regards, Authors