A Theoretically Grounded Extension of Universal Attacks from the Attacker's Viewpoint
摘要
评审与讨论
The paper proposes a new method for generating universal adversarial perturbations by optimizing multiple perturbations simultaneously and selecting the most optimal one.
优点
- The paper is easy to follow.
- The issue of universal adversarial perturbation is quite intriguing.
缺点
-
The theoretical contribution is incremental and offers limited guidance for algorithm design. The primary theoretical input is the application of Rademacher complexity to the generalization of adversarial perturbations, which is somewhat limited. Furthermore, this theoretical insight does not significantly influence the algorithm's design.
-
The experiments on adversarial transferability are not comprehensive enough. (1) Tests were conducted only on CIFAR-10, whereas most experiments concerning adversarial transferability are performed on ImageNet. (2) The choice of the target model is not reasonable; since the source model is already a ResNet, the target model should incorporate structures other than ResNet.
-
The motivation for Equation (9) is unclear. Regarding Equation (9), it is not explained why a set of perturbations is used. Also, the definition of problem 1 suggests there should be constraints on L; otherwise, when L equals n, Equation (9) essentially degenerates into Equation (1). Moreover, there is a lack of discussion in the experimental section regarding the appropriate size of the parameter L.
-
Comparative experiments are not entirely fair. The authors' method optimizes several perturbations simultaneously, but the baseline methods compared are all optimizing a single perturbation. This comparison's fairness needs further discussion.
问题
- The MI-FGSM is designed to improve adversarial transferability, but why does MI-FGSM not outperform PGD in Table 1?
The paper views the universal attack problem from a perspective of generalization bound with Radamecher complexity analysis and accordingly proposes an effective optimization procedure to learn the universally effective adversarial perturbations.
优点
-
The paper clarifies well, especially in the preliminary and related work part.
-
Viewing the universal attack from a generalization analysis perspective is novel to me.
缺点
-
The effectiveness of the L-UAP algorithm is not adequately explained. It basically designs a new step update procedure (inner loop of algorithm 1). How is that important? Is there any analytical or theoretical grounding?
-
The proof of Theorem 4 is missing. Although one paper is cited, and the theorem may be built from it, I do not think the paper considers Algorithm 1 proposed by you, so it was important to show the proof of it.
-
From the results in Table 1, why are multiple perturbations beneficial for source model attack but not that good for transfer attack?
-
Problem 1 and Algorithm 1 only do optimization with respect to one model, which is similar to a single model attack. What is the special component here for universal attack?
-
In proposition 1, the paper defines Radamecher complexity of a set of perturbations , which is weird to me since the complexity measure is defined for a function class.
-
In Theorem 2, consider the formula inside the probability operator, parameter does not appear in LHS. What the paper bounds here is exactly the same as that in Theorem 1.
-
Is it possible to consider the complexity measure (Rademacher complexity) as a regularization term to improve the transferability of universal attacks? If it is feasible, the idea looks more interesting.
-
Presentation: I do not even find the definitions of and in algorithm 1 in the main text, which is very confusing to readers.
问题
Please refer to the weakness part.
The authors consider a generalization of universal attacks in which, for any input, the attacker can choose the optimal attack from a fixed set of perturbations. Under this model, the authors prove generalization bounds on the fooling risk. Empirical results demonstrating the attack are then presented.
优点
The way universal attacks are generalized leads to a novel adversarial attacking framework. The ideas are presented clearly, and contextualized well.
缺点
It is unclear how good the generalization bounds are. As a result it is difficult to evaluate the theoretical results. (More details in the Questions section)
问题
On page 4, the claim that Equation 6 holds for any model we want to attack that is not necessarily the one we used to learn appears to be false.
The generalization bounds (Proposition 1 and Theorem 2) are given in terms of Rademacher complexity. What is the Rademacher complexity? Does it converge to 0? If so, how quickly? There is no indication of how large this term is. Are the generalization bounds vacuous?
On page 7, the estimation of the Rademacher complexity gives an approximate lower bound of the Rademacher complexity, but the Rademacher complexity is used as an upper bound. As a result I am not sure how useful this estimate is.
The bound given in Proposition 3 also seems quite bad. In particular, the term . It is claimed that "this term is small if we cannot find a set of perturbations that differs too much from the perturbations for ." But isn't it almost always possible to find such a set? As an example, you could set so that all of its elements have almost zero norm, so there is essentially zero perturbation.
This paper introduces an extension to the Universal Adversarial Perturbation (UAP) by developing a set of perturbations, termed L-UAP, which are jointly learned to optimize the likelihood of successfully attacking deep neural network models.
优点
Theoretically, the research broadens the scope of Rademacher complexity from a model learning perspective to encompass perturbation learning, aiming to gauge the generalization capability of universal attacks.
Experimentally, the paper provides comprehensive experiments. The trends of attack rates w.r.t the number of perturbation L provides information of readers to understand the ability of UAP.
缺点
The primary limitation of this research is the weak linkage between the theoretical framework and experimental results, even though it is a challenge commonly observed in uniform convergence analyses.
-
In both Theorem 1 and Theorem 2, the bounds are set for all perturbations within , rather than specifically for learned UAP attacks.
-
When contrasting Theorem 1 with Theorem 2, it becomes evident that the generalization abilities of UAP and L-UAP are evaluated by and respectively. Considering that represents a more extensive class, its Rademacher complexity would naturally be larger. As a result, the theorem might suggest that L-UAP's generalization is inferior to UAP, which contradicts the paper's primary objective. This is similar to a general criticism for previous rademacher complexity analysis: Rademacher complexity of large model is very large, yet large model generalize better.
Therefore, without a deeper analysis of and , the analysis is not informative. Given this, the assertion in Section 2 — "motivated by Theorem 2, we propose to maximize problem 1” — doesn't come across as compelling.
- Theorem 4's convergence results under the Lipschitz and KL conditions appear somewhat lacking in depth and specificity.
Given the weak theoretical analysis, increasing the number of perturbation candidates is not very novel.
问题
See weakness
Dear Reviewers,
Your time and considerate feedback on our submission are deeply valued. Every suggestion and concerns you provided has been meticulously examined. We have taken a comprehensive approach in addressing your comments and integrating them into our revised manuscript.
However, we believe that the required clarifications about both our theoretical bound and the applicability of the algorithmic solution cannot be integrated to meet the page limit.
We have made the decision to withdraw our paper.
Sincerely