Universally Amplifying Randomized Smoothing for Certified Robustness with Anisotropic Noise
A universal framework that boosts randomized smoothing methods with anisotropic noise.
摘要
评审与讨论
This paper proposes an approach to Certified Robustness via Randomized Smoothing, in which the noise distribution for the different data dimensions is allowed to vary, i.e. is anisotropic. Three example noise parameter generators are proposed, and experimental results given on standard datasets against SOTA methods.
优点
The paper is well presented, clearly structured and well written, with a reasonably comprehensive set of experiments performed.
缺点
My scoring reflects very much an issue raised in the Questions below. Clarification on that from the authors could substantially change my assessment. The paper presents only an incremental improvement on SOTA methods, and in many ways the main Theorem (3.2) is really no more than a simple Corollary from prior work with a simple affine transformation on the variables (which would usually be done anyway in dealing with non-image data with very different means and scales in each data dimension).
The main issues, written in questions to the authors, concern (a) the presentation of the comparison with isotropic SOTA, and (b) the validity of the input-dependent noise (the method introduced in section 4.3). This latter is also used as the basis for the main results in section 5.3. Hence my rating for the paper could be improved significantly in light of any author response/clarification to my questions.
Some more minor comments here though:
- The binary case rather than multi-class is used for Cohen. This gives poorer results, of course. Not an issue, as it is done consistently for the proposed and SOTA methods. But it should be at least clarified as the statement of Theorem 3.2 suggests that the p_B value will be used, but then in Table 1 it is not (ie it is replaced by p_B = 1 – p_A)
- The definition of Acc in the Metrics section has it defined as a function of V’_S. And yet V’_S does not appear on the RHS!! (instead you have replaced it with the dth root of the product of the sigmas times R)
- At small radii, the SOTA methods are better (on the graphs) than the proposed method. Some discussion would be welcome on this.
- In section 5.1, it is mentioned that Cohen gives a tight radius. Again, as per above, this is really only in the non-binary form.
问题
In the experiments, it seems that the results vs {min \sigma_i} R should be presented (per Corollary 3.3) for a fair comparison, i.e. it is not clear to me that the proposed technique certifies a strictly larger L_p ball in the same conditions than SOTA (and what “same conditions” may mean is not clear e.g. it may mean isotropic sigma=1 and \product \sigma_i = 1 in anisotropic case). That said, the results in section 5.3 apparently include a certified accuracy wrt radius. It is not clear if this may be an answer to my query as it is not clear exactly what is being reported here.
Re section 4.3, I am a bit confused. This input-dependent proposal would give an x-dependent sigma, and hence an x-dependent classifier (ie x-dependent noise, over and above the obvious dependence on x). Theorem 3.2 holds assuming the noise is constant in the ball around x. This proposal in 4.3 violates that surely. Hence it is no longer true that Theorem 3.2 guarantees that the classifier gives an unchanged output in the region claimed, and so the proposed classifier is not certified robust in the claimed reghion around x. I may have misunderstood. Clarification/explanation is welcome. Please see aeXiv paper 2110.05365 for an example of work done with input-dependent noise.
伦理问题详情
N/A
Thank you for the insightful comments and please find our clarifications as below. We also make more clarifications in the paper to avoid confusion/misunderstanding.
- On the Novelty of Theorem 3.2:
The paper presents only an incremental improvement on SOTA methods, and in many ways the main Theorem (3.2) is really no more than a simple Corollary from prior work with a simple affine transformation on the variables (which would usually be done anyway in dealing with non-image data with very different means and scales in each data dimension).
We respectfully clarify that the derived universal theories are not simple (by covering both strict robustness w.r.t. heterogeneous dimensions and universality), as detailed in Appendix A. Even if its conclusion might be considered as simple, as noted by Reviewer v1kJ, simple-but-effective solutions are not generally considered as ``weakness'', but would be desirable in practice.
- Clarification on Isotropic Results:
In the experiments, it seems that the results vs should be presented (per Corollary 3.3) for a fair comparison, i.e. it is not clear to me that the proposed technique certifies a strictly larger ball in the same conditions than SOTA (and what “same conditions” may mean is not clear e.g. it may mean isotropic and in anisotropic case). That said, the results in section 5.3 apparently include a certified accuracy wrt radius. It is not clear if this may be an answer to my query as it is not clear exactly what is being reported here.
In our experimental results, the “certified accuracy w.r.t radius” refers to certified accuracy w.r.t. , which equals to the radius (e.g., the Gaussian case). Therefore, we did achieve a strictly larger -ball than other SOTAs. In the entire paper (including Section 5.3), all the “radius” means traditional radius for the ball.
- Input-dependent Certification Validity:
Re section 4.3, I am a bit confused. This input-dependent proposal would give an x-dependent sigma, and hence an x-dependent classifier (ie x-dependent noise, over and above the obvious dependence on x). Theorem 3.2 holds assuming the noise is constant in the ball around x. This proposal in 4.3 violates that surely. Hence it is no longer true that Theorem 3.2 guarantees that the classifier gives an unchanged output in the region claimed, and so the proposed classifier is not certified robust in the claimed region around x.
Regarding the input-dependent certification, we noticed the discussion in Eiras et al. (2021). We respectfully clarify that our input-dependent noise method is still sound and does not breach certification principles.
The -dependent classifier is fixed in the certification on , that means with and , we are guaranteeing as described in Theorem 3.2 (note that noise is pre-computed and are actually constants). We do not change the -dependent classifier during the certification, thus we achieve consistent prediction over such a specific classifier. This does not violate the certification of randomized smoothing, instead, it sticks to the inherent characteristics of the RS-based certification, i.e., input-dependence (The RS certification is inherently input-dependent since the radius only works on the specific input and a specific classifier). Please also see the detailed explanation in our ``Common Concerns'' response.
- Responses to Minor Comments:
-
Binary Case: We appreciate this observation and have revised the caption of Table 1 accordingly.
-
Metrics Clarification: The metric named is referred to as ALM throughout the paper. We have amended the symbols to ALM for consistency and clarity.
-
Performance at Small Radii: The reason SOTA methods perform better at smaller radii is due to the noise variance-prediction accuracy trade-off. Our method, with single , competes with SOTA methods at various settings. Notably, our Input-dependent method outperforms these trade-offs even with a single setting.
-
Cohen's Tight Radius: We have revised the statement about Cohen's method to reflect its application more accurately.
Thank you for your responses. In line with reviewer v1kj I still am not convinced of the validity of the certification in the input-dependent case and maintain my score.
The authors interpret certified robustness from an anisotropic lens, with the aim of assessing how the performance of certification mechanisms within this context.
优点
Well written, comprehensive experiments that match community expectations, nice visualisations that really break apart the differences between fixed pattern, universal, and input dependent noise.
The input-dependent component of the noise is interesting.
缺点
My issues with this paper stem from two different directions, which I find to be significant hurdles to my ability to recommend this paper for publication.
The first is the fact that the contribution in constructing the anistropic noise measures (the core conceit of the paper) is essentially just a basic modification to extant techniques, with no other modifications being made. However, this is not in and of itself a reason for rejection - simple modifications can lead to impactful contributions.
My primary concern relates to the alignment of the chosen area of investigation to the broader problem space. Specifically, I do not think that there is any framework (either in the literature or suggested in the framework) that would case about the area of the region of certification. From the certifiers perspective, what information about the security of a model is gained by knowledge of the Lebesgue measure of the noise region (or any other measure of the area of certification)? The primary measure of risk is the nearest extant adversarial example - this is well established in the literature as a measure of the adversarial risk (see Gilmer's "Motivating the Rules of the Game for Adversarial Example Research", 2018), because this measures the effort required for an adversary to identify an adversarial example. Any other risk measure would need to be well justified and well posed, and this is not the case within this work.
Given that the certified distances to the nearest possible adversarial example are unchanged by your work, I do not see how this leads to an improved understanding of adversarial risk. I would argue that rather than significantly improving upon SOTA, you're introducing a new metric to mask the fact that you do not appear to produce any level of outperformance.
For a few minor issues: -The use of lambda as the scale parameter, and as the modification of the scale parameter. But is typically proportional to the standard deviation, so using as part of the notation is not as clear as it could be.
- A secondary minor issue is that Table 1, I believe the Lee et. al PDF should be proportional to , rather than .
- The idea of including a headline figure of an 182.6% improvement over SOTA - anyone reading this who was not familiar with this field would assume that this would be an apples-to-apples comparison, but it's not. There's no SOTA for certification that cares about area driven measures of certification, and so claiming a comparison to these prior techniques is not reasonable or well justified.
问题
Is there any justification for using the Lebesgue Measure as a proxy of adversarial risk?
伦理问题详情
n/a
Thank you for the insightful comments and please find our clarifications as below. We also make more clarifications in the paper to avoid confusion/misunderstanding.
- On the Contribution and Novelty:
The first is the fact that the contribution in constructing the anistropic noise measures (the core conceit of the paper) is essentially just a basic modification to extant techniques, with no other modifications being made.
We would like to respectfully clarify that our core contribution is not on the construction of the ALM measure but on the theoretical analysis of the robustness bound when injecting anisotropic noise, as well as the framework and novel methods for customizing anisotropic noise. ALM is one of our metrics for evaluating the certified robustness with anisotropic noise while we also derive the traditional radii (Corollary 3.3) for the robustness region with anisotropic noise.
My primary concern relates to the alignment of the chosen area of investigation to the broader problem space. Specifically, I do not think that there is any framework (either in the literature or suggested in the framework) that would care about the area of the region of certification.
The certified radius can be seen as a measure for the area of the regular and symmetric robustness region within the overall robustness region (sub-region in shape). On the other hand, ALM provides a full capture of the overall robustness region (a comprehensive understanding of the theoretical guarantees and boundaries of the certified defense). These two metrics complement each other, and we have provided the results for both radius and ALM as well as performed evaluations on both of them in the original submission.
From the certifiers perspective, what information about the security of a model is gained by knowledge of the Lebesgue measure of the noise region (or any other measure of the area of certification)? The primary measure of risk is the nearest extant adversarial example - this is well established in the literature as a measure of the adversarial risk (see Gilmer's "Motivating the Rules of the Game for Adversarial Example Research", 2018), because this measures the effort required for an adversary to identify an adversarial example. Any other risk measure would need to be well justified and well posed, and this is not the case within this work.
Thanks for the insightful comment. We know that the certified radius provides the knowledge that any perturbation within the distance will not succeed. In practice, the defense may tolerate the perturbation of a larger distance in some dimensions (see Figure 1 (b)), how to measure this gain of robustness would be a problem. Therefore, we propose the complementary ALM metric to measure the full area of the (certified) robustness region.
As you stated, researchers may only care about the nearest extant adversarial example (AE), however, this AE indeed only measures the minimum effort required for finding an AE, while the ALM can be seen as a measure for showing the average effort required for finding an AE. These two metrics just reflect two different aspects of the certified defense, so we provide both the nearest measure ( radius in Corollary 3.3) and the average measure (ALM). Our method significantly improves the certification performance under both measures.
Given that the certified distances to the nearest possible adversarial example are unchanged by your work, I do not see how this leads to an improved understanding of adversarial risk. I would argue that rather than significantly improving upon SOTA, you're introducing a new metric to mask the fact that you do not appear to produce any level of outperformance.
As discussed before, our method drastically improves the robustness against both the certified distance to the nearest possible AE (please see ``certified accuracy w.r.t. radius'' in Table 2-4 in our original submission) and the average distance measure (ALM).
The idea of including a headline figure of an 182.6% improvement over SOTA - anyone reading this who was not familiar with this field would assume that this would be an apples-to-apples comparison, but it's not. There's no SOTA for certification that cares about area driven measures of certification, and so claiming a comparison to these prior techniques is not reasonable or well justified.
The 182.6% gain over SOTA is under the evaluation of a standard certified radius (derived in Corollary 3.3, and measured via the ``certified accuracy w.r.t. radius'' in Table 2-4), which is an apples-to-apples comparison.
- Addressing Minor Technical Issues:
Thanks for your observations regarding the minor issues. We have revised the notation to $||z||_\infty$.
Thank you for response. I appreciate the effort placed into it, and have been following your discussions with the other reviewers closely. While you have resolved some of the issues I had originally held with the paper, the broader discussions around this paper do not leave me confident that changing my review would be justified, although I will keep an eye on this over the remainder of the review period.
There was one point in your rebuttal response that I wanted to directly address though. While I can see why you might argue that the ALM is a a measure of the average effort required for finding an AE, I think this is the kind of statement that would require significantly more evidence to establish - especially when adversarial examples often exist at a distance that can be well over an order of magnitude larger than the certified radii themselves (as shown in another ICLR paper under review). I would believe that the ALM only would measure the average effort required to attack a sample if the certifications were tight (which they aren't), or if the search algorithm was very naive.
Thanks for the response and the attention to our discussion. We have summarized the discussion with reviewer v1kJ for clarity and further clarified the concern about input-dependent noise in response Clarification on the input-dependent noise.
We understand the concern about the guarantee of ALM and would like to provide more information to clarify. The ALM is a measure of the volume of a high-dimensional space, although we don't have the tight certification w.r.t. the ALM measure, the high-dimensional space it measures, i.e., , has the tight certification as in Theorem 3.2. Therefore, we admit that the ALM is proportional to the average effort, but not an accurate measure of the exact average effort. When we are evaluating the certified robustness under the same conditions, ALM can serve as an auxiliary metric only for comparison.
This paper proposes to shift and re-scale the noise in randomized smoothing in order to generate anisotropic robustness guarantees. The approach is universal in the sense that any randomized smoothing-based method can be transformed into a model with such anisotropic guarantees. Experiments on benchmark image classification datasets demonstrate increased certified accuracy curves compared to past works.
优点
- The paper is very easy to read.
- The approach is simple to understand, and the resulting theoretical guarantee follows from past robustness guarantees in a very straightforward manner. This simplicity should be considered a strength of the method, not a weakness of the paper.
- The experiments are thorough, with comparisons to a wide range of prior methods and on high-dimensional image datasets (e.g., ImageNet).
缺点
See "Questions" section below.
问题
- "However, its theory is based on assumptions and the universality is relatively limited." What assumptions? Please at least briefly mention them and why they are stringent.
- Definition 3.1 does not really appear to be a "definition" in mathematical terms. It looks more like you are re-stating the certified radius theorems for general distributions and norms. So, this should probably be labeled as a "theorem".
- Please move the definition/review of alternative Lebesgue measure to Section 3.1, where Table 1 appears with Alt. Lebesgue Measure as a column.
- In Section 4.2, the loss function is solely a function of the NPG parameters , correct? If so, it would be good to explicitly write to emphasize to the reader what you are optimizing over. Furthermore, and would be functions of this parameter , right? If so, it would also be good to write and in the smoothing loss expression.
- MOST IMPORTANT PROBLEM: Your highest performing approach, using input-dependent anisotropic smoothing parameters that are optimized per-input, breaks the robustness certificates. Namely, randomized smoothing robustness certificates intimately rely on the same model being used to predict at the nominal point and all perturbed versions in the certified ball around . However, if you optimize at , then the smoothing-based certificate only says that will yield the same prediction if you also use the same parameters to define the prediction at . But, according to your scheme, you actually re-optimize at the perturbed test input to generate the prediction at , meaning you are using a different model than what smoothing certifies at . This mathematical breakdown of certified robustness for input-dependent smoothing has been noted before in past works, and is the reason why works like Eiras et al. (2022) augment their input-dependent scheme with "memory." In order for your input-dependent smoothing scheme to work, you would also need to appeal to some "fix" like this memory method, which comes with its own issues (e.g., relating to dependency on input order, and increased memory overhead costs). Either you should fix this issue (and hopefully your certificates still provide substantial improvement over state-of-the-art), or you should remove this input-dependent part of the paper (which, in my opinion, would significantly reduce the contributions of the paper).
Thank you for the insightful comments and please find our clarifications as below. We also make more clarifications in the paper to avoid confusion/misunderstanding.
- On the Presenting of Related Works:
"However, its theory is based on assumptions and the universality is relatively limited." What assumptions? Please at least briefly mention them and why they are stringent.
Our approach does not need the assumptions in Eiras et al., where the classifier's -Lipschitz continuity is a prerequisite. This assumption imposes limitations on the universality of their theorem. Conversely, our method is designed to work universally with any classifier, thereby offering broader applicability.
2 & 3. Regarding the Paper Revisions:
Definition 3.1 does not really appear to be a "definition" in mathematical terms. It looks more like you are re-stating the certified radius theorems for general distributions and norms. So, this should probably be labeled as a "theorem". Please move the definition/review of alternative Lebesgue measure to Section 3.1, where Table 1 appears with Alt. Lebesgue Measure as a column.
We appreciate your suggestions for improving the paper's structure and clarity. We have revised the paper accordingly, particularly in relabeling Definition 3.1 as a theorem and relocating the alternative Lebesgue measure definition to Section 3.1, as suggested.
- Clarifications in Section 4.2:
In Section 4.2, the loss function is solely a function of the NPG parameters , correct? If so, it would be good to explicitly write to emphasize to the reader what you are optimizing over. Furthermore, and would be functions of this parameter , right? If so, it would also be good to write and in the smoothing loss expression.
We have updated our paper to clarify that the loss function is optimized over both and . Additionally, we have made it explicit that and are functions of the parameter . These revisions should enhance the clarity of our methodology.
- On the Validity of Input-Dependent Randomized Smoothing:
Your highest performing approach, using input-dependent anisotropic smoothing parameters that are optimized per-input, breaks the robustness certificates.
We noticed the discussion in Eiras et al. (2022) and understand your concerns regarding our input-dependent Randomized Smoothing (RS) approach. However, we respectfully clarify the soundness of this method.
Namely, randomized smoothing robustness certificates intimately rely on the same model being used to predict at the nominal point and all perturbed versions in the certified ball around .However, if you optimize at , then the smoothing-based certificate only says that will yield the same prediction if you also use the same parameters to define the prediction at . But, according to your scheme, you actually re-optimize at the perturbed test input to generate the prediction at , meaning you are using a different model than what smoothing certifies at . This mathematical breakdown of certified robustness for input-dependent smoothing has been noted before in past works, and is the reason why works like Eiras et al. (2022) augment their input-dependent scheme with "memory."
In this case, with and where are pre-computed (independent of ) as constants, we are guaranteeing as described in Theorem 3.2. Then, the models used for predicting and are the same in the certification, thus do not break the certification. The entire certification process has nothing to do with and the noise optimization is not based on . Please also see the detailed explanation in our Common Concerns response.
- Notice that Eiras does not assume that the base classifier is Lipschitz, so their method is "designed to work universally with any classifier" as well. They prove that their smoothed classifiers are Lipschitz with respect to the anisotropic and norms, no matter what the base classifier is. This is a classical property of randomized smoothing: convolving a general function with an adequately smooth probability density function or "kernel" gives rise to a new function that inherits smoothness from the kernel. Their theoretical robustness certificate holds very generally for Lipschitz classifiers, of which smoothed classifiers via randomized smoothing are a special case. Therefore, your assertion that their "theory is based on assumptions that the networks are L-Lipschitz continuous and thus the universality is relatively limited" is not well-justified (since it is implying that they assume the base classifiers are L-Lipschitz, which they do not), and I suggest removing it.
2,3,4. Thank you for the revisions.
- I have read both your "Common Concerns" response, as well as your individual response to me. I am still not convinced that your robustness guarantee holds. Unless I'm misunderstanding something, then, at test time, the attacked input going to your model is . Therefore, your NPG will output and as the noise parameters. These will in general define a different smoothing distribution than that corresponding to the clean test input , which has parameters and output by the NPG. Therefore, the certification condition, as you call it, would need to be that , which is not what is guaranteed by the randomized smoothing framework. In other words, the actual output of your model at an attacked input is , and your certificates do not guarantee that this output coincides with the output generated by the clean input .
Given that my primary concern (shared with Reviewer SMHf) still remains, I maintain my original score.
Thanks for the response. Also, thanks for the clarification on Eiras et al. We will revise the description as suggested.
Based on the response, we would like to clarify that there is a misunderstanding on the setting of the randomized smoothing.
In general randomized smoothing, you are using the classifier's prediction probability on (empirical results) to guarantee its prediction on (theoretical results), you will not empirically testing but to theoretically guarantee this prediction. To further justify this, (1) if the prediction results on attacked/perturbed input rely on the empirical testing on the smoothed model (as mentioned in the response "at test time, the attacked input going to your model is "), what is the meaning of guaranteeing the prediction of ? (2) there exists an unlimited set of within the radius for , how to test all of them? We don't really infer in our algorithms/implementation, we theoretically guarantee its prediction.
The reviewer seems to misunderstand that we are doing , but actually, we are guaranteeing , please check our Theorem 3.2 and see the implementation details in our code. In the implementation, we can consider the classifier and the noise generation to be independent. Given the input , we fixed the noise for the whole certification. The noise parameters are generated once at the beginning of each certification based on , and fixed during the certification to guarantee the consistent prediction result for all kinds of within R. The certification is not violated in this case.
Thanks again for the prompt reply. First of all, with respect to your two proposed situations:
Assuming that you always know the clean input is an unrealistic and very stringent assumption. If I gave you , how would you know what the clean input is? If this were the case, then all of the recent work on certifiably robust machine learning in the presence of -norm bounded additive adversaries would be pointless, since one could simply reconstruct from , and then classify the clean input (which is much less likely to induce an error). So, this first situation is meaningless to consider in your setting of adversarial threats. Thus, the standard setting for papers in this area (and randomized smoothing, in particular), is to assume that all that you have at your disposal is whatever input you are given ( in the case of an attacked input), and you cannot distinguish whether this was attacked or not. In this standard setting, there is no possible way for you to take the attacked as an input, and compute for the clean input.
So, there is an essential question to consider that points to the flaw at hand: What does your "certificate" represent? To me, this says that, if you predict the output of using the smoothing distribution for , then the prediction will match that of . The problem is twofold: 1) you cannot compute (or even samples drawn from this distribution) unless you have access to the clean input , which you do not, and 2) is not even the prediction that your scheme would output if was treated as its own input, since the smoothing distribution would be re-optimized for the input .
Thanks for the interesting discussion.
It seems there is confusion between the setting of certified defense and the empirical defense. Let's assume that all we have at our disposal is whatever input we are given, then given arbitrary by an adversary, what information about robustness does randomized smoothing provide? The model owner can probably only provide the prediction without any guarantee, and this prediction can be either correct or incorrect, this is not a certified defense but an empirical defense. This example indicates that the certified robustness in randomized smoothing is a theoretical consistency of the prediction between two conceptual inputs. When it comes to reality, in your case, the adversary gives , then the model owner (in general RS) cannot guarantee anything to the prediction of since he/she doesn't know . The more realistic setting is the model owner holds a , and then computes a certified radius on it, then given any input by the adversary the model owner can compute and compare to to pre-determine the prediction without execution.
Therefore, in our setting, we hold and compute and to certify any potential to be correct or not on the smoothed classifier we set up. When we got , we compute the noise and to certify any potential on the smoothed classifier , the smoothed classifier is fixed for the input that we want to protect.
"Let's assume that all we have at our disposal is whatever input we are given, then given arbitrary by an adversary, what information about robustness does randomized smoothing provide?"
- Conventional randomized smoothing, with a smoothing distribution that is uniform over the entire input space, DOES give you that the classification of , namely, , is the same as that of the certified clean input , even if you don't know , so long as with being the certified radius. This result intimately relies on the fact that is computing using the same smoothing distribution as is computed.
"The more realistic setting is the model owner holds a , and then computes a certified radius on it, then given any input by the adversary the model owner can compute and compare to to pre-determine the prediction without execution"
- This is not how adversarial threat models are usually formulated. Can you point to a paper that assumes that they have access to the clean input at the time when they are predicting the class of an attacked input ? The two closest things that come to mind are [1] and [2], which use memory-based techniques for input-dependent randomized smoothing, in particular as a method to overcome the exact issue that I am saying your approach suffers from. However, you do not make clear in your paper that you are using such a memory-based approach, and furthermore those memory-based approaches have their own limitations. For example, the classification of an input may in general be dependent on the order of inputs that were previously classified+certified, and furthermore you incur a memory cost that continually grows as you predict+certify more and more inputs.
[1] "ANCER: Anisotropic Certification via Sample-wise Volume Maximization", Eiras et al.
[2] "Data Dependent Randomized Smoothing", Alfarra et al.
Are you generating your "certificates" for the inputs in the training set, or the test set?
We evaluate our method in the test set, please see the Experimental Settings in Section 5 in our original submission as follows:
"Following Cohen et al.(2019) on certification, we obtain the certified accuracy on the entire test set in CIFAR10 and MNIST while randomly picking 500 samples in the test set of ImageNet"
Thank you for the further discussion. I am still not convinced by your arguments, and therefore I maintain my overall score.
I encourage the authors to carefully read [1] (and in particular, Section 3.5) for more information on why robustness certificates for input-dependent randomized smoothing fail to hold without some careful modifications (like a memory-based approach), and to take these technicalities (and their associated limitations) into account in their future iterations of the project.
[1] Alfarra et al., "Data-Dependent Randomized Smoothing," UAI, 2022.
Thanks for the response.
We have carefully read through Alfarra et al. [1] and Eiras et al. [2] and found these papers' technique (memory-based certification) may not be necessary for our paper. The reasons are as follows:
-
The setting of our paper is different from these papers (i.e., input-dependent noise vs. input-dependent classifier). Please refer to the detailed explanation in response Clarification on the input-dependent noise.
-
As [1] and [2] point out, the memory-based certification wants to address the concern of the region overlapping, but it seems this concern does not exist in reality, evidenced by:
"While the memory-based certification is essential for a sound certification, empirically, we never found in any of the later experiments a case where two inputs predicted differently suffer from intersecting certified regions. That is to say while our sound certificate works on the memory enhanced data dependent smooth classifier, we found that the certified radius of the memory classifier for every input is the radius granted by the Monte Carlo certificates of Cohen et al. [2019] for the data dependent classifier. " in [1]
"Given the high dimensionality of the data, empirically, we never found a certificate in this situation within our experiments." in [2]
- We also followed the setting of [2] to evaluate our input-dependent method on CIFAR10 with memory-based certification and found none of the cases suffered from the intersection region problem. Therefore, our experimental results are still the same and valid even under the memory-based certification.
[1] M. Alfarra, A. Bibi, P.H. Torr, and B. Ghanem. Data dependent randomized smoothing. UAI, 2022.
[2] F. Eiras, M. Alfarra, P.H. Torr, M.P. Kumar, P.K. Dokania, B. Ghanem, and A. Bibi. ANCER: Anisotropic Certification via Sample-wise Volume Maximization. TMLR, 2022.
Thanks for the insightful discussion again.
DOES give you that the classification of , namely, g(x'), is the same as that of the certified clean input , even if you don't know , so long as with being the certified radius.
There may be a conflict in this response, given the fact that the is the certified radius computed with , if you don't know , how do you compute ? Note that is not a constant that works for all the inputs on a classifier (like L-Lipschitz continuous classifiers [1]). is an input-depend radius that only works for .
Our theorem is based on the formal definition of Randomized Smoothing (Cohen et al.[2]) different from previous conventional certified robustness. To make it more clear, let's look at how Cohen et al. defines the smoothed classifier and the certified robustness:
Randomized smoothing is a method for constructing a new, “smoothed” classifier from an arbitrary base classifier .
This indicates that the smoothed classifier is constructed by the model owner, so we can construct our own smoothed classifier as long as we can guarantee its robustness. For example, Cohen et al. construct different smoothed classifiers with , , and , these smoothed classifiers may have different predictions on the same input but maintains the certified robustness as long as they don't change the noise when guaranteeing with .
Let's further look at the Cohen et al.'s definition of certified robustness of RS:
Suppose that when the base classifier classifies , the most probable class is returned with probability , and the "runner-up" class is returned with probability . Our main result is that the smoothed classifier is robust around within the radius .
In our case, we can similarly state:
Suppose that when the base classifier classifies , the most probable class is returned with probability , and the "runner-up" class is returned with probability . Our main results is that the smoothed classifier where is robust around within the radius .
Note that, the smoothed classifier is based on the same noise that is computed when computing and and will not change within .
[1] "ANCER: Anisotropic Certification via Sample-wise Volume Maximization", Eiras et al.
[2] "Certified adversarial robustness via randomized smoothing", Cohen et al.
Thank you for your response. I believe you are still misunderstanding my point. Let me try to make it more clear:
I agree with you that, mathematically, the equality holds. However, my main point is that this equality does not reflect how your classifier makes its predictions in reality, and thus provides no meaningful robustness guarantee.
To see why this is the case, let be some fixed input. Your scheme then chooses a mean and covariance based on this . Then, to form the prediction for the input , you compute , where is some isotropic "base distribution," such as an isotropic Gaussian. Now, your equality says that for all perturbations inside of some region of the input space containing . This does not imply that your classification scheme's prediction is constant over this region.
Specifically, consider a NEW input (possibly different from , and therefore I put the "prime") for which you are to make a prediction. Well, substituting for in the above prediction rule (which we both have agreed upon) gives that the prediction is . In other words, to form a prediction for the input , you again compute the input-dependent mean and covariance, and perform the RS-based prediction at using its optimized and .
Now, suppose that I reveal to you that actually, I was an adversary, and the input that I gave to you and asked you to classify was for some perturbation that I chose within your region . Then your prediction for this perturbed input was computed as , which is not guaranteed to be equal to the prediction corresponding to . Therefore, the equality does not say anything about the prediction/output of an attacked input .
Thanks for the response.
If we are understanding correctly, your comments are based on an assumption that the noise and are changing according to the adversarial input during the certification. This is not the setting in our original submission, and we want to clarify more on this point.
Given the input that we want to certify, we compute the noise parameter and as constants and then construct the smoothed classifier with these fixed noises. During the certification on , no matter what is, the smoothed classifier is fixed on and , until the input that we want to certify changes. In other words, you can consider that we are constructing different smoothed classifiers for certifying different input .
Let's further take your example to illustrate how this can be achieved. Suppose you were the adversary, and we were the model owner. You gave me and asked me to classify for some perturbation within . There are two situations here:
-
suppose we both know and the corresponding , that means we are based on the knowledge of previous certification to classify the , then we can definitely compute and (since we know ) and predict or simply compare and .
-
suppose we both don't know and , then both Cohen et al. and our method cannot provide any guarantee on this prediction since the guarantee of the prediction on is based on the certification on . Without knowing , the radius cannot be computed.
We believe both Cohen et al. and our method are in situation 1). Thanks for the response and we are open to and welcome any further questions related to these points.
This paper aims at improving certified robustness by randomized smoothing with anisotropic noise. The universal theory for certification with anisotropic noise has been provided. The authors consider three kinds of customizing anisotropic noises, and provide corresponding noise generation methods. The authors conduct experiment to demonstrate that the proposed UCAN method achieve state-of-the-art performance compared to existing randomized smoothing-based methods for certified robustness.
优点
- The proposed method on smoothing with anisotropic noise is novel. It is interesting to see the expansion of RS-based methods from isotropic noises to anisotropic ones.
- This paper provides the theoretical guarantee of certified robustness under anisotropic randomized smoothing, and comprehensive analyses to transform existing randomized smoothing methods to anisotropic cases.
- Authors consider three different kinds of anisotropic noises, and provide a novel input-dependent one by optimizing and by a multi-layer neural network.
缺点
- My major concern of this paper is the potential unfairness in evaluation on UCAN and existing RS-based methods. If I am not misunderstanding, the evaluation criterion is based on scaled radius, which has different weight in each dimension.
I believe this is true because I am surprised to the evaluation results provided in Table 3 that the certification on CIFAR-10 reach over 70% even under radius 1.75. By simple calculation, for commonly-used norm with budget 8/255, it achieves at most , that means in your UCAN it achieves at least 70% robust accuracy for 8/255 attacks. This result is unbelievable, because existing SOTA performance of CIFAR-10 robustness may only achieve about 60% if no further data augmentations are conducted (like diffusion model), let alone UCAN is only a certified method based on norm. Therefore, although the paper said they evaluate certified accuracy w.r.t radius, I am doubtful of this claim and I think the authors only consider scaled radius robustness.
However, scaled radius certification seems not a fair criterion for certified robustness. It is reasonable that in some dimensions, the image is vulnerable to adversarial attacks, e.g., contour of an image. Reversely, in some dimension images are intrinsically robust to perturbations like background of the image. Therefore, I believe the corresponding variance is small when UCAN performs on these vulnerable dimensions, and gain robustness back in some ``unimportant’’ dimensions.
Overall, the evaluation setting of this paper seems differently from existing RS methods. It is a consensus that using norm as constraint for images, the authors should provide corresponding evaluation on standard norm, or at least, provide the explanations or practical scenario on why using scaled radius as the evaluation criterion.
-
In Theorem 3.2, your certification using the p-norm of but it seems that , is a one-dimension scalar as is the i-th dimension of perturbation . Furthermore, this theorem is seemingly a direct corollary from Theorem 3.1, because your certification divides the variance for each (not anisotropic anymore?).
-
There might be missing of some baselines for [a] and [b, c] certified robustness. It will be better to compare the UCAN with existing certified and methods.
[a] Levine et al. Improved, deterministic smoothing for L_1 certified robustness. In ICML 2021. [b] Zhang et al. Towards Certifying L_∞ Robustness using Neural Networks with L_∞-dist Neurons. In ICML 2021. [c] Zhang et al. Boosting the Certified Robustness of L-infinity Distance Nets. In ICLR 2022.
问题
- Why using 5-layers NN when generating universal/ input-dependent anisotropic noises? Is there some motivations or ablation studies for that?
- Could you provide more details on training of universal anisotropic noise? It seems that the variance loss is to optimize and smoothing loss containing when optimizing classifier . I believe the two losses are optimized alternately but not simultaneously.
- The authors said that randomized smoothing achieved great success for certified adversarial robustness. Could RS really make classifier robust? Can you provide comparison of RS based model to the SOTA methods for achieving robustness?
- In Response to the Theoretical Concerns:
In Theorem 3.2, your certification using the p-norm of but it seems that , is a one-dimension scalar as is the i-th dimension of perturbation .
The Hadamard division of over () is formally defined in Theorem 3.2 and Appendix A.
Furthermore, this theorem is seemingly a direct corollary from Theorem 3.1, because your certification divides the variance for each (not anisotropic anymore?).
We ensure that Theorem 3.2, while appearing to be straightforward, is a unique universal transformation from Definition 3.1, as detailed in Appendix A. We respectfully clarify that the derived theories in the theorem are not a direct corollary from Theorem 3.1 (since it covers both strict robustness w.r.t. heterogeneous dimension and universality). The guarantee in Theorem 3.2 is anisotropic w.r.t. , although our focus is not to derive an ``anisotropic radius'', but to extend isotropic noise to anisotropic noise for RS. The guarantee on the perturbation can be either anisotropic (Theorem 3.2) or isotropic (Corollary 3.3).
- As per the Suggestion to Compare with and Methods:
There might be missing of some baselines for [a] and [b, c] certified robustness. It will be better to compare the UCAN with existing certified and methods.
Our method, when benchmarked against and methods, shows significant improvements (see tables provided). This underscores our method's effectiveness across different metrics. We respectfully clarify that RS-based methods and -distance neural network-based methods may not be directly comparable (due to their differences, e.g., RS can certify any classifier). To further address this concern, we still compare with them, and our method also demonstrates superior performance in these scenarios (a huge improvement space brought by the anisotropic noise for RS).
| Methods | 0.0 | 0.5 | 1.0 | 1.5 | 2.0 |
|---|---|---|---|---|---|
| Yang et al. 2020 | 83% | 43% | 22% | 14% | 7% |
| Levine et al. 2021 | 79% | 71% | 61% | 54% | 49% |
| Ours ( radius) | 85% | 81% | 77% | 73% | 68% |
Table 1. Certified accuracy vs. perturbation (CIFAR10)
| Methods | 0/255 | 1/255 | 2/255 | 3/255 | 4/255 | 5/255 | 6/255 | 7/255 | 8/255 |
|---|---|---|---|---|---|---|---|---|---|
| Yang et al. 2020 | 83% | 63% | 48% | 36% | 27% | 20% | 16% | 13% | 10% |
| Zhang et al. 2021 | 51% | -- | -- | -- | -- | -- | -- | -- | 35% |
| Zhang et al. 2022 | 61% | -- | 54% | -- | -- | -- | -- | -- | 40% |
| Ours ( radius) | 85% | 83% | 82% | 80% | 78% | 77% | 75% | 73% | 70% |
Table 2. Certified accuracy vs. perturbation (CIFAR10)
- Responses to Specific Questions:
Why using 5-layers NN when generating universal/ input-dependent anisotropic noises? Is there some motivations or ablation studies for that?
The 5-layer MLP architecture in our Universal method is inspired by GANs (Goodfellow et al., Communications 2020), and the CNN structure in our Input-dependent method is designed per the dense blocks (Huang et al., CVPR 2017).
Could you provide more details on training of universal anisotropic noise? It seems that the variance loss is to optimize and smoothing loss containing when optimizing classifier . I believe the two losses are optimized alternately but not simultaneously.
For the training of NRG and the classifier, we employed simultaneous training, optimizing both for predictive accuracy and noise pattern enhancement. While alternate training is possible, our current approach promotes synergy between the classifier and NRG.
The authors said that randomized smoothing achieved great success for certified adversarial robustness. Could RS really make classifier robust? Can you provide comparison of RS based model to the SOTA methods for achieving robustness?
Please see the above experimental results and discussion. It is worth noting that RS methods can be universally applied to any classifier with different scales, and it is unfair to compare the robustness performance of different types of classifiers.
Thanks for the comments again. We hope these clarifications can address your concerns.
Thank you for the insightful comments and please find our clarifications as below. We also make more clarifications in the paper to avoid confusion/misunderstanding.
- Regarding the Evaluation Concerns Raised:
If I am not misunderstanding, the evaluation criterion is based on scaled radius, which has different weight in each dimension.
Our experimental evaluations are based on both 1) the radius from Corollary 3.3 (), which is a tight bound for the perturbation in anisotropic RS, and has the same weight in each dimension (please see Table 2-4 in the original submission); 2) the ALM metric (), which is a measure on the volume of the irregular asymmetric robust region, and has different weights in each dimension.
I believe this is true because I am surprised to the evaluation results provided in Table 3 that the certification on CIFAR-10 reach over % even under radius 1.75.
In Table 3, the 70% certified accuracy at on CIFAR10 is the performance based on the radius (), which is strictly derived in Corollary 3.3. This validates the significant improvement of RS with anisotropic noise over SOTA methods.
This result is unbelievable, because existing SOTA performance of CIFAR-10 robustness may only achieve about 60% if no further data augmentations are conducted (like diffusion model)
We agree with the reviewer that existing isotropic RS methods can hardly improve their performance to this level. However, we focus on a new dimension of RS (with anisotropic noise) that is orthogonal to existing RS methods, which can boost all the RS methods. Such results are based on radius, rather than scaled radius.
Therefore, although the paper said they evaluate certified accuracy w.r.t radius, I am doubtful of this claim and I think the authors only consider scaled radius robustness.
We understand the reviewer's concern as our results drastically outperform the SOTA methods (evaluated on both radius and ALM). To address the concern, we are providing our code in the supplemental material for transparency. Please see line 180-182 in certification_personalized.py for the implementation of our radius (and corresponding evaluations).
However, scaled radius certification seems not a fair criterion for certified robustness. It is reasonable that in some dimensions, the image is vulnerable to adversarial attacks, e.g., contour of an image. Reversely, in some dimension images are intrinsically robust to perturbations like background of the image. Overall, the evaluation setting of this paper seems differently from existing RS methods. It is a consensus that using norm as constraint for images, the authors should provide corresponding evaluation on standard norm, or at least, provide the explanations or practical scenario on why using scaled radius as the evaluation criterion.
We agree with the reviewer that the ALM (scaled radius) may result in some ``weak'' dimension for the adversary, which only provides the certified robustness in a specific shape (depends on ).
While working on the paper earlier, we also understood that reviewers may concern that solely evaluating on the ALM might be unfair (though ALM is defined as a full capture of the robustness region). Thus, we have also derived and presented the radii (derived in Corollary 3.3), and evaluated the performance of our method and SOTA using the ``certified accuracy w.r.t. radius'' (please see Table 2-4) in our original submission. We show that by customizing appropriate anisotropic noise, our method achieves significantly boosted performance on both the standard radius and the new ALM metric.
Thanks for your responses.
Although a part of my concerns has been solved, there are still some issues in the current version of this paper.
First, I am still doubtful of the soundness of your experimental results. Specifically, as shown in your Tables 2 and 3, the certified accuracy w.r.t. radius is greater than whose w.r.t. ALM is confusing. The guarantee w.r.t. radius should always less than those w.r.t. ALM because .
Furthermore, as reviewer SMHf and review v1KJ pointed out, the proposed certification may be not true. You said that what you guarantee is rather than . However, what is your (smoothed) classifier? I believe your classifier is , then the guarantee should be hold on for all .
Thanks for the response.
- We would like to clarify the settings in Table 2-4 further to address your concern. The smoothed classifiers for evaluating the "radius" and the "ALM" are trained with different variance loss term. For "ALM" we train the smoothed classifier with as the variance loss to maximize the , and for "radius", we train the smoothed classifier with to maximize the for larger radius. Please see the line 111-114 in the training code in "train_personalized_noise.py" as follows:
if args.IsoMeasure:
loss_variance -= torch.min(torch.abs(variance)) # this is for L_2 radius measure
else:
loss_variance-=torch.mean(torch.abs(variance))
In this case, the generated for the "radius" and the "ALM" are different, so it is sound that in some cases since the and are different. We will revise the paper to make this setting more clear.
- The noise for the smoothed classifier is fixed after generating according to each input, which means our smoothed classifier is unique for each input. To be more clear, let's look at one example: given x, we generate the noise parameters and as constant, and then construct the smoothed classifier , the guarantee is held on where and are constant but unique for each .
One question that may raised here is: Does it make sense that we have unique smoothed classifier , , ..., for different inputs compared with the traditional setting that we only have one for all the inputs? To answer this question, we should start with the definition of randomized smoothing. In the original work (Cohen et al.), the smoothed classifier is manually crafted by the model owner with his/her noise setting, which means the model owner guarantees the robustness of his/her own classifier defined by himself/herself (same as in our case). In addition, since the certified robustness (in RS) is tied to the specific input (radius only works on specific ), using -dependent smoothed classifier does not bring extra dependence for the certified robustness (still depends on ).
We trust that these clarifications adequately address your concerns. We are open to and welcome any further questions related to these points.
We appreciate all the reviewers' insightful comments. First, we would like to respectfully clarify some critical misunderstandings on this work. Then, more detailed explanation will be provided in the response to each comment.
- Missing Standard Metric (Reviewer mJxe, and eeVj)
Thanks for this comment. In the original submission, we have derived the isotropic certified radii (aka. traditional radius) in Corollary 3.3, and provided the evaluation on the radius. For instance, the ``certified accuracy w.r.t. radius'' in Table 2-4 refers to the evaluation results based on the traditional radius. It is worth noting that our method on the traditional radius can perform even better than on the ALM, e.g., % certified accuracy at on CIFAR10, and % improvement over SOTA methods on CIFAR10 (please see Table 3).
- The Fairness of ALM Metric (Reviewer mJxe, eeVj, and SMHf)
Thanks for this comment. The ALM metric serves as a complementary metric in addition to the traditional isotropic/ radius. Yes, existing isotropic radius, e.g., , measures the size of the robustness region in regular and symmetric shapes, which represents the sub-region in shape out of the entire certified robustness region. As a complementary metric, ALM measures the full irregular and asymmetric robustness region. We agree that it might be unfair to only use ALM as the metric. Thus, in Table 2-4 in the original submission, we have included the results for the radius to further validate the performance of randomized smoothing with anisotropic noise. Our method drastically outperforms the SOTA methods on both the ALM and the traditional radius (we mainly presented the results for certification since it is the most common setting with most related works). We add more clarifications in the updated manuscript for better clarity.
- Soundness of the Certification based on Input-dependent Noise (Reviewer v1kJ, and SMHf)
Thanks for this comment. We noticed the discussion on the soundness of the input-dependent noise in Eiras et al. (2022) earlier, but still believe our input-dependent method is sound. Specifically, let us first look at the workflow in our input-dependent randomized smoothing method and explain it with more detailed notations: for where holds.
Given the base classifier , smoothed classifier and the input for prediction, the noise (represented as and ) for input is pre-optimized based on . Then, the certification will give a radius for that is tied to , so we can guarantee the consistency for the prediction . The above guarantee strictly follows Theorem 3.2 and thus does not violate the certification (the pre-computed noises are constants when used for certification, and they are computed before/independent of the potential perturbation ). Per the definition of certified robustness, as long as the noise ensures consistency for prediction under the potential , the certification would not be violated.
We understand that the reviewer may question whether fixing the noise for and would violate our input-dependent setting in Section 4.3. Similar to the traditional RS with isotropic noise, the perturbed input is a conceptual input, and the RS algorithm/prediction will not be performed over it (and no noise will be optimized based on ). Thus, would be the certification condition for the RS with input-dependent noise while is not the certification condition in such case.
Dear Reviewers and Area Chairs,
We thank all the reviewers' efforts in the responses. During the multiple rounds of discussion with the reviewers, we addressed most of the concerns of reviewers, and the key concern that obstacles the reviewers in changing their ratings turned out to be the soundness of the input-dependent setting of the noise. Therefore, we want to summarize the discussions on this concern for clarity and further clarify the soundness of our approach.
Reviewer v1kJ has a strong opinion that our input-dependent noise method is not sound for the certification, and the core argument is that the changing noise within the certified radius will affect the soundness of the certification, which is supported by the following papers according to Reviewer v1kJ:
[1] F. Eiras, M. Alfarra, P.H. Torr, M.P. Kumar, P.K. Dokania, B. Ghanem, and A. Bibi. ANCER: Anisotropic Certification via Sample-wise Volume Maximization. TMLR, 2022.
[2] M. Alfarra, A. Bibi, P.H. Torr, and B. Ghanem. Data dependent randomized smoothing. UAI, 2022.
This statement is true in the settings of these two papers so they have to propose a memory-based certification to address this issue. However, our setting is very different from theirs and the above statement is not true under our setting.
To be more specific, after carefully reading these two papers, we found that these papers are based on the input-dependent classifier setting, where the classifier changes its noise according to the input, evidenced by:
" This is since the data dependent classifier does not enjoy a constant σ within the given certification region, i.e. tailors a new for every input including within the certified region of . " [2]
However, different from their settings, our approach is based on the input-dependent noise rather than the input-dependent classifier. This makes a significant difference in the certification process, where our noise is disentangled from the classifier, while their noise is tied to the classifier. Therefore, after computing the anisotropic noise parameters on the input that we want to certify, we can fix the noise of the classifier during the certification. Oppositely, their classifier keeps changing as the noise is changing across different within , making their certification unsound.
Therefore, since our setting is different from theirs, even if we are both input-dependent methods, our certification is sound under our setting, as illustrated in Theorem 3.2 and proved in Appendix A.
To be more objective, our input-dependent noise does not improve the robustness performance for free, the cost behind it (about which the reviewers may be curious) is the more specific condition of the robustness. In Cohen et al.'s RS, the certified radius is conditioned on and and , while in our case, the certified radius is conditioned on , , and , which is a more specific condition that may affect the generalizability of the certified radius. However, we believe the significance brought by the input-dependent noise is still valuable even at the cost of a small loss of generalizability.
We appreciate the reviewers' insightful comments (especially the insightful discussion with Reviewer v1kJ) and look forward to a fair justification of our paper. Further discussion and questions are always welcome.
Best,
Authors of Submission 6877
This paper aims at improving certified robustness by randomized smoothing with anisotropic noise. Four experts evaluated the paper and all recommended that the paper be rejected. The reviewers have highly questioned the contribution of this work and are not satisfied with the authors’ responses.
为何不给更高分
I reached this decision by evaluating the contributions and novelty of the work, taking into consideration both the reviews and the responses from the authors.
为何不给更低分
N/A
Reject
We appreciate the reviewers’ time and effort; however, we are disappointed that the conversation left some key aspects of our work misunderstood, especially the soundness of input-dependent randomized smoothing. Below, we address the main points of confusion and clarify why our approach remains theoretically valid.
1. Misunderstanding About “Knowing” vs. .
The reviewers repeatedly claimed that our method requires knowing the “clean” input even when the classifier is given an adversarially perturbed input . This is not the case. We emphasize the definition of randomized smoothing (RS):
- In certified RS, we fix a noise distribution once for the point that we wish to certify (same as in Cohen et al's). That same fixed noise distribution underlies the entire certification region.
- Our input-dependent strategy does not alter the noise parameters on any inside the certified radius. We do not re-optimize when the input is perturbed; the noise that was computed for stays fixed and guarantees consistent predictions throughout the region.
Hence, the claim “the classifier’s noise is re-optimized at ” is a misunderstanding. Once are computed for , they remain constants when certifying against any -perturbed . This is fully aligned with RS theory (e.g., Cohen et al. 2019) where the certified radius is computed upon each fixed .
2. Comparison to Memory-Based Methods.
The reviewers reference Eiras et al. and Alfarra et al. to argue that input-dependent noise is “unsound” unless a memory-based approach is used. Their settings differ:
- In those papers, the “classifier itself” is being changed per input, leading to overlapping certified regions that must be managed by memory.
- In our work, noise is tied to at certification time, but the “classifier” is not recalculated for every point in the neighborhood. We fix for each certified , satisfying the original RS assumption that the “same” smoothed model is used throughout the local region.
Since our approach never re-optimizes the noise distribution for each , we do not require memory-based solutions. Thus, the reviewer claim that “it breaks the certificates” is simply incorrect under our formalism.
3. Why The Method Is Sound.
Recall the statement in our paper:
then the smoothed classifier is guaranteed to predict class for all in the certified region. The variance (or covariance) used above is computed exactly once for and remains fixed. There is no re-optimization for . Hence, Theorem 3.2’s condition:
is sound because is defined with the same . Our “input-dependent” approach simply means different could have different , but each ’s certification still uses one .
4. Conclusion.
We stand by our theoretical derivations and the soundness of the input-dependent noise strategy. The notion that we “re-optimize” or “break” RS assumptions stems from a misunderstanding of how the smoothed classifier is fixed for each certified point . We hope that future discussions will not be misled by this confusion.
We remain open to further questions or clarifications and thank everyone for reading.
Sincerely,
The Authors