Fisher Divergence for Attribution through Stochastic Differential Equations
摘要
评审与讨论
The paper introduces a feature attribution framework for deep neural networks using Stochastic Differential Equations (SDEs) and Fisher Divergence. The method models continuous perturbations to explore input spaces and quantifies feature importance by linking Fisher Divergence to the time derivative of Kullback-Leibler (KL) Divergence.
The authors put forth different contributions, amongst which the development of an optimization framework based on the Information Bottleneck principle to identify informative features with minimal output change; the computation of Fisher Divergence and mutual information using diffusion models and time integration; and experiments showing how the method outperforms baselines, such as DeepLIFT, across various benchmarks..
给作者的问题
-
It is currently unclear to me how much does training a diffusion model weigh on the computational demands of the proposed method, and generally, how computationally demanding the method is with respect to varying model sizes or input dimensions. Can the authors explain? It would be worthwhile to provide both theoretical arguments as well as empirical evidence for this.
-
Can the authors expand on the limitations of this approach? Does the method ever underperform compared to traditional attribution methods?
-
Can the author provide a higher range of experimental evidence for their claims? (e.g. varying model architecture, model size, input dimensions, data distributions etc.). It would also be helpful to extend Figure 2 with additional examples.
论据与证据
The authors provide a good set of mathematical derivations from Fisher Divergence to the time derivative of KL Divergence for feature attribution. The few empirical results reported are also reasonable and show significant improvements over baseline methods. However, the experimental validation, in general, is quite limited, both in experimental breath and depth.
The framework’s efficiency claims lack runtime benchmarks or computational cost comparisons with simpler methods like saliency-based approaches. From what the authors present, it is unclear whether the proposed method is scalable, and can be applied to significantly larger-scale models.
方法与评估标准
The methods in the paper seem sound and tackle appropriately the feature attribution problem. The evaluations (e.g. DAUC and EHR) are reasonable and show improvements over baselines, but they are very limited in both breadth (e.g. model architecture, data distributions, data resolution etc.) and depth (e.g. model size). There is no evaluation of run-time performance or scalability of the proposed approach.
理论论述
Theorem 3.1 and Lemma 3.2. both correctly apply standard results from stochastic processes, no obvious errors were spotted.
实验设计与分析
The authors do not provide extensive methodology explanations for the experiments run, although they refer to code bases in prior work for the baselines. Generally, the metrics are sound with respect to the objectives in the paper.
补充材料
Exp. Setup, & additional qualitative results. While the additional results are useful they do not fully address concerns about scalability, computational costs, or a wider range of experiments.
与现有文献的关系
The paper builds on literature from stochastic differential equations, information theory, and diffusion models, and introduces a continuous-time perturbation framework. It also uses score-based generative models for score estimations. The novel contribution lies in the use of Fisher divergence and mutual information for more principled way to identify feature importance.
遗漏的重要参考文献
--
其他优缺点
The paper presents some interesting findings, but the work seems quite preliminary, and not yet ready for publication. The connection between Fisher divergence and mutual information is nice, and may be a promising direction for developing more principled attribution methods.
There are, however, many unanswered questions. There is minimal discussion on the sensitivity of the method to design choices, such as the specific form of the SDE or noise schedules used in diffusion processes. The authors don't discuss potential failure cases or limitations of their framework. The experimental results are very limited (see other answers) etc.
其他意见或建议
There are several minor issues such as misspellings, reverse quotes, etc.
Below we address your concerns one by one:
Experimental Designs Or Analyses: Our experiments are conducted on ImageNet (3224224). We also provide the detail optimization algorithm in the third reviewer's rebuttal section and the score network training algorithm at the end of this rebuttal.
Questions For Authors:
-
While training a diffusion model for an entire dataset can be expensive, our approach offers flexibility. We can train a diffusion model on a single image (3224224), which yields a score network that is sufficient for calculating pixel-wise mutual information for that image. In our experiments, generating an attribution map for one image takes about 4 minutes on a single GTX1080Ti. Moreover, when a pretrained diffusion model is available, the computation time is greatly reduced. Once the diffusion model is available (whether pretrained or efficiently trained on a single image), the attribution procedure itself mainly involves evaluating the score function and performing numerical integration (e.g., for mutual information). Our method employs closed-form expressions for the Variance Exploding (VE) and Variance Preserving (VP) processes, allowing efficient computation. These operations scale linearly with the input size and are highly parallelizable on modern hardware.
-
Our approach does have limitations. If a pretrained diffusion model is not available, one must train a model on a per-image basis—which, while less demanding than training on a full dataset, still requires significant computational resources (around 4 minutes per image on a GTX1080Ti). Although our method generally outperforms traditional attribution methods such as DeepLIFT, InputIBA, and IBA on metrics (as shown in our experiments and in the second reviewer's rebuttal), there are scenarios where the additional computational overhead may not justify the performance gains compared to simpler methods.
-
We acknowledge that our approach may be sensitive to the specific SDE form and noise schedule used. In the revised manuscript, we will include a sensitivity analysis to show how different choices of SDE parameters (such as varying in Equation (36)) affect the resulting attributions. We will also discuss potential failure cases and limitations of our framework to provide a balanced perspective. We appreciate the suggestion to extend Figure 2 with additional examples. We will supplement our current experimental section with further qualitative and quantitative evaluations, including more extensive visual comparisons and analyses across diverse datasets and network architectures.
Algorithm: Training Score Network for a Single Image
Input:
- : Original clean image
- : Score network with parameters
- : Maximum diffusion time
- : Number of training iterations
Function TrainScoreNetwork:
-
Initialize noise schedule:
- Define and (e.g., 0.01 and 50)
- Define noise schedule function for
-
For to :
-
Sample random time
-
Compute noise level
-
Sample noise
-
Create noisy image
-
Compute true score
-
Predicted score
-
Compute loss
-
Update using
-
-
Return
Thank you for addressing the raised concerns.
I have raised my score, contingent on the revisions to be included (i.e. the parameter ablations, and the clarifications and limitations included in the rebuttal).
The paper considers perturbation-based methods for feature attribution. In order to employ a large perturbation space, an SDE is defined. The paper derives a connection between Fisher divergence and the KL divergence, and proposes utilizing the information bottleneck principle for optimization.The paper provides a method to calculate feature importance, with empirical validations demonstrating the method's good performances.
update after rebuttal
The authors did provided further details on their proposed approach. However, I still have doubts concerning the validity of the surrogate loss of -t in place of mutual information. Therefore, I would like to retain my score.
给作者的问题
I find the presentation of the paper rather confusing, and I would appreciate it if the authors could clarify my questions. How is the loss in Equation 33 utilized? How is mutual information used to perform feature attribution?
论据与证据
The claims are supported by mathematical proofs or empirical experiments.
方法与评估标准
I am not familiar with the task. The empirical results do seem to suggest that the proposed method works well.
理论论述
The major theoretical result is Theorem 3.1. I briefly go through the proof of it. While I cannot determine its correctness for sure, it is quite likely to be a direct generalization of Theorem 1 in [1]. To me, the authors' presentation makes it difficult to follow what the proposed method actually is, as such I cannot verify its correctness.
[1] Interpretation and Generalization of Score Matching, Lyu, UAI 2009.
实验设计与分析
In my view, some experimental settings are missing from the paper and even the supplements. For instance, what are the datasets used in section 4.1 and section 4.2?
补充材料
I did not properly review the supplementary material.
与现有文献的关系
The paper is largely within the literature of perturbation based feature attribution methods. As claimed by the paper, it allows a large perturbation space.
遗漏的重要参考文献
N/A
其他优缺点
In terms of strength, using score-based approach for feature attribution is a good idea.
In terms of weaknesses, the authors' presentation is hard to follow. The proof of Theorem 3.1 takes up a lot of space, which could have been in the Appendix. Moreover, I have trouble following the proposed algorithm, as I could not find a place that explicitly give the proposed algorithm.
其他意见或建议
Page 4, second column, line 174 - 182: "And the objective is defined as", "And \beta >=0 is a trade-off", "And" seems to be redundant. Page 6, second column, Equation 32: Is this correct? Page 6, second column, line 312 (after Equation 36), "Songet al., 2020", broken reference name Page 8, first column, line 390 - 395, "Parameter Randomization Sanity Check", "Insertion and Deletion", "Segmentation-based Ratio", misused capitalizations
Below, we address your specific questions and concerns:
Theoretical Claims: Please see in 'Claims And Evidence' part of in the first reviewer’s rebuttal section.
Experimental Designs Or Analyses: Our experiments are conducted on ImageNet, as it is a standard dataset frequently employed in prior work on attribution (e.g., IBA[Schulz et al., 2020]). For the classifier, we use VGG16, which is widely used in attribution studies [Schulz et al., 2020, Zhang et al., 2021]
Questions For Authors:
-
How is the loss in Equation (33) utilized?
In our framework, the loss in Equation (33) is designed to approximate the mutual information between the perturbed input and the original input . Specifically, since increasing the noise level reduces the mutual information, we use as a proxy for the mutual information loss (). This term is then combined with the standard cross-entropy loss () for classification, yielding a total loss of the formMinimizing this combined loss encourages the model to find perturbations that significantly reduce mutual information—thus filtering out uninformative features—while ensuring that the classifier's predictions remain correct.
-
How is mutual information used to perform feature attribution?
Our method leverages the dynamics of mutual information along a continuous-time perturbation path defined by an SDE. By linking the time derivative of the mutual information (via its connection to the Fisher divergence) to the evolution of noise in the input, we obtain a quantitative measure of how much each feature contributes to preserving the model's output. Concretely, after complete optimization, we derive a tensor (with the same dimensions as the input) where each element indicates the “time-to-noise” required to significantly reduce the mutual information of that feature. Features with lower values are deemed more important because they are less robust to noise injection—i.e., perturbing these features causes larger changes in the mutual information and ultimately in the model’s output. Moreover, using Equations (49) and (50), we can further refine these scores by integrating the mutual information dynamics over time, thus arriving at a more precise attribution map.
We provide the detail algorithm for the optimaztion below for clarification.
Algorithm: Training for Diffusion-based Feature Attribution
Input:
- : U-Net model with parameters
- : Pre-trained classifier
- dataset: Training dataset of images and labels
- diffusion_type: Either "VE" or "VP"
Function TrainDiffusionAttribution(, , dataset, diffusion_type):
-
Initialize diffusion parameters:
- If diffusion_type == "VE":
- appropriate value for dataset
- If diffusion_type == "VP":
- If diffusion_type == "VE":
-
While not converged:
-
Sample batch from dataset
-
-
Sample noise
-
Generate perturbed image :
- If diffusion_type == "VE":
- If diffusion_type == "VP":
- If diffusion_type == "VE":
-
-
-
-
-
=
-
Update using
-
-
Return
I thank the authors for engaging in discussions.
Can the authors comment on how good a proxy -t is to the mutual information loss? I assume some approximation is employed here.
Furthermore, in the provided algorithm, \theta is updated using \nabla_{\btheta} L, but \theta does not appear in the computational graph of L_{CE}, so it is only trained using L_{MI}. How does this relate to the algorithm provided in response to reviewer 38Uj, in which case the score network seems to be trained using standard diffusion loss?
We thank the reviewer for the insightful comments and valuable feedback. Below are our responses to your questions:
Q1: Can the authors comment on how good a proxy -t is to the mutual information loss? I assume some approximation is employed here.
A1: Since increasing monotonically decreases the mutual information , we can use as a proxy for the mutual information loss term, thereby avoiding the costly integral-based computation during training. Once the optimization is complete and the final -tensor is obtained, we can still compute the exact mutual information using Equations (49) or (50) to produce the final attribution map. This two-stage approach—using as a surrogate loss and then performing an exact mutual information calculation—strikes a balance between computational efficiency and theoretical rigor.
Q2: Furthermore, in the provided algorithm, \theta is updated using \nabla_{\btheta} L, but \theta does not appear in the computational graph of L_{CE}, so it is only trained using L_{MI}.
A2: In our algorithm, although does not appear directly in the expression for , the loss is computed on the classifier output obtained from the perturbed input , and is generated based on , which in turn is produced by the network parameterized by . Thus, indirectly influences via the chain of dependencies . Consequently, both and contribute gradient signals to update in our training process.
Q3: How does this relate to the algorithm provided in response to reviewer 38Uj, in which case the score network seems to be trained using standard diffusion loss?
A3: In our approach using the Variance Exploding (VE) noise addition method, we require a score network to compute the mutual information with Equations (50) (second stage as explained in A1). If a pretrained score-based diffusion model suitable for our research scenario is unavailable, we train the score network ourselves. Furthermore, when performing attribution on a small number of images, it is unnecessary to employ a diffusion model trained on a large-scale dataset like ImageNet. In our response to reviewer 38Uj, we provided an algorithm for training a score network on a single image—a method that is similar to that presented in Song et al., 2020.
The paper studies the dynamics of the mutual information through SDE with the fisher divergence and the dynamics KL divergence. The computation process is proposed by discretization and the numerical studies apply the proposed framework in the feature attribution in the explainability of neural network.
给作者的问题
- how is the discretization error influence the mutual information integration.
- for more complex SDE (even maybe not explicit solution), can the proposed method work
- why ours show not signifcant improvement compared with InputIBA. I do not require this but just want to figure out the characteristics of the numerical studies.
- what if the sigma in Eq. (36) follows a larger range.
Update after rebuttal
Thanks for efforts of the authors. Most of my concerns are addressed but I believe the problem in the discretization error and range still needs further justification. And the theoretical contribution is not so significant. Therefore, I still decide to keep my score as WA.
论据与证据
Yes. The claim in the theoretical part in SDE and computation part of optimzation with the mutual informationis right.
方法与评估标准
The proposed analysis make senses for the dynamics of mutual information in perturbation for SDE and the computation method for the loss designed for mutual information is reasonable for the implementation.
理论论述
Yes. I check the dynamics for the KL divergence via Fish divergence, that's a mathematical manipulation. The KL and mutual information can link the fisher divergence and score function for the computation next.
实验设计与分析
Yes. I check the Parameter Randomization Sanity Check, Deletion and Insertion methods and Quantitative Visual check,
补充材料
I view Qualitative Comparision for the supplementary.
与现有文献的关系
The perturbations for mutual information can be another new metric to measure the influence. The proposed can be adopted to improve the explainability of the DL model.
遗漏的重要参考文献
Please compare with other work considering the perturbations about KL/Renyi divergence in SGLD like [Chourasia et al. 2022]
其他优缺点
Pros:
-
The paper provides clear analysis for the mutual information dyanmics in the perturbations.
-
The paper is well-written for understanding,
Cons:
-
the technique contribution needs to be clarified. It is not news for the perturbations and the mutual information is just a direct extension.
-
restate and verify the assumptions in the paper fot he thm 3.1 and Lemma 3.2
-
the loss of the mutual information is not direct from the theoretical results
其他意见或建议
Line 498 reference needs re-organized Line 312 the fonts are not right
Below are our responses to the specific concerns:
Other Strengths And Weaknesses:
-
Technical Contribution Clarification: While it is true that the mutual information dynamics under perturbations build on existing ideas, our work significantly extends these concepts by analyzing general SDEs (i.e., with time-dependent drift and diffusion ). This extension is nontrivial and enables us to integrate the KL–Fisher divergence relationship into a new optimization framework for feature attribution—offering both theoretical insights and practical advantages in explaining neural network predictions.
-
Restating and Verifying Assumptions for Theorem 3.1 and Lemma 3.2: We assume that for every , the densities and are twice continuously differentiable and decay sufficiently fast at infinity (i.e., there exists some such that , and similarly for ). These conditions ensure that all integrals are finite, boundary terms vanish under integration by parts, and the required interchanges of differentiation and integration are justified.
-
Mutual Information Loss Proxy (): Recognizing that increasing (i.e., injecting more noise) monotonically reduces mutual information, we adopt as an effective proxy for . This surrogate is both computationally efficient and consistent with prior works (e.g., IBA and InputIBA), even though the theoretical derivation of the mutual information dynamics is not directly reflected in the loss formulation.
Questions For Authors:
-
Discretization Error in Mutual Information Integration:
Once the optimization yields the tensor , the integration limits for the mutual information calculation are determined, and we can choose sufficiently small step sizes for the numerical integration. This ensures that discretization errors are negligible and do not affect either the computational efficiency or the accuracy of the mutual information integration. -
Applicability to More Complex SDEs:
Although our current experiments use SDEs with explicit solutions (e.g., the Variance Preserving or Variance Exploding processes), our framework is general. For more complex SDEs without explicit solutions, one can rely on established numerical SDE integration methods, albeit with higher computational cost, without compromising the overall approach. -
Comparison with InputIBA: Our experiments—particularly the Effective Heat Ratios (EHR) evaluation—demonstrate that our method clearly outperforms InputIBA. We observe a significantly higher EHR score for our approach, which indicates that our attributions are more concentrated in semantically meaningful regions. Furthermore, we provide more experimental results after the 'Other Comments Or Suggestions' sector.
-
The Range of in Equation (36):
Our approach is designed to explore the application of state-of-the-art diffusion models to feature attribution. At present, we employ the Variance Exploding (VE) and Variance Preserving (VP) processes, as these are the most representative and well-established diffusion models in the literature. In future work, we intend to investigate broader cases, including scenarios with larger ranges of in Equation (36), to further generalize and enhance our method.
Other Comments Or Suggestions: We appreciate the reviewer’s note regarding the formatting issues. These will be corrected in the final version.
More Experiments
- Bounding Boxes Evaluation [Schulz et al., 2020]
We leverage human-annotated bounding boxes from the ImageNet dataset to evaluate localization.
Box-Ratio Results:
| Method | Box-Ratio |
|---|---|
| IBA | 0.997 ± 0.001 |
| InputIBA | 0.998 ± 0.001 |
| Ours | 0.998 ± 0.000 |
| Integrated Gradients | 0.691 ± 0.006 |
| Guided-BP | 0.698 ± 0.002 |
| Deep-Lift | 0.695 ± 0.005 |
| HSIC-Attribution | 0.903 ± 0.002 |
- Segmentation-based Ratio Evaluation Using semantic segmentation masks from the FSS-1000 dataset, we replace bounding boxes with detailed segmentation regions and compute the Segmentation-based Ratio (SR) in the same manner as the Box-Ratio. | Method | SR | |----------------------|-------------------| | IBA | 0.488 ± 0.004 | | InputIBA | 0.468 ± 0.003 | | Ours | 0.501 ± 0.003 | | Integrated Gradients | 0.079 ± 0.006 | | Guided-BP | 0.078 ± 0.009 | | Deep-Lift | 0.080 ± 0.002 | | HSIC-Attribution | 0.377 ± 0.005 |
This paper proposes a perturbation-based feature attribution method, where the input features are perturbed based on a stochastic differential equation (SDE). The proposed framework optimizes an input such that the input has small mutual information with the unperturbed input, and large mutual information with a target class. The authors theoretically show that the mutual information between the perturbed and unperturbed inputs relate to the SDE score functions, which can be approximated by diffusion models--hence computable for optimization. The mutual information with a target class is replaced with the cross entropy loss for the model to be explained. Empirical results demonstrate that the proposed method achieves better Insertion and Deletion AUC compared to baselines, as well as having better overlap with known image segmentation.
给作者的问题
N/A
论据与证据
- Starting at line 107, it is claimed that a novel theoretical relationship between KL Divergence and Fisher Divergence is established. However, this result follows almost directly from Theorem 1 in Lyu et al., 2012. I wouldn't consider this as a novel contribution.
方法与评估标准
The proposed method and evaluation criteria make sense.
理论论述
-
I checked the proofs for Theorem 3.1 and Lemma 3.2. Both look correct to me. For Theorem 3.1, the authors should mention that most of the steps follow from the proof of Theorem 1 in Lyu et al., 2012.
-
What assumptions are used for ? They should be clearly stated. Overall, after integration by parts, some terms equal to zero. The assumptions used to get those terms to zero should be stated.
-
It is assumed that and are smooth and sufficiently decaying. These assumptions should be clearly formalized.
-
The symbol should be defined.
实验设计与分析
-
It is unclear how one would tune in Equation (5). It's also unclear why is dropped in the final loss function in Equation (33).
-
The definition for in Equation (32) is incomplete, and is not defined. An input should be optimized for in Equation (33) according to the conceptual framework, but on the right side of line 288 it is mentioned that a neural network is trained. The authors should clearly define and the optimization objective to prevent confusion.
-
Details about the diffusion model training should be included. Is the diffusion model trained on data for which attributions are computed? This has implication for the scalability and/or generalizability of the proposed method. If a diffusion model has to be trained for each dataset, then the proposed method might not be scalable. If a pretrained diffusion model is used, then the proposed method might not transfer well to images out of distribution for the diffusion model.
-
The proposed method learns a perturbed input . It is unclear how to go from to the attribution scores.
-
It's unclear what dataset was used to compute the Insertion and Deletion metrics. Also, more datasets should be included to demonstrate that the proposed approach is generalizable.
-
Shapley-based feature attributions should be included as baselines, since they are considered standard feature attributions.
补充材料
I reviewed the entire Appendix.
与现有文献的关系
This paper proposes a novel approach to perform perturbation for perturbation-based feature attribution.
遗漏的重要参考文献
N/A
其他优缺点
Overall, this paper proposes a novel approach for perturbing input features, with an optimization framework for getting perturbation-based feature attributions. However, in its current version, this paper suffers from clarity in terms of experimental details, such that a reader cannot reproduce the paper. Also, the experiments are limited in scope with respect to the number of datasets, the number of classifiers to explain, and the variety of diffusion models.
其他意见或建议
-
Much of Section 3.1 has been mentioned in Section 2. Consider merging the two sections.
-
Consider moving the proofs to the Appendix.
Below, we address each of your concerns in detail.
Claims And Evidence:
Comparison with Theorem 1 in Lyu et al. (2012):
Theorem 1 in Lyu et al. (2012) considers the simple case of the SDE
where the drift is and the diffusion coefficient is . This setup corresponds to a basic Wiener process.
In contrast, our work extends this result to the much more general case of
where both the drift and the diffusion are allowed to be time-dependent. This generalization is significant because:
- Broader SDE Models:
While Lyu et al. (2012) deal with a noise model where and , our formulation covers a wide range of stochastic processes. This is essential for practical applications, such as those in modern diffusion models DDPM likeand score-based diffusion models, where the noise characteristics evolve over time. - Handling of Additional Complexity:
Extending the proof in Lyu et al. (2012) to accommodate non-zero drift and time-varying diffusion requires new insights. For example, our proof first builds the intuition that the drift term can be canceled out under certain conditions—a step that is unnecessary in the Wiener process setting. Similarly, the treatment of involves more delicate handling due to its time dependency.
We follow the proof structure of Lyu et al. (2012) to enhance readability and make it easier for readers to identify which parts of our derivation are directly inherited from prior work and which parts constitute our innovations. We appreciate the reviewer’s suggestion to explicitly mention that our proof builds on the ideas of Lyu et al. (2012), and we will clarify this point in the revised manuscript.
Theoretical Claims:
- Please see 'Claims And Evidence' part.
-
- The boundary terms vanish due to suitable decay at infinity or zero boundary values for .
Hence,
- We assume that for every , the densities and are twice continuously differentiable and decay sufficiently fast at infinity (i.e., and similarly for so that all integrals are finite and boundary terms vanish.
- is the Laplacian operator.
Experimental Designs Or Analyses:
- We treat as a hyperparameter. In our experiments, we typically set , following the convention in prior works such as IBA and InputIBA. And we will clarify these details in the revised manuscript.
- in our formulation leverages the observation that increasing (i.e., injecting more noise) monotonically decreases the mutual information between the original input and the perturbed input; thus, we can use as an effective proxy for $L_{\mathrm{MI}}(L_{\mathrm{CE}}$ is the standard cross-entropy loss (please see before the Equation (33)), ensuring that the classifier’s predictions remain correct for the perturbed input. We will provided the training algorithm in the following.
- We can either train a diffusion model from scratch or use a pretrained one for efficiency. If a pretrained model is used, it must be sufficiently close to the domain of interest so that its learned score function remains accurate. it is possible to train a diffusion model on a single image, and we provide the optimization training algorithm in the third reviewer's rebuttal section and score network training algorithm in the fourth reviewer’s rebuttal section due to space constraints.
- After the complete optimization, we obtain a tensor with the same dimensions as the input image; this tensor effectively encodes the "time-to-noise" for each pixel, which we can interpret as a preliminary measure of feature importance—pixels with lower values indicate features that are more critical for preserving the model’s output. Moreover, by further computing the integrated mutual information along the diffusion path using Equations (49) and (50), we can yield a more precise attribution map.
- Currently, our experiments are conducted on ImageNet, as it is a standard dataset frequently employed in prior work on attribution (e.g., IBA[Schulz et al., 2020]). For the classifier, we use VGG16, which is widely used in attribution studies [Schulz et al., 2020, Zhang et al., 2021] and provides a reliable benchmark. Moreover, incorporating additional classifiers is possible, and we plan to include them and show the comparison between our methods and others in revised version.
- We will compare our method to at least one Shapley-value-based approach, acknowledging that Shapley methods are standard baselines in feature attribution.
The authors' response has addressed my clarification questions. I encourage the authors to include those clarifications. In the paper, it should be clearly indicated when a pretrained diffusion model is used and when a diffusion model is trained for each image.
I have accordingly raised my score. However, the updated score is contingent on the following.
- Clear details about experimental settings in the paper.
- More empirical results with additional classifiers, datasets, and (pretrained) diffusion models.
- Inclusion of Shapley-based attribution methods.
This paper introduces a novel perturbation-based feature attribution method that leverages stochastic differential equations (SDE) to model continuous perturbations and comprehensively explore the perturbation space. The proposed method employs SDE to define the perturbation space that is more flexible than that in most existing perturbation-based methods, and develops an optimization framework for feature attribution avoiding exhaustive search in the entire perturbation space. The proposed method is theoretically motivated and empirically evaluated.
Overall, most reviews are positive for this submission, although noting issues such as lack of clarity in some technical content, limited significance in the theoretical contribution, and insufficient empirical evaluation. The authors are encouraged to fully leverage the reviewer feedback to further improve the paper.