Sufficient and Necessary Explanations (and What Lies in Between)
This paper explores notions of sufficiency, necessity, and their unification for explaining how general machine learning models make predictions.
摘要
评审与讨论
This paper proposes a novel measure to interpret predictions of black box machine learning models. Given a masking strategy for a model the authors define -sufficient feature sets as sets whose masked prediction (mask= complement of S) deviates at most of the actual prediction with respect to some metric (deviation=). In contrast, -necessary feature sets are sets, where the masked prediction (mask=S) deviates at most from the fully masked prediction w.r.t (deviation=). Given these notions, the authors aim to find the set with the smallest deviations. Instead of solving each problem individually, the authors propose to find a set that minimizes the weighted average with a hyper-parameter , which they term "unified" problem, where the edge cases are the individual problems. The problem is further constrained, by finding a set with cardinality at most . The main theoretical contribution is Theorem 4.1 that states the existence of with sufficiently small , given sufficient and necessary solution and additional assumptions. Moreover, the solution satisfies conditional independence, if models the data-generating process (Corollary 5.1). Lastly, it is shown (Theorem 6.1) that the Shapley value of S treated as a joint player in the two-player game with S, [d]/S is bounded from below by . In the experiments, the proposed is evaluated for different hyperparameter configurations on tabular data (6.1) and for image classification (6.2). For the restrictive setting in image classification, the authors propose an optimization objective that solves a relaxed version, which is specific to image classification, since exactly computing the objective is NP-hard. The authors conclude that the two proposed notions identify distinct aspects of predictions, which is showcased on examples for image classification.
优点
- The proposed necessary and sufficient explanations are an intriguing concept that complements existing approaches, such as the Shapley value that summarize "average contributions". Enriching the toolbox of interpretability methods with simple concepts is an important extension of existing work.
- The paper is well-written, all theoretical claims are precisely stated and formally proven. The intuition behind the concepts are clear.
- The proposed approximation method for image classification is interesting, but not sufficiently evaluated.
- A first attempt to link the novel concepts to game-theoretic measures, such as the Shapley value, is promising.
缺点
- Limited discussion on computational aspects: The paper introduces two interesting concepts, but the main limitation in practice is the optimization problem, which optimizes over all possible subsets (2^d), similar to the Shapley value. For the Shapley value, however, there exist efficient approximation techniques by evaluating the target for a collection of sampled subsets, which are unbiased and known to converge. In contrast, in Section 6.2 the authors propose approximation strategies for which no theoretical guarantees are given. The paper would highly benefit from such theoretical guarantees and a general approximation approach. Moreover, for the given applications, it should be carefully evaluated why these approximations yield appropriate solutions (see Q1 below).
- Limited link to other concepts: Theorem 5.1 is a first step towards understanding connections to other methods, such as the Shapley value. However, this result is quite limited, since it considers a drastically reduced game of two players, instead of players, which does not give insights into the actual Shapley values that are usually computed for interpretability. The paper would strongly benefit from establishing such connections more carefully (see Q2 below), in particular given the claims made in the abstract.
- Comparison with existing attribution methods: The authors claim that necessary and sufficient explanations yield more insights into explanations, which are undermined by some interesting examples in image classification. It remains unclear, how these methods compare to existing attribution methods. For instance, it would be helpful to formally state the difference between these concepts and carefully evaluate empirically which questions can be answered by the novel concept, which could not be targeted with existing methods. Moreover, some choices in the comparison are unclear to me (see Q3 below).
- Theorem 4.1: The key assumption here is super-sufficient and super-necessary. I am not convinced that these properties are given in real-world examples (see Q4 below).
- Hyperparameters: The method requires two hyperparameters and , which need to be chosen in advance, and have a high impact on the explanations, as shown by the experiments. It remains unclear how to choose these parameters in practice. It would be helpful to understand these choices better with sensible defaults.
问题
- The proposed computation methods in Section 6.2. How well do they approximate the actual target? Could you do this analysis in a lower-dimensional setting, where ground-truths can be computed?
- Do you have any insights on the method compared to the actual Shapley values of the model?
- For the comparison of post-hoc explainability methods, what is the effect of the normalization? Is it negligible? Most methods already decompose the model's prediction, wouldn't it be sufficient to divide the attributions by the model's prediction to obtain normalized scores? Why don't you rely on the ranking of the attributions directly using ?
- Is super-necessary and super-sufficient actually observed in practice? It seems counter-intuitive to me. For instance, for images, unmasking some parts of the image could lower the prediction significantly, since they could reveal a novel concept or another concept that contains the previously predicted concept.
- The purpose of the metric is unclear to me. It seems more intuitive to me to consider the model's output directly. What is the benefit of using the metric ? Why don't you use the model output directly, which is the common choice for local explanations? For necessary explanations, this also seems somewhat problematic: Consider a binary prediction, where the prediction is , and the average prediction . Masking some features could substantially change the model's prediction to the opposite class, which would be captured with sufficient explanations. However, for necessary explanations, this would yield a similarly unsatisfying result (high \approx 0$) is equally far from the average prediction. Is this behavior intended? What is the reasoning behind it?
- In Section 6.2. what are the resulting sizes of explanation sets that you find? Could you give some statistics on these for the datasets? Do you consider these sizes also in the comparison with post-hoc attribution methods? If so, in which way are they accounted for? E.g. I would expect to choose attributions of features such that the number matches with ?
Minor
- typo line 414, missing blank after "demonstrate"
- in Appendix A, the navigation displays "Proof of Theorem 6.1", while the section title is correctly referring to Theorem 5.1.
Concern 1: Limited discussion of computational aspects
We thank the reviewer for the constructive comments. While it is true that the optimization problem is not tractable, we want to stress a couple of points:
-
Our contribution is not a tractable relaxation that identifies sufficient and/or necessary subsets. Instead, we provide precise and flexible notions of sufficiency and necessity, and bring to attention that these notions do not provide a complete picture of feature importance. To demonstrate our theoretical results in practice, we relied on computational methods that are already well-known in the literature [1,2,3,4].
-
Note that while the optimization problem may be intractable, the evaluation of and is not. Thus, we can still use these notions to evaluate how sufficient or necessary the explanations generated by different methods are. This allows us to provide a relative ranking of which method provides the best explanations in terms of sufficiency or necessity, which was not possible prior to our contribution.
Your point is well-taken, however, and we will expand on the limitations of the computational aspects of solving the proposed problems in the revised manuscript.
[1] "Interpretable Explanations of Black Boxes by Meaningful Perturbation," Ruth Fong, Andrea Vedaldi, ICCV 2017 [2] "Understanding deep networks via extremal perturbations and smooth masks," Ruth Fong, Mandela Patrick, Andrea Vedaldi, ICCV 2019 [3] "Cartoon explanations of image classifier," Stefan Kolek, et. al. ECCV 2022 [4] "Model Interpretability and Rationale Extraction by Input Mask Optimization," Marc Brinner, ACL 2023
Concern 2: Limited link to other concepts
First, we apologize if the claims in the abstract were misleading. Theorem 5.1 does establish a connection between the unified problem and the Shapley value, albeit for a two-player game. The reviewer is correct that is different from the way Shapley coefficients are typically used for approximating important features, but this does provide a connection between game-theoretic quantities and necessity/sufficiency methods which was unknown before. We do not find this different necessarily limiting, but rather a simpler and clearer picture of feature importance. By considering a subset and its complement , we are simply differentiating the important features from those that are not.Moreover, note that the feature scores obtained with traditional methods (including via Shapley) are often (though not always) thresholded to differentiate between important and important features. Our formulation addresses this case.
The comment is well-taken, and we will clarify these differences.
Concern 3: Comparison with existing attribution methods
We believe our empirical results do make a precise comparison of how different notions of importance differ. In particular:
-
We show that necessary and sufficient need not be the same, but that they are related. In particular, we show in the two image classification settings that the sufficient explanations are subsets of the necessary ones.
-
We show that common post-hoc methods fail in retrieving necessary features, and most simply estimate (approximately) sufficient ones (see the RSNA and CelebAHQ experiments).
-
We show that our notions do allow for solutions that balance the trade-off between these notions of importance (portrayed in the tabular examples).
Concern 4: Theorem 4.1 assuming super sufficiency/necessity
Thank you for this comment. While super-sufficiency and super-necessity will not always hold in every scenario, they are not as strict as they seem. Super-sufficiency simply states that, given sufficient set (e.g. a region of an image), this set is super-sufficient if supersets of it are also sufficient. This indeed holds in the examples we provide: the brain CT scans and CelebAHQ. For the CT scans, once we fix a region with at least one hemorrhage, any larger region that contains this hemorrhage will be sufficient for predicting the label. Similarly in the CelebAHQ example, once ones fixes the region of the image that contains the smile, every superset of this region still allows the predictor to predict smile as expected. One can reason similar situations for the super-necessary settings: given a necessary set in the brain CT examples that contains all of the hemorrhages present in an image (i.e. a necessary set), any super-set of it will also be necessary.
We hope these clarify our notions of super necessity and sufficiency, but we'd be happy to clarify further.
Concern 5: Hyperparameters
Thanks for bringing this up. We stress that, unlike other problems where the hyperparameters are hard to interpret, both and have precise meanings: the former controls the trade-off between sufficiency and necessity of the solution, whereas the latter determines the size of the reported features. As a result, there is no ``correct'' choice for these, but rather the choice should be determined by the specific problem domain or user preferences. For instance, if one is after a sufficient explanation, then is the correct choice -- correspondingly, ensures a necessary one. We demonstrate, through theory and examples, that sufficient explanations may not capture necessary ones and vice versa. As a result, if one wishes to capture all such features, then is the appropriate choice.
We will expand on this rationale in our revised version -- thanks for the comment!
Concern 6: Purpose of metric
Thank you for this question, from which we identify two key components: one pertains to the choice of requiring to be close to either or (for sufficient or necessary features, resp.) instead of being close to 1 or 0 (analogous to the reviewer's suggestion of using ``using the model output directly''); the second pertains to the choice of the general metric .
For the former: we argue that measuring the distance of the expected perturbed function with respect to the canonical choices of and is more general. Consider the case of searching for sufficient features, and consider as an example, the case where . Searching for a predictor (rather than one for which might provide features that are not sufficient for the predictor to produce the output (only sufficient for producing . This difference can be particularly important in cases where the predictions of are calibrated, and thus their specific values carry specific meanings that we want to account for. Likewise, for necessary features, note that requiring instead of (say, ) will result in a subset that is (on average) sufficient for predicting instead of necessary features for . This would result in features that are closer to counter-factual notions of explanations. Our notions, instead, simply require necessary features to be those without which the prediction is no better than a random guess.
Lastly, we use a general metric to accommodate different prediction settings: a regression task, an norm might be appropriate, whereas in a multi-class classification settings the difference in maximum scores might be better suited, etc.
Concern 7: Questions about section 6.2
Thanks for bringing this up, as we realize that this could have been presented more clearly. As we comment on Sec. 6.2, and to ensure a consistent analysis, we first normalize all generated attribution scores to the interval , and then obtain different important features by thresholding the scores at different values in . Each threshold results in a generated explanation with a specific cardinality, or size (which is a monotonically decreasing function of the threshold). As can be seen in Fig. 2(a) and Fig(3), our results are provided as a function of this threshold (i.e. for all choices of this parameter), and we further include the sizes of the reported features by reporting , where refers to the (relative) cardinality of the reported features. As a result, this provides a complete picture that allows us to compare these methods for any value of (i.e. for any size of the reported important features).
I thank the authors for answering my questions. I have some follow-up questions:
- Comparisons: Since the main contribution of your work is a novel concept for local explanations, I would highly appreciate to better understand the link to existing concepts, and theoretical differences. Specifically, since your novel concept is based on all possible perturbations/maskings, more extensive comparisons with the (actual) Shapley value (on feature level) as the leading concept of fair attributions would be beneficial. I think this would be best evaluated on a setting, where ground-truths can be computed, leaving the computational problem aside, e.g. evaluations across all choices of , are the identified sets consistent? If not, what are the consequences? In my view, as of now, your experiments, while interesting, focus too much on the application side of the problem, where multiple factors (concepts, hyperparameters, approximation) interplay, which makes it hard to distinguish your proposed concepts of sufficiency and necessity from other factors.
- super-sufficiency/necessity: While I understand your intuition, do you have any empirical evidence that super-efficiency/necessity indeed holds in your setting or a dimensionality-reduced setting in real-world applications?
- metric: Looking at necessity, this would still imply that a set of a specific size with would be preferred over another set with , given . In other words, the set of necessary features for the target class 1 is actually a set of features that, if removed/masked, outputs more likely the opposite class 0 than any randomized prediction. Am I missing something? A simple solution could be the use the prediction instead, and cap this at and , which would also resolve the issues you mentioned on sufficiency.
Thanks for the reply! Here are our responses to your comments/concerns.
Comparisons
We have added an additional synthetic experiment where ground truths can be computed, and extensive comparisons can be made. The details are below.
The experiment is the following:
We model features , where for . The remaining features and response follow:
where . For , the data-generating process is represented by the directed acyclic graph (DAG) shown below
with not connected to any variables. From the DAG, we can see that and . Thus, for and , the solutions to , , and with are:
In this experiment, we train a general predictor (a three-layer fully-connected neural network) to approximate and 1) validate that the sets listed above are the optimal solutions, and 2) demonstrate that common post-hoc interpretability methods do not recover these any of these sets.
Unfortunately, we cannot send figures through this forum, but we highlight the main takeaways from the experiments, and we’ll include the figures in the supplementary material.
Validating the solutions
For , , and 100 samples , we compute solutions, denoted as to the sufficiency, necessity, and unified problem. We find that:
- For 95% of the samples in , , the solution to the sufficiency problem.
- For 60% of the samples in , , the solution to the necessity problem.
- For 92% of the samples in , , the solution the unified problem.
These results indicate that the solutions computed via an exhaustive search do typically retrieve the correct solutions (the minor discrepancies are due to being an approximation of ). More importantly, this setting is a clear example of when one would not be able to identify the set as the most important one unless you directly solve the unified problem.
Comparison with other methods
For our model and 100 samples , we use Integrated Gradients, Gradient Shapley, DeepLift, and Lime to generate attribution scores. To identify whether these methods highlight sufficient and/or necessary features, and as done before in our manuscript, we perform the following steps on the attribution scores for a sample (so that the outputs of all methods are comparable)
- We normalize the scores to the interval [0,1] via min/max normalization.
- We generate binary masks by thresholding the normalized scores with thresholds .
- For , we compute , the hamming distance between and the true solutions to , , and
The main results from our analysis are the following:
-
There is no threshold in for which any method recovers the true solution to , . Furthermore, for the average hamming distance, , is for all methods, indicating that and disagree by at least one element.
-
There is no threshold in for which any method recovers the true solution to , . In fact for , the average hamming distance , is for all methods, indicating that and disagree by at least 2 elements.
-
For , integrated gradients and deeplift recover the true solution to , . However, for , the average hamming distance , is for all methods, indicating that and disagree by at least 2 elements.
We are in the middle of finishing this experiment, adding comparisons to the Shapley value, and updating the manuscript. We hope that the experiment and its results adequately address your concerns.
Metric
In the reviewer's example, set S is indeed more necessary than set T. We don't see any conflict between this and f_{S^c} < 0.5. Could the reviewer clarify why this is a concern?
Super-sufficiency/necessity
We provided weak evidence where super-sufficiency holds; e.g., any set that includes a brain hemorrhage in a CT scan will be super-sufficient for a 'good' predictor. However, we want to stress that these properties hold for a large class of data-generating processes. For example, they hold for distributions over that can be modeled with a Markov Random Field. If the process can be expressed with a Markov Random Field, then the smallest will be the parents of Y, and the smallest will be all variables connected that have a path to . For example, consider
X1 — X2 —Y — X3— X4
X5 X6 X7
Here and . From this, it’s clear that adding any additional elements to or retains the sufficiency/necessity. Thus, super-sufficiency/necessity holds. Markov random fields appear in many real-world applications such as imaging, social network analysis, and genomics [1,2,3]. Thus, there are many real-world settings where super-sufficiency/necessity indeed holds.
[1] Markov random field modeling in image analysis, Stan Z. Li Springer 2009
[2] A network-specific Markov random field approach to community detection, Dongxiao He et al. AAAI 2018
[3] A Markov random field model for network-based analysis of genomic data, Zhi Wei, Honghze Li, Bioinformatics 2007
Thank you again for the additional comments and experiments.
Comparison
I thank the authors for the additional computational experiment. This seems to be a promising setup to confirm that your approximation works. For a comparison of methods, I would suggest rather to take a traditional local interpretability setup, maybe tabular data with few features, where exhaustive search is still feasible together with a fixed modeling choice for the conditional expectations that your approach and the competitors both use. To be more clear, what I would actually would like to better understand is how these concepts are different from the Shapley values, e.g. questions like: can we construct sufficient and necessary sets from the Shapley values? If not, why not? (there seems to be a strong discrepancy, why?) For , is there a link to the TOP-1 Shapley value? If not, why not? A theoretical analysis of this would be highly beneficial, otherwise practitioners are left with two disconnected methods, one considering feature sets, and one individual attributions.
Minor comments regarding the experiment: Please include a more formal description how you deduce the necessary and sufficient subsets. Regarding the comparison with Shapley values, I think the selection of TOP-k values is good for the sufficient features, I think for the necessary set, probably a different approach is better suited, since you would like to identify the set of Shapley values such that they sum to the baseline prediction, and for uni the combination of both may be better suited. Moreover, for a better comparison, it would be more meaningful to compare with the results from the exhaustive search rather than the actual sets, to eliminate the computational aspect.
super-efficiency/necessity
Thank you again for this suggestion. I think it is valuable to link to cases where this theoretically holds. I would be very interested, if this also holds in well-established benchmarks of local interpretability XAI literature. Moreover, if it does not hold, what are the consequences? I am still in doubt that this holds in practice and I think, as of now, your claim is not convincingly supported by empirical evidence.
metric
I guess my concern here is that sufficient is depending on the class that is to be explained, whereas necessary (as a consequence of the example) is not. My understanding is now: Sufficient features are sufficient for the prediction of that class, whereas necessary features are necessary to make any prediction of any class (i.e. not predicting the baseline). I think there could be a problem arising from mixing these two into a single explanation, as effects could cancel.
Overall, I thank the authors for answering my questions and adding additional content. However, as of now, given the limitations of the current work, I decided to keep my score.
Thanks for the reply! Our responses to your concerns are below.
Sufficiency and necessity metric
We believe there is a misunderstanding here, and we are happy to clarify. Our notions of sufficiency/necessity are best understood when we take the model and . Then, for a sample , we say a set is sufficient for the prediction if . In other words, is sufficient if, by only using the features in , we can recover the original prediction that used all the features in . On the other hand, our notion of necessity defines a set as necessary if, by removing and only using the complement, , we have . We say is necessary because we need . Without it, our prediction is simply the naive prediction , which does not use any information about .
We are not sure how any issues could arise from combining these notions and how cancellations may occur, as the reviewer implies.
Super sufficiency/necessity
Thank you for the suggestion, we will indeed add the analytical example we provided to the revised version of the paper. Furthermore, we will provide empirical evidence of super sufficiency/necessity holding in real-world examples.
More comprehensive synthetic experiment
We have written a global message to all the reviewers that details a more comprehensive synthetic experiment we believe addresses your concerns. Please take a look and let us know if you have any questions!
The paper proposes to merge sufficency and necessity notions into one unified optimization problem and solving this combined optimization problem. Therein, the paper is interesting and shows that this unified approach retrieves explanations unlike baseline methods.
优点
-
Good Contribution: While I think the contribution can be strengthened (see weaknesses), the presented approach is very interesting. Joining and optimizing both conditions leads to very interesting explanations, which seem to be not retrievable by other methods. The work is sound and carried out well.
-
Theoretical Connections to Game Theory: The paper connects the novel optimization problem with well-established concepts like the Shapley value. While I think the discussion and interpretation is under-developed (see weaknesses) this connection can have impact on the line of research regarding Shapely values and hierarchical explanations.
-
Synthetic Experiments: I like that the paper studies the approach in synthetic environments! This is often not done anymore and can be very insightful of what the method is actually doing. The paper studies the stability of the method in this setting, which is good but leaves more to be desired.
-
Well written and structured: The paper is clearly structured and reads well. Particularly, the related work section is very strong.
缺点
I think this paper is borderline: The paper feels a lot about the what is being done, and not really about the why this may be useful. The paper does not really motivate the use of outside of the image domain and would benefit a lot from a wider evaluation in different domains such as language and tabular and the interesting insights that can be generated there.
-
Limited Contribution: The contribution of this paper is rather incremental in nature. The notions of sufficiency and necessity are quite established concepts. Here they are just added together (with a linear combination) and jointly optimized for. While the paper does make interesting findings because of this combined lens (see strengths). It falls short in motivating and placing the contribution well into the XAI research. It is unclear when from an XAI perspective this joint approach is superior to other techniques. The conducted experiments seem rather exemplary than confirmatory in this regard. Only a small selection of created images are presented without examples for the baseline explanation methods this work compares against. The contribution could be strengthened by providing a broader comparison in different domains and actually showing how the explanations of other methods differ with the the new perspective.
-
Doubt about the stability: The experiments in the work show that sufficiency and necessity optimized together can lead to interesting insights. However, a critical aspect of retrieving these explanations is by optimizing () in some form. This is again intractable like solving () and (). The authors acknowledge this fact (line 215, footnote 1). However, the authors point to two references where one is un-published work (Kolek et al. 2021), and other is a workshop article (Kolek et al. 2022). This makes me doubt that the retrieved solutions to () are easily retrievable for various models and feature spaces. Neither the paper nor the appendix discusses this issue further. While Experiment 6.1 is concerned with this fact, stability is evaluated only on a synthetic or tabular example with low feature dimensionality. While the remainder of the paper deals with image explanations, which seems to be the main motivation for this work, the stability is left undiscussed.
-
Synthetic Setting is misused: Given the above point stability is more interesting in higher-dimension settings (images, etc.). The synthetic environment would greatly benefit a comparative study with other explanation methods. It would be interesting to see how baselines fall short and be able to explain why and then to show how the unification does not. This is in the current draft not done (also not in the exemplary setting of image explanations).
-
The Shapley-Perspective is Disconnected: While I am glad about theoretical results linking this approach to game theoretic foundations (see strengths), Section 5.2 (connection of this method and Shapley) is quite disconnected to the rest of the paper. The results are presented technically but not put into context.
问题
-
How stable are the retrieved explanations for the image case?
-
What does it mean that one searches for a player with a large lower bound in a two player game in practice. Can you put the results regarding the link of into context. At the moment this connection is made and technically presented. It is not motivated or substantiated, why this is something good/bad or in-between.
Concern 1: Limited contribution
We thank the reviewer for voicing their concerns. We believe this work does provide a meaningful contribution to the XAI community for the following reasons:
- While previous works have established certain notions of sufficiency and necessity, we introduce a definition of necessity that is different and--we argue--that better reflects what necessity should mean (see global comment that elaborates on this).
- Importantly, we demonstrate that sufficient and necessary explanations need not be the same, but they are related (their intersection is not empty). This observation is what motivates the unified approach.
- In all previous works, it remained unclear how exactly these notions of sufficiency and necessity relate to other notions of importance, such as conditional independence and Shapley values. We provide theoretical results that provide precise relations between these seemingly different perspectives
- Lastly, through various experiments, we demonstrate how generating explanations along the necessity-sufficiency axis 1) allows us to detect important features that may otherwise be missed and 2) conclude that many post-hoc methods fall on the sufficiency side of this axis.
We do apologize that these contributions were not clearly presented in the paper. In the revised version we will make these points clear.
Concern 2: Doubt about stability
We appreciate the comment. We agree that it is critical to optimize in a manner that leads to efficient and stable solutions. The methods we employ are well-established in the literature [1,2,3,4] leading to solutions that are good and stable in practical settings spanning image and natural language settings. We apologize for not making this clear in the original version. We will address in the revised version by making this statement clearer in a new section before the experiments.
[1] "Interpretable Explanations of Black Boxes by Meaningful Perturbation," Ruth Fong, Andrea Vedaldi, ICCV 2017 [2] "Understanding deep networks via extremal perturbations and smooth masks," Ruth Fong, Mandela Patrick, Andrea Vedaldi, ICCV 2019 [3] "Cartoon explanations of image classifier," Stefan Kolek, et. al. ECCV 2022 [4] "Model Interpretability and Rationale Extraction by Input Mask Optimization," Marc Brinner, ACL 2023
Concern 3: Synthetic setting is misused
Thanks for the comment. In the synthetic example, we focused on how the optimal feature set changes as a function of with held constant, rather than comparing the solutions with other explanation methods (which do not depend on these parameters). Instead, we compare to other methods in more real and high-dimensional problems, including images. There, we indeed study how sufficient and necessary the ``important features'' provided by the different methods are by measuring their respective and .
Concern 4: Shapley-perspective is disconnected
We are glad the reviewer enjoys the theoretical result of Theorem 5.2. The purpose of Theorem 5.2 is to motivate why minimizing is a good idea at all, and we see our theoretical result as one such important justification. We agree with the reviewer that this was perhaps not clear enough, and we will make sure to stress it in revision.
Regarding the implication and meaning of the result, consider the following.
Denote the partition of into two disjoint subsets, and define the characteristic function to be . Then, the following result holds.
Recall that a cooperative game is specified by a tuple and since can be partitioned into 2 sets ways, there are potential games. For every game, the Shapley value assigns an importance score to each of the two players (i.e. partitions) in a way that is fair (and satisfies other axiomatic properties). Note that the inequality above holds for all games; i.e. for all partitions of . Thus, in solving for the with minimal , one is identifying the game in which has the largest lower bound on its Shapley value. This result is interesting because it motivates minimizing through a game-theoretic interpretation: this is equivalent to selecting the game between players that maximize their difference of Shapley values.
We hope this clarifies the result, and provide a stronger motivation to the story in manuscript. We will stress on this clarification in the revised manuscript!
Thank you to the authors for replying to my points. Unfortunately, my concerns with and view of the paper still remain.
Concern 4: Shapley result are not well substantiated/integrated
Yes, I get the technical side of the Shapley result, which is why I appreciate the result very much (c.f. strengths). However, the current paper or your elaboration on this topic in your reply does not go far enough. You write (bold added):
Thus, in solving for the with minimal , one is identifying the game in which has the largest lower bound on its Shapley value. This result is interesting because it motivates minimizing through a game-theoretic interpretation: this is equivalent to selecting the game between players that maximize their difference of Shapley values.
Yes! But why is this interesting? Your reply (and the additions in your updated manuscript) stops at this point where I would like to get some more intuition/guidance what novel things this theoretical results brings to the table apart from simply motivating your work by bringing in Shapley. What does identifying the game in which has the largest lower bound on its Shapley value really do?
Concern 3: Missed opportunity with synthetic experiments
I see what you did with the synthetic setting. However, my point still remains: The paper would have benefited a lot from actually comparing different XAI methods that do not necessarily depend on your methods parameters in synthetic settings where you can control what a model is/should be doing. It would be very interesting to see how your proposal compares to established methods and weather it actually helps understanding something the established ones do not retrieve.
Instead, we compare to other methods in more real and high-dimensional problems, including images. There, we indeed study how sufficient and necessary the important features provided by the different methods are by measuring their respective and. There, we indeed study how sufficient and necessary the important features provided by the different methods are by measuring their respective , and .
In this setting, you can't control for anything and we cannot really judge weather the results obtained by your method are substantially better or correct than the other methods. This discussion also touches my concern 1 (limited contribution), where I still do not see that the XAI community clearly learns when this unification is superior/preferred over other methods. However, I really think that this can be achieved by a proper validation in small-scale/synthetic experiments rather than high-dimensional settings.
Thanks for the quick response! Here are our responses to your comments/concerns.
Shapley result is not well substantiated/integrated
-
The result motivates identifying a subset that is equally sufficient and necessary. Recall in our unified formula that controls the sufficiency vs. necessity tradeoff. Our result, , only holds for , meaning that only under balanced sufficiency and necessity does the unified approach align with the Shapley value. Since the Shapley value is the only solution concept that is fair (i.e. satisfying the key properties of efficiency, symmetry, linearity, and null player), our result indicates that balancing sufficiency and necessity indirectly yields such favorable properties.
-
The result also motivates identifying a subset that is maximally and equally sufficient and necessary. In the XAI community, it is well regarded that features or sets of features with high Shapley values are more "important" since they provide larger contributions. Our result illustrates that, in a setting with 2 players, if one is after an important player (one with high Shapley value) then identifying one that is maximally sufficient and necessary (low $\Delta_{uni}) is a good strategy, as this player will surely have a large Shapley value. The result also implies that sets that are only sufficiency or necessity may not be as important (measured via the Shapley value) because optimizing for only sufficiency or necessity provides a smaller lower bound on the Shapley value.
Missed opportunity with synthetic experiments
To address this concern, we conducted a new synthetic experiment which we strongly believe highlights: 1) when/how solutions to the sufficiency, necessity, and unified problems differ, and 2) how current post-hoc methods fail to identify features that are sufficient and/or necessary.
The experiment is the following:
We model features , where for . The remaining features and response follow:
where . For , the data-generating process is represented by the directed acyclic graph (DAG) shown below
with not connected to any variables. From the DAG, we can see that and . Thus, for and , the solutions to , , and with are:
In this experiment, we train a general predictor (a three-layer fully-connected neural network) to approximate and 1) validate that the sets listed above are the optimal solutions, and 2) demonstrate that common post-hoc interpretability methods do not recover these any of these sets.
Unfortunately, we cannot send figures through this forum but we highlight the main takeaways from the experiments, and we’ll include the figures in the supplementary material.
Validating the solutions
For , , and 100 samples , we compute solutions, denoted as to the sufficiency, necessity, and unified problem. We find that:
- For 95% of the samples in , , the solution to the sufficiency problem.
- For 60% of the samples in , , the solution to the necessity problem.
- For 92% of the samples in , , the solution the unified problem.
These results indicate that the solutions computed via an exhaustive search do typically retrieve the correct solutions (the minor discrepancies are due to being an approximation of ). More importantly, this setting is a clear example of when one would not be able to identify the set as the most important one unless you directly solve the unified problem.
Comparison with other methods
For our model and 100 samples we use Integrated Gradients, Gradient Shapley, DeepLift, and Lime to generate attribution scores. To identify whether these methods highlight sufficient and/or necessary features, and as done before in our manuscript, we perform the following steps on the attribution scores for a sample (so that the outputs of all methods are comparable)
- We normalize the scores to the interval [0,1] via min/max normalization.
- We generate binary masks by thresholding the normalized scores with thresholds .
- For , we compute , the hamming distance between and the true solutions to , , and
The main results from our analysis are the following:
-
There is no threshold in for which any method recovers the true solution to , . Furthermore, for the average hamming distance, , is for all methods, indicating that and disagree by at least one element.
-
There is no threshold in for which any method recovers the true solution to , . In fact for , the average hamming distance , is for all methods, indicating that and disagree by at least 2 elements.
-
For , integrated gradients and deeplift recover the true solution to , . However, for , the average hamming distance , is for all methods, indicating that and disagree by at least 2 elements.
We are currently in the middle of updating the manuscript to include this experiment, but we hope that the experiment and the results adequately address your concerns. Thank you so much for engaging with us, and we look forward to hearing if this clarifies your concerns.
The authors consider the problem of feature importance, i.e., quantifying the influence of different input features in the context of supervised learning, which has become quite popular in explainable AI in the recent past. They propose new measures of sufficiency and necessity of feature subsets, as well as a convex combination between the two. They also consider the optimisation problem of finding (small) sufficient/necessary feature subsets.
优点
Interesting and up-to-date topic, well-written paper, thorough experimental study.
缺点
The first problem I have with this paper is that I find the definition of necessity flawed, or at least not very meaningful. The definition of sufficiency says that a feature subset S is sufficient if the function f projected to S remains eps-close to the original function using all features, which does make sense. Naturally, then, I would have expected that a subset S is called necessary if its complement is not sufficient. At least, this is the common duality between necessity and sufficiency/possibility also found in other branches of the literature (e.g., in modal/possibilistic logic).
Instead, the authors call a subset S necessary if f projected to the complement of S remains close to the default (average) prediction with no features. First, as already said, this does not establish a "duality", but apart from that, this definition is very questionable by itself. It somehow suggests that staying close to the default prediction is something bad, while moving away from it is good. Besides, it also leads to formal problems. For example, suppose that f is indeed a (close to) constant function. Then, according to the definition, all feature subsets are necessary, which is counter-intuitive.
The pathology outlined in the beginning of Section 4 is also a consequence of the flawed definition, I would say. Here, a feature subset is sufficient, but at the same time, its complement "contains important features". Again, this is completely counter-intuitive. How can a subset be sufficient if it misses important features?
Why is super-sufficiency an interesting property? Isn't it expected that adding more features will keep sufficiency? More interesting would be minimality: S is minimally sufficient if no feature can be removed from S without losing sufficiency. Analogously for super-necessity.
The "unification" is merely a convex combination of the two measures of sufficiency and necessity. Such convex combinations are routinely used in (multi-objective) optimisation, but why should we call them unification? Sure, one obtains both measures as special cases, for alpha=0 and alpha=1, respectively, but then it's more a generalisation than a unification.
The two perspectives in Section 5 are strange and in a sense again somewhat misleading. First, the notion of conditional independence is not defined in the standard way. Normally, conditional independence is a relation on random variables (used in probabilistic graphical models, for example). But why should we speak of conditional independence in the case of (5)?
The connection to the Shapley value is flawed as well. Normally, each feature is treated as a player and assigned a Shapley value. In Theorem 5.1, however, an entire feature subset S is treated as a single player, and its complement as a second player. Why should one consider such a partition as a game, and what does it help? A different game is then needed for every player. How do we connect this to the standard Shapley value? Eventually, with only two player S and S_c, the entire game is specified by the four values v(\emptyset), v(S), v(S_c), v([d]). These are also the values/approximations looked at in the definition of sufficiency/necessity, so it's not very surprising to find a relationship here.
问题
See above.
Concern 1: Flawed definition of necessity
We appreciate the comments. If we understand correctly, the reviewer is saying that
- Our notion of sufficiency is appropriate, i.e. is - sufficient if .
- Given this notion of sufficiency, necessity should be defined by saying that necessary if its complement is not sufficient, i.e . Or, formalizing this a little, the reviewer's proposal is that is -necessary if .
First, we note that our definition of necessity implies the one suggested by the reviewer: if then where . This is not surprising because if is close to , then it is far from as long as and are different.
Second, and more broadly, the reason we define the sufficiency and necessity of using the quantities and is because we want our notions to reflect the natural idea that, among all subsets , the complete set should be the maximally sufficient and necessary subset. This is indeed the case with our definitions: for , and and so is -sufficient and necessary as desired.
We hope this clarifies why we define sufficiency and necessity in this specific way, and we'd be happy to elaborate further.
Concern 2: Flawed definition of sufficiency
Thank you for this comment. We respectfully disagree, and we believe that there is a simple example that demonstrates that there is no contradiction in having two disjoint minimal sufficient features: Consider a scenario where the label is determined by the presence of certain features (as it happens in the brain CT example in our experimental section, Sec. 7): the presence of any individual hemorrhage in the scan is sufficient to classify the scan as "positive". Assume there are two hemorrhages in the scan, and , each of ``size'' of pixels (for simplicity of the argument). Then, any one of them individually is the smallest sufficient subset (of size ); i.e. . The same is true in many other settings where the presence of certain features determines the outcome of the task. Thus, in general, there is no reason why features in the complement of minimal sufficient ones can't provide useful information.
We hope this clarifies the confusion, and we'd be happy to elaborate further.
Concern 3: super sufficiency/necessity
While not immediately obvious, one can verify that super-sufficiency does not always hold, as conditioning on a superset of a sufficient region can significantly alter predictions. For example, a predictor might classify a small fur region of a dog image as a bear, but will eventually change towards predicting the presence of a dog as the region grows and reveals other dog-like features.
On the other hand, we indeed address the minimality of both sufficiency and necessity through the parameter , which controls the size of feature subsets. Smaller emphasizes identifying the smallest sufficient and necessary sets.
Concern 4: Use of "unification"
This is valid point. We find the term "unified" appropriate because as the reviewer states, our formulation is a combination of sufficiency and necessity. Indeed, there's nothing particularly revolutionary about this convex combination of both objectives. We'd be happy to rename the resulting problem if the reviewer has a specific suggestion in mind!
Concern 5: Confusion with conditional independence perspective
Thanks for bringing up this point. The reviewer is correct that conditional independence is a relation on random variables. To be precise for and , is independent of conditional on if, for all values of , we have . Informally, our result states that
Since this result pertains to conditional expectations for a fixed realization of , this is a local conditional independence relation on means. However, it is still a conditional independence relation, albeit weaker than the standard notions. We apologize for loosely using the term conditional independence, which we will correct in the revised version.
Concern 6: Confusion with Shapley perspective
We kindly disagree that the connection to the Shapley value is ``flawed''. The reviewer is correct in stating that traditionally every feature is treated as a player and Shapley values are computed for each feature. However, this is not necessarily "standard": in other work, see [1], Shapley values for sets of features are often defined, too. The motivation behind this is twofold. First, in most settings, a single feature barely contributes to a prediction (most notably, pixels in the case of images) and in reality, models often use sets of features (that may interact synergistically) to generate a prediction. Second, when considering sets of features as players, the computation of the Shapley value becomes tractable. Our result simply demonstrates that minimizing the unified approach is equivalent to identifying the two-player game for which one player has maximal Shapley value. While this result is simple and different from the traditional setting of -players game, it still provides a precise motivation for speaking for both sufficiency and necessity via game-theoretic justifications, which has never been shown before.
[1] "Feature importance: A closer look at shapley values and loco," Isabella Verdinelli, Larry Wasserman
This paper introduces a novel approach to sufficient and necessary subset explanations, proposing a method to identify subsets along a (symmetric) spectrum of necessity and sufficiency. The authors present theoretical results illustrating the properties of this unified approach and its connections to established importance measures, such as conditional independence and Shapley values. They demonstrate the effectiveness of their approach through experiments on both tabular and image data.
优点
- Clarity and Novelty: The paper is well-written and extends existing work with a creative new approach to subset explanations.
- Theoretical Contributions: The authors provide a thorough theoretical analysis of their method, detailing its properties and connections to existing notions of feature importance.
- Empirical Evaluation: The framework’s applicability is demonstrated through diverse experiments, supporting the practical relevance of the proposed subset explanations.
缺点
-
It would be helpful if the main paper, prior to the experiments, provided a more detailed overview of how solutions are computed.
-
In the experiments section, you mention using a relaxed optimization approach for image data. Does this imply that, for tabular data, the exact solution is found by examining all subsets? Additionally, is there any investigation into the guarantees or potential limitations of the relaxed approach compared to the original problem?
-
In the abstract (Line 15), "feature importance" is used to describe sufficient and necessary sets. However, "feature importance" may not be the most precise term for these concepts.
-
The implications of Theorem 5.1 require further elaboration. Comparing different subsets feels akin to comparing different players in different games or players competing in their own games. If I understand correctly, searching within P_{uni} involves finding the feature subset with the highest Shapley Value in its own game, evaluated against its complementary subset. This paragraph could benefit from further clarification to ensure readers fully grasp its significance.
问题
-
Why did you choose to use marginal distributions in defining sufficiency (L89)? Most literature on sufficient explanations relies on conditional distributions, and even in some experiments, you use conditional distributions.
-
What advantage does controlling minimality with a parameter like provide, rather than defining minimal sufficiency directly? Calibrating could be complex, especially as the existence of solutions for specific values is uncertain. If this choice were removed, how would it impact your approach?
-
Given the focus on local prediction, why not define necessity as an average out of a subset that diverges significantly from the current prediction of the considered observation (e.g., ), instead of using the average prediction? This definition might align more closely with interpreting necessity as the minimal feature set required to maintain the current prediction of the considered observation.
-
In all your results, would it not be necessary first to assume the existence of a solution to the problem given ?
-
Your unifying solution is defined as a combination of sufficient and necessary subsets using weights and . Why not allow for any convex combination where ? Would that significantly impact your results?
Missing Citation
In your related work on sufficiency, you may want to reference [1], a follow-up to Wang et al. (2021) published at NeurIPS 2022, which proposes a more tractable approach for tree-based models:
[1]: "Consistent Sufficient Explanations and Minimal Local Rules for Explaining Any Classifier or Regressor," Salim I. Amoukou, Nicolas J.B Brunel, NeurIPS 2022.
I am willing to increase if my questions and weakness comments are addressed.
Concern 1: Detailed overview of how solutions are computed
We agree with the reviewer. In the revised manuscript, we will include a section before the experiment that details how approximate solutions are computed.
Concern 2: Solutions for tabular vs. image settings
To answer the first question, the reviewer is correct. In the tabular example, exact solutions were identified by examining all subsets of a fixed cardinality . For the second question, the relaxed approaches we use are slight variations of methods introduced in [1,2,3,4]. None of these works provide any theoretical guarantees but the methods have been demonstrated to work well in practice. The main benefit of the relaxed approach is that it is tractable since it allows for the use of gradient based methods for optimization. We will definitely include a section before the experiments that comment on the these approaches and their benefits/limitations.
[1] "Interpretable Explanations of Black Boxes by Meaningful Perturbation," Ruth Fong, Andrea Vedaldi, ICCV 2017 [2] "Understanding deep networks via extremal perturbations and smooth masks," Ruth Fong, Mandela Patrick, Andrea Vedaldi, ICCV 2019 [3] "Cartoon explanations of image classifier," Stefan Kolek, et. al. ECCV 2022 [4] "Model Interpretability and Rationale Extraction by Input Mask Optimization," Marc Brinner, ACL 2023
Concern 3: Abstract line 15
Thank you for pointing this slight issue. We have reworded the abstract as follows:
"To address this, we introduce and formalize two precise concepts—sufficiency and necessity—that characterize how sets of features contribute to the prediction of a general machine learning model."
We think it is more useful to measure the sufficiency and necessity of sets of features (rather than the importance scores of individual features) because in most settings a single feature does not contributes to a prediction (most notably, in the case of images). In reality, models often use sets of features (that may interact synergistically) to generate a prediction. We hope the reviewer finds this edit more precise, which we have incorporated to our revised version.
Concern 4: Implications of Theorem 5.1
We appreciate the comment, and we will further elaborate on Theorem 5.1. In short, the reviewer is correct. Denote by {} the partition of , and define the characteristic function to be . Then, the following result holds.
A cooperative game is specified by a tuple and since can be partitioned into 2 sets ways, there are games. For every game, the Shapley value assigns an importance score to each of the two players in a way that is fair and satisfies other desirable properties. For each game, the above inequality holds. Thus, a clearer way to interpret the result is that, in solving for the with minimal , one is identifying the game in which has the largest lower bound on its Shapley value. This result is interesting because it motivates minimizing through a game-theoretic interpretation by selecting the game with one player with the largest large Shapley value. We will stress on this clarification in the revised manuscript.
Concern 5: Notation in line 89
We thank the reviewer for pointing this out. This is actually a typo and we did not indent to use the term marginal distributions. What we intended to say is
To define feature importance precisely, we use the average restricted prediction,
where is fixed, and is a random vector drawn from an arbitrary reference distribution , that may/may not depend on . For example, two commonly used reference distributions are the marginal distribution and conditional distribution .
The revised manuscript incorporates this correction.
Concern 6: Advantage of
First, note that some constraint on the cardinality of the subsets is needed -- otherwise, without it, the solution to the problem is achieved at , i.e. using all features.
Perhaps the reviewer meant to switch the objective of the optimization problem with the constraint on its size; i.e, posing a problem like s.t . This problem minimizes the cardinality of directly, but it has the equivalent difficulty of setting the appropriate parameter . Lastly, one could consider in the problem above, but this is not very useful for real settings since exact sufficiency (or necessity) might not be achieved.
Concern 6: Notion of necessity
We appreciate the question. We agree with the reviewer that the definition of necessity should align with the the idea that "a set is necessary if we cannot generate the original prediction without it, i.e. ." More formally, is necessary if for some . Note, while our definition seems to differ from this, it is in fact more general in that it implies this condition. If then where . Intuitively, if is close to then it is far from as long as and are different.
The reason we define necessity differently is because we want our notion of necessity to align with the intuitive principle that, "the set should be maximally sufficient and necessary for " With our definition, for , we have , indicating that is -necessary (maximally necessary) as desired. In the revised version we further elaborate on this in Section 2 and provide a detailed comparison of our notion with classical definitions, along with its advantages, in the Appendix.
Concern 7: Existence of solution for given
Recall the unified problem is to minimize subject to . For any , we can always minimize . Now, for very small , the minimizer may not be a good one in the sense that may not be small, meaning is neither very sufficient nor necessary.
If the reviewer is referring Lemma 4.1, Theorem 4.1 and Corollary 5.1, in these results we always do make a statement of the form "let be a solution for ...", and so here we are assuming we have a solution. Note here we are not assuming a solution exists (in fact it always does) but instead assuming we were able to solve for it. We hope this adequately addresses your question but, if not, please feel free to elaborate.
Concern 8: Any convex combination using and
What the reviewer suggested is analogous to what we do: Note that you can define and . Thus, one can use these weights ( and ) and they will always satisfy . Our definition just simplifies this process by using a single parameter that controls the trade-off.
Concern 9: Missing Citation We thank the reviewer for bringing this to our attention! We will include the new reference in the revised manuscript.
To address vague notions of feature importance in many XAI methods, the paper introduces formal definitions for sufficiency and necessity for local feature-based explanations. The authors further present a unified notion of sufficiency and necessity through a joint convex optimization problem to generate explanations that are both sufficient and necessary. The empirical results show that while existing feature importance methods can identify sufficient features, they fall short in finding necessary features.
优点
The paper is mostly easy to read, every definition and theorem/results are relevant to the discussion and do not feel like the authors have included them for the sake of including math and notation. I also appreciate the authors for recognizing the need for auxiliary user studies in their limitations to highlight the fact that theoretical desiderata does not necessarily translate to real world impact/performance.
缺点
Having said that, the paper could improve on a couple of things:
- Natural language explanation of "sufficiency" and "necessity" in Section 2 (I know this is in the introduction). Since they are the central concepts of the paper, it would be beneficial to really drive home what these mean in plain English for the reader.
- The experiments section (detailed below)
The experiment setup was difficult to follow, a lot of missing details that I had to assume. For example, I couldn't find what the metric is defined (I am assuming that is the cardinality of the resulting for each method at thresholds). Similarly, the role of the threshold was unclear. I am assuming features with attribution/importance higher than will be included in , but the paper does not explicitly mention that. It might be worth answering what the effect of adjusting is.
Moreover, motivation for the tabular dataset experiment is very weak. What is the significance of investigating stability (and a synthetic dataset)? I know in the introduction, the authors ask the question "when do necessary and sufficient sets differ"? Perhaps this is something that the authors should mention in this section too. Note that this is not the case for the image classification experiment, which explicitly outlines its purpose (line 346).
Most importantly, I, as a reader, am unsure of the main takeaway from the experiments apart from the technical contributions of the paper at work. For example, what should we make of the result that necessary explanations are sufficient? This is an important question to answer since the abstract mentions that the paper demonstrates how strictly sufficient and/or necessary explanations fall short in providing a complete picture.
问题
Questions:
- Line 89: Is a typo? I'd imagine one of them should be from the complement?
- Theorem 4.1 shows existence of , do you have any theoretical results on uniqueness? Or can there be a unique solution in the first place?
- What happens when ? (i.e. prediction is close to baseline)
Notes:
- I would suggest following a more standard notation for set complements: , though I know having a superscript in a subscript may not be ideal
- Figure references in Section 6.2.2 "Sufficiency vs. Necessity" paragraph may be incorrect?
- Line 414: "demonstrateLemma"
- Line 424: is a really big typo. Only found out when I looked at A.2.
Concern 1: Natural language explanation of "sufficiency" and "necessity" in Section 2
We agree with the reviewer. We have reworded the first paragraph of section 2.1 to clarify the meanings of sufficiency and necessity with the following text:
"We now present our proposed definitions of sufficiency and necessity. At a high level, these definitions were formalized to align with the following guiding principles:
- is sufficient if it is enough to generate the original prediction, i.e. .
- is necessary if we cannot generate the original prediction without it, i.e. .
- The set should be maximally sufficient and necessary for .
The principles P1 and P2 are natural and agree with the logical notions of sufficiency and necessity. Furthermore, because the full set of features provides all the information needed to make the prediction , it should thus be regarded as maximally sufficient and necessary (P3). With these principles laid out, we now formally define sufficiency and necessity."
We hope this clarifies our definitions. We have also added further clarifications after the definitions of sufficiency and necessity in a revised version.
Concern 2: Experimental Setup
We apologize for the lack of clarity. 1) The refers to the relative cardinality of (|S|/d). 2) The threshold is used to generate the subsets and . For a fixed , all normalized importance scores higher than are included in and those lower than are not. As a result, controls the sensitivity of the choice of reporting important vs un-important features, for all methods.
Concern 3: Tabular Data Motivation
Thanks for pointing this out. Both tabular examples aim to demonstrate how optimal sets may change as we vary the levels of sufficiency and necessity we require. We argue this is an important question since, if they were very stable, one shouldn't be too concerned with the specific trade-off between sufficiently and necessity. In short, our experiments demonstrate that this is not the case: it is evident that as we demand higher levels of sufficiency (via increasing ) the features in the optimal solution are constantly changing (measured via the Hamming distance to optimal necessary ).
To make this point clearer, we have added a small description before the tabular examples that clearly outlines the objectives of these experiments. Thanks for the suggestion!
Concern 4: What are the experimental takeaways
This is a great point and we would be happy to explain. We believe the experimental results yield a few key important conclusions:
-
Sufficient and necessary sets differ: One and other notions of importance convey different observations about the response of a model (on sufficiency and necessity of the studied features, respectively). Our experiments demonstrate that, while one could have a situation where necessary and sufficient features coincide, in most common experimental settings and common post-hoc methods, these two notions differ.
-
There exist domains where the sufficient sets are subsets of the necessary sets (our image classification settings highlight this). As a result, a minimal sufficient explanation will not highlight necessary features, thus falling short in reporting all important features more broadly. Without our results (cite which one in particular) one would not be able to draw this conclusion.
-
Our definitions allow us to conclude that many current post-hoc methods identify small sufficient subsets, but not necessary sets. This finding is important as it allows us to better understand the limitations of current methodology (and to propose our unification strategy as a solution).
Concern 5: Theorem 4.1 Uniqueness:
We don't have a complete answer to the uniqueness of , but we can provide a partial answer: Arguments for uniqueness are best understood when we consider the expected predictor under the true conditional distribution, i.e. and . From our results in section 5, a solution to the unified problem then satisfies:
If the joint distribution satisfies the Markov properties, then it is known that is unique [1]. However, this known result is strong because it is a global characterization (i.e. holds for all ). Nonetheless, we believe this is a direction worth exploring, and we will incorporate it into our manuscript.
Additional Question: Is a typo? I'd imagine one of them should be from the complement?
Yes, the reviewer is correct. The equation should be . Thanks for pointing this out!
[1] Pearl, Judea. Causality: Models, Reasoning, and Inference. Cambridge University Press, 2009.
Additional Question: What happens when ? (i.e. prediction is close to baseline)
We appreciate the reviewer asking this question. For simplicity, let . Then, if we have sample for which the prediction of the model is equal to the baseline prediction, this indicates that there is nothing particularly "interesting" about this . More precisely, in these cases, the empty subset of features will be both sufficient and necessary. To see this, note that is the smallest sufficient subset because by assumption which implies is -sufficient. Likewise, is the minimal necessary set because , and so by assumption, which implies that is -necessary. Thus, for , we see that is approximately a good sufficient and necessary set, indicating that there are no "distinctive" features in that generate the prediction .
We will incorporate this discussion to our manuscript.
Additional notes/small errors
We appreciate the reviewer for pointing out grammar and notion errors. We have corrected them in the revised version.
I thank the authors for their response! I still have two concerns:
- Definition of Necessity
- Main Takeaway of the Paper
Definition of Necessity
The reason I asked about what happens if the prediction is close to the baseline is that I thought the result did not make much sense. As the authors pointed out in their response, is the minimally necessary set according to the paper's definition. However, there also exist subsets of that are necessary as well (most notably itself).
From the authors' response to Reviewer 1XKc, it seems like being necessary is intentional. Why is this a "natural" idea? To me, it seems to go against the logical notion of necessity because it is possible to obtain the prediction without .
In addition, as noted in the updated manuscript, the definition of necessity is analogous to if and differ. Why is this a reasonable assumption?
Main Takeaway of the Paper
I understand that (1) sufficient and necessary sets differ, (2) there are domains where sufficient sets are subsets of necessary sets, and (3) existing explainability methods don't return necessary sets. I was hoping the authors would go beyond this and discuss the significance of these results.
- Why should we care that sufficient and necessary sets differ?
- What are the consequences of current explainability methods returning small sufficient sets? (i.e., what effect does this have on downstream tasks)
- Should we always look to return necessary sets? Is the answer domain-dependent?
Unfortunately, there were no meaningful changes to Section 7 in the updated manuscript.
Thank you for the reply! Our responses to your concerns are below:
Definition of Necessity
We think there is a misunderstanding here. In our response, we stated is the smallest necessary set when . In general, is not the smallest necessary set.
We also want to state that the complete set being necessary is not intentional. We strongly believe a notion of necessity should reflect that is not only necessary but maximally necessary. This is because, for a sample and model prediction , if we do not have all the feature values , then it should not be possible to generate the original prediction value . If this were the case, then this implies , which further means that is both sufficient and necessary.
Lastly, regarding the statement that and differ. We believe that one should be interested in generating sufficient and necessary explanations for such that these quantities differ. Otherwise, as stated earlier, is both sufficient and necessary, implying there are no distinctive features in .
Main Takeaway of the Paper
Our paper's main takeaway is that sufficient and necessary explanations can often provide an incomplete picture of which features are important. They both have their utility, but one should not expect a sufficient set to be necessary and vice-versa. In turn, if one desires both, our unified approach provides such an explanation. The paper then provides a theoretical analysis of unified solutions along with different interpretations of sufficiency and necessity through notions of conditional independence and game theory. We then demonstrate how current methods often fall on the sufficiency side of the sufficiency-necessity axis.
We believe our finding that current methods return small sufficient sets is important because it informs us as a community about what properties the explanations have. Our results highlight that many methods will highlight the small set of features that is enough to reconstruct the prediction and, in turn, may not highlight all the important features. Take, for example, our CT scan example. Our results show that most post-hoc methods will highlight a single hemorrhage but not all hemorrhages. Yet, we also identify that all hemorrhages are necessary for the prediction. This suggests that these common explainability methods look for the "smallest" or "simplest" explanation as opposed to the "complete explanation," which we believe is very interesting. This can be useful to individuals using these methods because they know the features the methods do not highlight are not necessarily unimportant.
We believe that one should not always look for necessary sets. We think the choice of sufficiency vs. necessity is domain-specific, and both have their benefits. A simple example highlighting their utility is in loan approval. Suppose a bank uses a model to approve loans based on features like income, credit score, and employment history, and we identify that a high income and good credit score are sufficient for loan approval. This is useful as we can explain to applicants which factors guarantee loan approval, enabling better transparency. On the other hand, suppose we identify that a high credit score is necessary. This is equally important information to applicants because it informs them that, regardless of all the other details on the application, if the credit score is low, the applicant will never be approved for a loan. Here, a necessary explanation provides an actionable item for the applicant: they should first improve their credit score. Thus, a necessary explanation can also be useful. Overall, the choice between which one is more desirable is domain-specific. However, if you are unsure, then generating an explanation that is both, as our unified approach does, could be the safe course of action.
Lastly, the changes to our paper, namely a new synthetic experiment, are located in the appendix and not directly in Section 7. We also have written a global note detailing this experiment for all reviewers to read. We apologize for not making this clear.
We would like to thank all the reviewers for the insightful questions and comments about our work. We have uploaded a revised version of the manuscript that includes changes suggested by the reviewers (highlighted in blue). In short the major changes are the following.
- We have rephrased parts of Section 2 to clearly motivate our proposed definitions of sufficiency and necessity. In this section we provide simple and intuitive "guiding principles" that motivated our definitions. We introduce the following text at the beginning of section 2.
Definitions We now present our proposed definitions of sufficiency and necessity. At a high level, these definitions were formalized to align with the following guiding principles:
P1. is sufficient if it is enough to generate the original prediction, i.e. .
P2. is necessary if we cannot generate the original prediction without it, i.e. .
P3. The set should be maximally sufficient and necessary for .
The principles P1 and P2 are natural and agree with the logical notions of sufficiency and necessity. Furthermore, because the full set of features provides all the information needed to make the prediction , it should thus be regarded as maximally sufficient and necessary (P3). With these principles laid out, we now formally define sufficiency and necessity.
- The second change is we have included a short section before the experiments that details the methods we used to generate exact or approximate solutions to the sufficiency, necessity, and unified problems in the experimental section.
We believe these changes have greatly improved our work and we appreciate all the help! We encourage the reviewers to take a look and to let us know if they have any additional questions/concerns.
We have noticed that one of the main concerns raised by many reviewers was that the synthetic experiments were not comprehensive enough and/or lacked comparisons with traditional methods. To address this, we conducted the following experiment that we will add to the paper (Note: For simplicity, we use , , and to refer to , , and .)
The main findings are the following:
- The optimal solutions to , , and for a fixed need not be the same.
- The sets recovered by many common feature attribution methods are not optimal solutions to the unified problem
As a result, these findings highlight the utility of the unified framework because our approach can recover a small sufficient and necessary set, something other methods are not capable of.
The experiment is the following,
We model features where for . The remaining features and response follow
where . The entire data-generating process is represented by the directed acyclic graph (DAG) shown below
with the remaining features not connected to any other variables. In this setting, and , thus for solutions to and are
Furthermore, the distribution is a multivariate normal. As a result, we can exactly compute for all . With this setup, we solve , , and for and accomplish the following
- We validate that the optimal solutions for and are and respectively
- Demonstrate that there is a subset of samples for which minimizes the unified objective, , for
- We demonstrate that, for , many common post-hoc methods do not identify as an important set.
Validating solutions
For a holdout set of 1,000 points, we perform an exhaustive search to generate solutions to and , which we denote as and respectively.
For , this entails calculating for all and identifying the set for which this is minimal. Similarly, for we pick the that minimizes . Upon doing so, we identify that for all 1,000 points, and as expected
Solution to unified problem
For this holdout set, we also perform an exhaustive search to generate solutions to for . In doing so, we identify that, for , there is a subpopulation of samples for which is nearly sufficient and necessary and accordingly, the optimal solution to . We denote this subset of samples as .
Comparison with post-hoc methods
Following the reviewers’ recommendation, for every in , we use the following attribution methods to compute importance scores for every feature
- Integrated Gradients
- Gradient Shapley
- Deep Lift
- Lime
- The Leave Out Covariate (LOCO) value,
- The Shapley value, for 3 different contribution functions,
- , the negative loss in information when using features in
- , the gain in information when using features in
We select the three features with the highest scores for each method to create a set . In doing so, we come to the following conclusions:
Conclusion 1: For Integrated Gradients, Gradient Shapley, Deep Lift, Lime, and LOCO, for all , we have . In simple terms, all of these methods assign high scores to the features that comprise . This implies the rankings of scores generated by these methods can be used to deduce the optimal sufficient set. Identifying why this is the case for most of these methods is a matter of future work. For LOCO, this is not very surprising. Since , then for any not in , we have
In other words, the LOCO parameter for all features not in will be 0. The remaining features, those in , will have non-zero scores, and so selecting the features with the top scores is equivalent to identifying features that makeup .
Conclusion 2: Using Shapley values, for the samples in , approximately 70% of samples have . For the other 30%, or In other words, Shapley often assigns high scores to the features comprising . In conclusion, the set created by picking the features with the highest Shapley values can often, but not always, be used to deduce the set that is a solution to the unified problem (for ). This finding is interesting, as it suggests that combining information about how a feature contributes to all subsets , as the Shapley value does, is equivalent to measuring whether a feature is a member of a set that is both sufficient and necessary. Exploring why this happens is a matter of future work.
I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.