PaperHub
6.7
/10
Poster3 位审稿人
最低5最高8标准差1.2
7
5
8
3.3
置信度
正确性3.7
贡献度3.0
表达3.3
NeurIPS 2024

Improving Decision Sparsity

OpenReviewPDF
提交: 2024-05-15更新: 2024-11-06

摘要

关键词
sparse explanationdecision sparsity

评审与讨论

审稿意见
7

The paper extends the notion of decision sparsity called the sparse explanation value (SEV). Cluster-based and tree-based SEV are introduced, as well as some algorithms to optimise the decision sparsity are considered. The core of the paper -- SEV -- is defined as the number of factors that need to be changed to a reference values in order to change the decision.

优点

The problem considered is important for real (e.g., medical, criminology applications). The example (Table 2) is very helpful to understand the problem. The approach is mathematically sound and in general well-described.

缺点

The method is promising but its current version is hardly scalable.

问题

Line 85: "humans have no intuition for why a point belongs to one class or the other". I cannot completely agree with the statement, and it would be helpful to provide some examples. For me, on the contrary, medical doctors have often an intuition how to classify a patient, however, they do not always know how to explain their intuition.

How the sparse explanation for the sample x (mentioned on line 171) is formally defined? Is the number of features of x (original) and x (sparse) different?

局限性

The method is not scalable.

作者回复

Thank you so much for your review! We really appreciate it! See below for our response to your questions and concerns.

Line 85: "humans have no intuition for why a point belongs to one class or the other". I cannot completely agree with the statement, and it would be helpful to provide some examples. For me, on the contrary, medical doctors often have an intuition how to classify a patient, however, they do not always know how to explain their intuition.

There’s a good example from Figure 3 in the paper of Keane [1] where he shows images on either side of the decision boundary that look visibly identical (we have provided the figure in the attached pdf file). There’s just no intuition why the point belongs to one class or another. We agree with you that for some decisions that, particularly where there are discrete variables, the decision boundary could be clearer. We’ll adjust the wording of that sentence. However, even for medical decisions by doctors, it’s not also clear how to diagnose someone, which is why AI can be helpful. (We have multiple projects on exactly this topic.)

[1] Delaney, Eoin, et al. "Counterfactual explanations for misclassified images: How human and machine explanations differ." Artificial Intelligence 324 (2023): 103995.

How the sparse explanation for the sample x (mentioned on line 171) is formally defined? Is the number of features of x (original) and x (sparse) different?

The number of features of the original query x and the sparse explanation x are the same. The sparse explanation for a specific query is defined as a p-dimensional vector that has changed the smallest number of features from its query value to the reference value to flip the prediction, while the other features remain unchanged. Four examples are shown in Table 1. The explanations (lower rows) are the same size as the Query (top row). The gray colored values are the unchanged query values, while the black ones are changed. This is the same type of definition as other counterfactual explanation methods.

The current version is hardly scalable.

SEV measures decision sparsity, so it is calculated for each point separately, therefore in practice, when we provide an SEV for one loan applicant at a time, the calculation is very fast for that applicant (fraction of a second to a few seconds, more detailed calculation time is shown in Table 7-9 in Appendix H). If you want to calculate SEV for all points at once for a huge dataset, like in the experiments of our paper, it can be computed in parallel for each point.

If you want to compute it for each point and for a large number of features, we could apply the gradient-based method from our paper first in order to reduce the mean SEV of the model to be close to 1 without sacrificing the model performance. This could speed up SEV calculation for all points.

Thank you so much once again for your review!

评论

I acknowledge the rebuttal. Thank you for the detailed answer.

审稿意见
5

The authors build on top of the Sparse Explanation Value approach by Sun et al and provide improvements in terms of closeness and credibility.

优点

Sensible problem, well presented solution.

缺点

The main limitation I can see in the work is its very incremental nature with respect to the approach by Syn et al.: going from one negative reference point to a set by clustering negatives (cluster-based SEV) is a trivial extension, while tree-based SEV only works if the underlying model is (or can be approximated as) a decision tree, substantially restricting the applicability of the approach.

Additionally, the superiority of sparsity with respect to distance in terms of acceptability for humans is intuitive but not always guaranteed. When dealing with recourse, for instance, which is the setting used in the experiments, the main problem is the cost of the change, and slightly modifying two features could be less expensive than modifying a single one by a larger value. This should be better discussed in the paper.

问题

How can you turn a DT leaf into a reference point when dealing with continuous features? what is the actual value of the continuous variables in the leaf?

局限性

The approach is very incremental, and the main extension substantially restricts the applicability of the method.

The rebuttal of the authors did shade additional light on the novelty of the contribution.

作者回复

Thank you so much for your review! We really appreciate it! See below for our response to your questions and concerns.

The work is its very incremental nature with respect to the approach by Sun et al. cluster-based SEV is a trivial extension, while tree-based SEV only works if the underlying model is a decision tree, substantially restricting the applicability of the approach.

In terms of your question about the contribution of our paper, we understand that some of the ideas seem trivial (going from one negative reference to a set), and that you think the extension to decision trees is not broadly useful because it applies only to trees. Indeed the idea of going from one reference to many is easy. Working out how to do it, and to add credibility to the computation, is not. In terms of trees, there are many papers at NeurIPS each year that are only focused on decision trees! Trees are some of the most popular algorithms for interpretable machine learning, since the 1960’s. The fact that we have a substantially better way to define and compute SEV for trees is important for a lot of applications! Previous studies have also taken tree-based models for separate discussion. For instance, there exists a lot of papers discussing about fast shap values computation in tree-based models, also known as treeSHAP.

While it’s tempting to believe ideas are incremental in retrospect, they are often not obvious until after seeing them. Many of our ideas are not obvious at all, for instance looping through all good sparse decision trees to find one that optimizes SEV + accuracy, which makes what would have been a practically impossible computation now take seconds. Many important ideas in ML seem obvious in retrospect, e.g., placing skip connections in neural networks. If you glance through the orals in NeurIPS from last year, they are almost all variations on existing topics. Our paper is only the second paper on SEV, which is a new notion of sparsity, and it generalizes the definition and works out how to optimize it and make its computations credible.

We have actually made a lot of improvements and generalizations with respect to the original SEV paper after generalizing the framework of SEV to make it support instance-wise reference selection.

  • We proposed cluster-based SEV and its variants in order to handle the two objectives of instance-wise selection: the flexible reference solves the issue of higher SEVs (cluster SEVs are larger than standard SEVs), and the credibility approach enables the explanations to be located in a high density region. The credibility setup is definitely not obvious.
  • For tree-based models, we introduced SEV-T which directly uses negative leaf nodes as the reference points, and introduced a fast calculation procedure for SEV-. That fast procedure is definitely not obvious and is comparable to content within an algorithms course. The bound on credibility in SEV- in Theorem 4.3 isn’t obvious either - tree-based SEV- computations are always sufficiently credible.
  • We generalized the original gradient-based methods and introduced new search-based methods specifically for tree-based models in order to optimize the specific model type to have low SEV. The methods for optimizing SEV are not obvious, particularly Sec 5.2 where we optimize SEV by scanning all good trees using the new Rashomon set algorithms and finding the one with the best SEV. (In contrast, think about how many papers optimize some combination of ell1ell_1 loss with accuracy, or optimize neural networks using some variation of gradient descent!)

The superiority of sparsity with respect to distance in terms of acceptability for humans is intuitive but not always guaranteed. When dealing with recourse, for instance, slightly modifying two features could be less expensive than modifying a single one by a larger value.

Remember that SEV is not recourse, it is a measure of sparsity only. That said, it’s possible to change the metric to a recourse metric if desired. We did use multiple distance metrics. In our experiments we also used the LL_\infty metric to evaluate the cost of change, which would handle the case you mentioned - when we care about how much we change the features. Table 6 in Appendix G provides the mean LL_\infty for each explanation. Based on Table 6, SEV-C achieves better sparsity, relatively low LL_\infty change (compared to DiCE under the same sparsity level), and meaningfulness (high median log likelihood) simultaneously, while for the other methods except DiCE, they use almost all features in their explanations, which is not only too complex to interpret, but also is not actionable in practice, and would yield worse recourse scores. This is not the case of modifying two features instead of one: it is modifying all instead of one. So already we would be better in terms of most recourse loss functions. It is possible to specialize the computation to a particular recourse loss, but that’s not our goal here, we’re concerned with sparsity of the explanation, not other costs. We will put a note for future work by others.

How can you turn a DT leaf into a reference point when dealing with continuous features? What is the actual value of the continuous variables in the leaf?

Interestingly, the reference can be any point x within the leaf. So if the leaf is defined by x1>5x_1>5 and x3>0x_3>0, then any point with those conditions is a viable reference. For a query, the algorithm needs to flip its feature values to satisfy those of the leaf conditions to make an opposite prediction.

Since you can choose any point in the leaf as a reference value, you could choose the median/mean values of points in the leaf. That choice won’t influence the fast calculation of SEV-T, so the user can choose any actual value inside the leaf that they believe is most meaningful as the leaf’s reference.

Thank you so much once again for your review!

评论

Thanks for your detailed feedback, I appreciate the clarifications and your arguments in favour of the novelty of the contribution, I encourage you to better clarify these aspects in the manuscript. This said, I am happy to raise my score to a borderline accept.

评论

Thank you for the insightful comments and willingness to raise your score! We will add those clarifications to the revised manuscript.

评论

Dear reviewer brvf,

As the discussion period is approaching its conclusion in two days, we would like to kindly remind you that we have addressed your comments in our rebuttal. We would greatly appreciate any additional feedback you may have before the deadline. If you have any further questions or concerns, please do not hesitate to reach out, and we will do our utmost to respond promptly.

Thank you for your time and consideration.

Best regards,

The Authors of Submission 11304

审稿意见
8

This paper proposes several ways to create closer, sparser, and more credible explanations for the SEV, along with two optimizing models. The results of the experiments on various datasets support the paper's claims.

优点

  1. Before reading this paper, I was unfamiliar with the sparse decision field. However, this paper is well-written and enjoyable to read.
  2. I think this paper is quite creative by simultaneously considering closeness, sparsity, and credibility.
  3. The comprehensive experiments address most of the claims they proposed in the introduction section.

缺点

It could be better to provide the complexity analysis and the analysis of time expenditure for different variants of SEV. For example, the computational benefits of using tree-based SEV.

问题

  1. In Table 1, what do the gray numbers represent?
  2. In line 199, there are duplicate "and"s.

局限性

I'm concerned about the depth of the tree if the model applies to large-scale data. In section A, the number of observations is up to 100K. If we encounter a much larger data set, the complexity of the model would exponentially increase with the depth of the tree. That's why I'm interested in time analysis.

作者回复

Thank you so much for your review! We really appreciate it! See below for our response to your questions.

In Table 1, what do the gray numbers represent?

Thank you for pointing out the missing explanation for the gray numbers. The gray numbers in Table 1 represent the query feature values that haven’t been changed, while the black numbers are the sparse explanations that have been made in order to make the prediction flipped. You can observe that the gray feature values are the same as the query feature values. In SEV, we can only use the changed features as the explanations, which is the same approach as DiCE.

In line 199, there are duplicate "and"s.

Thanks for pointing out the typo! We have already corrected the typo in our paper.

It could be better to provide the complexity analysis and the analysis of time expenditure for different variants of SEV. For example, the computational benefits of using tree-based SEV. I'm concerned about the depth of the tree if the model applies to large-scale data. In section A, the number of observations is up to 100K. If we encounter a much larger data set, the complexity of the model would exponentially increase with the depth of the tree. That's why I'm interested in time analysis.

Thanks for your question about the time complexity for tree-based SEV. The time complexity of tree-based SEV calculation is based on the structure of the tree instead of simply the depth of the tree. The worst case for an tree-based SEV calculation is O(n)O(n), where n is the number of nodes in the decision tree. The worst case does not scale as the number of data points. Moreover, SEVT\text{SEV}^T will be stopped early when the query has a sibling leaf node with negative prediction (shown in theorem 4.1), or the query already finds an explanation with SEVT\text{SEV}^T = 1 during the search (see Line 28-29 in Algorithm 3). Therefore, the calculation of SEVT\text{SEV}^T remains fast with the growth of observations.

Thank you so much once again for your review!

评论

Thanks for your answers!

作者回复

Thank you to all reviewers. The following pdf goes with the response for Reviewer xH4u.

评论

Hi all, the authors have provided detailed responses to the reviews. Please read the responses, and take advantage of the reviewer-author discussion period to ask for additional clarifications if needed.

In particular, reviewer bvrF has given the lowest score due to a number of issues she/he identified with the submission. The authors addressed these issues in their response. Does the reviewer agree with the response?

最终决定

The authors present enhancements on the Sparse Explanation Value (SEV) framework, proposing cluster-based and tree-based variants that improve the explanations provided. The contributions of this work are significant from a technical and practical point of view. After the author feedback, all reviewers recommended accepting the paper.