Faster Approximation of Probabilistic and Distributional Values via Least Squares
摘要
评审与讨论
The paper proposes two generic estimators of the importance value of the -th data point for the whole family of probabilistic values. Both estimators require evaluations to obtain -approximation, which is the optimal among known bounds. Further, the paper designs a framework for computing the distributional value estimator by connecting it to the obtaining least square regression. Vast experimental results show the faster convergence of the estimator over other benchmarks.
优点
The paper extends the current results to general probabilistic values by resolving the main limitation in computing the estimator of importance value. Further, the connection between estimating distributional value and the least square estimator in (13) saves computational costs drastically. These results have a potentially significant impact on data valuation.
缺点
(1) One major concern I have is whether (which is defined in the proof of Proposition 3 and 5) is a constant independent of for any probability values . It seems that setting and for yields . As the appears in the bound [(34) and (39)] for the number of evaluations, this seems critical in the first main contribution.
(2) The intuitive explanation of how the generalization to general is possible should be included in the main text.
(3) Discussion on the cases when in Theorem 1 is close to would be helpful.
问题
Q1. Is defined in the proof of Proposition 3 and 5 is a constant independent of ? Q2. How the generalization to a general family of probability values is possible? Q3. When can converge to ? Could any convergences results be derived?
We greatly appreciate your careful review! Below we address your concern.
Q: is constant and How the generalization to probabilistic values possible?
A: Thank you for pointing out this technical issue! Indeed, may not be a constant.
-
To simplify matters, in Appendix I we restrict to semivalues defined by a probability measure over to clarify our convergence analysis. Note that all probabilistic values in our experiments are semi-values. Specifically, the in Proposition 4 is defined by , which is if (e.g., Beta Shapley with ) and if has a bounded density (e.g., Beta(4,1)). Our results provide faster estimators for the Beta Shapley values (with integers ) that proved useful in both data valuation [2] and feature attribution [3]. More specifically, for the Beta Shapley value Beta(), we achieve time complexity if and if .
-
In Appendix H, we added an interpretation for . To sum, where is the probability of picking by Algorithm 1. Looking into Algorithm 1, it takes each to update the estimates of data. That means, is the average rate of reusing utility evaluations while running Algorithm 1. Therefore, higher average rate of reusing utility evaluations leads to more efficient ranking estimator.
-
As the reviewer pointed out, the worst case of the ranking estimator happens if, to be extreme, if and otherwise (corresponding to leave everthing else out). In this case, (equivalently, ) and our complexity bound becomes , which is much worse. Interestingly, in this case, our ranking estimator in fact takes one utility evaluation to update the estimate of one data point. This indicates that our complexity bound may still have room for further improvement (at least in some cases).
Q: When can converges to ? Could any convergences results be derived?
A: If the distance between the empirical data distribution and the authentic data distribution converges to 0 as N increases, then we think such a convergence could be possibly established. For example, Theorem 2.7 in [1] shows that can be bounded by the Wasserstein distance of and . Empirically, to have as close distance as possible, the size of (note our framework assumes ) should be large enough, and that is why we set for verifying the efficacy of Algorithm 3.
[1] Ghorbani, Amirata, Michael Kim, and James Zou. "A distributional framework for data valuation." International Conference on Machine Learning. PMLR, 2020.
[2] Kwon, Yongchan, and James Zou. "Beta Shapley: a Unified and Noise-reduced Data Valuation Framework for Machine Learning." International Conference on AI and Statistics. 2022.
[3] Kwon, Yongchan, and James Y. Zou. "WeightedSHAP: analyzing and improving Shapley based feature attributions." Advances in Neural Information Processing Systems 35 (2022): 34363-34376.
Thank you for the response and updates. The response is clear and my major concern is resolved. I will change my score accordingly.
This paper proposes an algorithm that estimates the probabilistic value of each data point , i.e., , where is a given utility function that maps any subset of data points to a value, and the weights 's can be chosen arbitrarily by the user, as long as they satisfy . The well-known Shapley value is a special case of the probabilistic value. Their estimator uses evaluations of the utility function , an improvement over the previous work [Kwon and Zou 2022] which uses evaluations.
Their estimator is based on a simple observation: because , we have .
Hence, their method uses random sampling to simultaneously estimate for all and , and then substracts the later from the former to get the estimates of for all data points .
The advantage of their method is that a random sample can be used to estimate for all simultaneously, using just a single evaluation of the utility function -- . In comparison, if we use the sample to estimate the for all directly, we would need to evaluate for all . Thus, they get an improvement from evaluations to evaluations.
优点
- The idea is simple and cute.
- The writing is fine overall.
缺点
- The result is technically a bit thin.
- The least squares interpretation seems a bit artificial and redundant.
- From the experiments in appendix, it is not obvious to me that the proposed estimator is any better than previous works (e.g., the well-known SHAP for estimating Shapley value).
问题
Could you elaborate the advantage of your estimator over previous works in terms of practical performance?
Thank you for your feedback! Below we address your concerns.
Q: The result is technically a bit thin.
A: We agree that the idea behind Algorithm 1 is simple, but the analysis for its convergence as well as extending the idea up to Theorem 1 is not trivial. In Appendix I, we clarify conditions on which the proposed estimators enjoy the currently best complexity bound in terms of -approximation. Empirically, in Figures 3 and 5, the proposed framework of training value estimator towards distributional values is verified to be significantly efficient and effective.
Q: The least squares interpretation seems a bit artificial and redundant.
A: We agree that the least square interpretation may not be needed if our goal is to directly approximate the valuation of a fixed dataset. However, to cope with unseen data points through training a value estimator , the least squares interpretation offers a clean optimization objective, which we exploited in Section 3.3: compare Eq (13) and Eq (15).
Q: It is not obvious that the proposed estimator is any better than previous works (e.g., the well-known SHAP for estimating Shapley value).
A: Our goal is to design generic estimators for all probabilistic values, instead of specializing to any particular choice, such as the Shapley value that SHAP is for. To our knowledge, before our work, the (weighted) sampling lift estimator is the only one that can apply to all probabilistic values.
To our knowledege, there is no complexity bound for SHAP in terms of -approximation. In contrast, we proved our ranking estimator enjoys the currently known best complexity . Emprically, our generic estimators are overall slightly slower than the specialized SHAP, and they are faster than the (weighted) sampling lift estimator.
Q: Could you elaborate the advantage of your estimator over previous works in terms of practical performance?
A: We would like to answer this question separately for i) the two proposed estimators that approximate probabilistic values, and ii) the framework of training value estimator towards distributional values.
-
In both data valuation [1] and feature attribution [2], the Beta Shapley values (restricted to integer ) tend to perform significantly better than the Shapley value. It is highly likely that the best choice often is dataset and task dependent. Thus, it is important to design faster generic estimators that work for all probabilistic values. Empirically, on Beta(4,1) (see Figures 9 and 14) and Beta(2, 2), our proposed estimators are indeed faster than the (weighted) sampling lift estimator. Note that AME applies to Beta(2,2) but not Beta(4,1).
-
Though probabilistic values were seen useful in data valuation, one big issue is that Eq. (1) cannot naturally extend to any unseen data. In contrast, the distributional value in Eq. (2) is more scalable as its definition already accounts for every possible data point. Previously, the distributional value of any unseen data point can only be approximated (and there is only one estimator for this end). Our proposed framework of training value estimators towards distributional values (Algorithm 3, which is derived from Theorem 1) removes this hurdle, i.e., the distributional value of any unseen data point can be evaluated in one single forward-pass using a well-trained value estimator. The first two columns of Figures 3 and 5 demonstrate that the trained value estimators using Algorithm 3 learn the distributional values much better than approximation using 10,000 utility evaluations per data point. Moreover, the removal experiment, i.e., the last column of Figures 3 and 5, shows that the trained value estimators are more capable of selecting the most influential data on training, compared with the baselines.
[1] Kwon, Yongchan, and James Zou. "Beta Shapley: a Unified and Noise-reduced Data Valuation Framework for Machine Learning." International Conference on AI and Statistics. 2022.
[2] Kwon, Yongchan, and James Y. Zou. "WeightedSHAP: analyzing and improving Shapley based feature attributions." Advances in Neural Information Processing Systems 35 (2022): 34363-34376.
[3] Jethani, Neil, et al. "Fastshap: Real-time shapley value estimation." International Conference on Learning Representations. 2021.
Thanks for your answers!
Probabilistic values, rooted in cooperative game theory, play a pivotal role in data valuation. However, their computation poses significant challenges, especially as dataset size (N) increases. Many current estimators either come with high computational overheads or cater only to specific instances.
Here are the main contributions of this paper:
- The authors introduce two versatile estimators, anchored in least squares regression, capable of efficiently approximating a broad range of probabilistic values, transcending the confines of earlier methods typically designed for specific values like the Shapley value. (Refer to Sections 3.1 and 3.2)
- Through Propositions 3 and 5, the authors establish that both novel estimators necessitate only O(N log N) utility evaluations to achieve a (ε,δ)-approximation, thereby aligning with the best-known computational complexities for certain cases.
- Section 3.3 unveils a pioneering approach to cast the distributional value—a more consistent alternative to probabilistic values—as a least squares problem. This breakthrough facilitates the training of machine learning models that can swiftly gauge distributional values of unseen data in a single pass.
- Validating the efficacy of their proposals, Section 4 presents experiments that substantiate the accelerated convergence of their estimators compared to preceding methods. Moreover, they successfully illustrate the potential of training models to adeptly predict distributional values for novel data points.
优点
- The paper introduces efficient methods for both ranking and value estimation of data points. Through Propositions 1 and 2, they've laid out a mechanism to estimate the relative ranking underlying any probabilistic value. Notably, this method is computationally advantageous.
- They've established a framework, as seen in Theorem 1 and the related discussions, for training models to serve as value estimators. This is highly valuable, as trained models can swiftly evaluate any unseen data point from similar distributions in just a single forward pass, making the evaluation process much more scalable.
缺点
- There are computational challenges associated with training value estimators. When evaluating the distributional values, even with approximations, the computational costs are significant.
- The paper mentions that the faster convergence of certain estimators like SHAP, SHAP-paired, and the complement comes at the cost of using Θ(N^2) memory storage instead of Θ(N). However, the implications of this memory increase, especially for large-scale applications, are not discussed in depth.
问题
- While the paper presents a novel approach to data valuation using distributional Shapley values, could you shed light on the practical applications where this approach might be most beneficial?
Q: Could you shed light on the practical applications where this approach might be most beneficial?
A: We would like to answer this question separately for the proposed estimators and the framework of training value estimators.
-
In our perspective, faster estimators for probabilistic values could potentially benefit the field of feature attribution. We have included the details in Appendix G. To be precise, there are three related works to mention:
-
[1] FastSHAP was introduced as a framework for training explainers, which replaces in the problem (7) (whose uniquely optimal solution is the Shapley value) by trainable models. Before our work, how to cast other probabilistic values into solving optimization problems was an open question. Precisely, we answered this question by Proposition 4 (Algorithm 2 is derived from solving the problem inside this proposition). Note that AME only partially answered this question as it is restricted to a subfamily of probabilistic values which does not include, e.g., Beta(4, 1).
-
[2] showed that other candidates of the Beta Shapley values (restricted to integers ) tend to be significantly better than the Shapley value in feature attribution. Note that [2] did not train explainers for the Beta Shapley values.
-
Very recently, [3] showed that FastSHAP is equivalent to training explainers towards a biased Shapley value, and thus suggested it is better to train explainers towards the unbiased one, i.e., any estimators that converge to the Shapley value. In this view, faster estimators could potentially improve the training.
-
-
We think the framework of training value estimators towards distributional values makes data valuation methods based on marginal gains more scalable and practical.
- Even though probabilistic values proved handy in data valuation, their main issue is that Eq. (1) cannot be naturally extended to unseen data, which makes it intractable to train value estimators for probabilistic values.
- The value assinged to each data point indicates its quality of some type (e.g., the extent of being an outlier), and the type of quality that our framework determines is through the aggregation of marginal gains . In other words, the choice of utility function is critical for the final performance of the considered task. Therefore, there are possibilities considering that the utility function in our framework only serves as a module, such as detecting outlies, the most influential data, etc..
[1] Jethani, Neil, et al. "Fastshap: Real-time shapley value estimation." International Conference on Learning Representations. 2021.
[2] Kwon, Yongchan, and James Y. Zou. "WeightedSHAP: analyzing and improving Shapley based feature attributions." Advances in Neural Information Processing Systems 35 (2022): 34363-34376.
[3] Zhang, Borui, et al. "Exploring Unified Perspective For Fast Shapley Value Estimation." arXiv preprint arXiv:2311.01010 (2023).
Thank you for the response and updates. I left the score unchanged.
Thank you for your efforts! Below we address your concerns.
Q: The paper mentions that the faster convergence of certain estimators like SHAP, SHAP-paired, and the complement comes at the cost of using Θ(N^2) memory storage instead of Θ(N). However, the implications of this memory increase, especially for large-scale applications, are not discussed in depth.
A: We note that this increase in memory does not apply to our methods, and we provide the following explanation for the prior methods:
-
For SHAP, it is derived from solving the problem (7). Precisely, the objective is rescaled to be an expectation, and then the objective is approximated using multiple sampled subsets. The estimate is yielded by solving the approximated objective with the efficiency constraint. The approximation includes a matrix of dimension , which leads to memory storage. Particularly, the inverse of this matrix is calculated at the end for aggregation, which induces time complexity.
-
SHAP-paired just uses a more efficient sampling procedure to do approximation, and thus memory storage and another time complexity are still there.
-
For complement estimator, they decompose Eq. (6) into where . The procedure is to approximate for every , and the estimate is calculated at the end by aggregating all , which leads to storage memory. The time complexity of the aggregation is .
Q: There are computational challenges associated with training value estimators.
A: The computation time can be saved a lot by parallel computing. For MNIST, using 800 cpu cores, it takes less than two days to well approximate the ground-truth on 200 data points; less than a day to generate used for training. To train a value estimator, it takes less than two days using 5 cpu cores. All in all, we believe these costs are reasonable and we hope future work will further reduce them.
The paper proposed two generic probabilistic values estimators (one for ranking and the other for exact probabilistic values) that achieves utility evaluations, improving upon previously known generic estimators that require . The authors prove theoretically convergence rates of their estimators, and present their performance with numerical experiments.
优点
The result is interesting and novel to me, though I am not an expert in this area. The estimators proposed is general enough, and comparisons with previous literature is thorough. The paper provides both good theoretical and numerical evidence.
缺点
- Lack of direct comparisons to previous estimators specific to special cases achieving the same rate. I would like to see more discussions on SHAP, MSR and AME estimators. Why they cannot be generalized? What are their main ideas comparing to this paper, what is the fundamental differences?
- It may be that I'm not familiar with the field, but the presentation does not properly introduce the background. It would be nice if the authors can provide a few more demonstrations in the introduction, i.e. examples for Shapley and Banzhaf. I did not see proper explanations on those probabilistic values while reading the paper.
- The numerical plots could look better in log-scale, for now it's hard for me to distinguish between lines when they converge.
- Lack of expanding on the theoretical results. Some explanations on the intuition on why an convergence rate is expected and what makes it different from previous estimators would be nicer.
问题
In conjunction with 1-4 in the weakness part.
Q: It would be nice if the authors can provide a few more demonstrations in the introduction.
A: We added Appendix F on this topic. For each probabilistic value Eq. (1), the only constraint is that is non-negative and . But for semi-values (a subfamily of probabilistic values), we have where is a probability measure on . Conversely, each probability measure leads to a semi-value through the same formula. To our knowledge, all probabilistic values in the previous references are all semi-values. For example, uniform produces the Shapley value, i.e., ; (Dirac delta distribution) generates the Banzhaf value, i.e., ; beta distribution with pdf leads to Beta (notation for Beta Shapley values), i.e., . In [2-3], the employed Beta Shapley values restricted to integers .
Q: Lack of expanding on the theoretical results. Some explanations on the intuition on why an convergence rate is expected and what makes it different from previous estimators would be nicer.
A: We have added Appendix I to clarify our convergence analysis. We mainly compare our proposed estimators to the (weighted) sampling lift estimator as (to our knowledge) it is the only generic estimator before our work. To compare, ours use one utility evaluation to update the estimates of (in expectation , see Appendix H) data, while the (weighted) sampling lift estimator takes two utility evaluations to update only one data point. In this view, ours are more efficient in reusing each utility evaluation, which explains its faster convergence. Specifically, for Beta(), ours achieve if and if . Previously, to our knowledge, only the (weighted) sampling lift (which requires instead) and AME (which only works with ) can be used for the Beta Shapley. Though it was proved that AME enjoys under certain assumption, the Beta Shapley dose not verify this assumption.
[1] Wang, Jiachen T., and Ruoxi Jia. "Data banzhaf: A robust data valuation framework for machine learning." International Conference on Artificial Intelligence and Statistics. PMLR, 2023.
[2] Kwon, Yongchan, and James Zou. "Beta Shapley: a Unified and Noise-reduced Data Valuation Framework for Machine Learning." International Conference on AI and Statistics. 2022.
[3] Kwon, Yongchan, and James Y. Zou. "WeightedSHAP: analyzing and improving Shapley based feature attributions." Advances in Neural Information Processing Systems 35 (2022): 34363-34376.
Thank you for your helpful feedback. We have re-plotted all the convergence results for comparing estimators in log-scale in the revision. Below we address your comments.
Q: Lack of direct comparisons to previous estimators specific to special cases achieving the same rate. I would like to see more discussions on SHAP, MSR and AME estimators. Why they cannot be generalized? What are their main ideas comparing to this paper, what is the fundamental differences?
A: We detail the difference below.
-
MSR: It does sampling based on Eq. (4). Precisely, it repeats the procedure of sampling a subset by including each data point with probability . With sampled subsets , the estimate for the -the data point is where and . In terms of -approximation, its time complexity is , ours enjoy the same time complexity as proved in Corollary 1.
- Difference: MSR only applies to probabilistic values that satisfy for every (see Appendix C.2 in [1]) while ours are for all probabilistic values.
- Why it cannot be generalized? Not every probabilistic vlaue corresponds to a such that Eq. (1) can be rewritten as Eq. (4).
-
SHAP: It is derived from solving the estimate of the problem (7). Precisely, the objective there is scaled into an expectation (which does not modify its uniquely optimal solution, the Shapley value). Then, subsets are sampled according to the scaled weights to approximate the objective, after which the estimate is calculated by solving the estimated objective with the efficiency constraint.
- Difference: i) To our knowlege, there is no complexity bound for SHAP in terms of -approximation, while our ranking estimator has complexity . ii) SHAP is to approximate the Shapley value while our methods work with any probabilistic value.
- Generalizations: Ours may be seen as generalizations of SHAP as they are derived from solving the problems (10) and (11). Precisely, the ranking estimator is to approximate (up to some scalar) in which represents the uniquely optimal solution to the problem (10).
-
AME: It is derived from solving the estimate of the problem (8). Its objective is already an expectation. So, it uses multiple sampled subsets to approximate the objective, with which the approximate optimal solution is obtained. This solution represents the estimate of the underlying probabilistic values.
- Differnce: i) For every probability measure over , the definition for in Eq. (1) yields a probabilistic value, which is the uniquely optimal solution to the problem (8) while using to sample and ; if , the correponding probabilistic value cannot be approximated by AME. Examples in our experiments include the Shapley value and Beta(4,1). ii) By assuming that has support where and , they proved AME has complexity bound . Under this assumption, ours achieve the same bound. However, the Beta Shapley value Beta(), with integers , does not verify this assumption, but we can still prove that ours achieve if and if .
- Why it cannot be generalized? For each sampled subset , is defined to be if and otherwise, whereas . Then, the optimal solution to the problem (10) is . Specifically, the diagonal entries of are all , which will blow up if .
Thank you for the update and it addresses my concerns. I'm raising my score accordingly.
We thank all reviewers for their efforts and constructive feedback! Here we summarize the updates in the revision. All changes in the main paper and appendix were marked in blue.
-
Figure 1, Figures 6 - 14 that compare the convergence of estimators have been re-plotted in log-scale.
-
Appendix I is to clarify our convergence analysis, i.e., Propositions 4 and 5. As pointed out by Reviewer 3wRG, our complexity bounds contain terms and that may depend on . In Appendix I, we restrict to semivalues defined by a probability measure on and we show that if (e.g., Beta-Shapley for ) while if has a bounded density (e.g., Beta(4,1)). We note that all probabilistic values employed in our experiments and in the literature are semivalues. Moreover, faster estimators (except AME) than the straightforward sampling lift method are not designed for the Beta Shapley values, a gap that we fill with the best known complexity bound.
-
Appendix H is to provide an intuitive interpretation on that appears in Proposition 4. To conclude, it is the inverse square of the average rate of reusing utility evalutions while running Algorithm 1.
-
Appendix J is to add more introduction to probabilistic values used in practice.
-
Appendix G is to comment on why faster estimators for probabilistic values (instead of for specific cases) could benefit the field of feature attribution.
This paper proposes generic probabilistic values estimators that incur less utility evaluations to achieve the same approximation quality as existing ones.
STRENGTH
The technical contributions are novel, interesting, and non-trivial.
WEAKNESS
(1) There are some clarification issues (e.g., explanations for the theoretical results, direct comparisons with existing estimators, discussion of practical applications, gamma being a constant, etc) that have been addressed well by the authors in the rebuttal.
(2) I would also like the authors to discuss in their revised paper whether and how well the corresponding axioms can be satisfied after applying their proposed approximations. See, for example, the following missing reference:
Probably Approximate Shapley Fairness with Applications in Machine Learning. AAAI 2023.
The authors should also comment on the difference in approximation quality of their proposed estimator from that in the above reference, and discuss any implications.
为何不给更高分
Though the results are new, the practical significance may be limited since the utility evaluations remain computationally costly, especially when scaling up to large data.
为何不给更低分
The proposed estimators are non-trivial improvements over the existing ones.
Accept (poster)