PaperHub
7.8
/10
Poster5 位审稿人
最低4最高6标准差0.7
4
4
5
5
6
2.8
置信度
创新性3.0
质量3.0
清晰度3.2
重要性2.6
NeurIPS 2025

Efficiently Verifiable Proofs of Data Attribution

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29
TL;DR

We demonstrate an efficient interactive verification protocol that allows resource-constrained parties to trust and reliably check the quality of expensive attribution models provided by powerful, untrusted parties.

摘要

Data attribution methods aim to answer useful counterfactual questions like "what would a ML model's prediction be if it were trained on a different dataset?" However, estimation of data attribution models through techniques like empirical influence or "datamodeling" remains very computationally expensive. This causes a critical trust issue: if only a few computationally rich parties can obtain data attributions, how can resource-constrained parties trust that the provided attributions are indeed ``good,'' especially when they are used for important downstream applications (e.g., data pricing)? In this paper, we address this trust issue by proposing an interactive verification paradigm for data attribution. An untrusted and computationally powerful Prover learns data attributions, and then engages in an interactive proof with a resource-constrained Verifier. Our main result is a protocol that provides formal completeness, soundness, and efficiency guarantees in the sense of Probably-Approximately-Correct (PAC) verification (Goldwasser et al., 2021). Specifically, if both Prover and Verifier follow the protocol, the Verifier accepts data attributions that are $\varepsilon$-close to the optimal data attributions (in terms of the Mean Squared Error) with probability $1-\delta$. Conversely, if the Prover arbitrarily deviates from the protocol, even with infinite compute, then this is detected (or it still yields data attributions to the Verifier) except with probability $\delta$. Importantly, our protocol ensures the Verifier's workload, measured by the number of independent model retrainings it must perform, scales only as $O(1/\varepsilon^2)$; i.e., independently of the dataset size. At a technical level, our results apply to efficiently verifying any linear function over the boolean hypercube computed by the Prover, making them broadly applicable to various attribution tasks.
关键词
Data AttributionComponent AttributionVerifiable MLInteractive ProofsPAC-Verification

评审与讨论

审稿意见
4

This paper introduces a novel interactive protocol that allows computationally limited verifiers to efficiently verify whether training data attribution vectors computed by a powerful prover are approximately optimal. The verification is grounded in a PAC framework, aiming to estimate the gap between the candidate attribution and the best possible attribution (in terms of MSE). The key technical contributions include a transformation of a previously known residual estimation algorithm into a two-message interactive protocol with spot-checking, reducing verifier cost from O(1/ε3)O(1/\varepsilon^3) to O(1/ε2)O(1/\varepsilon^2).

优缺点分析

Strenths

  • Introduces elegant and efficient PAC-verification protocol for data attribution.

  • Solid integration of Boolean harmonic analysis, noise stability, and residual estimation.

  • Robust theoretical analysis with completeness, soundness, and efficiency guarantees.

Weaknesses

  • No experimental results. Given the strong practical motivation (e.g., verifying attributions in sensitive settings like data pricing or medical diagnosis), this absence is glaring. Even synthetic experiments could have helped ground the theoretical claims. Moreover, real-world GPU training often struggles to meet non-deterministic conditions, potentially undermining completeness/reliability.

  • Lemma 1 assumes that erroneous samples are independent and uniformly distributed. If a malicious Prover selectively tampers with the subset that contributes significantly to noise-stable estimation, it can substantially shift B^2B̂_{≥2} within the corruption budget, breaking the ε/4 error bound.

  • The symbol ρ is not explained when it first appears in the main text.

问题

See weaknesses.

局限性

  • No experiments or simulation support.

  • Dependence on a strong determinism assumption.

  • Possible limited practicality in modern deep learning settings.

最终评判理由

The authors’ rebuttal address my concerns and some of the minor typos are also corrected, therefore, I will keep my score.

格式问题

  • Typos: "straight forward" to "straightforward", and the usage of "ϵ" and "ε" should be unified.
作者回复

Dear Reviewer Hayo,

Thank you for your time and feedback. Below, we address the raised weaknesses and questions.

Experimental Results and Practicality

We agree that experimental validation is a natural next step. The primary contribution of this work is to establish the theoretical foundations for efficiently and verifiably proving data attribution, with formal guarantees of completeness and soundness against powerful, malicious Provers. Given focus on provable security and verifiability, which is difficult to simulate meaningfully, we deferred a full empirical study to future work.

Regarding the non-determinism in GPU training, we discussed in the paragraph titled "Equivalence Checking Procedure" (Lines 536-543) that, while the theorems assume perfect equivalence checking procedure, the check-equiv procedure can be instantiated practically by, for instance, comparing model outputs or final loss values on a held-out set, rather than requiring a bit-for-bit identity of model weights. This accommodates the realities of modern deep learning frameworks, and any error in equivalence checking can be wrapped into the probabilities for soundness and completeness via union bounds.

On the conditions in Lemma 1

We would like to clarify that the robustness guarantee of Lemma 1 is stronger than the review suggests. The lemma does not assume that bad samples are chosen uniformly or independently by the cheating Prover. Instead, our analysis considers the adversarial setting, where the Prover can corrupt any mm function evaluations of its choosing, in order to to most impact the Verifier's estimate (see Line 571: "...an adversary corrupts up to mm ... of the nn function evaluations..."). The proof in Appendix D.1 is constructed specifically to bound the worst-case error under these adversarial, non-uniform corruptions. Therefore, the scenario described by the reviewer is precisely the threat model that Lemma 1 is proven to be robust against.

Typos

Thank you for your careful reading and help with notation and typos. We will be sure to make the edits in the final version, especially defining the symbol ρ\rho at the right time.

Thank you again for your time and valuable comments.

Sincerely,

Authors

审稿意见
4

The paper addresses the challenge of verifying data attributions in machine learning models, particularly in scenarios where a computationally powerful entity (the Prover) provides attributions to a resource-constrained party (the Verifier). The authors propose an interactive verification protocol to ensure that the attributions are "good" (i.e., close to optimal) without requiring the Verifier to perform the computationally expensive task of computing the attributions themselves.

优缺点分析

Strengths:

  • Well-Structured: The paper is logically organized, with clear motivation, problem formalization, protocol description, and proofs.
  • Good Intuition: The authors provide intuitive explanations (e.g., the "naive verification fails" example) before diving into technical details.
  • Full Proofs: The main text outlines key ideas, while detailed proofs are deferred to appendices, improving readability.

Weaknesses:

  • The heavy reliance on Fourier analysis and Boolean function notation might make the paper less accessible to readers without a strong theoretical background.
  • While Algorithm 1 is detailed, a high-level pseudocode summary in the main text could improve clarity for practitioners.

问题

Add a flowchart or simplified pseudocode in the main text to visually summarize the interactive protocol (beyond Algorithm 1 in the appendix). This would help readers quickly grasp the key steps (e.g., challenge setup, spot-checking, residual estimation).

局限性

The paper acknowledges that retraining-based attribution methods (like datamodels) are computationally expensive, creating a trust gap between powerful entities (e.g., corporations) and resource-limited users (e.g., individuals or small labs). This is a key motivation for the work.

格式问题

N/A

作者回复

Dear Reviewer on4T,

Thank you for your time and feedback!

In regards to the following weakness statement in your review:

While Algorithm 1 is detailed, a high-level pseudocode summary in the main text could improve clarity for practitioners.

We would like to clarify that Algorithm 1 does have a pseudocode/summary in the main text, at the beginning of section 4. If we have a misunderstanding, please let us know and we are happy to clarify further.

Thanks again for your review.

Sincerely,

Authors

审稿意见
5

Suppose that we train an ML model on an NN size dataset, and we would like to know how much each data element (or subset of elements) influenced the training. This is modeled by a function f ⁣:1,1NRf \colon \\{-1,1\\}^N \rightarrow \mathbb{R}, where each x1,1Nx \in \\{-1,1\\}^N encodes a subset (xi=1x_i = 1 iff the jj’th element is included in the subset), and this function can be estimated for every xx by retraining the model on the subset xx and then computing the prediction quality or test error of the new model. The problem is that computing this function is very expensive (each subset requires a fresh training), and we would like to compute an approximation of ff, hopefully using only a few retrainings. The work of Saunshi et al. (2022) showed how to compute a linear model aa such that f(x)<a,x>f(x) \approx <a,x>, where the number of retrainings to obtain aa is only 1/ϵ3\approx 1/\epsilon^3, which is independent of NN (ϵ\epsilon is the required accuracy measured by the MSE distance from the optimal linear model). This is done using Residual Estimation.

This paper considered a model where we have two entities: a prover, which has more computational resources, and the verifier, which is more restrictive but still would like to get the estimation of ff. The problem arises when the prover is not necessarily honest, so the verifier must have a way to check whether the linear model aa produced by the prover is indeed close to optimal. The main result of this paper is to provide a 2-message interactive protocol between the prover and the verifier, where the verifier, using only 1/ϵ2\approx 1/\epsilon^2 retrainings (and not 1/ϵ3\approx 1/\epsilon^3 as in the non-interactive version of Saunshi et al. (2022)), can catch an unhonest prover with high probability. The solution relies on the same Residual Estimation of Saunshi et al. (2022): The verifier asks the prover to perform the 1/ϵ31/\epsilon^3 retrainings, but checks only ϵ\epsilon fraction of them by doing the retrainings itself (where the exact ϵ\epsilon fraction is unknown to the prover). If the prover lied in more than 1/ϵ1/\epsilon out of the 1/ϵ31/\epsilon^3 requests, the verifier will catch it with high probability. Otherwise, there could be at most 1/ϵ1/\epsilon wrong values, so the main technical contribution of this paper is to prove that the approach of Saunshi et al. (2022) is error-robust against such an amount of errors.

优缺点分析

I haven’t read the robustness proof that appears in the appendices (Lemma 1, page 15), but assuming it is correct, I think it is a nice paper that provides a nice solution to a well-motivated problem while addressing a real technical challenge (the robustness of the residual estimation). Therefore, I recommend acceptance.

问题

No questions

局限性

yes

最终评判理由

I had no questions for the authors, and the other reviews were also supportive, so I'm keeping my positive score.

格式问题

No formatting concerns

作者回复

Dear Reviewer BYJZ,

Thank you for your time and effort in completing your review. We appreciate the vote for acceptance.

Sincerely,

Authors

评论

Thank you for your response.

审稿意见
5

This paper studies the "verifiable" problem for data attribution, an important growing field in data-centric AI. Specifically, the paper defines the verifying problem in data attribution and proposes an "efficient" interactive protocol that is PAC-verifiable.

优缺点分析

Strengths

  1. With the growing interest in data attribution, with its potential societal impact on data compensation (or data economy), the problem studied in this work is indeed important and well-motivated.
  2. The writing is professional, consistent, and self-contained.
  3. The proposed algorithm and interactive protocol are clear and well-motivated (with the appropriate introduction to the prior works), which makes them easy to follow overall.

Weaknesses

  1. If my understanding is correct, the notion of "efficiency" is the number of retrainings required. It is, hence, a bit misleading to say that the "cost" is independent of dataset size NN. I suggest that the authors can elaborate and clarify this point in the next iteration.
  2. The scope of the present work is a bit restricted and not justified: for instance, the current vast data attribution literature has moved beyond retraining-based datamodel style methods to explore settings such as "single-model TDA" via, e.g., dynamic methods. I think the design of the present work is largely due to the prior work by Saunshi et al., which makes sense, but it'll be helpful to discuss this more in detail.

Writing Nothing really significant. Just want to point out:

  1. Page 13 can probably be handled more elegantly..

问题

  1. Can the authors discuss what is the role of the final attribution score aa' produced by the prover? Since, at the end of the day, if the prover is honest/passes the spot-checking, the verifier will use the provided models and curate the final attribution scores using the residual estimation algorithm. It seems to me that the attribution score aa' provided by the prover is useless in this sense.
  2. (Weakness 1) I suggest that the authors can elaborate and clarify the notion of efficiency used in this paper more explicitly.
  3. (Weakness 2) I understand that due to the space limits, no related works are presented in the paper. I suggest that (if published) the authors can add a dedicated related work section to carefully discuss the potential impacts and some design choices to help justify and position the paper in the literature.

局限性

While no specific limitations were discussed in the paper, the reviewer does not have additional concerns on potential limitations.

最终评判理由

I maintain my positive score, and no further concerns are raised after reading all the discussions with other reviewers.

格式问题

NA

作者回复

Dear Reviewer cTDk,

Thank you for your kind and thoughtful review!

Notion of "Efficiency"

Indeed, you are correct. We definitely didn’t mean to be misleading and will clarify/correct it in the final version. Currently, our use of "efficiency" refers to the Verifier's sample complexity (the number of retrainings), which is itself independent of the dataset size NN. The cost of each individual retraining does depend on NN, and we will make this distinction explicit in the final version.

Scope/Relation to Other Methods

We feel that focusing on datamodels and empirical influences in the context of predictive attribution will always be a worthwhile focus because they are the provably optimal linear attributions for minimizing predictive Mean Squared Error (MSE) (which is our metric). This is proved by [1]. Verifying w.r.t. empirical influence / datamodels thus addresses the fundamental question of whether a party has computed the best possible linear explanation. We agree that while related work is discussed as necessary throughout the body of the paper, a dedicated section would also be valuable, especially mentioning other lines of work like single model TDA. We will add this for the final version.

Prover's Attribution Score

The Prover's attribution vector a\*a^{\*} is very important. The core of the Verifier's protocol is to check if the error of this submitted particular a\*a^{\*} is close to the optimal error (Algorithm 1). Without a\*a^{\*} from the Prover, there would be no candidate solution for the Verifier to check.

Thank you again for your valuable comments.

Sincerely,

Authors


References

[1] Saunshi, N., Gupta, A., Braverman, M., and Arora, S. (2022). Understanding influence functions and datamodels via harmonic analysis. arXiv preprint arXiv:2210.01072.457

评论

Thank you for your response, and I have no further questions. Great work by the way : ) Really enjoy it.

审稿意见
6

This paper proposes a technical solution to the trust problem centered around data attribution vectors computed from datamodels. In this context, datamodels refer to mappings that quantify the relationship between subsets of a training dataset and the resulting performance of a model trained on those subsets. For example, given a fixed training set of size NN, we can define a mapping f:±1N[0,1]f: \\{\pm 1\\}^N \to [0,1], where the selection vector x±1Nx \in \\{\pm 1\\}^N indicates which data points are included in the subset and f(x)f(x) gives the final 0/1 loss of the classifier trained on that subset.

Current approaches compute data attribution vectors from datamodels. They are defined as the linear coefficients aRNa \in \mathbb{R}^N of the optimal linear approximation to ff with respect to an inner product. For simplicity, we work in the inner product space induced by the uniform distribution over ±1N\\{\pm 1\\}^N. The analysis extends essentially unchanged to the case of i.i.d. products of pp-Bernoulli distributions. These vectors have important practical applications, such as data pricing, allowing one to compensate individual data points within a training set according to their “influence” on the final performance of the model.

Evaluating the datamodel ff on each training subset xx is extremely expensive, and therefore computing honest data attribution vectors creates significant asymmetry and trust issues. Only a few parties have the resources to perform these computations, while most others must rely on, and therefore trust, this small group of powerful actors. The paper addresses this asymmetry by introducing an interactive protocol in which the Verifier can check the optimality of the Prover’s data attribution vector with much less computational cost. Here, the Verifier’s computational cost, defined to be the number of datamodel evaluations, is shown to be independent of the dataset size NN.

A key technical insight comes from previous work by Saunshi et al., who observed that computing certain Fourier coefficients of the residual g(x)=f(x)a,xg(x) = f(x) - \langle a, x\rangle, given aa is much cheaper than computing the attribution vector from scratch. The novelty of this paper lies in formulating the trust problem as a PAC verification task and offloading the computation of the residual’s Fourier coefficients to the Prover. With the spectral computation pushed to the Prover, the Verifier only needs to perform random spot checks on the Prover’s results to verify their validity. The offloading reduces the Verifier’s computational cost from 1/ϵ31/\epsilon^3 to 1/ϵ21/\epsilon^2 (ignoring other parameters of the interaction), where ϵ\epsilon denotes the (additive) suboptimality of the Prover’s candidate attribution vector relative to the optimal one.

优缺点分析

Strengths

  1. Compelling narrative and motivation. The paper presents a creative and well-motivated application of interactive proofs, clearly explaining how asymmetry in computational cost creates trust problems in data attribution, and highlighting the need to resolve them for practical applications such as data pricing. Formulating this trust problem as an interactive protocol feels natural and provides an elegant framework for a real, practical issue.
  2. Insightful use of tools from theoretical computer science. The paper elegantly combines Fourier analytic ideas from Saunshi et al. and techniques from interactive proofs to arrive at a quantitative improvement. Because random spot checks are sufficient to detect the Prover’s deviations with high probability, the Verifier can shift the computational burden of residual estimation to the Prover, an approach very much in the spirit of probabilistically checkable proofs (PCPs).
  3. Clear, concise, and precise exposition. The paper is exceptionally well-written, presenting the technical content in an accessible way without compromising rigor. Moreover, the narrative and technical tools are tightly integrated, resulting in a seamless presentation that effectively connects the motivation to the methodology.

Weakness

  1. Missing assumptions on equivalence checking. The paper lacks precise details about the assumptions required for the equivalence checking procedure. I assume all PAC verification guarantees hold if model training is deterministic, with CheckEquiv(a_1, a_2) returning true if and only if a_1 = a_2. However, if there is significant randomness or noise in training outcomes, completeness and soundness guarantees may fail to hold. For example, if CheckEquiv always passes any input pair, the protocol would fail to satisfy soundness guarantees. A clearer specification of the conditions under which equivalence checking ensures both soundness and completeness would strengthen the paper.

  2. Other choices of computational cost for the Verifier. The claimed independence of Verifier’s cost from NN is somewhat misleading. While the number of datamodel evaluations required for the protocol is indeed independent of NN, the cost of each individual evaluation can grow with NN. For example, even with a fixed inclusion probability pp, the expected training subset size scales with pNpN, and the cost of training a model on these subsets may increase accordingly. Moreover, in high-dimensional settings where the data point dimension dd scales linearly or superlinearly with the number of samples NN, even parsing a single data sample can scale with NN. In this sense, a more budget-relevant accounting of computational costs may not be truly independent of NN. Clarifying these caveats and considering alternative, reasonable notions of the Verifier’s computational cost would strengthen the paper.

问题

  1. Could the authors discuss the shortcomings of the current definition of Verifier’s computational cost?
  2. What general assumptions on the equivalence checking procedure are sufficient to preserve the PAC verification guarantees?
  3. How sensitive are the attribution vectors to the choice of pp in the pp-Bernoulli marginal distribution over selection vectors? For example, are there empirical studies examining how the ranking of data point influence varies with pp?

局限性

yes

最终评判理由

Rating remains the same. Reasons are provided in my response to the authors' rebuttal.

格式问题

no

作者回复

Dear Reviewer fc4o,

Thank you for your kind and thoughtful review!

About the Notion of "Efficiency"

Your point is correct. We certainly don't mean to mislead and will clarify it in the final version. Currently, our use of "efficiency" refers to the Verifier's sample complexity (the number of retrainings), which is itself independent of the dataset size N. The cost of each individual retraining does of course depend on N, and we will make this distinction explicit in the final version of our paper.

Non-determinism in GPU training

We would like to point out that we did discuss this, in the paragraph titled "Equivalence Checking Procedure" (Lines 536-543) of the submission. While the theorems as stated do assume a perfect equivalence checking procedure, the check-equiv procedure can be instantiated practically by, for instance, comparing model outputs or final loss values on a held-out set, rather than requiring a bit-for-bit identity of model weights. Any error in equivalence checking can be wrapped into the probabilities for soundness and completeness via union bounds.

Regarding your question:

How sensitive are the attribution vectors to the choice of pp in the pp-Bernoulli marginal distribution over selection vectors? For example, are there empirical studies examining how the ranking of data point influence varies with pp?

In general, the choice of pp does affect the attributions. Different values of pp were used and studied in, for example, [1]. However, due to high linearity observed in practice on typical benchmark settings (e.g., ResNet-18 trained on CIFAR-10), the values of attributions are not "very" sensitive to pp.

Thanks again for your kind review!

Sincerely,

Authors


References

[1] Ilyas, A., Park, S. M., Engstrom, L., Leclerc, G., and Madry, A. (2022). Datamodels: Predicting predictions from training data. arXiv preprint arXiv:2202.00622.

评论

Thank you for the clarification and for being receptive to my comment about notions of efficiency for the interactive protocol. Since my questions were only minor clarifications, they did not affect my (strongly positive) assessment of the paper. Therefore, my rating remains unchanged. Great work!

最终决定

Estimating data subset influence on an ML model is computationally expensive due to repeated retraining. Building on Saunshi et al. (2022), who proposed an efficient linear approximation, this work introduces an interactive protocol where a weak verifier outsources this task to a powerful but untrusted prover. The verifier spot-checks a small fraction of the prover's computations. The paper's main technical contribution is proving that the approximation method is robust against a small number of errors, which allows the verifier to efficiently detect a dishonest prover. All reviewers are supportive, and the paper would make a great addition to NeurIPS.