PaperHub
4.9
/10
Poster4 位审稿人
最低2最高4标准差0.8
4
2
2
3
ICML 2025

Optimal Fair Learning Robust to Adversarial Distribution Shift

OpenReviewPDF
提交: 2025-01-24更新: 2025-07-24

摘要

关键词
FairnessRandomizationClassificationBayes-OptimalRobustnessDistribution ShiftAdversarial NoiseCorrupted Data

评审与讨论

审稿意见
4

This paper demonstrates that randomized fairness-aware classifiers have the local Lipschitz property, which makes them (somewhat) robust to adversarial perturbations of their training data. The authors demonstrate that randomization is crucial for this property (confirming a previously known result), but also demonstrate that it may be confined to a single point; finally, they bound the Lipshitz constant for several popular fairness criteria.

给作者的问题

None

论据与证据

Yes.

方法与评估标准

There are no experiments in this paper.

理论论述

Yes; I did not find any issues.

实验设计与分析

There were no experiments.

补充材料

I read the appendix.

与现有文献的关系

The key contributions present some properties of randomized fairness-aware classifiers that, as far as I know, are novel. As such, this work adds to the literature that studies randomized fairness-aware classifiers and demonstrates their advantages over deterministic ones in cases of fairness thresholds.

遗漏的重要参考文献

None, to my knowledge.

其他优缺点

I really appreciated the mostly clear writing of the paper and the intuitions provided. The authors’ positioning of their work in a sociological context by discussing the advantages and disadvantages of randomized classifiers vis-a-vis the sociological goals was also very nice.

其他意见或建议

The intuition in Section 1.2. was not immediately very clear, it would have been helpful to have more explanation for how the 1-skeleton of the intersection of the hyperplane and the hypercube relates to the classifier and the datapoints.

作者回复

Thank you for your thoughtful feedback and for appreciating the novelty of our results and the clarity of presentation. We recognize the importance of making the intuition in Section 1.2 as clear as possible for better readability of the rest of our paper. Acting on your suggestion, we will update our manuscript with a diagram (similar to the visualization here https://cococubed.com/images/raybox/five_shapes.pdf).

We provide a brief verbal explanation here: The hypercube represents the space of all classifiers, with each dimension corresponding to a data point. The fairness constraint is modeled as a hyperplane. The intersection of any hyperplane with the hypercube results in a bounded region whose vertices are either (i) vertices of the hypercube itself or (ii) points formed by the intersection of the hyperplane with an edge of the hypercube. In other words, all vertices of this intersection lie on the 1-skeleton (or polytope graph) of the hypercube. Accuracy can be understood as a directional objective, and the fair BOC corresponds to an extremal point in that direction on the intersection of the hyperplane and the hypercube. Since, in every direction, an extremal point must be a vertex, and all vertices lie on the 1-skeleton, it follows that the fair BOC necessarily lies on the 1-skeleton of the hypercube.

We hope that our response adequately addresses your questions, and we sincerely request your support towards a consensus for acceptance of our paper.

审稿人评论

I thank the authors for the additional context. I will keep my original rating for this paper. While I wouldn't call this work groundbreaking, I think that it's a solid contribution in terms of the new results that it contributes regarding adversarial robustness, as well as providing a nice and helpful intuition for fair classifiers and the role that randomization plays.

审稿意见
2

This paper analyzes/bounds the robustness of the optimal fair classifier ff (w.r.t. a distribution P\mathcal P) under distribution shifts in terms of both accuracy and performance. The specific setting considered is binary class, binary group, in the attribute-aware setting, and the fairness criteria considered are SP, EO, EOpp (can be extended to other similar metrics).

In other words, what would the increase in the accuracy and unfairness of ff be if we evaluate it on PP\mathcal P' \neq\mathcal P?

给作者的问题

See Other Strengths And Weaknesses

论据与证据

Yes

方法与评估标准

N/A

理论论述

Proofs are not checked, but the theorem statement/results are reasonable.

实验设计与分析

N/A

补充材料

Did not review supplementary material.

与现有文献的关系

This paper considers the robustness of fair classifiers subject to distributional shifts; the exact contribution is the bound/sensitivity analysis for accuracy and unfairness of an optimal fair classifier evaluated on a shifted distribution.

But the reviewer is aware of similar results in prior publications that appear to subsume the results in this paper: see "Other Strengths And Weaknesses"

遗漏的重要参考文献

See Other Strengths And Weaknesses.

其他优缺点

Weaknesses.

While the results are concrete and rigorously established, the reviewer would like the authors to comment on the similarities/differences between these results and the sensitivity analysis in prior work for optimal fair classifiers:

  • [Xian & Zhao, 2024]: Theorem 3.1 in Section 3.2 Sensitivity Analysis bounds the accuracy and unfairness of a classifier that is optimal and fair under a different distribution (specified via r^r\hat r\neq r for a shift in the label distribution, and g^g\hat g\neq g for a shift in the group membership distribution); this analysis applies to SP, EO and EOpp under attribute-blind setting for multigroup and multiclass classification, which is more general than the setting considered in this paper. They also seem to show that their bound has a matching lower bound by an example.
  • [Chen et al., 2024]: Section E contains a similar sensitivity analysis, for the same setting except binary group and binary class.
  • [Chen et al., 2022]: They also consider fairness under distribution shift, with a similar result that involves a Lipschitz condition. Could the authors comment on any similarities/differences between the two works?

[Xian & Zhao, 2024] A Unified Post-Processing Framework for Group Fairness in Classification
[Chen et al., 2024] Post-hoc Bias Scoring Is Optimal For Fair Classification
[Chen et al., 2022] Fairness Transferability Subject to Bounded Distribution Shift

其他意见或建议

N/A

作者回复

Thank you for your time and insightful feedback. We will cite and discuss the novel results in the papers you pointed out. However, our results are incomporable to those of the papers mentioned. Fundamentally, the underlying settings are different, and these results complement each other.

1. Comparison with [Xian & Zhao, 2024]:

Significantly, our setup assumes PXP_X is a discrete distribution (over either a discrete or continuous domain XX), and we will make this explicit in our revision (note that this is a standard assumption for realistic tabular datasets with features like height, income, race, etc.). On the other hand, [Xian & Zhao, 2024] implicitly assume XX is a continuous domain, and their results in general hold only for continuous distributions PXP_X over XX. See Remark 2.4 (fail cases of Assumption 2.2) in [Xian & Zhao, 2024], where the results don’t hold when the push-forward distribution r#PXr\#P_X contains atoms and the randomized Fair BOC splits their mass, corresponding to the case our paper deals with, i.e., discrete distributions (See Claim 1 in our paper).

There exist many examples of discrete distributions where the optimal fair classifier must be randomized, and resorting to a deterministic classifier must necessarily come with a sharp drop in accuracy (even with the perturbation step described in Proposition 2.3 of [Xian & Zhao, 2024]); see the example in Claim 1 in our paper, where perturbing the scores/risks of each point would not do away with the need for randomization. The technical challenges in the discrete setup are very different from those in the continuous setup. For example, in most discrete distributions, the optimal fair classifier must be randomized, whereas in continuous distributions a deterministic classifier suffices under Assumption 2.2 in [Xian & Zhao, 2024]. Unlike the continuous setting, enforcing determinism in the discrete setting typically leads to non-robustness, for example in Claim 1.

We also note that the approach mentioned in Remark 2.4 of [Xian & Zhao, 2024] to handle discrete distributions, i.e., smoothing the discrete distribution and then applying the algorithm in the paper with a deterministic classifier, does not apply to discrete domains; furthermore, for continuous domains it will almost certainly lead to a drop in accuracy.

To summarize, while our submission and [Xian & Zhao, 2024] share the focus on robustness and fairness, our respective underlying frameworks are fundamentally different and complementary.

Finally, we emphasize that our robustness guarantees (Theorems 1, 2, & 3) cover adversarial distribution shifts over X×Z×YX\times Z \times Y, whereas the sensitivity analysis in Theorem 3.1 of [Xian & Zhao, 2024] covers either a shift in label distribution over X×YX\times Y, or a shift in group distribution over X×ZX \times Z.

2. Comparison with [Chen et al., 2024]: Unlike us, they do not deal with adversarial distributions shifts, but only label distribution shifts and/or group distribution shifts. In addition, our setups are fundamentally different, theirs being the continuous case, and ours being the discrete case. Moreover, their sensitivity analysis in Theorem 2 is looser, and has an extra additive error term, unlike ours and that of [Xian & Zhao, 2024]. Besides, they do not deal with the case of perfect fairness, and require δ>0\delta > 0.

3. Comparison with [Chen et al., 2022]: Their result is fundamentally different, and essentially shows that the fairness of a fixed hypothesis class on two similar distributions is similar. This is essentially what we show in Claims 2/5/6, however, they only deal with label and covariate shifts, while we tackle the more general case of adversarial distribution shifts.

We hope that our response has addressed your concerns, and if so, we respectfully request you to consider increasing your rating.

审稿人评论

The reviewer would like to thank the authors for the response.

Regarding point 1 [Xian & Zhao, 2024]:

  • The reviewer would like to point out that although their proposed algorithm requires the continuity assumption, their sensitivity analysis in theorem 3.1 does not... in fact, the result covers arbitrary distributions, discrete, continuous, or mixed, whereas the present work is only discrete. There is also this difference in generality that their result handles attribute-blind, multi-class, and multi-group, whereas the present only attribute-aware, binary class, and binary group. This, in reviewer's opinion, remains a major weakness...
    • Although now irrelevant, the reviewer would like to point out that the smoothing technique used in theirs incur arbitrarily small drop in performance (see their discussions in section 2.2)
  • The reviewer agrees with the authors that their theorem 3.1 only handles label shift and not covariate shift. Although, in the reviewer's humble opinion, an extension to handle covariate shift should be straightforward...

The reviewer's main concern regarding the limitation in the settings covered by the results in present work still holds, and as such, would like to keep the current rating—but considering that there is some novelty in handling covariate shifts, the reviewer would not be upset if the paper is accepted. In either case, the reviewer encourages the author to discuss and include these comparisons to prior work in the revision.

作者评论

Thank you for reading our rebuttal and several thoughtful comments regarding comparison with the recent, unpublished work of [Xian-Zhao 24] (arxiv, 23 Dec 2024). We will definitely cite and discuss their work, and also explain why our results are not subsumed by [Xian-Zhao 24]. We highlight some strong points of our results below.

  1. Adversarial distribution shifts: The sensitivity analysis (Theorem 3.1) in [Xian-Zhao 24] only holds for label or group shifts, whereas our robustness guarantee works for adversarial distribution shifts. Adversarial or arbitrary distribution shifts are strictly more general than label/covariate shifts, and moreover, they cannot be simulated by any combination of label/covariate shifts.
  2. Tighter analysis: In the sensitivity analysis (Theorem 3.1, 2nd result) in [Xian-Zhao 24], the change in accuracy due to group distribution shift, is a constant in the case of perfect fairness (i.e., α=0\alpha = 0). Then the expression min(1α,ϵgα+ϵg)\min\left(1 - \alpha, \frac{\epsilon_g}{\alpha + \epsilon_g}\right) evaluates to 11, so the excess risk is an additive 2r2||r||_{\infty}, which is a constant independent of the amount of distribution shift. We prove a stronger Lipschitzness guarantee, where the excess risk goes to 00 as distance between the distributions becomes arbitrarily small.
  3. Characterization and minimal randomness: Theorem 2.1, and the perturbation/smoothing algorithm in [Xian-Zhao 24] does not hold for discrete domains. While the LP can be used to compute a randomized optimal classifier for discrete domains, there is no description of this classifier. On the other hand, we provide a complete characterization of the same, and show the existence of a nearly-deterministic optimal solution with minimal randomness, whereas in [Xian-Zhao 24], the optimal solution of the LP could use an arbitrary amount of randomization.
  4. Complexity: Our algorithm is very simple and efficient, running in O(Xlog(X))O(|X| \log (|X|)), while the algorithm in [Xian-Zhao 24] solves a large linear program with O(X)O(|X|) constraints in O(X)O(|X|) variables, requiring a much higher complexity.
  5. Multiple sensitive groups: We recognize that [Xian-Zhao 24] applies to attribute-blind, multi-class, multi-group fair classification. Our algorithm and robustness guarantee can also be extended to the case of multiple sensitive groups, as stated in our Footnote 1 (page 3).
  6. Cost-sensitive risk: As stated in our Footnote 2 (page 3), in addition to the standard 00-11 loss 01\ell_{0-1}, our results also hold for the more general loss function α\ell_{\alpha} known as cost-sensitive risk [Menon-Williamson, 2018] that assigns weight α\alpha to False Positives and weight (1α)(1-\alpha) to False Negatives.
  • Menon, A. K. and Williamson, R. C., The cost of fairness in binary classification, FAccT 2018.
审稿意见
2

The paper studies the fairness-aware classification problem, where the goal is to maximize accuracy subject to the demographic parity or equal opportunity fairness constraint. The paper first discusses Claim 1 that a deterministic classification rule can have high sensitivity to perturbations to the data distribution. An example is discussed under Claim 1 to support this claim. Next, the discussion focuses on evaluating the sensitivity of the randomized fair classifier and the main result in Theorem 1 (and Corollary 1) shows an ϵ\epsilon TV-perturbation to the data distribution leads to a worst-case ϵ\epsilon-times-a sensitive attribute probability constant difference in the accuracy of the optimal fair classifier.

给作者的问题

1- Can the authors provide standard fairness-aware classification settings (like COMPAS and Adult) where the observation in Claim 1 would be relevant?

2- Why the example in the proof of Claim 1 can hold in non-degenerate scenarios where the space of variable XX can include a large number of outcomes?

3- Can the proof of Theorem 1 be shortened and is Lemma 1 necessary to prove the lemma?

论据与证据

Only partially. My main concern with this submission is that the authors do not present any numerical and enough theoretical evidence that their theoretical claims are relevant to real non-synthetic data distributions. As I said, the paper does not provide any numerical evaluation of the claim that "deterministic fair classifiers are sensitive to adversarial noise".

Concerning the theory results, I think Claim 1 gives a rather degenerate setting where the deterministic fair classifier is sensitive to noise. The example discussed in the proof of Claim 1 focuses on a Binary input XX, and from the construction of the example, it seems to me that the case is degenerate and highly depends on a binary input xx that is independent of the sensitive attribute variable ZZ and then perturbed to exhibit correlation with the sensitive attribute ZZ.

I am wondering if such an example would be the case for a standard fairness-aware classification setting, such as COMPAS and Adult data. A significant difference in those real-world examples is that the input variable XX is multi-dimensional and the cardinality of possible input vector xx is much greater than the binary setting in Claim 1. Therefore, I am highly unsure if the issue highlighted in Claim 1 would be relevant to actual fairness-aware tasks. I believe the paper should provide enough evidence that this issue is more significant than a degenerate scenario, particularly because the entire paper is dedicated to showing the issue exists in the deterministic case, while it is not the case with randomized classifiers.

方法与评估标准

The paper has no numerical experiments on fairness-aware classification. Regarding the evaluation criteria in the analysis, the paper is missing the discussion of approximately fair classifiers. In practice, the demographic parity condition is not fully enforced in the classification, and instead a dependence quantity, such as mutual information or correlation metric, is regularized in the fairness-aware learning process. Therefore, the independence condition in DP does not strictly hold in real-world applications.

Studying the paper's main results and argument, it seems to me that the Binary-input XX example in Claim 1 can be easily addressed by relaxing the DP constraint to a bounded mutual information or maximal correlation. As said above, this has not been discussed in the paper.

理论论述

The main text includes the proof of Claim 1 and Theorem 1, and I think the theorems are correct. Still, I was wondering if the proof of Theorem 1 could have been shorter. On another note, I was wondering why the authors use about 3 pages to prove Theorem 1 before stating the result. I suggest discussing the theorem right after Claim 1 and postponing the proof to the supplementary material. The saved space can be used for numerical evaluation of the results and further theoretical discussion.

实验设计与分析

The paper presents no experimental results on fairness-aware classification.

补充材料

Not in detail, I read the theorems on NP-completeness of deterministic fair classifier and the extension of the results to equal opportunity and predictive equality.

与现有文献的关系

I think the literature review and paper's discussion is missing a large body of related works on approximate fairness in fairness-aware classification. This includes references suggesting to replace the hard demographic parity constraint with regularizing dependence measures such as mutual information, Pearson correlation, HGR correlation, ff-mutual information, and its special cases. In practice, the demographic parity and equal opportunity are not enforced with hard constraints, and instead there is an optimization constraint or regularization penalty term to penalize significant values of dependence between the variables. I understand that the paper focuses on the extreme case of completely fair classifier, but the discussion should still include some clues on how the results would theoretically or empirically relate to more practical scenarios.

遗漏的重要参考文献

Please see my answer to the above question.

其他优缺点

Please see my previous answers.

其他意见或建议

Please see my previous answers.

作者回复

Thank you for your time and insightful feedback. We respond to your comments below.

1. Real-world datasets: Our setup assumes a distribution over a discrete domain space XX, which is a standard assumption for real-world tabular datasets (e.g., COMPAS) with features such as age, race, DoB, prior counts etc. We do not make any other assumptions on XX. In particular, XX could be large, and multidimensional. The same holds for the non-robustness phenomenon is Claim 1. While experiments could be insightful, we emphasise that the focus of our paper is to establish the theoretical foundations of robust fair learning, building upon the theoretical papers of Konstantinov-Lampert (2022), and Blum et al. (2024).

2. Larger, non-degenerate example: In Claim 1, we provided a small example, independent of the sensitive attribute, just to demonstrate the phenomenon of non-robustness that occurs in optimal fair deterministic classifiers. The example can be made much larger (in fact, it can be made arbitrarily large). In addition, XX can also be multidimensional, and can depend on the sensitive attribute. The below example assumes X=10|X| = 10, with XX also depending on the sensitive attribute. Consider the following distribution PP.

P,S(x1,A)=(0.055,0.9)P,S(x_1, A) = (0.055, 0.9) | P,S(x1,D)=(0.045,0.91)P, S(x_1, D) = (0.045, 0.91) | P,S(x2,A)=(0.055,0.8)P, S(x_2, A) = (0.055, 0.8) | P,S(x2,D)=(0.05,0.78)P, S(x_2, D) = (0.05, 0.78) | P,S(x3,A)=(0.045,0.73)P, S(x_3, A) = (0.045, 0.73) | P,S(x3,D)=(0.045,0.71)P, S(x_3, D) = (0.045, 0.71) | P,S(x4,A)=(0.04,0.67)P, S(x_4, A) = (0.04, 0.67) | P,S(x4,D)=(0.06,0.63)P, S(x_4, D) = (0.06, 0.63) | P,S(x5,A)=(0.055,0.53)P, S(x_5, A) = (0.055, 0.53) | P,S(x5,D)=(0.05,0.51)P, S(x_5, D) = (0.05, 0.51) | P,S(x6,A)=(0.045,0.45)P, S(x_6, A) = (0.045, 0.45) | P,S(x6,D)=(0.05,0.46)P, S(x_6, D) = (0.05, 0.46) | P,S(x7,A)=(0.055,0.33)P, S(x_7, A) = (0.055, 0.33) | P,S(x7,D)=(0.06,0.32)P, S(x_7, D) = (0.06, 0.32) | P,S(x8,A)=(0.04,0.28)P, S(x_8, A) = (0.04, 0.28) | P,S(x8,D)=(0.04,0.24)P, S(x_8, D) = (0.04, 0.24) | P,S(x9,A)=(0.06,0.17)P, S(x_9, A) = (0.06, 0.17) | P,S(x9,D)=(0.055,0.15)P, S(x_9, D) = (0.055, 0.15) | P,S(x10,A)=(0.05,0.09)P,S(x_{10}, A) = (0.05, 0.09) | P,S(x10,D)=(0.045,0.7)P, S(x_{10}, D) = (0.045, 0.7)

Consider ff, which classifies {x1,...,x5}×{A,D}\{x_1,..., x_5\} \times \{A, D\} as 11, and {x6,...,x10}×{A,D}\{x_6,...,x_{10}\} \times \{A, D\} as 00. ff satisfies DPDP, and has accuracy much more than 0.50.5. Consider the neighboring PP^′ which differs with PP on (x5,A)(x_5, A), and (x5,D)(x_5, D), as follows.

P,S(x5,A)=(0.055+ϵ,0.53)P', S(x_5, A) = (0.055 + \epsilon, 0.53) | P,S(x5,D)=(0.05ϵ,0.51)P', S(x_5, D) = (0.05 - \epsilon, 0.51),

where ϵ>0\epsilon > 0 is arbitrarily small. There are only 2 deterministic classifiers satisfying DP on PP’, either the constant 1 classifier f1f_1, or the constant 0 classifier f0f_0, both with accuracy close to 0.5. Hence, the difference in accuracy of the deterministic DP-fair BOC on arbitrarily close P,PP, P^′ is large, demonstrating non-robustness.

3. Proof length: Perhaps the proof can be shortened, we will try and simplify it to aid readability. Lemma 1 played a crucial role in simplifying the proof overall, but it can be done without Lemma 1 as well.

4. Approximate fairness: We show through the example below that the non-robustness phenomenon highlighted in Claim 1 also holds when we only require approximate fairness. In particular, this can hold in the case where sensitive group populations are highly imbalanced, for example when the mass of group AA is much larger than the mass of group DD, i.e., P(A)>>P(D)P(A) >> P(D). We define a δ\delta-approximately fair classifier as follows, “If rr denotes selection rate, a classifier ff is δ\delta-approximately DP-fair if r(f,A)r(f,D)<δ|r(f, A)−r(f, D)| < \delta". We set δ=0.25\delta=0.25, and slightly modify the example in Claim 1, where we skew the probability mass towards Group AA (in Claim 1, the group masses are balanced). Consider a distribution P, where

P,S(x1,A)=(0.4,1)P, S(x_1, A) = (0.4, 1) | P,S(x1,D)=(0.1,1)P, S(x_1, D) = (0.1, 1) | P,S(x2,A)=(0.4,0)P, S(x_2, A) = (0.4, 0) | P,S(x2,D)=(0.1,0)P, S(x_2, D) = (0.1, 0)

Consider the (deterministic) classifier ff, with f(x1,A)=f(x1,D)=1,f(x2,A)=f(x2,D)=0f(x_1, A) = f(x_1, D) = 1, f (x_2, A) = f (x_2, D) = 0. ff satisfies DP, and Acc(f)=1\text{Acc}(f) = 1. Consider the neighboring distribution PP^′ differing only on (x1,D),(x2,D)(x_1, D), (x_2, D), as follows.

P,S(x1,D)=(0.1+0.05,1)P', S(x_1, D) = (0.1 + 0.05, 1) | P,S(x2,D)=(0.10.05,0)P', S(x_2, D) = (0.1 - 0.05, 0)

If we apply ff on PP’, it does not satisfy approximate DP for any δ<0.25\delta < 0.25, even though TV(P,P)TV(P, P’) is small (0.050.05). There are only 2 deterministic classifiers satisfying approximate DP for any δ<0.25\delta < 0.25, either the constant 1 classifier f1f_1, or the constant 0 classifier f0f_0, with Acc(f1)=1/2+0.05\text{Acc}(f_1) = 1/2 + 0.05, and Acc(f0)=1/20.05\text{Acc}(f_0) = 1/2 − 0.05. Hence, the difference in accuracy of the deterministic (approximate) DP-fair BOC on closeby P,PP, P^′ is almost 0.5, demonstrating non-robustness.

We will also discuss fairness regularizers based on mutual information, Pearson correlation, HGR correlation, and f-mutual information. These are relevant for broader literature review but not directly related to the problem we study.

We hope we have addressed all your concerns, and we sincerely request you to reconsider our rating.

审稿人评论

I thank the authors for their responses to my comments. I appreciate theoretical contributions to the machine learning literature, and I found the discussion comparing randomized and deterministic classification rules in the paper to be interesting.

However, I am still unconvinced that the issues raised in the paper have significant implications for real-world applications of fairness-aware classification. I would appreciate the authors' further clarification on the following points:

  1. Can the authors provide any numerical evidence using real datasets where the theoretical discussion described in the paper leads to any demonstrable form of lack of robustness in deterministic fairness-aware classification rules? There is no such numerical evidence in the paper or the authors' response.

  2. Is there a general class of distributions on X×Y×ZX \times Y \times Z under which the robustness issue for deterministic fairness-aware classification rules provably holds? The arguments in the paper and the authors' response are based on specific, constructed examples, making it difficult to see the actual severity of the robustness issue.

作者评论

Thank you for reading our rebuttal and your insightful comments on our responses. We respond to your questions below.

  1. Robustness of fair classifiers on real-world data: Robustness of fair classifiers under bias/shift in the data distribution is a well-known issue in fair machine learning literature. Akpinar et al. (https://arxiv.org/pdf/2204.10233) empirically study the robustness of BOC and fair BOC on synthetic data distributions and provide a sandbox tool for stress-testing fair classifiers. Sharma et al. (https://arxiv.org/pdf/2302.05906) and Ghosh et al. (https://arxiv.org/pdf/2307.03306) empirically study robustness of fair classifiers under data bias on semi-synthetic real-world datasets (i.e., real-world datasets with synthetically injected bias/shift). In both these papers, Exponentiated Gradient Reduction (EGR) or ExpGrad (https://arxiv.org/pdf/1803.02453) stands out for its better robustness under data bias/shift, and it is inherently a randomized classifier. We can discuss these works, if you think it is not a digression from our main result. However, reconciling our theoretical results with empirical observations on real-world datasets is an important direction for future work, but not the focus of this paper.
  2. General class of distributions: Apart from the empirical evidence in the above papers, here is a uncountably infinite general class of distributions on which we can provably show non-robustness of deterministic fairness-aware classifiers. Consider a domain set as (x_1, x_2, … x_n)$$\times$$(A,D), where nn is an arbitrary positive integer. Now, for each point (xi,z)(x_i, z) in the domain (where z(A,D)z \in (A,D)), choose its probability mass and score to be random number from the uniform distribution over [0,1][0,1] (or any continuous distribution over [0,1][0,1]). Make sure this forms a probability distribution by normalizing the probability masses suitably. Sort the points in each group by score, and choose i,ji, j i.i.d. from (1,,n1)(1, …, n-1). We now want the boundaries of the first ii elements of AA, and first jj elements of DD to align. Choose rr from the uniform distribution over [0,1][0,1] (or any continuous distribution over [0,1][0,1]), and normalise suitably so that the first i(j)i(j) elements of group A(D)A(D) comprises of rr fraction of the total mass of that group. Call this transformed distribution PP. Now, randomly perturb the score/mass of each element by a small amount (for example, from the uniform distribution over [0,ϵ][0, \epsilon], for small ϵ>0\epsilon > 0), and call this neighbouring distribution PP’. For distribution PP, the (deterministic) classifier that accepts the first ii elements of AA, and first jj elements of DD, and rejects the others, satisfies DP, and is significantly more accurate than the constant 00 and constant 11 classifier. Also, it is easy to see that (almost surely) none of the boundaries will align for PP’. Hence, the only (deterministic) classifiers satisfying DP on PP’ are the constant 00 and constant 11 classifiers, resulting in a significant drop in accuracy, and demonstrating non-robustness.
审稿意见
3

The authors study the robustness of the Fair Fair Bayes Optimal Classifier (BOC) to adversarial noise in the data distribution. They show that the robustness guarantee for BOC breaks down when fairness constraints are aded and propose a randomized Fair BOC that is robust to malicious noise in the data distribution. They demonstrate this with various fairness constraints such as Demographic Parity, Equal Opportunity and Predictive Equality.

给作者的问题

While the randomized Fair BOC is shown to be nearly deterministic, the existence of stochasticity may limit it's applicability over BOC. Is this a concern?

论据与证据

The proofs for the theoretical claims are thorough. I would like to see additional empirical results as well to solidify the claims.

方法与评估标准

Yes

理论论述

Yes, I reviewed the proofs and claims but it is possible that I missed something.

实验设计与分析

There is no direct experimental design. An empirical analysis is missing.

补充材料

Yes, it was very thorough.

与现有文献的关系

Yes, fairness issues have grown in importance and ensuring that adding fairness constraints preserves robustness is an important contribution.

遗漏的重要参考文献

Not that I noticed.

其他优缺点

The motivation and the writing are very clear and easy to follow. The theoretical contributions and proofs are also clear.

其他意见或建议

I would suggest validating the contributions with empirical results as this would make the paper stronger.

作者回复

Thank you for your insightful feedback, and for appreciating the importance of our contribution, and the clarity of presentation. We respond to your comments below.

1. Experiments: We emphasise that the focus of our paper is to advance the theoretical foundations of robust fair learning, building upon the theoretical works of Konstantinov-Lampert (2022) and Blum et al. (2024). While experiments would be valuable, we leave that as future work.

2. Randomization: It is established that on most distributions, (exact) fairness is non-trivially achievable only by randomized classifiers [Agarwal et al, Hardt et al.]. In addition, the randomization in our result is almost nil, and is the minimum amount required (only on one element). If we strictly enforce determinism, the drop in accuracy is significant, as demonstrated in Claim 1. Hence, there is a clear tradeoff between randomization and accuracy (subject to fairness), and for a tiny amount of randomness, we obtain a big jump in accuracy.

  1. Moritz Hardt, Eric Price, and Nathan Srebro. Equality of Opportunity in Supervised Learning. NeurIPS 2016
  2. Alekh Agarwal, Alina Beygelzimer, Miroslav Dudík, John Langford, and Hanna Wallach. A Reductions Approach to Fair Classification, ICML 2018.

We hope that our response has addressed your concerns, and if so, kindly request you to consider increasing your score, and support towards a consensus for acceptance of our paper.

审稿人评论

I confirm that I have read the author response. I maintain my score. I still feel that the addition of experimental results to validate the theoretical contributions are needed to increase the impact of the paper. I am still recommending a weak accept as the theoretical contributions are strong and useful.

作者评论

Empirical validation: Below we discuss some recent works that empirically study robustness of fair classifiers under data bias/shift and explain the limitations and challenges in doing similar empirical validations for our paper. Akpinar et al. (https://arxiv.org/pdf/2204.10233) study robustness of Bayes Optimal Classifier (BOC) and fair BOC on synthetic data distributions by injecting bias/shift in a stylized set-up. They use a simple Gaussian data distribution (Subsection 4.2 (1) in their paper), so that BOC and fair BOC are linear because these classifiers are NP-hard to compute on general distributions. Sharma et al. (https://arxiv.org/pdf/2302.05906) and Ghosh et al. (https://arxiv.org/pdf/2307.03306) study robustness of fair classifiers under injected data bias/shift on semi-synthetic data (i.e., real-world datasets with synthetically injected bias/shift). Both these papers observe that the Exponentiated Gradient Reduction (EGR) or ExpGrad (https://arxiv.org/pdf/1803.02453) stands out for better robustness under data bias/shift, and it is a randomized classifier by construction (i.e., a random ensemble of classifiers). Our work studies deterministic-vs-randomized BOC and fair BOC under a more general adversarial data bias/shift that is much harder to construct empirically on real-world data. As you mentioned, reconciling our theoretical results with empirical observations on real-world datasets is important. However, we consider it as independent future work and cannot address it here given the focus of our current paper and limited time. We greatly appreciate your constructive criticism nevertheless.

最终决定

The papers study the robustness of fair Bayes optimal classifiers under malicious noise.

There was a good amount of discussion between us about the pros and cons of this paper. Some reviewers appreciated the clear motivation of the paper and highlighted the importance of randomness in the robustness of fair qualifiers. Other reviewers, on the other hand, found the example provided in Claim 1 to be unrealistic or requested empirical study. After going back and forth, I sided with the former group. I consider the example to demonstrate solely the claim that randomness is necessary. Also, this is a theory paper, and ICML papers are not required to contain an experimental section. Therefore, I would be happy to recommend this paper for acceptance.

Finally, although the work of Xian-Zhao 24 pointed out by Reviewer 9j8M counts as concurrent work based on ICML policy, I recommend the authors cite this paper as concurrent work and describe the similarities and differences now that they are aware of it.