PaperHub
5.5
/10
Rejected3 位审稿人
最低3最高3标准差0.0
3
3
3
ICML 2025

EA-PS: Estimated Attack Effectiveness based Poisoning Defense in Federated Learning under Parameter Constraint Strategy

OpenReviewPDF
提交: 2025-01-20更新: 2025-07-26
TL;DR

EA-PS method minimizes the long-lasting backdoor attack effect with a parameter constraint strategy to enhance stability.

摘要

关键词
"Backdoor Poisoning Defense""Robust Federated Learning"

评审与讨论

审稿意见
3

This paper proposes a client-side defense methond in federated learning, EA-PS, that constrains the pertubation range of local parameters while minimizing the impact of attacks by forming the problem into an optimization problem. This paper further provides convergence and robustness analysis. This paper validates its algorithm through experiments.

给作者的问题

  1. why can we assume \lambda is a linear set of A? What will be sacrifice with this assumption?
  2. How is \tilde{H} calculated?

论据与证据

This paper claims that with EA-PS, combined with server-side defense method, can achieve robust and stable performance under attack. The claims are clearly supported through theoretical results and empirical results.

方法与评估标准

The experiments are extensive with varying parameters of the algorithms, supporting the claim of the paper.

理论论述

I have checked the proof of Theorem 4.1, no issues discovered.

实验设计与分析

I have checked the experiments compared proposed methods against other baselines, varying \beta, varying \alpha, different \lambda and different \gamma. No issues discovered.

补充材料

I have reviewed Appendix A.1, A.2. Appendix B, C and D. No issues discovered.

与现有文献的关系

The key contributions of the paper is to constraint the local parameter updating in federated learning under attack, which reduce the variance during learning process.

遗漏的重要参考文献

There are no essential references not discussed to my knowledge.

其他优缺点

Strengths: Experiments are extensive with strong theoretical guarantees. Weaknesses: Preliminary knowledge isn't explanied enough, make it hard to follow the paper. For example, why A_t - A_{t-1} can be interpreted as long-lasting attacks? what is the definition of long-lasting attacks? Figure 2 didn't illustrate the idea of parameter constraint strategy.

其他意见或建议

None.

作者回复

Thank you for your thorough analysis and constructive feedback on our paper. We appreciate the opportunity to clarify the points raised and to provide additional insights into our research.

  1. What is the definition of long-lasting attacks? Why AtAt1A_t - A_{t-1} can be interpreted as long-lasting attacks?

Response:

  • We appreciate the opportunity to clarify the definition of long-lasting attacks. In our work, we follow the FL-WBC's observations on the long-lasting attack effect. The definition of "long-lasting" in the long-lasting attack is to describe the effects of an attack in the current round that can persist through multiple rounds of training. Therefore, there exists a slight misleading in our work. The correction is as follows: “long-lasting attack” to "attack with long-lasting effects ".

  • AtAt1A_t-A_ {t-1} in our work is "the difference of different rounds in the coefficient of attack effects". AtA_t is defined as "the coefficient of attack impact between two rounds". AtAt1A_t-A_ {t-1} can form a chain structure to better measure the accumulation of attack impact in different periods (i.e., long-lasting attack effects). For example, in our work, Theorem 4.1 shows that minimizing AtAt1A_t-A_ {t-1} yields a smaller optimization upper bound than a traditional method such as LeadFL; Theorem 5.2: Certified Radius analysis shows that the reduction of AtAt1A_t-A_ {t-1} improves the robustness of the model against long-term attack effects. The experimental results also support the above proof.

  1. Figure 2 didn't illustrate the idea of parameter constraint strategy.

Response:

  • We believe that Figure 2 illustrates the idea of the parameter constraint strategy. Based on the last paragraph of section 3, we will provide additional insights into our strategy. The goal of the parameter constraint strategy is to enhance the stability of poisoning attack defense by constraining the perturbation range of model parameters. The key components are 1)Optimized Manifold Space AA; 2) Unit Space II; 3) Rank Constraint λλ.

  • Optimized Manifold Space AA represents the unconstrained parameter space of the model, which may involve high-dimensional or complex parameter distributions. In this space, malicious attacks (e.g., backdoor attacks) can create long-lasting effects through parameter perturbations. As illustrated in equations (formalized as equation (16): I=B1ABI=B^{-1}AB), the manifold space AA is mapped into a simpler, low-dimensional unit space (Unit Space II). By constraining parameter perturbations within a bounded region (Rank Constraint λλ), the strategy suppresses the cumulative effects of adaptive or persistent attacks (formalized as equation (17): AB=λBAB=λB). This ensures stable defense performance under long-lasting attack effects.

  1. Why can we assume λ\lambda is a linear set of AA? What will be sacrifice with this assumption?

Response: For the reason why we can assume λ\lambda is a linear set of AA, based on the references and descriptions on page 5 of our work, we will provide additional insights into it. Firstly, we assume that λλ is a linear set of AA mainly because the linear decision rule can transform complex uncertainty descriptions into a more tractable linear form, thus yielding a computationally solvable robust optimization model. Specifically, the linear assumption simplifies the complex parameter constraints into a linear combination of historical information. Moreover, linear approximations facilitate the simplification of proofs for convergence and robustness guarantees. However, setting λλ as a linear combination of AA means ignoring possible nonlinear relationships, which may result in suboptimality according to the linear decision rule (Bertsimas et al., 2019), which has noted on page 5 of our work.

  1. How is HH calculated?

Response: We appreciate the opportunity to add details about how HH is calculated. We will change the equation of Ht,ekH_{t,e}^k in our work (section 4.2) to illustrate the details of how HH is calculated, as follows.

Ht,ek=2F(θt,ek)=(θt,e+1kθt,ekθt,ek)/ηt.H_{t,e}^k \overset{\bigtriangleup }{\underset{}{=}}\bigtriangledown ^2 F( \theta _{t,e}^k)= (θ _{t,e+1}^k-θ _{t,e}^k-∆θ _{t,e}^k)/ η_t.

We hope this response adequately addresses your points and welcome the fruitful discussion. We are thankful for the contribution to the manuscript's refinement.

审稿意见
3

This paper proposes EA-PS (Estimated Attack Effectiveness-based Poisoning Defense with Parameter Constraint Strategy), a client-side defense designed to constrain the perturbation range of local parameters while minimizing the impact of attacks. The authors prove that our methods have an efficiency guarantee with a lower upper bound, a robustness guarantee with a smaller certified radius, and a larger convergence upper bound.

给作者的问题

See above

论据与证据

  • efficiency guarantee with a lower upper bound Evidence: 4.2 and Appendix A.2

  • a robustness guarantee with a smaller certified radius 5.3 and Appendix A.4 Theoretical analysis

  • a larger convergence upper bound 5.2 and Appendix A.3 Theoretical analysis

方法与评估标准

  1. introduce an enhanced objective function (EA-PS−)
  2. propose a client-based defense approach named Estimated Attack Effectiveness based Poisoning Defense method under Parameter Constraint Strategy (EA-PS). It minimizes the long-lasting backdoor attack effect with a parameter constraint strategy to enhance stability by constraining the perturb range in the parameter space

Evaluation: main task accuracy(MA), backdoor accuracy(BA) I think the evaluation metric is reasonable Dataset: FashionMNIST and CIFAR10 datasets under both IID and non-IID settings. I think the datasets, although widely used in FL, do not contain real-world noniid and is too simple. can the author also consider dataset like FEMINIST, which contains nature non-iid? The attack method used here is only one pattern. I encourage the author to check the performance also for untargeted attack.

理论论述

  • efficiency guarantee with a lower upper bound Evidence: 4.2 and Appendix A.2

  • a robustness guarantee with a smaller certified radius 5.3 and Appendix A.4 Theoretical analysis

  • a larger convergence upper bound 5.2 and Appendix A.3 Theoretical analysis

I don't see flaws in the theoretical analysis , yet I'm not an expert of theory, please refer to other reviewer's suggestions.

实验设计与分析

As mentioned before: Dataset: FashionMNIST and CIFAR10 datasets under both IID and non-IID settings. I think the datasets, although widely used in FL, do not contain real-world noniid and is too simple. can the author also consider dataset like FEMINIST, which contains nature non-iid? The attack method used here is only one pattern. I encourage the author to check the performance also for untargeted attack, and other targeted attack, such as follows:

[1] ] Xiaoyu Cao and Neil Zhenqiang Gong. 2022. Mpaf: Model poisoning attacks to federated learning based on fake clients. [2] DBA: Distributed Backdoor Attacks against Federated Learning

补充材料

Yes, theoretical analysis.

与现有文献的关系

It supplements the client-side poisoning defense with theoretical guarantees.

遗漏的重要参考文献

I don't know well the related literature for client-side defense. But I think the attacks evaluated are limited, as mentioned above

其他优缺点

Table 1 caption is not quite clear benign accuracy / (attack success rate)?

其他意见或建议

See above.

作者回复

Thank you for your recognition of our work and for your insightful comments.

  1. Table 1 caption is not quite clear benign accuracy / (attack success rate)?

Response: The metric used in Table 1 is backdoor accuracy, which is the attack success rate for backdoor attacks. We will change it to "backdoor accuracy" in the revision.

  1. Can the author consider FEMINIST, which contains nature non-iid? I encourage the author to check the performance for untargeted attack (MPaf) and other targeted attack (DBA).

Response:

  • We appreciate your suggestion to include the nature non-iid dataset (FEMINIST), MPaf, and DBA. In addition to the methods suggested above, we also added Spectrum (targeted) and Label-Flip (untargeted) to further enhance the experiment. The results are as follows.
(%)FEMINIST
1-pixel9-pixelSpectrum(our)DBA(suggested)Label-Flip (our)MpAf(suggested)
clientserverMABAMABAMABAMABAMAMA
EA-PSMultiKrum87.7643.288.4853.8487.924.2487.5718.7987.9588.01
Bulyan86.2549.4384.7666.4386.844.08287.677.0387.8187.47
Lead-FLMultiKrum88.3165.7588.3859.0188.574.6188.3225.2487.787.94
Bulyan88.2565.0987.8176.6288.534.3287.179.3587.4487.17
  • We also apply the suggested attack methods and our added methods on the CIFAR-10 dataset to further illustrate the performance. The results are as follows.
(%)CIFAR-10
Spectrum(our)DBA(suggested)Label-Flip (our)MpAf (suggested)
ClientServerMABAMABAMAMA
IIDEA-PSMultiKrum32.4173.0634.5176.7213.3914.15
Bulyan32.1849.7633.7280.2614.814.94
Lead-FLMultiKrum33.9576.7735.6877.0810.910.47
Bulyan33.4440.8534.9449.1414.4214.27
Non-IIDEA-PSMultiKrum33.7546.1234.1838.9415.915.67
Bulyan34.8140.1335.2430.0216.0815.94
Lead-FLMultiKrum33.9555.4235.1845.9513.4213.19
Bulyan32.141.3335.0141.6416.2415.35
  • It's important to note that none of the existing client-side defense methods focus on untargeted attacks. Through added experiments, we found that although our method struggles to defend against untargeted attacks, it still slightly outperforms the state - of - the - art client - side defense methods.

  • For new target attacks, our method outperforms the state - of - the - art client - side defense methods with server-side defense methods.

The added code will still open-source to the original link in the manuscript.

We hope this response has addressed your concerns effectively. We are grateful for your valuable input.

References

[1] Wang, Tong et al. “An Invisible Black-Box Backdoor Attack Through Frequency Domain.” European Conference on Computer Vision (2022). (Spectrum)

[2]Zhang, Mengmei et al. “Adversarial Label-Flipping Attack and Defense for Graph Neural Networks.” 2020 IEEE International Conference on Data Mining (ICDM) (2020): 791-800. (Label-Flip)

审稿意见
3

To combat persistent adaptive attacks, the authors propose EA-PS, a client-side defense that enhances server-side methods for robust, stable performance. By limiting attack impact and constraining local parameter perturbations, EA-PS mitigates backdoor poisoning. Theoretically, it achieves a lower upper bound, smaller certified radius, and larger convergence upper bound. Evaluations on FashionMNIST and CIFAR-10 show EA-PS reduces attack success rates by up to 14.9% and improves stability with up to 40% lower variance compared to other client-side defenses.

给作者的问题

The attack impact measures the differences between two rounds. I wonder why multiple rounds are not considered, as in a real-world FL system, malicious clients may not be selected every round. Additionally, I suggest measuring the tradeoff between security and training speed, as the current optimization goal may slow down the training process.

论据与证据

The paper presents strong empirical and theoretical evidence supporting the effectiveness of the proposed method. However, two key aspects are missing: (1) The cost of implementing client-side defense, such as communication overhead, should be compared to pure server-side defense. (2) The efficiency of the proposed method is not thoroughly evaluated—while convergence results provide some insight, empirical experiments are needed. Specifically, how much additional time does the client-side defense require compared to standard FedAvg? Additionally, it would be beneficial to show model accuracy and backdoor accuracy throughout the FL process to illustrate whether this defense slows down main task training, which is just as crucial as security in practice. Overall, most of my concerns are from empirical perspective, I appreciate the authors offer theoretical guarantees.

方法与评估标准

  1. I will suggest adding different type of backdoor attacks (e.g., distributed trigger, adaptive backdoor) as baselines, except for only using fixed pattern since the theoretically results suggest a general defense.
  2. Multikrum and Bulyan are designed to defend against model poisoning attacks. Defenses that include a post-training stage, such as CRFL, should also be considered as baselines.

理论论述

N/A

实验设计与分析

See "Methods And Evaluation Criteria"

补充材料

I haven't checked the proof details

与现有文献的关系

N/A

遗漏的重要参考文献

N/A

其他优缺点

Minor: Backdoor attacks are typically considered a specific type of targeted attack in previous FL security papers, which is slightly inconsistent with Section 2.1.

其他意见或建议

It is better to clarify the defender's knowledge and ability considering the difference between client-side defense and server-side defense. To make it more practical, how clients and/or sever exchange knowledge, information should be specified.

作者回复

Thank you for your thorough review and valuable feedback on our work.

  1. The communication overhead should be compared to pure server-side defense.

Response: We'd like to address the concern regarding the communication overhead in our work. Since nothing but the parameter constraint strategy is used in this work, the communication overhead is the same as pure server-side defense per round.

  1. I will suggest adding distributed trigger, adaptive backdoor, and CRFL as baselines.

Response: For distributed trigger and adaptive backdoor, we added DBA[2] and A3FL[1]. We also added CRFL to compare with Multi-Krum and Bulyan. The results are as follows.

(%)CIFAR-10
DBAA3FL
ClientServerMABAMABA
IIDEA-PSMultiKrum34.5176.7232.1557.96
Bulyan33.7280.2632.8759.37
CRFL27.4912.5327.2926.42
Lead-FLMultiKrum35.6877.0833.9658.49
Bulyan34.9449.1432.6552.8
CRFL26.8924.7327.1839.83
Non-IIDEA-PSMultiKrum34.1838.9432.6447.49
Bulyan35.2430.0234.7441.38
CRFL26.2910.5925.8213.67
Lead-FLMultiKrum35.1845.9534.5353.62
Bulyan35.0141.6433.5949.86
CRFL26.4111.8926.2517.84
  1. How much additional time does the client-side defense require compared to standard FedAvg?

Response: As convergence analysis proved, our work is slightly inefficient compared with other methods. We appreciate your suggestion to add time overheads (seconds average round), and the experimental results on CIFAR-10 with IID distribution and FEMNIST with natural non-IID distribution are as follows.

Time(s/r)CIFAR-10FEMNIST
FedAvgKrumBulyanCRFLFedAvgKrumBulyanCRFL
EA-PS27.9828.6628.2145.1380.7582.6378.89118.51
Lead-FL22.5523.1122.8340.6256.5158.7458.02109.36
FL-WBC20.9521.3421.2738.3244.1750.4449.3898.27
NULL19.7420.9420.6137.6843.1747.5546.5797.46
  1. Backdoor attacks is slightly inconsistent with Section 2.1.

Response: We appreciate the opportunity to change the "slightly inconsistent" expression that "one of the specific types of targeted attacks (known as backdoor attacks )" in section 2.1.

  1. It would be beneficial to show model accuracy and backdoor accuracy to illustrate whether this defense slows down main task training.

Response: Details of MA (Main-task Accuracy) are in the Appendix. We will extract the MA to the experiment section.

  1. The attack impact measures the differences between two rounds. I wonder why multiple rounds are not considered, as in a real-world FL system, malicious clients may not be selected every round.

Response: We appreciate the opportunity to clarify the "multiple rounds attack impact". Our work follows previous works setting as in Lead-FL and FL-WBC, where "In each adversarial round, malicious clients are randomly selected and participate in the training". We will highlight this in the experimental setting.

  1. It is better to clarify the defender's knowledge and ability considering the difference between client-side defense and server-side defense. To make it more practical, how clients and/or sever exchange knowledge, information should be specified.

Response: 1) Only aggregation information is exchanged between clients and the server. 2) To clarify "the defender's knowledge and ability, considering the difference between client-side defense and server-side defense", we will add the comparisons in a table as follows.

ComponentClient-Side Defense (Ours)Server-Side Defense
KnowledgeLocal model parameters and gradients;Local training data distributionGlobal aggregated model;Aggregated update statistics (e.g., gradient norms)
CapabilityCan apply local parameter masking/smoothing; Cannot modify server aggregation logicCan modify aggregation rules (e.g., clip gradients, weight averaging)
AssumptionsClients may be malicious ;Server is honestServer is fully trusted;Clients may be malicious

The added code will still open-source to the original link in the manuscript. We are grateful for the chance to discuss our work's improvement, and wish to thank you again for your valuable input.

Reference:

[1]Zhang, Hangfan et al. “A3FL: Adversarially Adaptive Backdoor Attacks to Federated Learning.” Neural Information Processing Systems (2023).

[2] DBA: Xie, et al. "Distributed Backdoor Attacks against Federated Learning." International Conference on Learning Representations (2020)

最终决定

This paper proposes EA-PS (Estimated Attack Effectiveness-based Poisoning Defense with Parameter Constraint Strategy), a client-side defense designed to constrain the perturbation range of local parameters to minimizing the impact of targeted attacks. The paper provide both theoretical and empirical evidence for the effectiveness of the proposed defense.

Feedback for the authors: please add the new experiments to the paper, as well as discussions regarding computational cost, (in)-effectiveness against untargeted attacks.

With the above changes, this paper can make a valuable contribution to the literature of provably efficient defenses against poisoning attack in federated learning.