PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
4
4
5
4
3.0
置信度
创新性2.5
质量2.5
清晰度2.8
重要性2.5
NeurIPS 2025

FedRW: Efficient Privacy-Preserving Data Reweighting for Enhancing Federated Learning of Language Models

OpenReviewPDF
提交: 2025-04-07更新: 2025-10-29

摘要

Data duplication within large-scale corpora often impedes large language models' (LLMs) performance and privacy. In privacy-concerned federated learning scenarios, conventional deduplication methods typically rely on trusted third parties to perform uniform deletion, risking loss of informative samples while introducing privacy vulnerabilities. To address these gaps, we propose Federated ReWeighting (FedRW), the first privacy-preserving framework, to the best of our knowledge, that performs soft deduplication via sample reweighting instead of deletion in federated LLM training, without assuming a trusted third party. At its core, FedRW proposes a secure, frequency-aware reweighting protocol through secure multi-party computation, coupled with a parallel orchestration strategy to ensure efficiency and scalability. During training, FedRW utilizes an adaptive reweighting mechanism with global sample frequencies to adjust individual loss contributions, effectively improving generalization and robustness. Empirical results demonstrate that FedRW outperforms the state-of-the-art method by achieving up to $28.78\times$ speedup in preprocessing and approximately $11.42$% improvement in perplexity, while offering enhanced security guarantees. FedRW thus establishes a new paradigm for managing duplication in federated LLM training.
关键词
LLM PrivacyData DeduplicationFederated LearningSecure Multi-Party Computation

评审与讨论

审稿意见
4

This paper introduces Federated ReWeighting (FedRW), a privacy-preserving framework designed to address the challenges of data duplication in the federated training of large language models (LLMs). Traditional deduplication methods in federated learning often rely on trusted third parties and involve deleting redundant data, which can lead to the loss of valuable information and introduce privacy risks. FedRW proposes a "soft deduplication" approach that reweights data samples based on their global frequency, rather than deleting them, and does so without requiring a trusted third party.

优缺点分析

Strengths

  • The paper focuses on the data duplication issue in federated learning, an important topic in the field.
  • The paper clearly illustrates the framework in the method part.

Weaknesses

  • The paper does not clearly show the difference between the proposed method and the baseline, EP-MPD.
  • The paper does not explain the FL experiment settings. What FL settings are used currently? Is the proposed method still effective under different FL settings, such as IID, non-IID, or personalized FL?
  • In the experiments, only GPT-2 Large and DistilGPT-2 are used for the method. It would be better to include some more recent LLMs in the experiments.

问题

See above.

局限性

There is no discussion on limitations in the paper.

最终评判理由

The rebuttal has addressed my concerns. I updated the score accordingly.

格式问题

None.

作者回复

We appreciate the reviewer for the valuable and constructive comments. We are encouraged by the positive feedback on the significance of topic and the clarity of the paper. We hope the following discussions can address the reviewer's remaining questions.

1. Clarity of Difference From Baseline and FL Experimental Settings

We appreciate this feedback on the clarity of the difference between FedRW and the baseline (EP-MPD), and the FL experimental settings. We want to clarify that we have detailed the key information in our manuscript:

  • Section 1 (Introduction): We introduced EP-MPD, the SOTA work for federated hard deduplication, and illustrated three key challenges unresolved, followed by the introduction of our proposed method, FedRW. We also presented an overview of EP-MPD and its potential limitations in Figure 1.
  • Section 2 (Related Work): In the subsection on Distributed Deduplication in existing work, we discussed the mechanism and limitations of EP-MPD.
  • Section 4 (Framework): This section offered a detailed explanation of FedRW. In Figure 2, we presented the architecture framework of FedRW and provided an intuitive comparison on protocol runtime, perplexity, and third-party reliance. We consider these presentations to be sufficient in illustrating the essential differences between the two methods.
  • Section 5 (Experiments): The experimental results on preprocessing and model performance clearly showed the difference between FedRW and EP-MPD.
  • Section 5.1 (Experimental Settings): This section declared the datasets, client number, data preprocessing and distribution strategies, evaluation metrics, and overall experimental setup. For completeness, Appendix C includes a detailed description of the datasets, the partitioning standard, the experimental environment, and the hyperparameter configurations.

We sincerely hope that the above clarifications effectively resolve your concerns.

2. Effectiveness Across FL Settings

We appreciate your concerns for more detailed effectiveness under various FL settings. Regarding the three specific FL settings:

  • Current FL settings: In this study, we followed the baseline setup for unsupervised fine-tuning. As described in Lines 203-204, "To simulate redundancy, duplicates are synthetically added into the training set at different rates and distributed uniformly across 10 clients." The data follows an IID distribution.

  • Non-IID: We appreciate the suggestion to explore the Non-IID setting, and we have conducted experiments on Qwen3-0.6B to evaluate FedRW's effectiveness.

    1. Setup 1: We categorized the Rotten Tomatoes dataset by label and set up 5 clients. To simulate Quantity Skew, we distribute the data according to the proportion of the total train set as [40%, 20%, 20%, 10%, 10%]. For Label Skew, we set the distributions for the binary (0/1) labels as: [(0.5, 0.5), (0.6, 0.4), (0.4, 0.6), (0.9, 0.1), (0.1, 0.9)].

    2. Result 1 (under 25% duplication)

      MethodIIDQuantity SkewLabel Skew
      Baseline1.712.022.44
      FedRW (Ours)1.591.961.66
    3. Setup 2: To simulate Feature Skew, we allocated datasets Poetry, Sonnets, and Plays to separate clients. These datasets differ distinctly in terms of text structure, sentence length, and lexical and syntactic complexity.

    4. Result 2

      MethodFeature Skew
      Baseline3.43
      FedRW (Ours)2.70

    The experimental results indicate that FedRW consistently maintains its performance advantages in the Non-IID settings.

  • Personalized FL: We appreciate the suggestion to explore personalized FL. While the current FedRW framework primarily focuses on enhancing a global LLM, we acknowledge that language models benefit significantly from personalization to adapt to individual user preferences. A deeper exploration may involve diverse model architectures or personalization strategies, beyond the scope of this foundational privacy-preserving soft deduplication framework. We consider adapting FedRW to integrate with this pattern a valuable and exciting avenue for future research.

3. Evaluations on Mainstream Models

We appreciate your valuable suggestion, and have conducted additional perplexity evaluations on Qwen3-0.6B, Qwen2.5-0.5B-Instruct, and Llama-3.2-1B-Instruct under 30% duplication.

  • Qwen3-0.6B

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.472.611.712.544.078.21
    FedRW (Ours)2.362.441.592.213.627.23
  • Qwen2.5-0.5B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.212.481.582.284.1111.77
    FedRW (Ours)2.122.361.552.033.849.92
  • Llama-3.2-1B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.142.341.652.394.1118.35
    FedRW (Ours)2.092.211.541.994.0016.03

As shown in the tables, FedRW continues to show its advantage in addressing the impact of duplication on model performance, particularly under challenging conditions such as data sparsity or complexity.

4. Discussion on Limitations

We appreciate your concern regarding the lack of a detailed discussion on limitations. While we have discussed FedRW’s boundaries and potential extensions in Section 6 (Lines 284–286), as well as certain limitations within the Experiments section, we commit to providing a clearer and more comprehensive discussion of these aspects in the revised manuscript.

We sincerely hope our replies have clearly addressed your concerns. We are deeply grateful for the valuable feedback and the broader perspectives you offered, which have helped us better position the significance of this work within the broader research landscape.

评论

Thank you for the detailed rebuttal and it has addressed my concerns. I will update the score.

审稿意见
4

In FL settings, there is a fundamental dilemma for deduplication: local fails to detect inter-client duplication. Existing method may remove some domain-specific information and introduce additional overheads. This paper propose a reweighting method to softly deduplicate samples without a trusted third party. This framework reweight based on the frequency for a more nuanced control based on the multi-party computation.

优缺点分析

Strength:

  1. FedRW introduces a pioneering soft deduplication method for FL, avoiding data deletion and preserving dataset integrity, which is a significant advancement over hard deduplication techniques.
  2. By eliminating the need for a trusted third party and using secure 2PC with PSI, FedRW ensures robust privacy guarantees, addressing critical vulnerabilities in decentralized settings.

Weakness: See questions

问题

  1. Due to the different data collection fields or geographical locations of different clients, cooperation is meaningful. In this case, it is difficult for two clients to have exactly the same samples. Therefore, when the two repeated samples are not exactly the same, is this method still effective?
  2. I also have doubts about the significance of data deduplication in real FL scenarios. Is there such a problem in real datasets? Have different clients collected a large number of similar datasets?
  3. The experimental settings mention hyperparameters (e.g., learning rates, epochs), but the paper does not discuss sensitivity analyses or how hyperparameter choices impact performance, which could affect generalizability.
  4. This paper also did not conduct experiments with the current mainstream models. Therefore, whether partial repetition will still be harmful when using mainstream models is also an unknown question

局限性

See above

最终评判理由

I am still concerned about the importance of this issue if the model size gets larger, so I will keep my score.

格式问题

No

作者回复

We appreciate the reviewer for the valuable and constructive comments. We are encouraged by the positive feedback on novelty, privacy and security guarantees, significance of topic, and the overall clarity of the paper. We hope the following discussions can address the reviewer's remaining questions.

1. Non-Exact Duplicates

We greatly appreciate this insightful question. Indeed, there are scenarios where non-exact duplicates are more commonly encountered. We have conducted perplexity evaluations on the dataset, Twitter Sentiment Analysis [1], which contains non-exact repeated patterns due to the nature of social media.

  • Result

    MethodQwen3-0.6BLlama-3.2-1B-Instruct
    Baseline3.303.74
    FedRW (Ours)2.982.49

These results demonstrate the effectiveness of our method in addressing non-exact repetitive patterns.

2. Significance of Deduplication in Real FL Scenarios

We truly appreciate this critical concern. Exact and near-exact duplicates are indeed a common issue in large-scale real-world datasets for LLMs. The presence of duplicates among private client data stems from several prevalent mechanisms:

  • Influence of Shared Public Information: Private data often contains content derived from or heavily influenced by widely shared public information. Similar to open-source datasets like Twitter Sentiment Analysis, which inherently exhibits extensive repetition due to the nature of social media, private data collected from users or organizations also tends to contain duplicated or near-duplicate patterns. Specific common patterns may appear across the private data of different clients.

  • Inherent Data Collection: Beyond external influences, duplicates can occur within sensitive personal information. For instance, GPS locations in personal travel information can exhibit duplicates due to recurring travel patterns or shared high-traffic areas [2], which can be collected to analyze user behavior patterns.

  • Data Augmentation: The data augmentation strategies [3] used in private datasets, including those based on generative models or synthetic data, may also produce a large number of similar or identical samples with duplicated patterns.

These mechanisms indicate that data duplication is a pervasive challenge in real-world, distributed private datasets in FL, necessitating solutions like FedRW.

References

[1] passionate-nlp, "Twitter Sentiment Analysis Dataset." Kaggle.

[2] Wu, Zhaomin, et al. "Federated transformer: Multi-party vertical federated learning on practical fuzzily linked data." Advances in Neural Information Processing Systems 37 (2024): 45791-45818.

[3] Hou, Charlie, et al. "Pre-text: Training language models on private federated data in the age of llms." arXiv preprint arXiv:2406.02958 (2024).

3. Evaluations on Mainstream Models

We appreciate your suggestion, and have conducted supplementary perplexity evaluations on Qwen3-0.6B, Qwen2.5-0.5B-Instruct, and Llama-3.2-1B-Instruct under 30% duplication.

  • Qwen3-0.6B

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.472.611.712.544.078.21
    FedRW (Ours)2.362.441.592.213.627.23
  • Qwen2.5-0.5B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.212.481.582.284.1111.77
    FedRW (Ours)2.122.361.552.033.849.92
  • Llama-3.2-1B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlays
    Baseline2.142.341.652.394.1118.35
    FedRW (Ours)2.092.211.541.994.0016.03

As shown in the tables, FedRW consistently maintains its advantage in mitigating the impact of redundancy on model performance, particularly under challenging conditions such as data sparsity or complexity.

We agree that the impact of partial repetition may vary with model advancements, while data quality remains a fundamental factor in model performance. Deduplication is still an essential preprocessing step to ensure high-quality data.

4. Hyperparameter

We appreciate this valuable feedback. Regarding epochs, we typically utilize a small number (Appendix C, Line 516) as standard practice to prevent overfitting. Hence, our primary sensitivity analysis focused on the learning rate to assess FedRW's robustness.

  • Setup: We evaluated perplexity on DistilGPT2 and Qwen2.5-0.5B-Instruct under learning rates of 1e-3, 5e-4, 3e-4, 1e-4, 5e-5, and 3e-5. We selected the Plays dataset, which exhibited a significant performance gap, to investigate FedRW's generalizability.

  • DistilGPT2

    Method1e-35e-43e-41e-45e-53e-5
    Baseline16.1216.3816.1314.2115.0714.18
    FedRW (Ours)14.4214.7914.7413.1714.5012.76
  • Qwen2.5-0.5B-Instruct

    Method1e-35e-43e-41e-45e-53e-5
    Baseline12.8611.0710.6311.5011.7710.67
    FedRW (Ours)11.489.358.158.149.929.81

The experimental results indicate that FedRW consistently maintains its performance advantages and exhibits stable training behavior across this range of learning rates.

We are deeply grateful for your thoughtful consideration, which has enriched our understanding and improved the quality of this work.

评论

Thanks to the author for their efforts during the rebuttal period, but I am still concerned about the importance of this issue if the model size gets larger, so I will keep my score.

评论

Dear Reviewer d6ex,

Thank you again for raising the vital concern. As model scales, memorization of specific patterns due to duplication becomes more pronounced and critical, leading to overfitting, degraded generalization, and increased privacy risks [1]. Our method is specifically designed to mitigate these issues in the FL setting.

To validate the issue in larger models, we have conducted perplexity evaluations on Qwen2.5-3B-Instruct and Qwen2.5-7B-Instruct.

  • Qwen2.5-3B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlaysTwitter
    Baseline1.692.092.202.334.149.173.35
    FedRW (Ours)1.551.942.011.853.297.532.46
  • Qwen2.5-7B-Instruct

    MethodHaikuShort JokesRotten TomatoesPoetrySonnetsPlaysTwitter
    Baseline1.682.071.742.094.528.822.24
    FedRW (Ours)1.491.951.611.813.436.541.35

While larger models may exhibit lower perplexity on certain datasets, the results show that FedRW continues to outperform the baseline. Under challenging conditions such as data sparsity and complexity in Poetry, Sonnets, and Plays, or extensive near-duplicate patterns in Twitter, FedRW achieves an average relative reduction of approximately 25.77% in perplexity compared to the baseline.

We hope these new results and our explanation clarify the significance of our method in larger models, and we are happy to address any remaining questions during the discussion period.

References

[1] Carlini, Nicholas, et al. "Quantifying memorization across neural language models." ICLR'23.

评论

Thank you for the efforts you have made to solve my problem. I will revise my score.

审稿意见
5

FedRW has proposed an innovative privacy preserving federated learning framework that replaces trusted third parties with a frequency aware sample reweighting mechanism, securely calculates global sample frequencies using the PPMPR protocol, and combines a layered parallel strategy to reduce preprocessing complexity, achieving a preprocessing acceleration of up to 28.78 times while improving model generalization ability.

优缺点分析

Strengths:

  1. Balancing privacy and efficiency
  2. By using logarithmic functions to dynamically adjust sample weights and preserve low-frequency sample information, an average confusion reduction of 11.42% was achieved on datasets such as Haiku and Sonnets, which is particularly better than the performance of hard deletion methods in data scarcity scenarios.
  3. MPC protocol based on semi honest model provides provable security Weaknesses: 1.The innovation of FedRW is mainly reflected in the first soft deduplication implementation and efficiency optimization in federated scenarios (such as pre-processing acceleration), but its protocol, weight mechanism, and parallel strategy heavily rely on existing technologies (PSI, centralized soft deduplication). 2.This framework provides an efficient and scalable solution for privacy sensitive federated training, but further optimization is needed for semantic redundancy processing and dynamic client adaptation capabilities.

问题

See Weakness.

局限性

Yes, the proposed method, while promising, reveals certain limitations in the Experiments section.

最终评判理由

After the rebuttal and discussion process, I am happy to change my decision to accept this paper.

格式问题

This paper has no major formatting issues.

作者回复

We appreciate the reviewer for the valuable and constructive comments. We are encouraged by the positive feedback on novelty, efficiency and effectiveness, privacy and security guarantees, significance of topic, and the overall clarity of the paper. We hope the following discussions can address the reviewer's remaining questions.

1. Rely on Techniques

We sincerely appreciate your recognition of the innovation and originality of FedRW. We agree that FedRW draws inspiration from the centralized soft deduplication. However, the same problem remains unresolved in the FL setting, where simply applying the existing techniques introduces new privacy challenges. Compared with multiple rounds of key agreement and encryption in SOTA work, we leverage the PSI primitive as a foundational component to implement a lightweight variant, tailored for the defined 2PC functionality, while achieving the stronger, third-party-free privacy assumption. We also present an innovative parallel orchestration strategy to overcome the inherent computational overhead of MPC protocols in FL. The proposed innovations achieve up to 28.78× speedup in preprocessing.

Thus, we respectfully argue that the core novelty of our work lies in the innovative integration, adaptation, and optimization of these techniques to address a previously unexplored and challenging problem: enabling efficient and privacy-preserving soft deduplication in FL without relying on a trusted third party.

2. Future Optimization

We appreciate the reviewer's insightful comments for optimization, and we fully agree that these are vital directions for future research.

Regarding the semantic matching, this is indeed a significant issue in large-scale real-world datasets for LLMs. We may leverage the representation learning capability of transformer-based architectures to extract semantic duplication.

Regarding the dynamic client adaptability, we acknowledge the importance of dynamic scenarios in which clients may join, leave, or exhibit varying computational capabilities. A preliminary idea is to employ Updatable PSI [1] to address the changes in sample weights caused by client arrivals and departures, and this may trade off between functionality and computational overhead. We plan to explore these valuable aspects in our future research.

References

[1] Agarwal, Archita, et al. "Updatable private set intersection from structured encryption." Cryptology ePrint Archive (2024).

We sincerely appreciate your thoughtful and well-justified suggestion, as well as the valuable inspiration you have provided regarding future optimization of our work.

评论

Thanks for the discussion and it addressed some of my previous concerns.

审稿意见
4

This paper proposes FedRW, a novel framework for privacy-preserving soft deduplication in federated LLM training. The PPMPR protocol avoids the need for a trusted third party, scales efficiently through a parallel orchestration strategy, and integrates sample-level loss weighting in training.

优缺点分析

Strengths

The technical contributions are clearly defined, and the paper is well-written with a logical structure.

Weaknesses

  1. It would be helpful if the authors could include training loss curves or convergence comparisons to illustrate whether the proposed reweighting mechanism slows down or accelerates model convergence relative to deletion-based approaches.

  2. There are a few writing-related issues that could improve the clarity of the manuscript. First, using vague terms such as “nuanced” is discouraged in scientific writing, as it does not convey precise meaning. Second, Figures 1 and 2 are visually engaging but lack adequate explanation of key elements. Third, the system formulation in Section 3 is not fully adapted to the federated learning setting, it would benefit from a clearer illustration.

  3. The evaluation is limited to synthetic duplication. It would strengthen the paper to include results or discussion on how FedRW performs in real-world scenarios with naturally occurring duplication, such as instruction-tuning data, which often contains repeated patterns.

  4. The protocol assumes a semi-honest threat model. However, in practice, malicious clients may attempt to manipulate the protocol by misreporting sample frequencies or injecting misleading data into the intersection process. It would be valuable to discuss whether FedRW could be extended or adapted to remain robust under such adversarial behavior.

问题

Check Weaknesses

局限性

Yes

最终评判理由

Thanks for the clarification. I will remain positive and stick with my original score.

格式问题

None.

作者回复

We appreciate the reviewer for the valuable and constructive comments. We are encouraged by the positive feedback on novelty, efficiency and effectiveness, privacy and security guarantees, significance of topic, and the overall clarity and logical structure of the paper. We hope the following discussions can address the reviewer's remaining questions.

1. Convergence Comparisons.

We appreciate your suggestion and have included the average training loss of FedRW compared with the baseline under 30% duplication in the tables.

  • GPT-2 Large

    DatasetMethodRound 0Round 1Round 2Round 3
    HaikuBaseline1.8420.8390.8060.811
    FedRW (ours)4.6710.8350.7250.705
    Rotten TomatoesBaseline1.1240.4440.4260.379
    FedRW (ours)3.0380.4250.3780.375
    Short JokesBaseline0.9940.7200.6220.641
    FedRW (ours)1.2630.6790.5950.547
    SonnetsBaseline2.8351.4661.3661.730
    FedRW (ours)4.3291.3491.2431.168
  • DistilGPT2

    DatasetMethodRound 0Round 1Round 2Round 3
    HaikuBaseline3.2251.2831.2671.274
    FedRW (ours)2.9501.2741.2551.252
    Short JokesBaseline1.5110.751--
    FedRW (ours)1.3880.745--
    Rotten TomatoesBaseline1.7710.607--
    FedRW (ours)1.3380.437--
    IMDBBaseline4.2341.964--
    FedRW (ours)3.8381.939--
    PoetryBaseline10.8582.9891.3131.107
    FedRW (ours)9.9081.9991.1221.049
    SonnetsBaseline17.5893.4432.0191.870
    FedRW (ours)15.9013.0481.9711.802
    PlaysBaseline23.8483.3792.8852.815
    FedRW (ours)20.4603.2342.8542.767

The datasets, including Short Jokes, Rotten Tomatoes, and IMDB, were trained for fewer rounds to mitigate overfitting. Notably, on GPT-2 Large, FedRW exhibited a higher initial loss compared with the baseline, whereas the DistilGPT-2 model showed a lower one. This consistency is related to the loss reweighting mechanism and the variance of model capacity. In summary, FedRW demonstrates faster model convergence and better final performance.

2. Writing-related Issues.

We sincerely appreciate your thorough reading, and we will revise the issues according to your comments in the final version.

To further clarify Figures 1 and 2, we offer the following supplementary explanation:

  • Fig. 1: (a) illustrates the inherent challenges of global deduplication in a decentralized setting, where privacy constraints prevent direct data sharing.
    (b) describes the paradigm adopted by the state-of-the-art solution (also used as our baseline), which relies on a trusted third party for deduplication over encrypted data, followed by decryption. The right part of Fig.1 indicates that this pattern may lead to the loss of key training samples.
  • Fig. 2: The left panel of Fig.2 illustrates the parallel framework of FedRW, where each small square represents a single 2PC protocol interaction. The middle section depicts our reweighting scheme, which preserves the integrity of the dataset. The colors reflect the different weights assigned to samples. The right panel provides a comparison between FedRW and the baseline approach.

As you pointed out, Section 3 and Appendix A.3 illustrate the security and threat model. Due to the space constraints, we were unable to illustrate adaptation to FL in detail. We will include a brief description in the revision as follows:

  • FedRW operates under the semi-honest adversary model in federated learning (FL), meaning all participants will follow the protocol honestly but may attempt to infer additional information from observation. For the scope of this work, we assume that the parties do not actively collude with one another. While we acknowledge that other forms of more active or malicious attacks exist in FL – such as inference attacks, backdoor attacks, or reconstruction attacks – these threats are considered orthogonal to the primary objective of this study, which focuses on designing the first privacy-preserving soft deduplication via sample reweighting. These additional and vital security considerations will be explored in our future research to enhance FedRW's robustness.

Following your valuable suggestion, we will revise the manuscript with the above information.

3. Evaluation in Real-World Scenarios.

We appreciate your valuable suggestion, and have conducted experiments on real-world datasets with repeated patterns on two mainstream models, Qwen3-0.6B and Llama-3.2-1B-Instruct.

  • databricks-dolly-15k [1] is an open-source dataset of instruction-following records generated by thousands of Databricks employees.
  • Twitter Sentiment Analysis [2] is an open-source dataset of tweets labeled with sentiment categories, which inherently contains repeated patterns due to the nature of social media.
  • Results (Perplexity)

    • Qwen3-0.6B

      Methoddatabricks-dolly-15kTwitter Sentiment Analysis
      Baseline5.113.30
      FedRW (Ours)4.652.98
    • Llama-3.2-1B-Instruct

      Methoddatabricks-dolly-15kTwitter Sentiment Analysis
      Baseline4.793.74
      FedRW (Ours)4.512.49

As presented in the table, FedRW consistently outperforms the baseline under the same experimental settings, showing the effectiveness and generalization ability of our method.

References

[1] databricks, "databricks-dolly-15k." Hugging Face.

[2] passionate-nlp, "Twitter Sentiment Analysis Dataset." Kaggle.

4. Extension to Malicious Model.

Thank you for raising the thoughtful concerns about malicious clients. Extending FedRW to resist malicious adversaries would be an interesting and important research direction.

As mentioned in 2, we acknowledge that the FedRW primarily operates under a semi-honest threat model, which is sufficient for the vast majority of real-world scenarios, given its high efficiency and practicality.

Regarding the malicious model, it is feasible to design specific verification based on FedRW’s existing architecture, such as Zero-Knowledge Proofs (ZKPs), to verify the consistency of parties’ behaviors during preprocessing and training. However, such approaches would come at a significant computational and communication cost.

These potential schemes trade off between model accuracy, data privacy, and computational overhead. Given the scope and depth of these problems, we have deferred this research to our future follow-up work.

We are deeply grateful for the thought-provoking discussions and valuable insights that have enriched our understanding and improved the quality of this work.

评论

Dear Reviewer L3R8,

Please leave some feedback on the authors' rebuttal.

With kind regards,

Your AC

评论

Thank you for the thorough response, it resolves my concerns.

最终决定

There is a fundamental dilemma for deduplication in FL: local devices fail to detect inter-client duplication. Existing method may remove some domain-specific information and introduce additional overheads. This paper propose a reweighting method to softly deduplicate samples without a trusted third party.

Strenghts: FedRW introduces a new soft deduplication method for FL. By eliminating the need for a trusted third party and using secure 2PC with PSI, FedRW ensures robust privacy guarantees.

Weaknesses: Whether deduplication is a real problem in FL is not fully addressed.

The main reason for my decision is that the reviewers seem to agree on the merits of this paper: proposing a novel solution to deduplication in FL that does not require a trusted 3rd party.