An Effective and Secure Federated Multi-View Clustering Method with Information-Theoretic Perspective
In this paper, we propose an effective and secure federated multi-view clustering method, which is designed to alleviate the trade-off between privacy preservation and performance improvement in the field of federated multi-view clustering.
摘要
评审与讨论
Focusing on federated multi-view learning, this paper presents a novel method to alleviate the dilemma between privacy protection and multi-view clustering performance improvement. The authors conduct both theoretical analysis and empirical evaluations, demonstrating superior performance over baseline methods while offering enhanced privacy protection.
给作者的问题
- In order to analyze their impact to the final performance, could the author provide the hyperparameter experiment about and of Equation 12? Even if the author do not tune hyperparameter in their experiments.
- How does the proposed method address potential threats from malicious clients, as it currently only demonstrates effectiveness against reconstruction attacks on the server side?
- Could the authors integrate the discussion on extending the method to cross-device scenarios (currently in the appendix) into the main text and provide more in-depth analysis?
- Section 4, titled "Discussion and Analysis," seems to lack discussion on key findings—could the authors elaborate further to strengthen this section?
论据与证据
Yes, the authors provide the detailed description of the method and multiple experiments to support the points.
方法与评估标准
Yes, the proposed methods and evaluation criteria make sense for the problem.
理论论述
Yes, I have verified the privacy analysis in this paper.
实验设计与分析
All experiments and their analysis have been checked.
补充材料
The supplementary material has been reviewed.
与现有文献的关系
This methodology and theoretical results of this paper are mainly about federated learning and multi-view clustering.
遗漏的重要参考文献
There are no Not Discussed Essential References have been found.
其他优缺点
The paper is well-structured, with a clear motivation and thorough theoretical and experimental support. The proposed method makes a meaningful contribution to the advancement of the multi-view learning community.
However, several aspects require further clarification and improvement:
- The proposed method only demonstrates its effectiveness against reconstruction attacks on the server side, without addressing potential threats from malicious clients.
- The authors extends the method to cross-device scenarios in appendix, which is valuable. It would be beneficial to integrate this discussion into the main text and provide further analysis.
- Section 4 is titled "Discussion and Analysis," but lacks discussion on key findings.
其他意见或建议
See weakness
Thank you for your valuable comments and suggestions.
Q1: In order to analyze their impact to the final performance, could the author provide the hyperparameter experiment about and of Equation 12? Even if the author do not tune hyperparameter in their experiments.
A1: Thank you for your suggestion. We redefine Equation 12 as and analyze the hyperparameter . The table below shows the sensitivity of this parameter by varying the range on the BDGP dataset in incomplete scenarios. The results show that the model achieves the best performance at , reflecting the fact that the two parts of Equation 12, and , are similarly weighted and are both important for model optimization, so the hyperparameter is not tuned in our experiments.
| ACC | 87.48 | 91.88 | 94.92 | 93.76 | 89.28 |
| NMI | 78.08 | 80.67 | 84.93 | 82.94 | 72.27 |
| ARI | 75.33 | 81.20 | 87.83 | 85.05 | 75.45 |
Q2: How does the proposed method address potential threats from malicious clients, as it currently only demonstrates effectiveness against reconstruction attacks on the server side?
A2: Thank you for your careful reminder. Our proposed method primarily addresses model attack threats caused by semi-honest participants. Although we did not explicitly delve into defenses against data attacks initiated by dishonest participants such as malicious clients, our method remains effective in such cases. Specifically, if there exists malicious clients with false or manipulated data, their local clustering structures are likely to significantly deviate from that of other clients. By comparing the local cluster assignments or centroids uploaded by each client, the server can set a malicious threshold to flag outliers. If a client's data exceed this threshold, its shared information can be disregarded, and the client can be marked as malicious, effectively mitigating such attacks.
Q3: Could the authors integrate the discussion on extending the method to cross-device scenarios (currently in the appendix) into the main text and provide more in-depth analysis?
A3: Thank you for your suggestion. We will include experimental results and a discussion on extending to cross-device scenarios in Section 5.3. We believe that enhancing the scalability of the method to adapt to various scenarios is an essential area of exploration. In the new version, we will present additional experimental results on more multi-view datasets and provide further analysis of these results. Furthermore, we have envisioned a solution for the scenario where the sample size per client becomes insufficient for effective training as the number of clients increases. To maintain good performance, we will consider strategies such as continuing training based on existing models or sharing partial local model information to alleviate the issue of insufficient samples per client, enabling collaborative training across multiple clients.
Q4: Section 4, titled "Discussion and Analysis," seems to lack discussion on key findings—could the authors elaborate further to strengthen this section?
A4: Thank you for your feedback. In the new version, we plan to add two parts discussing model generalization and privacy protection. Regarding model generalization, we will focus on the scalability of the proposed method, including extensions to incomplete and cross-device scenarios. These scenarios primarily aim to limit shared information, only relying on the model’s generalization ability. For example, for incomplete scenarios, each client uploads clustering-related features for overlapping samples to the server, while non-overlapping samples are clustered locally. For cross-device scenarios, each client uploads clustering-related features extracted from local data, with the server aligning overlapping samples. Our method leverages generalization to perform well with a few extensions. Regarding privacy protection, we will discuss the privacy-preserving scenarios addressed by our method and potential strategies for further enhancing privacy. Our method is primarily designed for environments where all participating parties are semi-honest, meaning they faithfully execute the training protocol but may attempt privacy attacks. Currently, our feature splitting strategy is sufficient to defend against common model inversion attacks in such settings. For further privacy enhancement, we could integrate commonly used privacy-preserving techniques in federated learning, such as differential privacy or homomorphic encryption, to offer additional privacy protection.
The paper proposes ESFMC which aims to address the privacy concerns and performance trade-offs in federated learning for multi-view clustering. The main idea is to allow all clients to do collaborative clustering without leaking sensitive data, and they follow a privacy-preserving strategy based on information theory by only sharing the clustering-related features, instead of the raw or sample-related features to minimize the risk of privacy leakage. Additionally, ESFMC is extended to handle incomplete multi-view clustering by introducing a collaborative alignment strategy. The paper also conducts extensive experiments to show that ESFMC outperforms existing state-of-the-art methods in terms of clustering accuracy and privacy preservation
给作者的问题
- What is the advantage of the collaborative alignment strategy over the cross-view alignment strategy based on adaptively calculate alignment matrices in FCUIF (Ren et al., 2024)?
- To optimize \omega^m_{t,k}, why traditional SGD is not suitable?
- For privacy preservation, if the work aims to minimize the information to be shared, a straightforward way is to minimize I(X^m;Z_c^m), but it seems that the method uses I(Z_x^m, Z_c^m) instead.
论据与证据
Yes, the paper provides several experiment results and theoretical analysis to support its claim.
方法与评估标准
The methods and evaluation criteria(ACC,NMI, and ARI) are well-suited to the problem of federated multi-view clustering.
理论论述
Yes, I have checked the proofs in the appendix, including the proof of Lemma 3.1, the generalization analysis, and the privacy analysis.
实验设计与分析
Yes, the authors conducted several experiments on the popularly used datasets, and the results demonstrate the effectiveness.
补充材料
Yes
与现有文献的关系
The work provided a privacy solution for federated multi-view clustering based on information theory, and extended the work to incomplete scenarios.
遗漏的重要参考文献
None
其他优缺点
- The paper's main contribution is further solving the privacy-preservation in federated multi-view clustering. While there are existing methods for federated learning and multi-view clustering, the information-theoretic feature splitting used in ESFMC is a novel contribution. By only sharing the clustering-related features, the paper successfully balances the privacy and performance in federated learning.
- The collaborative alignment strategy to deal with incomplete data in a federated setting is another significant innovation. This strategy extends its application.
- The paper is well-organized. The ablation studies and theoretical analysis are detailed, helping to validate the contributions effectively. The extensive experiments conducted on various datasets and the ablation study provide support for the method's effectiveness. Weaknesses:
- While the paper focuses on information-theoretic privacy preservation, it does not provide a detailed comparison with other commonly used privacy-preserving techniques in federated learning.
- The authors do not provide a clear definition or description of the privacy under the federated multi-view clustering scenario
其他意见或建议
Refer to the above comments.
伦理审查问题
None
We thank the reviewer for valuable comments and suggestions that have greatly improved our paper.
Q1: While the paper focuses on information-theoretic privacy preservation, it does not provide a detailed comparison with other commonly used privacy-preserving techniques in federated learning.
A1: Thank you for your valuable feedback. Our method can be integrated with commonly used privacy-preserving techniques in federated learning to further enhance privacy protection. Specifically, for clustering-related features shared across clients, our feature splitting strategy already prevents common attacks, such as model inversion attacks, preventing attackers from reconstructing the original data. However, for stronger privacy guarantees, additional privacy-preserving techniques can be added. For instance, as referenced in our response to Reviewer qNsU Q3, we report the impact of differential privacy on our method’s performance under different privacy budgets. Additionally, our method aims to balance privacy protection and clustering performance. Excessive privacy constraints inevitably lead to performance degradation, as confirmed by our experimental results. Similarly, homomorphic encryption could be integrated with our method to further enhance privacy protection, though at the cost of increased computational overhead in both training and inference.
Q2: The authors do not provide a clear definition or description of the privacy under the federated multi-view clustering scenario.
A2: Thank you for your feedback. We describe the privacy protection scenario of ESFMC in Lines 416–418: we assess whether all semi-honest participants can reconstruct the original data through certain attack methods based on shared information. However, this may not be explicitly stated. To clarify, we will restate ESFMC’s privacy protection scenario in the problem statement: we assume that all participating parties are semi-honest and do not collude. An attacker follows the training protocol but may attempt privacy attacks to infer private data from other parties.
Q3: What is the advantage of the collaborative alignment strategy over the cross-view alignment strategy based on adaptively calculate alignment matrices in FCUIF (Ren et al., 2024)?
A3: First, in terms of performance, our proposed method demonstrates superior effectiveness in handling incomplete scenarios compared to FCUIF on the same datasets (BDGP and Scene), achieving better results. Second, from a methodological perspective, FCUIF leverages sample commonality and view versatility, enabling the server to adaptively compute alignment matrices for cross-view alignment. In contrast, our collaborative alignment strategy with an information-theoretic perspective, aligns features by maximizing mutual information. Compared to FCUIF, our method serves as a generalizable interface module that can be integrated into other methods. It offers greater adaptability and scalability while also achieving improved performance.
Q4: To optimize , why traditional SGD is not suitable?
A4: Traditional SGD seeks a single optimal point estimate of the parameters by minimizing the loss function. In contrast, our method focuses on the posterior distribution of the parameters rather than a single estimate. To achieve this, we employ SGLD, which integrates SGD with Langevin dynamics by introducing noise into the gradient updates. This added noise facilitates sampling from the posterior distribution rather than converging to a single mode.
Q5: For privacy preservation, if the work aims to minimize the information to be shared, a straightforward way is to minimize , but it seems that the method uses instead.
A5: This is an interesting perspective. Minimizing is indeed a more direct approach to reducing the amount of sensitive information from in the shared information . However, we choose to minimize to emphasize the feature splitting strategy better. We aim to ensure that the extracted features, and , serve distinct purposes with minimal redundancy, thereby achieving high-quality feature splitting. We believe that both optimization strategies serve a similar goal, differing primarily in their formulation and optimization approach.
The paper proposes a novel federated multi-view clustering (FedMVC) method, Effective and Secure Federated Multi-View Clustering (ESFMC), which aims to address the privacy-performance trade-off in federated learning settings. The key contribution of this work is an information-theoretic feature-splitting mechanism, where clients retain sample-sensitive features locally and share only clustering-related features with the central server. This design effectively mitigates privacy risks while ensuring high-quality clustering results. To extend its applicability, ESFMC introduces a collaborative alignment strategy that ensures consistency across non-overlapping samples in incomplete multi-view scenarios, where certain data samples are missing across different clients. The paper provides theoretical guarantees on privacy protection and generalization performance, as well as extensive empirical evaluations on six real-world multi-view datasets. Experimental results demonstrate that ESFMC outperforms state-of-the-art centralized and federated multi-view clustering methods in both clustering accuracy and privacy preservation.
update after rebuttal
The authors have addressed my concerns and I would like to keep my rating.
给作者的问题
- Have you considered evaluating ESFMC on real-world federated datasets, such as medical imaging data, financial records, or IoT sensor networks, to better assess its applicability in practical scenarios?
- Can you provide additional quantitative results on privacy guarantees, such as empirical differential privacy noise impact analysis?
论据与证据
The main claims made in the paper include:
Claim 1: ESFMC mitigates the privacy-performance trade-off by using feature splitting to retain privacy-sensitive information locally while sharing only clustering-relevant features. Evidence: The paper provides theoretical privacy guarantees, including differential privacy analysis and empirical validation against model inversion attacks. The results show that ESFMC successfully prevents sensitive information leakage while maintaining superior clustering performance compared to traditional FedMVC approaches.
Claim 2: ESFMC is extendable to incomplete multi-view settings using collaborative alignment to ensure feature consistency across clients. Evidence: The collaborative alignment strategy is evaluated on incomplete datasets, where certain clients have missing views. The results demonstrate that ESFMC maintains high clustering performance even when data is incomplete. Additionally, ablation studies confirm that the collaborative alignment mechanism plays a crucial role in improving global clustering quality and robustness.
方法与评估标准
The proposed method is well-motivated, particularly for federated multi-view clustering in privacy-sensitive scenarios where direct data sharing is infeasible. The evaluation covers six real-world datasets with varying sample/view completeness, ensuring robustness across different conditions. Comparisons against five centralized and four federated clustering methods provide a strong performance benchmark. Standard metrics, including Accuracy, Normalized Mutual Information, and Adjusted Rand Index, ensure fairness in evaluation.
理论论述
I reviewed the Methodology section, which included generalization analysis, privacy analysis, and complexity analysis. The theoretical derivations appear correct, demonstrating ESFMC’s ability to scale effectively across multiple clients through upper bounds on clustering error. The privacy analysis provides differential privacy guarantees, theoretically proving that feature splitting minimizes privacy risks by ensuring that only clustering-relevant information is shared. The complexity analysis confirms that ESFMC remains computationally efficient and scalable for large-scale federated deployments. There are no major issues.
实验设计与分析
Yes, I checked the experimental settings, results, and analysis. ESFMC consistently outperforms existing methods in clustering performance while preserving privacy, as demonstrated by extensive evaluations across multiple datasets.
补充材料
Yes, I reviewed the supplementary material.
与现有文献的关系
This work contributes to federated multi-view clustering by tackling the fundamental challenge of balancing privacy protection and clustering performance. It builds on prior research in federated learning and privacy-preserving techniques by integrating differential privacy and information-theoretic principles.
遗漏的重要参考文献
None
其他优缺点
Strengths
- The paper introduces a feature-splitting strategy that distinguishes clustering-related features from sample-related features, allowing privacy-sensitive data to remain local while sharing only task-relevant information. This significantly enhances privacy protection without compromising clustering performance. The proposed collaborative alignment strategy effectively addresses challenges arising from incomplete multi-view scenarios, where clients have non-overlapping samples.
- The method is computationally efficient and scalable, making it highly suitable for large-scale federated learning applications. Unlike methods that require extensive client-server communication or computationally expensive privacy-preserving mechanisms, ESFMC optimizes information sharing without introducing significant computational overhead. The feature-splitting mechanism reduces the amount of information transferred between clients and the server, enhancing communication efficiency.
- The paper provides strong experiments, demonstrating that ESFMC outperforms state-of-the-art methods in both clustering performance and privacy preservation.
Weaknesses
-
The experimental validation primarily relies on synthetic and public benchmark datasets, which may not fully capture the complexities of real-world federated learning environments. In practice, data distributions in federated settings tend to be non-IID, with significant noise, missing information, and domain-specific constraints. While the inclusion of six multi-view datasets provides a strong foundation for evaluation, additional experiments on real-world federated datasets (e.g., healthcare, financial transactions) would strengthen the practical impact of ESFMC and demonstrate its applicability beyond academic benchmarks.
-
While the paper provides theoretical guarantees for privacy preservation using differential privacy analysis, the empirical validation of privacy guarantees is somewhat limited. The experiments focus primarily on clustering performance, with only basic privacy attack simulations. More extensive empirical experiments—such as testing ESFMC against more sophisticated adversarial attacks (e.g., gradient inversion, membership inference attacks) would provide stronger evidence that the method effectively protects sensitive information in real-world deployment scenarios.
其他意见或建议
NA
We sincerely appreciate your constructive comments and suggestions.
Q1: Have you considered evaluating ESFMC on real-world federated datasets,to better assess its applicability in practical scenarios?
A1: Thank you for your suggestion. We have considered evaluating ESFMC on real-world datasets and have obtained some preliminary results. The Organ{A,C,S}MNIST [1] dataset is derived from the liver tumor segmentation benchmark (LiTS), which consists of 3D computed tomography (CT) scans. We generate 2D images by extracting center slices from the 3D bounding box in the axial, coronal, and sagittal planes, corresponding to three visual views, resulting in a total of 13,000 samples. Below are our experimental results:
| Method | IMVC-CBG (2022) | DSIMVC(2022) | AGDIMC(2024) | FedDMVC(2023) | FCUIF (2024) | ESFMC(Ours) |
|---|---|---|---|---|---|---|
| ACC | 37.35 | 46.33 | 47.92 | 44.80 | 49.28 | 53.62 |
| NMI | 24.68 | 53.69 | 55.25 | 50.73 | 57.26 | 59.88 |
| ARI | 34.32 | 30.44 | 43.28 | 18.31 | 46.53 | 48.96 |
[1]Yang J, Shi R, Wei D, et al. MedMNIST v2-A large-scale lightweight benchmark for 2D and 3D biomedical image classification. Scientific Data, 2023, 10(1): 41.
Q2: More extensive empirical experiments—such as testing ESFMC against more sophisticated adversarial attacks (e.g., gradient inversion, membership inference attacks).
A2: Thank you for your valuable comment. Our primary focus is on privacy verification against model inversion attacks, which are among the most common threats targeting shared intermediate results. These attacks leverage model outputs to reconstruct the original input data and are a type of sophisticated adversarial attack. They are also one of the most prevalent and effective attacks in federated learning settings that share intermediate results (such as the features in our paper).
Regarding the other attacks you mentioned, they may not be applicable to our federated learning scenarios. For instance, gradient inversion attacks reconstruct original data by exploiting gradient information, which is commonly observed in federated learning settings that share model parameters or gradients. However, in our method, all client model training parameters remain strictly local, making such an attack infeasible. Similarly, membership inference attacks aim to determine whether a specific sample was part of the training dataset, thereby probing data privacy. However, since our training data are entirely unlabeled and does not include explicit sample membership information, attackers would gain no meaningful insights merely by inferring whether a sample participated in the training.
Additionally, we wish you to refer to our response to Reviewer pjmK Q2, where we elaborate on how our proposed method addresses potential threats from malicious clients, further strengthening its ability to safeguard sensitive information across different scenarios. Lastly, it is important to highlight that our method specifically tackles the trade-off between privacy preservation and performance improvement. Unlike existing works, which typically focus on one aspect at the cost of the other, our approach strives to maintain strong performance while minimizing privacy leakage as much as possible.
Q3: Can you provide additional quantitative results on privacy guarantees, such as empirical differential privacy noise impact analysis?
A3: Yes, we have included additional quantitative results on privacy guarantees. Specifically, we incorporate differential privacy by adding noise to the clustering-related features uploaded from clients to the server. The table below presents the clustering performance of ESFMC under different privacy bounds on Caltech dataset. We observe that ESFMC achieves both high performance and privacy at . However, as the level of noise increases at , the performance of ESFMC unavoidably degrades.
| Privacy Bound | No Privacy | ||
|---|---|---|---|
| ACC | 91.50 | 90.78 | 86.29 |
| NMI | 84.54 | 83.21 | 74.03 |
| ARI | 83.45 | 81.98 | 73.13 |
This paper introduces an effective and secure federated multi-view clustering method from an information-theoretic perspective. The proposed approach preserves privacy while effectively mining complementary global clustering structures. Additionally, the paper provides theoretical analyses of its generalization bounds and privacy guarantees.
给作者的问题
The effectiveness of the proposed method depends on the accuracy of feature splitting. How do the authors evaluate the correctness of their feature-splitting strategy?
论据与证据
Yes
方法与评估标准
Yes
理论论述
Yes, I checked the theoretical proof of the work.
实验设计与分析
Yes, I reviewed the experimental analyses for this paper. The multi-view datasets chosen for this paper are common but simple, and I would like the authors to add experimental results of the method on large-scale dataset.
补充材料
Yes
与现有文献的关系
This paper can inspire researchers to use multi-view methods for privacy-constrained scenarios, which is meaningful for the development of the federated multi-view learning field.
遗漏的重要参考文献
None
其他优缺点
Strengths: 1.The paper is well-written and clearly presents the proposed method. 2.The method is supported by both theoretical analysis and experiments, demonstrating its effectiveness in balancing privacy protection and clustering performance. 3.The method is extendable to incomplete scenarios and cross-device scenarios.
Weaknesses: 1.The multi-view datasets used in the experiments, while common, are relatively simple, and the paper lacks results on large-scale datasets. 2.In the incomplete multi-view setting, only different sample overlapping rates among clients are considered. More scenarios of data heterogeneity, such as quantity skew, should also be explored.
其他意见或建议
Refer to the weakness
We sincerely thank the reviewer for these valuable comments.
Q1: The paper lacks results on large-scale datasets.
A1: We conduct further experiments on the large-scale YoutubeVideo dataset, which contains 101,499 samples across 31 classes, where each sample has three views of cuboids histogram, HOG, and vision misc. The clustering results of ESFMC and several comparison methods in incomplete scenarios are shown below:
| Method | IMVC-CBG (2022) | DSIMVC (2022) | AGDIMC (2024) | FedDMVC (2023) | FCUIF (2024) | ESFMC(Ours) |
|---|---|---|---|---|---|---|
| ACC | 18.32 | 15.01 | 24.42 | 21.52 | 23.04 | 25.52 |
| NMI | 11.83 | 8.11 | 20.25 | 16.96 | 18.46 | 23.40 |
| ARI | 2.04 | 1.20 | 5.22 | 3.42 | 3.72 | 5.65 |
We select the YoutubeVideo dataset, which is 20 times larger than the Scene dataset with 4,485 samples. The results demonstrate that our method adapts well to large-scale datasets and outperforms other methods in terms of performance, ensuring the proposed method's robustness and broader applicability.
Q2: In the incomplete multi-view setting, more scenarios of data heterogeneity, such as quantity skew, should also be explored.
A2: Thank you for your suggestion. To explore the scenario of data heterogeneity, such as quantity skew, we introduce Dirichlet distribution when constructing incomplete datasets. A smaller Dirichlet parameter leads to more heterogeneous splits, resulting in highly imbalanced sample sizes among clients. The table below presents three levels of heterogeneity by setting to (high), (moderate), and (none) on BDGP dataset in incomplete scenarios. The results demonstrate that ESFMC maintains strong performance even under high heterogeneity, with only a slight performance drop.
| Levels of Heterogeneity | None | Moderate | High |
|---|---|---|---|
| ACC | 94.92 | 94.40 | 93.67 |
| NMI | 84.93 | 84.11 | 82.86 |
| ARI | 87.83 | 86.67 | 84.64 |
Q3: How do the authors evaluate the correctness of their feature-splitting strategy?
A3: This is indeed an issue worthy of attention. Our feature splitting strategy is guided by an intuitive training loss function. Specifically, ensures that sample-related features accurately reconstruct the original data, encourages clustering-related features to capture meaningful clustering structures, and minimizes redundancy between different feature types, promoting high-quality feature splitting. In the ablation studies (Table 3), removing and in the variants results in performance degradation, demonstrating that the feature splitting strategy effectively splits different features, thereby ensuring the effectiveness of our method. Furthermore, in Table 5, we analyze the impact of sharing different types of features on clustering performance. The results indicate that the clustering-related features extracted through our feature splitting strategy are more effective and accurate in capturing clustering structures, further validating the strategy’s accuracy and effectiveness. Overall, these ablation studies confirm that our feature splitting strategy successfully separates sample-related and clustering-related features, significantly enhancing clustering performance.
After the rebuttal and discussion phases, the paper received scores of 4, 4, 4, and 3, which exceed the expected threshold for acceptance. All reviewers are satisfied with the paper and the authors' response. After briefly reviewing the comments and the authors' responses, I believe the paper meets the acceptance criteria for ICML.