Learning Personalized Causally Invariant Representations for Heterogeneous Federated Clients
We propose a provable shortcut discovery and removal method under personalized federated learning to develop the optimal personalized invariant predictors that can generalize to unseen OOD data for the heterogeneous clients.
摘要
评审与讨论
The paper proposes a novel approach to mitigating shortcut learning in personalized federated learning, which is a challenging problem in real-world settings. The proposed method, FedSDR, utilizes structural causal models to discover and remove shortcuts while preserving personalization knowledge.
优点
-
This paper provides extensive background information, offering readers significant convenience in understanding the topic.
-
The authors conducted extensive experiments on some real-world datasets, validating the excellent performance of the proposed method.
缺点
-
I believe the main issue with this paper is that the research motivation seems weak. The paper claims to primarily address the scenario where training sets are Non-IID among clients, and where within each client, the training and test sets are also Non-IID. However, I think existing Robust Federated Learning and Federated Domain Generalization methods are capable of handling the aforementioned scenario. Although these two methods primarily focus on the issue of Non-IID training and test sets within each client.
-
The paper states, "To the best of our knowledge, we are the first to consider the shortcut trap problem in personalized federated learning and analyze it by formulating the structural causal models for heterogeneous clients." While it's true that this paper is indeed the first to study the use of PFL+SCM to address the shortcut trap problem, existing Robust Federated Learning and Federated Domain Generalization methods can also address the shortcut trap problem. Therefore, the contribution of this paper appears to be limited.
-
In Figure 1, I find the classification of the scenarios where RFL&FedDG are applicable not very accurate. In the scenarios where RFL&FedDG are applicable, the Test-test relation can also be IID or Non-IID.
-
In the experimental section, the proposed model is only compared to FL and PFL methods. However, I believe it should also be compared to RFL and FedDG methods to provide more convincing experimental results.
-
The experimental section lacks detailed information about the configuration of the model used in this paper. More detailed information is needed.
-
The paper lacks a thorough discussion of the limitations of the proposed FedSDR method. While the authors mention some potential limitations in passing, a more detailed discussion of the assumptions and constraints of the method would be helpful for readers to better understand its applicability in different scenarios.
-
Although the authors provide some high-level explanations of how FedSDR works, a more detailed discussion of the causal modeling framework and how it is used to address shortcut learning would be helpful for readers who are not familiar with this area of research.
问题
Please see above
Dear Reviewer Qvuh,
Thanks for your valuable feedbacks. As the deadline for author-reviewer discussion is approaching, we are concerned you might miss the clarifications that we make for your questions. We would appreciate it very much if you could reassess our work according to the clarifications. Of course, more discussions are welcomed if you still have any concerns or questions for this work.
Thanks for the reviewer's valuable feedbacks and suggestions. The detailed responses to the reviewer's questions and concerns are listed as follows:
1. I believe the main issue with this paper is that the research motivation seems weak. The paper claims to primarily address the scenario where training sets are Non-IID among clients, and where within each client, the training and test sets are also Non-IID. However, I think existing Robust Federated Learning and Federated Domain Generalization methods are capable of handling the aforementioned scenario. Although these two methods primarily focus on the issue of Non-IID training and test sets within each client.
Answer: Unfortunately, both robust federated learning and domain generalization methods cannot handle the scenarios well where target datasets are Non-IID across local clients. Since both Federated Domain Generalization (FedDG) and Robust Federated Learning (RFL) generate a shared global model, they abandon the important personalization information on each client. Therefore, the relation between our method and RFL/FedDG is same as the relation between the existing personalized federated learning (FL) and traditional federated learning (FL).
2. The existing Robust Federated Learning and Federated Domain Generalization methods can also address the shortcut trap problem. Therefore, the contribution of this paper appears to be limited.
Answer: The existing Robust Federated Learning (RFL) and Federated Domain Generalization (FedDG) methods will abandon the important personalization information on each client since they output a shared global model, although they can also mitigate the shortcut features. In many real-world applications (e.g., medical diagnosis and automatic driving), personalization information is of vital importance and cannot be overlooked. The main contribution of this paper is to fully exploit the significant personalization information, and in the meanwhile eliminate the shortcut features. It's rather challenging because the personalized and shortcut features are usually entangled in the real-world scenarios.
3. In Figure 1, I find the classification of the scenarios where RFL&FedDG are applicable not very accurate. In the scenarios where RFL&FedDG are applicable, the Test-test relation can also be IID or Non-IID.
Answer: The existing Robust Federated Learning (RFL) and Federated Domain Generalization (FedDG) methods will abandon the important personalization information on each client since they output a shared global model, when target datasets are Non-IID across local clients. Therefore, we conclude that RFL&FedDG can't cover the scenario well. The relation between our method and RFL/FedDG is same as the relation between the existing personalized federated learning (FL) and traditional federated learning (FL).
4. In the experimental section, the proposed model is only compared to FL and PFL methods. However, I believe it should also be compared to RFL and FedDG methods to provide more convincing experimental results.
Answer: Actually, we included three RFL&FedDG methods as baselines in the experimental section. The considered competitor DRFA is a typical RFA method while FedSR and FedIIR are two state-of-the-art FedDG methods. The overall experimental results of these baseline methods can be found in Table 1.
[1] Yuyang Deng, Mohammad Mahdi Kamani, and Mehrdad Mahdavi. Distributionally robust federated averaging. Advances in Neural Information Processing Systems, 33:15111–15122, 2020.
[2] A Tuan Nguyen, Philip Torr, and Ser-Nam Lim. Fedsr: A simple and effective domain generalization method for federated learning. In Advances in Neural Information Processing Systems, 2022.
[3] Yaming Guo, Kai Guo, Xiaofeng Cao, Tieru Wu, and Yi Chang. Out-of-distribution generalization of federated learning via implicit invariant relationships. In Proceedings of the 40th International Conference on Machine Learning, pp. 11905–11933, 2023.
5. The experimental section lacks detailed information about the configuration of the model used in this paper. More detailed information is needed.
Answer: The detailed information about the adopted models were provided in section 5.1. For CMNIST and CFMNIST, we adopt the deep neural network with one hidden layer of (128, 128) as feature extractor and an subsequent fully-connected layer as classifier. As regard to Waterbird and PACS, ResNet-18 is used as the learning model where the part before the last fully-connected layer works as feature extractor and the last fully connected layer works as classifier.
6. The paper lacks a thorough discussion of the limitations of the proposed FedSDR method. While the authors mention some potential limitations in passing, a more detailed discussion of the assumptions and constraints of the method would be helpful for readers to better understand its applicability in different scenarios.
Answer: We believe the outstanding limitation of this work is that the provided theoretical guarantee is restricted to linear feature space and we mentioned it in conclusion section. We will include more discussions in the updated version. Of course, extending the theoretical guarantee to more complex cases is rather challenging and we will push ahead with it in the future work.
7. Although the authors provide some high-level explanations of how FedSDR works, a more detailed discussion of the causal modeling framework and how it is used to address shortcut learning would be helpful for readers who are not familiar with this area of research.
Answer: Actually, a reasonable structural causal model can be regarded as some prior knowledge. We propose the structural causal models in federated learning by adding the user/client indicator and deconstructing the invariant features into two separate parts: the personalized invariance and the shared/global invariance . Other parts of the causal models are same as what are commonly adopted in the literature on centralized invariant learning.
As for how they are used to address shortcut trap in federated learning, the key is two provable causal signatures that we observe from the proposed structural causal models. As listed in Lemma 1, the first causal signature indicates that we can discover the shortcut features using training environments across local clients even if the data generating mechanisms are heterogeneous among them. The second causal signature makes it possible to develop the optimal personalized invariant predictors with the discovered shortcut features even though there is just one training environment on each client, since the relationships between , and are independent of environment . With these two causal signatures, we design the provable shortcut discovery and removal method for federated learning, i.e., FedSDR.
Thanks again for your valuable feedbacks. More discussions are welcomed if you still have any concerns or questions for this work.
The paper proposes a new personalized federated learning approach named FedSDR. Considering generalization on unseen test data, the paper utilizes invariant learning in the federated setting. Specifically, FedSDR first extracts shortcut features that are irrelevant to the task and them remove it to extract the most informative personalized invariant features by carefully designing the objectives. Experiments show that FedSDR outperforms the other baselines in the settings where the test data distribution shifts.
优点
-
Applying invariant learning in the federated setting is interesting and promising.
-
The organization of the paper is clear.
-
FedSDR significantly outperforms the other baselines.
缺点
-
My main concern is about the experimental setting. Experiments are based on the simulated setting where the authors manually add shortcut features and change the test data distributions. Based on the motivation, the shortcut features should naturally exist in datasets. Experiments on real-world natural datasets are necessary. Otherwise, the application of FedSDR may be very limited.
-
The theoretical analysis has strong assumptions, e.g., logistic regression, linear function, etc. The analysis may not be applicable in the experimental settings.
-
The number of clients used in the experiments is missing. Experiments to evaluate the scalability of FedSDR are not provided.
-
Typos: 1. Page 4: “Theorem 1” -> “Lemma 1”; 2. Page 6: “guarantee” -> “guarantees”
问题
-
Can you add a synthetic dataset with a simple model in the experiments? It can be used to verify the theorems by satisfying the assumptions.
-
Can you add experiments without manually adding shortcut features? It is quite important. Currently, I’m not clear what are the real applications of FedSDR.
-
Can you add experiments that increase the number of clients and adopt client sampling?
Thanks a lot for the reviewer's valuable feedbacks and suggestions. The detailed responses to the reviewer's questions and concerns are listed as follows:
1. Can you add a synthetic dataset with a simple model in the experiments? It can be used to verify the theorems by satisfying the assumptions.
Answer: We generate a synthetic dataset using the same strategy as in [2]. Specifically, it is a logistic regression task and the data instance is generated by , where the dimensionalities of , and are , and respectively. The linear function is implemented by one fully-connected layer which has 12 neurons. The latent variables , and are subject to , and respectively. Target variable is taken from the distribution . Both and are randomly sampled from while is randomly sampled from . To make the shortcut representation easier to learn, we choose and as in [2]. Each fixed value of indicates one environment. We generate 10 training environments and 5000 test environments to evaluate the out-of-distribution generalization performance. Each (training/test) environment contains 10000 data samples (X, y) and the training data samples are distributed onto totally 100 clients. The training and test data samples on each client are generated with an identical value of . Besides, we choose the client sampling rate as 0.1. The experimental results on this synthetic dataset are shown in the following table:
| Algorithm | FedAvg | DRFA | FedSR | FedIIR | FTFA | pFedMe | Ditto | FedRep | FedRoD | FedPAC | FedSDR |
|---|---|---|---|---|---|---|---|---|---|---|---|
| worst-case (%) | |||||||||||
| average (%) |
In particular, when we manually select the causal features as the discriminating features, we find the optimal personalized classifiers achieve an stable accuracy around in different test environments. Therefore, the results shown in Table 1 can demonstrate the effectiveness of our FedSDR on developing the optimal personalized invariant predictors, compared with the state-of-the-art FL and PFL methods.
[2] Rosenfeld, E., Ravikumar, P., & Risteski, A. (2021, July). The Risks of Invariant Risk Minimization. In International Conference on Learning Representations.
2. Can you add experiments without manually adding shortcut features? It is quite important. Currently, I’m not clear what are the real applications of FedSDR.
Answer: Actually, PACS is a real-world dataset that is commonly used in related papers. It's a multi-class classification task and there doesn't exists explicit (or manually adding) shortcut features. Specifically, It consists of 7 classes distributed across 4 environments (or domains). The results on PACS dataset are also shown in Table 1.
3. The number of clients used in the experiments is missing. Experiments to evaluate the scalability of FedSDR are not provided. Can you add experiments that increase the number of clients and adopt client sampling?
Answer: Due to the space limitation, we put the complete experimental setups in Appendix B.1. In particular, we simulate 8 clients in the experiments on CMNIST, CFMNIST and WaterBird. The experiments on PACS are conducted on 6 clients. To further evaluate the scalability of FedSDR, we firstly partition CMNIST dateset into 8 subsets using the same strategy adopted to simulate 8 clients. And then , we randomly distribute each subset onto 10 clients. In this way, we totally construct 80 clients for CMNIST dataset. Similarly, we construct 80, 80, 60 clients for CFMNIST, WaterBird and PACS respectively. When evaluating the model performance on these four datasets, we adopt a client sampling rate of 0.1. The experimental results are shown in the following table:
| Dataset | CMNIST | CFMNIST | WaterBird | PACS | ||||
|---|---|---|---|---|---|---|---|---|
| Test accuracy (%) | worst-case | average | worst-case | average | worst-case | average | worst-case | average |
| FedAvg | ||||||||
| DRFA | ||||||||
| FedSR | ||||||||
| FedIIR | ||||||||
| FTFA | ||||||||
| pFedMe | ||||||||
| Ditto | ||||||||
| FedRep | ||||||||
| FedRoD | ||||||||
| FedPAC | ||||||||
| FedSDR | 50.41 | 51.85 | 52.81 | 57.14 | 59.96 | 68.09 | 48.07 | 51.55 |
The results show that FedSDR can still outperform the competitors when there are a large number of clients in the federated learning system, which can validate the scalability of the proposed FedSDR.
4. The theoretical analysis has strong assumptions, e.g., logistic regression, linear function, etc. The analysis may not be applicable in the experimental settings.
Answer: Actually, the algorithm itself doesn't require the assumption of linear feature space. In particular, the experiments on WaterBird and PACS are conducted with ResNet as learning model. Therefore, feature spaces of these two datasets are non-linear with regard to input . Moreover, PACS is a multi-class classification task rather than logistic regression task. The results on thses two datasets can demonstrate the effectiveness of FedSDR in general scenarios (including non-linear cases and multi-class classification).
Of course, the theoretical guarantee is still restricted to linear cases in the current version. Extending the theoretical guarantee to more complex cases is rather challenging and we will push ahead with it in the future work.
5. Typos: 1. Page 4: “Theorem 1” -> “Lemma 1”; 2. Page 6: “guarantee” -> “guarantees”
Answer: We will fix them in the updated version.
Thanks again for your valuable feedbacks. More discussions are welcomed if you still have any concerns or questions for this work.
Thanks for your response. Most of my concerns have been addressed. I have raised my score to 6.
Thanks a lot for your reassessment of our work. Thanks again for your valuable feedbacks which help to improve the completeness of this work.
The paper focuses on addressing the "shortcut trap" issue within personalized federated learning (pFL) by introducing FedSDR. This solution aims to identify and eliminate shortcut features, leading to enhanced performance for individual clients within pFL on their respective local datasets. The authors support their approach with theoretical proofs and comprehensive experiments, demonstrating the effectiveness of FedSDR across multiple scenarios.
优点
-
Originality: The paper pioneers the exploration of the shortcut trap problem in pFL.
-
Solid Theoretical Support: The inclusion of strong theoretical foundations.
-
Sufficient Experiments: The comprehensive set of experiments strengthens the paper's contributions.
缺点
-
Notation Conciseness: The notation could be more concise, potentially enhancing the paper's readability.
-
Confusing Training Process: A section of the training process appears confusing and requires clarification.
-
Inconclusive Experimental Results: Certain aspects of the experimental outcomes lack conviction and need further clarification for robustness.
问题
-
The paper’s algorithmic approach (Algorithm 1) deviates from the conventional practice of client selection in Federated Learning (FL). Could the authors elucidate why they've chosen this approach over the typical FL client selection methodology?
-
Concerns arise from certain experimental results in Table 1, particularly the unexpected lower accuracies observed for specific baseline methods on CMNIST and CFMNIST. Can the authors provide an explanation to address these concerns?
Thanks a lot for the reviewer's valuable comments and suggestions. The detailed responses to the reviewer's questions and concerns are listed as follows:
1. The paper’s algorithmic approach (Algorithm 1) deviates from the conventional practice of client selection in Federated Learning (FL). Could the authors elucidate why they've chosen this approach over the typical FL client selection methodology?
Answer: We adopted the same federated learning framework as in pFedMe [1] while the conventional client selection is conducted before local updates. In the adopted framework, the global model is generated by aggregating the updates from the selected clients and the updates from the clients that are not selected are never used. Then on each client, the local model will be initialized by the global model before local update. Therefore, the final output models of these two client selection methods are exactly identical. Our method FedSDR can also be implemented using the conventional client selection methodology and the experimental results will keep unchanged.
[1] T Dinh, C., Tran, N., & Nguyen, J. (2020). Personalized federated learning with moreau envelopes. Advances in Neural Information Processing Systems, 33, 21394-21405.
2. Concerns arise from certain experimental results in Table 1, particularly the unexpected lower accuracies observed for specific baseline methods on CMNIST and CFMNIST. Can the authors provide an explanation to address these concerns?
Answer: As shown in Table 1, our FedSDR achieved around 6.5%, 9%, 3.5% and 2% higher worst-case accuracy than the second best algorithm among the baselines and in the meanwhile reaches the highest average accuracy on CMNIST, CFMNIST, WaterBird and PACS, respectively. So, what do you mean by "the unexpected lower accuracies observed for specific baseline methods on CMNIST and CFMNIST"? In addition, we will make the project codes public soon to enhance the reproducibility of the conducted experiments.
Thanks again for your valuable feedbacks. More discussions are welcomed if you still have any concerns or questions for this work.
Thanks a lot for your response. I would like to inquire further about the reasons behind the significantly low worst-case results obtained by pFL baseline methods. It would be greatly appreciated if the authors could provide a reasonable explanation.
Glad to receive your further responses.
Good question. The prevalent PFL methods develop the personalized model on each client with the guidance of a shared global model (e.g., FTFA, pFedMe and Ditto) or some shared knowledge (e.g., FedRep, FedRoD and FedPAC). Since both personalization information and environment-dependent shortcut features are varied across local clients, the shared global model/knowledge excludes both personalized and shortcut features. When these methods conduct local adaptation to develop the personalized model on each client, the developed model can readily pick up the shortcut features if local training data is biased because these PFL methods lack a strong mechanism for shortcut identification and removal. However, relying on shortcut features can degrade the performance of the developed model to any extent when the distribution shift between training and test datasets is very large. In other words, the worst-case accuracy can be significantly low. In comparison, the proposed FedSDR can effectively extract the shortcut features and then eliminate the identified shortcut features during the local adaptation. Therefore, FedSDR can achieve a much stabler and higher test accuracy than the PFL baseline methods.
Thanks again for your further responses. More discussions are welcomed if you still have any concerns or questions for this work.
This paper introduces FedSDR, a method for addressing shortcut learning in personalized federated learning (PFL). PFL faces challenges in real-world settings due to the assumption of in-distribution test data, leading to shortcut learning. FedSDR utilizes structural causal models for federated clients, offering provable shortcut discovery and removal. It collaboratively discovers shortcuts using training data and develops an optimal personalized causally invariant predictor for each client. Theoretical analysis and experiments on diverse datasets validate FedSDR's superiority over state-of-the-art PFL methods in out-of-distribution generalization performance.
The paper got controversial comments from the reviewers. While most of the reviewers recognize the technical novelty and effectiveness of this paper, one reviewer has some concern on the comparison with robust federated learning and federated domain generalization. As discussed in the rebuttal, the reviewer ignored such discussion and comparison in the paper. Therefore, despite the overall low rating, mainly influenced by one significantly low score, I lean towards recommending the acceptance of the paper.
为何不给更高分
overall rating is not high
为何不给更低分
The paper makes a good contribution to causal treatment of federated learning. The technical quality is good and the experiments are solid.
As the overall score is low, due to a bias in one reviewer, I am open to discussion if there are some concerns.
Accept (poster)