Enhancing Personal Decentralized Federated Learning through Model Decoupling
摘要
评审与讨论
This work proposes a personalized federated learning (PFL) framework called DFEedMDC, which pursues robust communication and better model performance with a convergence guarantee. Besides, to promote the shared parameters aggregation process, the authors propose DFedSMDC via integrating the local Sharpness Aware Minimization (SAM) optimizer to update the shared parameters.
优点
This work designs personalized local models and training schemes for decentralized federated learning. The authors present theoretical analyses for the convergence, which shows the negative influence of the statistical heterogeneity and the communication topology. Extensive experiments are conducted to evaluate the effectiveness of the proposed methods.
缺点
What do the "Grid" and "Exp" represent in Fig. 3? It would be easier for the readers to understand different communication topologies by visualizing them in the main test or in the Appendix.
In light of Theorem 1 and Theorem 2, the communication topology (i.e., the eigenvalue ) has an impact on the DFedMDC and DFedSMDC methods. The reviewer suggests the authors report the values of different communication topologies in Fig. 3 and discuss the influence of on the test accuracy.
问题
The proposed DFedSMDC method, a variant of DFedMDC, achieves better performance by integrating the SAM optimizer into the local iteration update of shared parameters. The reviewer is curious if the incorporation of this optimizer could similarly enhance the performance of other baseline methods.
We greatly appreciate your suggestions and comments, which will make our work better than this version. Thanks for your positive comments, and we address the weaknesses and minor comments below.
W1: What do the "Grid" and "Exp" represent?
Thank you for your suggestion and we added the visualization of topologies in Appendix B.3 in the revision. The explanations of Grid and Exp are as follows:
- Grid: Grid topology refers to a network configuration where clients are arranged in a two-dimensional grid-like pattern. That means clients are organized in rows and columns, forming a rectangular or square grid structure. In a 100-client network, the exponential neighbors of client 10 are client 11(10+1), client 9(10-1), client 19(10+9), client 1(10-9).
- Exp: Exponential(Exp) topology refers to a network configuration where clients are arranged in an exponential manner. In a 100-client network, the exponential neighbors of client 10 are client 11(10+), client 12(10+), client 14(10+), client 18(10+), client 26(10+), client 42(10+), client 74(10+).
W2: The value and the influence of in Fig.3.
Thank you for your suggestion and we have added more analysis of the spectral gap of different communication topologies to the revision. From the relationship between the spectral gap and the participation number of clients ( in our experiment), as follows, we can see that the convergence bounds of different topologies are ranked as follows: Fully-connected > Exponential > Grid > Ring, which is consistent with our experiment results.
| Graph Topology | Spectral Gap |
|---|---|
| Fully-connected | 1 |
| Disconnected | 0 |
| Ring | |
| Grid | |
| Exponential Graph |
Q1: Incorporation of SAM optimizer with other baseline methods.
We add more experiments about the incorporation of SAM optimizer with other baseline methods on CIFAR-10 as follows.
| Algorithm | FedAvg | FedPer | FedRep | FedBABU | FedRod | Ditto | DFedAvgM | Dis-PFL | DFedMDC |
|---|---|---|---|---|---|---|---|---|---|
| Dir0.3 | 79.66 | 84.06 | 84.50 | 83.26 | 85.68 | 73.51 | 82.60 | 82.71 | 86.50 |
| Dir0.3+SAM | 80.02 | 86.81 | 87.19 | 85.61 | 85.33 | 71.07 | 77.36 | 82.66 | 88.32 |
| Pat2 | 85.04 | 90.94 | 91.16 | 91.17 | 90.10 | 84.96 | 90.72 | 88.19 | 91.26 |
| Pat2+SAM | 84.99 | 91.73 | 91.80 | 91.67 | 89.52 | 83.36 | 90.14 | 87.89 | 92.21 |
From the experimental results, SAM could similarly enhance the performance of the partial model personalized methods, but whether it could enhance the full model personalized methods is doubtful. For the partial model personalized methods FedPer, FedRep, FedBABU and DFedMDC, SAM adds proper perturbation in the direction of the shared models’ gradient to make the shared part more robust and flat, which will benefit the average progress for each client. For the full model personalized methods FedAvg, FedRod, Ditto and DFedAvgM, pursuing the full model flatteness will decrease the classifier’s adaptation to local distribution.
Thank you for the responses. After careful consideration, I decided to keep my initial score.
This paper studies the personalized federated learning problem under fully decentralized setting. The framework of the considered personalized learning is the commonly used model decoupling with a globally shared model and personalized local models. DFedSMDC, an algorithm via integrating the local Sharpness Aware Minimization (SAM) optimizer to update the shared parameters, is proposed. Theoretical convergence results and numerical experimental results are both presented.
优点
This paper studies the personalized federated learning problem under fully decentralized setting, and proposed DFedSMDC, an algorithm via integrating the local Sharpness Aware Minimization (SAM) optimizer to update the shared parameters.
缺点
-
The reviewer is quite doubt about the final results as shown in Theorem 1 and Theorem 2. I’ve checked the theoretical proof in the appendix and do not find the exact expressions for the final convergence results, but only the expression. The first questionable part is that the right-hand side of Eqs. (3)-(4) will goes to 0 as the number of rounds goes to infinity, while in reality, this is not true for non-i.i.d scenarios. There will exists some constant terms related with heterogeneity that are irrelevant to . Please explain this.
-
The second part that may not be true in the theoretical results is that the convergence speed is monotonically related with the spectral gap . If this is true, it solves the challenging topology design problem of decentralized federated learning, since a fully-connected topology is the optimal topology according to the theoretical results in this paper. There is no discussion about this point in current manuscript and this leads to a doubtful result.
-
Why is the convergence results not related with the number of workers? This is also a weird part.
-
Why Theorem 1 is related with the cross Lipschitz constant , and Theorem 2 is related with ? How about ?
-
The results in Fig.3 are questionable according to the second comment. The reviewer is not sure if a fully-connected topology is the best.
-
What is the meaning of Fig. 4? Are multiple local epochs good or bad? How is it related with the theoretical results?
问题
See the weakness above.
Thanks for your comments and we reply to them below. Please reconsider the contributions of our work.
W1: The doubts about Theorem 1 and Theorem 2.
- Only the expression. is the final results after assuming the learning rates and . Similar results can be found in [1-3]. Due to the limited space, the initial convergence analysis has been presented in Formula (19) and Formula (33).
- The right-hand side of Eqs. (3)-(4) will goes to 0 as the number of rounds goes to infinity.
- The gradient that goes to 0 is one of the optimality conditions in non-convex optimization. As an optimization problem, we consider minimizing problem: and find the global minima , that satisfies . In non-convex optimization, the minima can be defined by . So we prove the convergence from the gradients going to 0, which satisfies . More details about non-convex optimization can be referred to [4,5].
- We assume that and , which means that the learning rate and tend to 0 as the number of rounds goes to infinity, and the gradients eventually tend to 0 correspondently. We deduce that the reviewer wants to say that the proposed methods will converge to a variance under the constant learning rate. Actually, our methods converge to a variance domain with the sublinear order convergence speed.
- Moreover, in a non-iid scenario, each client will achieve convergence and their gradient will approximate to 0 after sufficient large communication rounds. However, this does not mean that all clients have the same solution. For PFL, every client will own its unique solution , the of which is different from each other.
W2: The convergence speed is monotonically related to the spectral gap.
- The convergence speed is monotonically related to the spectral gap. The convergence speed is related to the statistical heterogeneity , the smoothness , , of loss functions, the local epochs and the communication topology (1 − ) in the theoretical results. Due to the limited space, we only present the final results in Formula (3) and Formula (4). Please check the details in Formula (20-21) and Formula (34-36).
- Whether Fully-connected topology is the optimal topology according to the theoretical results.
-
Theoretically, Fully-connected topology is the optimal topology among the comparison topologies and the comparison of the spectral gap of each topology is as follows ( refers to the partition number of clients):
Graph Topology Spectral Gap Fully-connected 1 Disconnected 0 Ring Grid Exponential It is clear to see that the convergence bounds of different topologies are ranked as follows: Fully-connected > Exponential > Grid > Ring. But it contains more communication cost compared with other topologies. So it is a trade-off between the convergence and communication cost in the real world. We have added the comparison to the revision.
-
Empirically, we conduct experiments on different topologies (i.e. Ring, Grid, Exponential and Fully-connected)in Fig 3, which suggests the personalized performance of different topologies is ranked as follows: Fully-connected > Exponential > Grid > Ring.
-
W3: The convergence results are not related to the number of workers.
Refer to the table above, the effects of the participation workers have been involved in the spectral gap of communication topologies. We have added it to the revision.
[1] Decentralized federated averaging, ICML2022.
[2] Improving the Model Consistency of Decentralized Federated Learning, ICML2023.
[3]Federated Learning with Partial Model Personalization, ICML2022.
[4] Non-convex Optimization for Machine Learning.
[5] First-order methods in optimization.
W4: Theorem 1 and Theorem 2 are all related to the cross Lipschitz constant .
-
First, based on the Lipschitz Smooth assumption, we define is --Lipschitz with respect to and --Lipschitz with respect to in Assumption 1, which means and . The same for is --Lipschitz with respect to and --Lipschitz with respect to . More details about Lipschitz Smooth assumption can be found in [6].
-
Then, we conduct
in the second line in Formula (15). -
The theoretical analysis progress is consistence with the algorithm, which first updates the personalized parts and then the shared parts . So Theorem 2 and Theorem 1 are associated with not .
W5: Is a fully-connected topology is the best?
Fully-connected topology achieves the best performance in the comparisons under different topologies. But it needs more communication costs for each client. So it is a trade-off of the test performance and communication cost to choose the topology.
W6: The meaning of Fig. 4.
- Local epochs for the personal parameters is an important hyperparameter in our algorithm, which will significantly influence the final experiment results. From the comparison of FedBABU(=0) and FedRep(=10), we deduce is an important hyperparameter in the alternately partial personalization. represents the trade-off of the contribution of the shared representation and personalized classifiers to the convergence. Bigger means the local convergence relied more on the fitting ability of the personalized classifiers, while smaller means the local convergence relied more on the generalization of the shared representation. This point has never been discussed before in the partial personalization works, which is one of our surprising findings as well.
- Theoretically, multiple local epochs speed up the convergence but may be prone to over-fitting. This will decrease the generalization accuracy and it is a common phenomenon in FL. Similar to it is the large batch-size training. The convergence will be accelerated with large batch sizes, but the generalization will become worse. Multiple local epochs good or bad are related to the data distribution in our experiments and similar performance appears in [7] as well.
- has been cancelled out from the assumption in Formula (19-20).
[6] Nonlinear Systems 3rd.
[7]On Large-Cohort Training for Federated Learning, Neurips 2021.
I read the authors' response and will keep my original score.
Thanks for your reply. We acknowledge your effort in reviewing our paper and providing valuable feedback.
If you have checked our response, can you enumerate where else we can improve?
Thanks
This paper proposes interesting methods DFedMDC and DFedSMDC for PFL, which simultaneously guarantee robust communication and better model performance with convergence guarantee via adopting decentralized partial model personalization based on model decoupling.
优点
- The study of personalized FL on decentralized FL is meaningful.
- The experiments demonstrate that the proposed method is useful.
缺点
- The proposed algorithm seems trivial and common in PFL. It seems its idea is the adoption of the method in DFL. Can you clarify what is the main novelty of this method?
- Why introduce the SAM? It is unclear about the advantage of introducing this optimizer. Can you elaborate on it intuitively and theoretically?
- In the theorem, why is it , instead of , and what does it mean?
- The experiment results are a bit weird, in Table 1. Why do all baselines achieve better performance under larger heterogeneity? As I know, larger heterogeneity will usually lead to worse performance [1].
- Regarding ``The test performance will get a great margin with the participation of clients decreasing’’: What will happen when the client number is less than 10, even 1? Does it mean no collaboration is the best?
[1]Karimireddy S P, Kale S, Mohri M, et al. Scaffold: Stochastic controlled averaging for federated learning[C]//International conference on machine learning. PMLR, 2020: 5132-5143.
Minors:
- It seems the hyperparameters of the proposed methods are finetuned (like and local epoch for the personal part). Are the baselines’ results well finetuned? What's the used hyperparameter for baselines?
- What is the definition of in Theorem 2?
问题
- Could you give more explanation on Theorem 2? What is the difference/advantage compared with Theorem 1 as you introduce SAM into shared parameters?
- Can you provide baseline results with more hyperparameter settings?
- Could the authors provide more details about the experiment settings?
M1: What's the used hyperparameter for baselines? Are the baselines’ results well fine-tuned?
The baselines’ results have been well-finetuned and the hyperparameter of the baselines can be found in Appendix B.3 (MORE DETAILS ABOUT BASELINES). We show the finetuned processing as follows.
- Setting: For all methods, we keep the experiment setting as the same as follows. We perform 500 communication rounds with 100 clients. The client sampling radio is 0.1 in CFL, while each client communicates with 10 neighbors in DFL. The batch size is 128. The learning rate is 0.1 with 0.99 exponential decay. The local epochs are set to 5 for full model personalized methods and the shared models of the partial model personalized methods. The local optimization is SGD with momentum 0.9 and 5e-4 weighted decay.
- Finetuning: For FedAvg, FedPer, FedBABU, Fed-RoD, DFedAvgM, there is no more hyperparameters. We show the finetuning progress for FedRep, FedSAM, Ditto, Dis-PFL on CIFAR-10 as follow:
-
FedRep: We finetune the local epochs for the personalized parts and set in our experiments.
1 4 7 10 15 Dir-0.3 88.50 90.71 90.81 91.09 90.79 Pat-2 80.85 83.93 84.17 84.50 84.27 Theoretically, multiple local epochs will speed up convergence but may be prone to over-fitting. It is a trade-off between convergence and generalization. The difference of the best epochs for the personal part demonstrates the difference between CFL and DFL.
-
FedSAM: The extra hyperparameter in FedSAM is the perturbation radius and we set in our experiments.
0.01 0.1 0.3 0.5 0.7 0.9 Dir-0.3 78.78 79.08 79.20 79.61 80.02 79.52 Pat-2 83.26 83.69 84.11 84.57 84.99 83.48 With a larger , the flatter models that benefit the aggregation average can increase the global generalization. But for PFL, it is a trade-off between the personalization and the generalization.
-
Ditto: The extra hyperparameter in Ditto is the interpolated coefficient and we set in our experiments.
0.1 0.4 0.75 1.25 2 Dir-0.3 64.77 70.63 73.51 72.73 69.50 Pat-2 80.10 83.64 84.96 84.78 76.71 offers a trade-off between the personalization and the generalization. When is set to 0, Ditto is reduced to training local models; as grows large, it recovers the global model objective.
-
Dis-PFL:We finetuned the sparsity ratio in Dis-PFL and set it as 0.5 in our experiments.
Sparsity Ratio 0.2 0.4 0.5 0.6 0.8 Dir-0.3 80.17 82.63 82.72 82.56 81.94 Pat-2 87.77 88.07 88.19 87.99 87.82 The sparsity ratio is a trade-off between generalization and personalization. A higher sparsity ratio may bring more generalization benefits when the sparsity is small. But with the sparsity ratio increasing, less information exchange will lead to performance degradation.
-
M2: What is the definition of in Theorem 2?
is a variate associate with the gradient perturbation in SAM. We define in Formula (24-26). And when assuming , .
Q1: What is the difference/advantage compared with Theorem 1 as you introduce SAM into shared parameters?
Could you give more explanation on Theorem 2? What is the difference/advantage compared with Theorem 1 as you introduce SAM into shared parameters?
- The complete final bound in Theorem 1 is in Formula (20). And complete final bound in Theorem 2 is in Formular (34). SAM introduce the gradient perturbation in Theorem 2.
- From the comparison, the dominant terms are the same in Theorem 1 and Theorem 2, which means SAM doesn't improve the convergence bound in theoretical analysis. However, it makes the shared models more robust and flatter, which will benefit the average process for each client.
Q2: Provide baseline results with more hyperparameter settings.
Please check the reply in M1.
Q3: Provide more details about the experiment settings.
Please check more details about the experiment settings in 5.1 (EXPERIMENT SETUP) and Appendix B.2 & B.3 in the revision paper.
Thanks for your comments and we reply to them below. Please reconsider the contributions of our work.
W1: Clarify the main novelty of this method.
We clarify more details about our background, motivation and contribution. Please check them in the top official comment.
W2: Why introduce the SAM?
- Intuitively, SAM adds proper perturbation in the gradient direction to make the shared part become more robust and flatter, which will benefit the model average between clients.
- Theoretically, we give the effect on bound after adding the operation of SAM in Remark 2, where assuming . Due to the limited space, we have presented the bound before assuming in Formula (34) in Appendix C.2.2 on Page 26 as follows:
- before using :
- after using :
- Empirically, the comparison between DFedSMDC and DFedMDC in different heterogeneity and different topologies can significantly demonstrate the effectiveness in enhancing the consistency of the shared models in decentralized PFL.
W3: not in the theorem.
Thanks for your suggestion and we have polished it in the revision paper.
W4: "Why do all baselines achieve better performance under larger heterogeneity? "
What we focus on are the PFL problems, not the FL problems. In PFL tasks, the higher heterogeneity of data distribution means it owns fewer classes of data locally, which makes the classification task easier and clients will achieve better performance. For example, in the Pathological-2 setting, the local task is a binary classification task, which is easier than the five classification tasks in the Pathological-5 setting, so the average test performance in Pathological-2 is better than that in Pathological-5. The same phenomenon can be seen in most PFL works [1, 2, 3, 4].
W5: "The test performance will get a great margin with the participation of clients decreasing.’’
We have updated more experiments when the client number is 10 and 5 in Fig 5 in the revision paper. The test performance still improves with the participation of clients decreasing. And we clarify our opinion theoretically and empirically as follows:
-
Theoretically, the convergence bound is related to the topology, which is associated with the participation of clients. The relationship between the topology and the participation clients is as below:
Graph Topology Spectral Gap 1- Fully-connected 1 Disconnected 0 Ring Grid Exponential Graph When the participation clients decrease, the spectral gap increases and the convergence bound is tighter.
-
Empirically, with the participation of clients decreasing, the number of local data increases. More data will help to achieve a better performance.
-
When the client number is 1, the problem will become a classical ML problem based on a centralized data assumption, not an FL problem based on a decentralized data setting. We aim to collaboratively train the personalized models without transmitting data.
[1]On bridging generic and personalized federated learning for image classification. ICLR2022.
[2]Fedbabu: Towards enhanced representation for federated image classification. ICLR2022.
[3]Personalized federated learning with feature alignment and classifier collaboration. ICLR2023.
[4]FedProto: federated prototype learning across heterogeneous clients AAAI2022.
Dear Reviewer A4M3,
We sincerely appreciate the time and effort you have invested in reviewing our work. As the deadline for the discussion period is approaching, we kindly request that you inform us of any remaining questions you may have.
We are confident that our response has adequately addressed your concerns, and we would be grateful for your feedback. Should you require further clarification, we would be delighted to answer any additional questions and provide more information.
Best wishes,
The Authors
In this paper, the authors present an innovative framework known as DFedMDC, which leverages model decoupling to address these issues and aims to provide robust communication and superior model performance while guaranteeing convergence. DFedMDC achieves this by personalizing the "right" components within modern deep models through alternate updates of shared and personal parameters, facilitating the training of partially personalized models in a peer-to-peer manner. To enhance the shared parameters aggregation process, the authors introduce DFedSMDC, which incorporates the local Sharpness Aware Minimization (SAM) optimizer to update shared parameters. SAM optimizer introduces proper perturbations in the gradient direction to mitigate inconsistencies in the shared model across clients.
The paper provides a thorough theoretical foundation, offering a convergence analysis of both algorithms in a general non-convex setting with partial personalization and SAM optimizer for the shared model.
优点
- The paper is well-written and exhibits a high degree of clarity, making it accessible and easy to comprehend.
- The paper's strength is further underscored by its meticulous convergence analysis, enhancing its overall robustness.
- The paper substantiates its claims with an exhaustive array of experimental results, effectively confirming the effectiveness of the proposed method.
缺点
-
A significant concern revolves around the novelty of the proposed method. The concept of model decoupling in personalized federated learning [1] and the application of Sharpness Aware Minimization (SAM) [2] to address model inconsistencies in decentralized federated learning have both been extensively explored in the literature. As such, the proposed method may appear to be a fusion of existing ideas (resembling an 'A+B' approach). It is essential for the authors to underscore their distinctive contributions in a more prominent manner.
-
In terms of experimental baselines, it is recommended that the authors include the most recent decentralized federated learning method ([2]) for a comprehensive comparison. This will enhance the paper's completeness and relevance in the context of the current state of the field.
-
Regarding the convergence analysis, it would be valuable to incorporate a discussion that compares the proposed method's convergence rate with the state-of-the-art (SOTA) approaches.
[1] Exploiting Shared Representations for Personalized Federated Learning [2] Improving the Model Consistency of Decentralized Federated Learning
问题
See weaknesses section above.
Thanks for your comments and we reply to them below. Please reconsider the contributions of our work.
W1: The novelty of our work.
We give more details about our background, motivations and contributions. Please check them in the top official comment box.
W2: Add the most recent decentralized federated learning method for a comprehensive comparison.
Thanks for your suggestion. We provide a comparison with DFedSAM in the following and we have added this comparison in the revision.
| Dataset | | Cifar-10 | | | | Cifar-100 | | |
|:-------------:|:-------|:------------:|:-----:|:-----:|:-------:|:-------------:|:-----|:------|
| Algorithm | Dir-0.1 | Dir-0.3 | Pat-2 | Pat-5 | Dir-0.1 | Dir-0.3 | Pat-5 | Pat-10 |
| DFedSAM|84.96| 77.36|90.14|83.05|58.21|47.80|74.25|67.34|
| DFedAvgM | 87.39 | 82.60 | 90.72 | 84.69 | 59.76 | 54.98 | 76.70 | 71.08 |
| DFedMDC| 88.85 | 86.50 | 91.26 | 86.85 | 66.26 | 57.66 | 78.78 | 72.19 |
| DFedSMDC | 91.08 | 87.67|92.20|88.34|67.03|58.73|80.82|74.50|
From the comparison between DFedAvgM and DFedSAM we can see that adding the gradient perturbation to the full model will increase the model consistency but hurt the personalized performance. On the contrary, adding the perturbation to the shared representation parts in DFedSMDC outperforms DFedMDC clearly, which shows that both improving the shared representation parts consistency and keeping the private classifiers locally are both the key points to facilitate the personalized performance in decentralized PFL.
W3: compares the proposed method's convergence rate with the state-of-the-art (SOTA) approaches.
Compared with the SOTA bounds ( is the upper bound of the gradient) of existing work DFedAvg and of DFedSAM in decentralized works, our algorithms reflect the impact of the L-smoothness and the gradient variance of the shared model and personalized model on convergence rate. On the other hand, compared with the SOTA bound of FedAlt in personalized works, our algorithms reflect the impact of the communication topology (the value of that increases when connectivity is more sparse, which has been studied by existing work [1].)
Therefore, the main contribution of our convergence analysis is the first to analyze the impact of model decoupling and the decentralized networks on the convergence rate instead of delivering the SOTA bound compared with others. Meanwhile, this theoretical analysis can be mutually verified with experimental results.
[1] Topology-aware Generalization of Decentralized SGD. ICML2023.
Dear Reviewer iEM9,
We sincerely appreciate the time and effort you have invested in reviewing our work. As the deadline for the discussion period is approaching, we kindly request that you inform us of any remaining questions you may have.
We are confident that our response has adequately addressed your concerns, and we would be grateful for your feedback. Should you require further clarification, we would be delighted to answer any additional questions and provide more information.
Best wishes,
The Authors
Background about DFL:
- DFL is different from both CFL and decentralized/distributed training.
- VS. CFL. DFL discards the central server to avoid the communication congestion problem and privacy leakage risk of the central server. DFL also has some unique problems: (i) severe model inconsistency problem is caused by locality aggregation without a central server, which may cause severe over-fitting; (ii)the network connection topology has a great impact on model training, especially on heterogeneous data or sparse communication; (iii)the analysis of the specific impact of the topology (measured by the spectral gap 1- of gossip matrix ) is more complex than CFL.
- VS. decentralized/distributed training. One-step local iteration is used in decentralized/distributed training. But in general FL, clients adopt multi-step local iterations to relieve the communication burden. The technical difficulty is that local updates fail to be an unbiased gradient estimation after multiple local iterates, thereby merging the multiple local iterations is non-trivial.
- The advantage of DFL:
- Enhances decentralization by facilitating model aggregation at multiple nodes, drastically reducing dependence on a single central server and enabling direct, pairwise model sharing;
- Employs asynchronous communications, which contributes to system resilience and ensures continued learning even if encountering nodes' delays or disconnections;
- Local information transmissions allow for feeding the models of adjacent clients, generating a more intelligent private collaborative model.
- Challenges for Decentralized PFL(DPFL):
- Effect of network topology on decentralized SGD;
- The theoretical analysis of schemes that perform several local updates than those using a single SGD step;
- Overfitting and insufficient generalization of the model.
Motivations:
- Based on the advantages above, the study of PFL on DFL is meaningful.
- We reckon training and aggregation with a full model in most of the existing DPFL works lead to over-fitting in local rounds but “catastrophic forgetting" in global aggregation, which is one of the most important challenges in DPFL under data heterogeneity. Based on the works in model decoupling and partial model personalization, we consider that partial model personalized can alleviate the over-fitting problem due to the inexistence of the central server and make it outperform the centralized PFL methods under the same limited communication bandwidth.
Contributions:
Although both model decoupling for centralized PFL and SAM for DFL have been studied independently, the contributions in our work are non-trivial and beneficial to the PFL community. In this paper, we solved the following problems and proposed an insight into improving the partial model personalization :
- We first analyze that the full model training and aggregation may lead to over-fitting in local rounds but “catastrophic forgetting" in global aggregation in DPFL. Then we propose the first extensible decentralized partial personalization framework DFedMDC, which keeps the linear classifiers locally and exchanges the shared representation in a peer-to-peer manner. It can alleviate the over-fitting problem due to the inexistence of the central server, and achieve better convergence and performance than the centralized PFL methods under the same limited communication bandwidth.
- We further explore how to improve the partial personalization in DPFL and proposed DFedSMDC, which adds proper perturbation in the gradient direction to alleviate the shared model inconsistency across clients. It demonstrates that enhancing the consistency and generalization of the shared representation parameters and keeping the classifier fitting the local distribution can both improve the personalized performance in DPFL. The insight inspires us to improve the model decoupling methods for PFL no matter in CFL or DFL via improving the shared model consistency, enhancing the personalized classifiers' adaptability to local distribution, and increasing the alignment between them. The insight and the try are non-trivial.
- We give the first convergence analysis of the partial model personalization in DPFL and carefully consider the ill impact of the statistical heterogeneity , the smoothness of loss functions and the communication topology . This is the first detailed exploration of partial personalization in DPFL, which fully considers the special convergence problems in DFL. The analysis contribution is non-trivial.
- We conduct extensive experiments to prove the effectiveness of DFedMDC and DFedSMDC under different data heterogeneity and different communication topologies, which is tightly associated with our convergence analysis.
Three out of four reviewers recommended rejection, one recommended acceptance. The reviewers did not change their mind after rebuttal and discussion. I have read all reviews and discussion. Moreover, I have read the paper myself. Here are some remarks:
- The authors claim that "Theoretically, we give the first convergence analysis of the partial model personalization in DPFL, which achieves the best convergence speed in DPFL among the SOTA decentralized methods. Moreover, we first reveal the relationship between the personalized convergence and decentralized communication topology, which carefully considers the special convergence problems in DFL and it will be the basis for convergence analysis in DPFL. The convergence analysis contribution is non-trivial." However, decentralized PFL reduces to centralized (= distributed but not decentralized) PFL for a fully connected network, and I do not see any comparison of the new results obtained in this paper and prior theoretical results in the distributed setting. I believe this is absolutely necessary since without such a comparison, I find it impossible/hard to appreciate these more general results. Indeed, generalized results need to be compared and appreciated in the context of the results they are generalizing. The authors already cited some works on partial model personalization (e.g., FedSim and FedAlt). These works seem to be missing though:
- Hanzely et al, Federated learning of a mixture of global and local models, arXiv:2002.05516
- Mishchenko et al, Partially Personalized Federated Learning: Breaking the Curse of Data Heterogeneity, https://arxiv.org/abs/2305.18285
- Hanzely et al, Lower bounds and optimal algorithms for personalized federated learning, NeurIPS 2020
- Hanzely et al, Personalized federated learning: A unified framework and universal optimization techniques, https://arxiv.org/abs/2102.09743
How do your results compare to these works? In particular, the last work provides minimax optimal rates for CPFL. In particular, do your results recover these results as a special case? I believe this is not the case, and this points to a problem with the current work. In other words, I find it hard to appreciate a generalization which is not tight (unless the authors give very clear and convincing reasons for why tightness can't be obtained).
-
The obtained results give rate even in the noise-less regime, i.e., when . However, in this regime, it is possible to obtain rate for decentralized non-personalized gradient-based methods. For example, see Zhao et al, BEER: Fast O(1/T ) Rate for Decentralized Nonconvex Optimization with Communication Compression, NeurIPS 2022. Since local steps (use in the reviewed paper) are an alternative communication saving strategy to communication compression (used in BEER), it makes sense to check whether the proposed method can beat BEER in communication efficiency. It seems it can't. Moreover, the complexity should improve with personalized FL. This is not the case here. Again, I find it hard to appreciate these results.
-
Assumption 2 is too strong. For example, it is not necessarily satisfied when subsampling, which is the key algorithmic way of obtaining stochastic gradients. Due to these reasons, more modern works do not rely on this assumption. Example: Hanzely et al, Personalized federated learning: A unified framework and universal optimization techniques, https://arxiv.org/abs/2102.09743 (see their Assumption 5). A reference focusing on explaining this phenomenon: Khaled and Richtarik, Better theory for SGD in the nonconvex world, TMLR 2022. It seems the only mechanism guaranteeing Assumption 2 holds for stochastic gradients is this: compute the gradient first, and then add zero mean bounded variance noise. Obviously, this kind of a model is too simplistic and does not capture the way stochasticity appears in applications.
-
I also believe Assumption 3 is too strong, despite the fact that the authors say "The above assumptions are mild and commonly used in the convergence analysis of FL". Indeed, many works use such assumptions. However, this does not make them weak. The reason why this assumption is strong is this: it is possible to obtain better rates than the ones you obtain in your work even without this assumption (just like gradient descent does not need it). The state-of-the-art theoretical results in CFL and CPFL do not use this assumption.
I could add a few other observations made by the reviewers to this list.
In summary, I do not believe the results are interesting / strong enough to be published.
为何不给更高分
Due to the issues I mentioned above, the paper can't be accepted, so I can't suggest a higher score than a "reject".
为何不给更低分
N/A
Reject