PaperHub
6.8
/10
Rejected4 位审稿人
最低3最高6标准差1.3
5
3
6
3
4.5
置信度
创新性3.0
质量3.0
清晰度2.8
重要性3.0
NeurIPS 2025

A Stochastic Approximation Approach for Efficient Decentralized Optimization on Random Networks

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29
TL;DR

Primal-dual decentralized algorithms incorporating random graph communication through a stochastic Lagrangian formulation.

摘要

A challenging problem in decentralized optimization is to develop algorithms with fast convergence on random and time varying topologies under unreliable and bandwidth-constrained communication network. This paper studies a stochastic approximation approach with a Fully Stochastic Primal Dual Algorithm (FSPDA) framework. Our framework relies on a novel observation that the randomness in time varying topology can be incorporated in a stochastic augmented Lagrangian formulation, whose expected value admits saddle points that coincide with stationary solutions of the decentralized optimization problem. With the FSPDA framework, we develop two new algorithms supporting efficient sparsified communication on random time varying topologies --- FSPDA-SA allows agents to execute multiple local gradient steps depending on the time varying topology to accelerate convergence, and FSPDA-STORM further incorporates a variance reduction step to improve sample complexity. For problems with smooth (possibly non-convex) objective function, within $T$ iterations, we show that FSPDA-SA (resp. FSPDA-STORM) finds an $\mathcal{O}( 1/\sqrt{T} )$-stationary (resp. $\mathcal{O}( 1/T^{2/3} )$) solution. Numerical experiments show the benefits of the FSPDA algorithms.
关键词
Decentralized OptimizationTime Varying GraphRandom NetworkPrimal-dual LagrangianStochastic Approximation

评审与讨论

审稿意见
5

This submission is concerned with the problem of the distributed optimization over time-varying nondirected communication graphs of a smooth (possibly nonconvex) function given as a sum F(x)=1ni=1nfi(x)F(x) = \frac{1}{n} \sum_{i=1}^{n} f _i(x), where and each fi:RdRf _i : \mathbb{R}^d \mapsto \mathbb{R} is exclusively available to agent ii though noisy measurements of its gradient (i=1,,ni=1,\dots,n). The problem is reformulated as the constrained optimization problem

F(x)=1ni=1nfi(xi)F(x) = \frac{1}{n} \sum_{i=1}^{n} f _i(\mathbf{x} _i) subject to x1==xn\mathbf{x} _1 = \cdots = \mathbf{x} _n,

and two algorithms are proposed: FSPDA-SA and FSPDA-STORM. Both algorithms use stochastic approximation (SA) (in primal-dual form) to reach a saddle point x=(x1,,xn)\mathbf{x}^\ast = (\mathbf{x}^\ast _1, \dots , \mathbf{x}^\ast _n) of a (stochastic) augmented Lagrangian for the constrained problem, which implies finding a stationary point xx^\ast of the initial problem with consensus on the individual variables x1==xn=x\mathbf{x}^\ast _1= \cdots = \mathbf{x}^\ast _n = x^\ast. The two algorithms can be implemented asynchronously for the agents, and FSPDA-STORM has the specificity that momentum is used in order to reduce the variance due to inhomogeneity in the local objective functions, measurement noise and changing graph topologies. Analysis shows that FSPDA-SA matches existing methods in terms of convergence rate (O(1/T)O(1/\sqrt{T})), while offering more flexibility in implementation, and FSPDA-STORM reaches the faster rate O(1/T2/3)O(1/T^{2/3}). Numerical experiments are provided for benchmark problems.

优缺点分析

I thought the submission was interesting and well written. In my view (and in the light of Table 1), the technical contributions are significant. The discussions are also rich and useful (model and assumptions, interpretation of the results, comparison with other methods, proof sketches, etc.), which makes the paper accessible to non-experts on the topic. The overall organization of the paper is satisfying, as the main text summarizes well the many pages of computations that were relegated to the appendices. A strong point point of the submission is the extensive tests on several benchmark problems and from many viewpoints (convergence, communication costs, impact of noise, graph sparsity, asynchrony, the acceleration technique, etc.).

I was not able to proofread all the derivations in equal detail, but the appendices I did review were carefully written and technically sound. Most of the developments rely on common concepts from graph theory and function analysis, and on inequalities and techniques typically used in convergence analysis.

My comments below are mostly concerned with the presentation. I saw a few typos but no real weakness in the submission.

问题

  1. I was confused by Assumptions 3.3 and 3.4. As the assumptions are stated, it looks like one can find constants σi(xi)\sigma _i(\mathbf{x} _i) and σA(x)\sigma _A(\mathbf{x}) for every new xi\mathbf{x} _i and x\mathbf{x}. Are these not uniform bounds for all xi\mathbf{x} _i, x\mathbf{x} (as one would imagine and as apparent from the developments)? One way to reformulate the assumptions then would for instance be "For i[n]i\in [n], there exists σI0\sigma _I \geq 0 such that \dots holds for all xiRd\mathbf{x} _i \in \mathbb{R}^d \dots". Could the authors please clarify the statements?

  2. In (6), (8b), I believe there is a sign issue and the variables xj\mathbf{x} _j and xi\mathbf{x} _i should be permutated. Could the authors please double-check these equations?

  3. In (10c) and (10d), I think three ±\pm signs should be \mp here too.

  4. In Corollary 3.7, it would be useful to give the expression for FtF _t (even with unspecified parameters), which I believe is only available in the appendix. And in (20), what is Et\mathbb{E} _t? Is it the same conditional expectation as E[ξ:t]\mathbb{E} [ \cdot | \xi^{:t}] in Section 3.1? I do not remember these notations being introduced anywhere in the paper.

  5. On Line 211, how did you choose the value for the stepsize α\alpha in (20)? I was unable to reproduce its derivation based on Theorems 3.5 and C.5. And what is cc on Line 212? Is is the same parameter as c\mathbf{c} in (45)?

  6. In Algorithm 3 (asynchronous implementation), I understood c^i\hat{c} _i as an estimate of the relative frequency of gradient updates at node ii. Hence my intuition was to find the factors 1/c^i1/\hat{c} _i (in place of c^i\hat{c} _i) in (32) and (35). Could the authors please explain the c^i\hat{c} _i factors in the two equations? Also, does this technique rest on the assumption of stationary update frequencies for all agents?

  7. In (58), I got lost at the time of deriving the last two terms. The reader might welcome a word of explanation here. Did you use Hölder's inequality and the property that KK=K\mathbf{K}\mathbf{K}=\mathbf{K}?

Minor comments:

On Line 80: time varying topology -> time-varying topology

On Line 107: Ni(ξ)\mathcal{N} _i(\xi) the neighborhood of the iith agent as -> Ni(ξ)\mathcal{N} _i(\xi), the neighborhood of the iith agent, as

In (29), (30): II -> IdI_d

On Line 750: The neighbor jj whose communicated with -> The neighbor jj who has communicated with

局限性

yes

最终评判理由

The proposed FSPDA algorithm is an extension to time-varying networks of the EXTRA algorithm for nonconvex optimization, where the linear constraint enforcing consensus from which the augmented Lagrangian is derived now has the form of an expectation. The specific requirements of the algorithm are that the graph must be connected only in expectation (not at each iteration), and no bound is needed on the heterogeneity of the local objective functions.

In my sense, the proposed algorithm and the various techniques that are used are not particularly innovative, but the analysis looks technically sound to me, and it is original for nonconvex optimization in the considered time-varying setting (nondirected time-varying graphs that are only connected in expectation) and without more stringent assumptions. In particular, no bounds on heterogeneity are required in the analysis, which is a valuable feature of FSPDA, compared to the methods that rely on such bounds and tend to slow down in the case of highly heterogeneous local objective functions (as apparent from the numerical experiments in Appendix E.1).

格式问题

I did not detect formatting issues.

作者回复

Thank you for your careful review for our paper and constructive comments. Please find our point-to-point responses below.

  1. I was confused by Assumptions 3.3 and 3.4. As the assumptions are stated, it looks like one can find constants σi(xi)\sigma_i(\mathbf{x}_i) and σA(x)\sigma_A(\mathbf{x}) for every new xi\mathbf{x}_i and x\mathbf{x}. Are these not uniform bounds for all xi\mathbf{x}_i, x\mathbf{x} (as one would imagine and as apparent from the developments)? One way to reformulate the assumptions then would for instance be "For i[n]i\in [n], there exists σI0\sigma_I \ge 0 such that ... holds for all xiRd\mathbf{x}_i \in \mathbb{R}^d". Could the authors please clarify the statements?

The statements were indeed confusing. We will update the statements according to your suggestions.

  1. In (6), (8b), I believe there is a sign issue and the variables xj\mathbf{x}_j and xi\mathbf{x}_i should be permutated. Could the authors please double-check these equations?
  1. In (10c) and (10d), I think three ±\pm signs should be \mp here too.

We apologize for the typos in (6), where the correct sign should be permuted as (xixj)(\mathbf{x}_i - \mathbf{x}_j). Similarly, (8) should be (xixj)(\mathbf{x}_i - \mathbf{x}_j), (10c) is correct in the submission, (10d) should be (xixj)(\mathbf{x}_i - \mathbf{x}_j). We will implement these corrections in the revision.

  1. In Corollary 3.7, it would be useful to give the expression for FtF_t (even with unspecified parameters), which I believe is only available in the appendix. And in (20), what is Et\mathbb{E}_t? Is it the same conditional expectation as E[ξ:t]\mathbb{E}[\cdot | \xi^{:t}] in Section 3.1? I do not remember these notations being introduced anywhere in the paper.

Thank you for the suggestion, we will include the expression for FtF_t when extra space is allotted in the camera-ready version. Et\mathbb{E}_t in (20) is the conditional expectation conditioned on ξt\xi^t.

  1. On Line 211, how did you choose the value for the stepsize in (20)? I was unable to reproduce its derivation based on Theorems 3.5 and C.5. And what is on Line 212? Is is the same parameter as c\mathbf{c} in (45)?

The choice of α\alpha in (21) is due to the result of Appendix C.6, where we can telescope the inequality (83) and upper bound the geometric sum of k=0t(1δ)k1/(1(1δ))=1/δ\sum_{k=0}^{t}(1-\delta)^{k} \leq 1/(1-(1-\delta)) = 1/\delta where δ=O(α)\delta = O(\alpha) due to the condition in line 886. We then obtain FTf=O((1δ)T(F0f)+Cσα2σˉ2/δ)F_{T} - f_\star = O((1-\delta)^T(F_0 - f_\star) + \mathbb{C}_{\sigma} \alpha^2 \bar{\sigma}^2 / \delta) and δ=O(α)=O(ln(T)/(n2T))\delta = O(\alpha) = O(\ln(T)/(n^2T)) is the optimal choice for the above bound.

The constant cc represents the omitted constants in the above big-O\mathcal{O} notations and is different from c\mathtt{c} in (45). For clarity, we will rewrite line 211 as "By setting α=Θ(ln(T)/(n2T))\alpha = \Theta( \ln(T) / (n^2 T)) in (20), ..." in the revision.

  1. In Algorithm 3 (asynchronous implementation), I understood c^i\hat{c}_i as an estimate of the relative frequency of gradient updates at node ii. Hence my intuition was to find the factors 1/c^i1/\hat{c}_i (in place of c^i\hat{c}_i) in (32) and (35). Could the authors please explain the c^i\hat{c}_i factors in the two equations? Also, does this technique rest on the assumption of stationary update frequencies for all agents?

Thank you for the careful observation. You are right that the factors regarding c^i\hat{c}_i should be inversed. c^i\hat{c}_i is introduced to maintain the unbiasedness of stochastic gradient. We will correct the typo by rewrting (34) as c^i=(ti+1)/gi\hat{c}_i = (t_i' + 1) / g_i if fi()\nabla f_i(\cdots) is ready. The same applies to (32). This unbiased scaling does depend on the assumption that the update frequency is stationary, i.e., c^iD_ci\hat{c}_i \sim \mathcal{D}\_{c_i} for a fixed distribution D_ci\mathcal{D}\_{c_i} for the sake of analysis.

  1. In (58), I got lost at the time of deriving the last two terms. The reader might welcome a word of explanation here. Did you use Hölder's inequality and the property that KK=K\mathbf{K}\mathbf{K}=\mathbf{K}?

You are right. The last two terms in (58) are derived from applying Young's inequality and using the fact KK=KK=K{\bf K}^\top {\bf K} = {\bf K} {\bf K} = {\bf K}. Particularly, for any α>0\alpha>0, we observe that

2(IγL)xtegt_K2 \braket{(\mathbf{I} - \gamma \mathbf{L})\mathbf{x}^t| \mathbf{e}^t_g}\_{{\bf K}}

=2K(IγL)xtKegt= 2 \braket{{\bf K} (\mathbf{I} - \gamma \mathbf{L})\mathbf{x}^t| {\bf K} \mathbf{e}^t_g}

α(IγL)xt_K2+1αegt_K2\leq \alpha || (\mathbf{I} - \gamma \mathbf{L})\mathbf{x}^t ||\_{\bf K}^2 + \frac{1}{\alpha} || \mathbf{e}^t_g ||\_{\bf K}^2

A list of typos suggested by the reviewer.

We will correct them in the revision. Thank you.

评论

Thank you very much for your detailed feedback. You have addressed all my questions and comments.

审稿意见
3

This paper proposes the FSPDA framework for decentralized optimization on random networks. While tackling time-varying topologies is relevant, the work suffers from fundamental flaws in novelty, technical rigor, and validation that preclude publication. The claims of outperforming existing methods are unsupported and theoretical analyses contain critical gaps.

优缺点分析

Strengths

  • Effectively addresses the critical gap in decentralized optimization for random time-varying topologies (§1.5), which is highly relevant to real-world dynamic networks (e.g., UAV swarms, edge computing).
  • Proposes a unified stochastic augmented Lagrangian framework (Eq. 4) that concurrently models:
  • Data randomness (ξi)\left(\xi_i\right)
  • Topology randomness (ξa)\left(\xi_a\right).

Weaknesses

  • vtv^t variable inconsistency: Definition vt=Aλt+αηf(1xˉt)v^t=A^{\top} \lambda^t+\frac{\alpha}{\eta} \nabla f\left(1_{\otimes} \bar{x}^t\right) (Eq. 30) requires global gradient knowledge, violating decentralization principles.
  • Unverifiable assumptions: σA2xK2\sigma_A^2 \propto\|x\|_K^2 (Eq. 15) lacks empirical/theoretical justification and fails for sparse graphs.
  • Infeasible parameter conditions: Step-size constraints (Eq. 46) become unsolvable when σA>1\sigma_A>1 (common in sparse networks).
  • Ambiguous convergence claims: O(1/nT)O(1 / \sqrt{n T}) rate (Eq. 19) hides σA4\sigma_A^4 dominance (p. 204)
  • Critical notation overload: L,vt,Cij(ξ)\mathrm{L}, v^t, C_{i j}(\xi) redefined without consolidation

问题

1.The paper positions FSPDA as the "first" solution for random topologies, but prior works (Kovalev et al., 2021; Li & Lin, 2024) explicitly handle time-varying graphs. The distinction between gradient tracking (Qu & Li, 2017) and primal-dual methods (Hajinezhad & Hong, 2019) is superficial.

2.Assumption 3.4 Unverifiable: The bound σA2xK2\sigma_A^2 \propto\|x\|_K^2 (Eq. 15) is non-standard and lacks proof. For sparse graphs, σA\sigma_A likely scales with 1/E1 / \sqrt{|\mathrm{E}|}, contradicting the analysis.

3.Parameter Infeasibility: Conditions (46) for Theorem 3.5 require δ18\delta_1 \geq 8 and γO(σA2)\gamma_{\infty} \sim \mathrm{O}\left(\sigma_A^{-2}\right). When σA\sigma_A is large (e.g., highly sparse graphs), γ\gamma_{\infty} becomes vanishingly small, stalling convergence. No feasible configuration is demonstrated.

4.Omitted Dependencies: The O(1/nT)\mathrm{O}(1 / \sqrt{n T}) rate (Eq. 19) misleadingly hides ρmin1\rho_{\min }^{-1} and σA4\sigma_A^4 terms (Eq. 204), voiding linear speedup guarantees.

  1. The emphasis on "fully stochastic" networks ignores that most real-world systems (e.g., federated learning) exhibit structured time-variation (e.g., periodic updates), handled better by local-GT.

局限性

  • Appendix A notes that asynchronous FSPDA has "suboptimal" convergence but provides no quantitative analysis (e.g., convergence rate degradation under delays).
  • §3.2 mentions σA\sigma_A (topology noise) without examining real-world feasibility (e.g., how σA\sigma_A scales with edge sparsity pp in dynamic networks).
  • Local updates are called "suboptimal" (§1.3) but lack comparison to SOTA (e.g., ProxSkip).

最终评判理由

Thank you very much for your detailed feedback. You have answered all my questions and opinions. I will improve my score.

格式问题

I didn't find any major formatting issues in this article.

作者回复

Thank you for the critical and constructive comments. We believe there are some misunderstandings by the reviewer, and wish to clarify them in the point-to-point responses below.

  1. The paper positions FSPDA as the "first" solution for random topologies, but prior works (Kovalev et al., 2021; Li & Lin, 2024) explicitly handle time-varying graphs. The distinction between gradient tracking (Qu & Li, 2017) and primal-dual methods (Hajinezhad & Hong, 2019) is superficial.

The novelty of FSPDA lies on being a "fast" decentralized algorithm for time-varying (TV) graphs under the stochastic non-convex optimization setting - as illustrated in Table 1 - note that by "fast" algorithm, we refer to ones that converge without restrictions such as bounded heterogeneity. As hinted by prior works on static graphs, such "fast" algorithm design can only be achieved using gradient tracking (GT) or primal-dual (PD) methods, note GT can be viewed as a special case of PD [Chang et al., 2020].

The analysis in [Kovalev et al., 2021, Li & Lin, 2024] (and [Nedic et al., 2017]) are applicable to GT on TV graphs, but are limited to strongly convex functions under the deterministic gradient setting. These are fundamentally different settings from our non-convex & stochastic cases as the iterates are expected to converge to a unique solution. Indeed, it was explicitly pointed out in [Koloskova et al., 2021] that analyzing the latter case with GT is highly non-trivial (in the non-convex setting) as it involves "the non-symmetric product of two mixing matrices", and therefore it only analyzed the simpler DSGD algorithm whose convergence are restricted by the bounded heterogeneity condition.

  1. Assumption 3.4 Unverifiable: The bound σA2xK2\sigma_A^2 \propto \|x\|_{K}^2 (Eq. 15) is non-standard and lacks proof. For sparse graphs, σA\sigma_A likely scales with 1/E1 / \sqrt{|E|}, contradicting the analysis.

§3.2 mentions σA\sigma_A (topology noise) without examining real-world feasibility (e.g., how σA\sigma_A scales with edge sparsity pp in dynamic networks).

We should have highlighted Assumption 3.4 as one of our technical novelty to overcome the challenges in analyzing "fast" algorithm over TV graphs - as the latter is applied in Lemma C.2 for controlling the consensus error recursion. As an illustration, the assumption can be verified easily by the crude bound below:

E[AA(ξa)xARAx2]\mathbb{E}[|| \mathbf{A}^\top \mathbf{A}(\xi_a) \mathbf{x} - \mathbf{A}^\top \mathbf{R} \mathbf{A} \mathbf{x} ||^2]

=E[AI_nd(ξa)AxARAx2]=\mathbb{E}[|| \mathbf{A}^\top {\bf I}\_{nd}(\xi_a) \mathbf{A} \mathbf{x} - \mathbf{A}^\top \mathbf{R} \mathbf{A} \mathbf{x} ||^2]

=E[(AI_nd(ξa)AR)(Ax)2]= \mathbb{E}[|| (\mathbf{A}^\top {\bf I}\_{nd}(\xi_a) - \mathbf{A}^\top \mathbf{R} ) ( \mathbf{A} \mathbf{x}) ||^2]

E[AI_nd(ξa)AR2]Ax2\leq \mathbb{E}[ || \mathbf{A}^\top {\bf I}\_{nd}(\xi_a)- \mathbf{A}^\top \mathbf{R} ||^2 ] \cdot || \mathbf{A} \mathbf{x} ||^2

E[I_nd(ξa)R2]A2ρˉmaxx2_K\leq \mathbb{E}[ || {\bf I}\_{nd}(\xi_a)- \mathbf{R} ||^2 ] \cdot || {\bf A} ||^2 \cdot \bar{\rho}_{\max} \cdot || \mathbf{x} ||^2\_{\mathbf{K}}

where R=E[Ind(ξa)]\mathbf{R} = \mathbb{E}[{\bf I}_{nd}(\xi_a)] and the last inequality is due to Assumption 3.2. This shows that

σA2E[I_nd(ξa)R2]ρˉmaxA2\sigma_A^2 \leq \mathbb{E}[ || {\bf I}\_{nd}(\xi_a)- \mathbf{R} ||^2 ] \cdot \bar{\rho}_{\max} \cdot || {\bf A} ||^2.

Note the above bound can be improved. In the case when a pp-sparse graph are used for the communication steps of FSPDA that selects an edge from G\mathcal{G} with probability pp, we have E[I_nd(ξa)R2](1p)2\mathbb{E}[ || {\bf I}\_{nd}(\xi_a)- \mathbf{R} ||^2 ] \leq (1-p)^2 and thus σA2=(1p)2ρˉmaxA2\sigma_A^2 = (1-p)^2 \bar{\rho}_{\max} || {\bf A} ||^2. Note one has A2O(E)|| {\bf A} ||^2 \leq O( \mathcal{E} ).

The reviewer is right that σA2\sigma_A^2 is likely to scale with E|\mathcal{E}| especially in the extreme case of p=1/Ep = 1/|\mathcal{E}|. However, we politely disagree that it contradicts with our analysis. In practice, the graph size is finite and so is E|\mathcal{E}|, and it remains feasible to find a stepsize configuration attaining fast convergence rate for FSPDA.

  1. Parameter Infeasibility: Conditions (46) for Theorem 3.5 require δ18\delta_1 \ge 8 and γO(σA2)\gamma_\infty \sim \mathcal{O}(\sigma_A^{-2}). When σA\sigma_A is large (e.g., highly sparse graphs), γ\gamma_{\infty} becomes vanishingly small, stalling convergence. No feasible configuration is demonstrated.

It remains feasible to satisfy (46) by picking α=Θ(1/T)\alpha = \Theta(1/\sqrt{T}) for large TT. A possible choice is to have γ=γ=O(ρmin/σA2)\gamma = \gamma_{\infty} = O(\rho_{\min}/\sigma_A^2), η=η=O(ρmin3/σA2)\eta = \eta_{\infty} = O(\rho_{\min}^3/\sigma_A^2). With α=Θ(1/T)\alpha = \Theta(1/\sqrt{T}) such that for large enough TT, it satisfies αα=O(ρmin6/σA4)\alpha \leq \alpha_{\infty} = O(\rho_{\min}^6/\sigma_A^4). The above configuration is feasible, as long as σA<\sigma_A < \infty is finite and independent of TT, where they guarantee the convergence of FSPDA-SA.

We understand the concern that the above choices may lead to small parameters when random sparse graph sampled from G\mathcal{G} with large E|\mathcal{E}| are used in FSPDA. Besides that they remain feasible for finite E|\mathcal{E}|, we wish to point out that the effects of σA,ρmin\sigma_A, \rho_{\min} vanishes as T1T \gg 1 (also see the clarification to Q4 below). It can be seen in Theorem C.5 (full ver of Theorem 3.5), where one has a=O(1/T),α=O(1/T)`a` = O(1/\sqrt{T}), \alpha = O(1/\sqrt{T}) and it will make the last terms in (50) vanish.

Additionally, we notice that such step size choices are common for the TV graph setting. For example, in Theorem 1 of [Li & Lin, 2024], the step size for GT is O(1/B4)O(1/B^4) for BB-connected TV graph. With extremely sparse TV graph, one has BEB \propto |\mathcal{E}|, their analysis also suggests a vanishingly small step size parameter.

Lastly, our experiment results demonstrate extremely sparsified cases where the random graphs are drawn as one-edge graphs in Fig 1, 2. In Fig 5 of Appendix E, different degrees of sparsification are also demonstrated. All of our experiment results demonstrate convergence in practical neural network training problems. The hyperparameter configurations of every experiments are provided in Appendix F.

  1. Omitted Dependencies: The O(1/nT\mathcal{O}(1/\sqrt{nT} rate (Eq. 19) misleadingly hides ρmin1\rho_{\min}^{-1} and σA4\sigma_A^4 terms (Eq. 204), voiding linear speedup guarantees.

We believe that the reviewer referred to L204 on p.6. The sentence actually discusses the convergence rate when the gradients used in FSPDA-SA are deterministic with σˉ=0\bar{\sigma} = 0, where it is irrelevant to the linear speedup rate in (19) as the latter is discussed for cases with stochastic gradient σˉ>0\bar{\sigma} > 0. As discussed in [Lian et al., 2017], such linear speedup behavior (such that the decentralized algorithm matches the performance of a centralized one) is only relevant when stochastic gradients are used.

  1. The emphasis on "fully stochastic" networks ignores that most real-world systems (e.g., federated learning) exhibit structured time-variation (e.g., periodic updates), handled better by local-GT.

Local updates are called "suboptimal" (§1.3) but lack comparison to SOTA (e.g., ProxSkip).

In contrary, it is a unique advantage of FSPDA to handle fully stochastic networks with unstructured time variation while performing local updates asynchronously - see Algo. 2 & 3 - note Algo. 2 & 3 also includes structured time variation update as special cases. In addition to offering strictly better flexibility, the achieved convergence rate remains comparable to those with specialized analysis in the dominant O(1/nT)O(1/\sqrt{nT}) term. We admit that our analysis, which is obtained without structured TV graph, maybe less tight than prior works such as [R1] and is thus said to be "suboptimal" - We apologize for the unclear wordings. Yet we found that the differences in rate mostly appear in the transient terms that vanish as T1T \gg 1, see Theorem C.5 & Theorem D.8.

Lastly, we refer the reviewer to the experiments in Fig 1. Here, we show that using an unstructured time variation updates, FSPDA is able to achieve comparable performance, or outperform algorithms utilizing specialized structured TV / periodic updates, e.g., DecenScaffnew/ProxSkip in [Mishchenko et al., 2022] & [R1].

[R1] L. Guo, S.A. Alghunaim, K. Yuan, L. Condat, J. Cao, Achieving Linear Speedup with ProxSkip in Distributed Stochastic Optimization, arxiv/2310.07983.

Appendix A notes that asynchronous FSPDA has "suboptimal" convergence but provides no quantitative analysis (e.g., convergence rate degradation under delays).

For the convergence rate under delays with stochastic gradients, it can be quantified as having an increased gradient variance σˉ\bar{\sigma} as discussed in Section 2.1. Similar to how we handle random graphs, our model of asynchrony is only in concern with the rate of successful access to stochastic gradient bˉi\bar{b}_i and the rate of edge activation σA,ρ_min\sigma_A, \rho\_{\min}, rather than the more restrictive deterministic delay bound that is commonly adopted in the existing asynchronous algorithms. In particular, by Algorithm 1 in Appendix A, each agent always apply the latest stochastic gradient (see (32), (35)), instead of stale gradient from past iterates. This distinguish our asynchronous algorithm from the existing delay-bounded algorithms that utilizes stale gradient updates.

We look forward to discussing with the reviewer if any further clarifications are needed.

评论

Dear Reviewer TNhj,

As the end of discussion period is approaching, we would like to see if our rebuttal has addressed your concern. We kindly remind the reviewer that you can follow up with new questions to our rebuttal, or raise the rating of our submission if you are satisfied with our response.

Best regards,

Authors of Paper 26853.

评论

Thank you very much for your detailed feedback. You have answered all my questions and opinions. I will improve my score.

审稿意见
6

The authors present a stochastic, primal-dual approximation algorithm that can be used with random, time-varying networks under unreliable communication bandwidth. The randomness in time-varying networks is formulated as a stochastic augmented Lagrangian function whose expectation yields saddle points that coincide with stationary solutions of the decentralized optimization. Two variants of the proposed algorithms developed from this framework are (1) FSPDA -- which allows agents to execute multiple local gradient steps depending on the time-varying topology, and (2) FSPDA-STORM, which incorporates variance reduction steps to improve sample complexity. Empirical results are presented on the MNIST and ImageNet datasets using an extreme setting for the realization of the time-varying topology; baselines are presented with SOTA algorithms such as CHOCO SGD, Swarm-SGD and others.

优缺点分析

The paper is very well written and addresses a novel problem -- that of achieving fast convergence on time-varying random networks. Prior related work in the domain of interest has been discussed clearly.

The problem is formulated as a stochastic augmented Lagrangian function; however, extension to time-varying random networks is not straightforward due to inconsistency of dual variables. This challenge is overcome by proposing a stochastic equality constrained reformulation, and two related algorithms FSPDA and FSPDA-STORM are presented.

The implementation of 8a and 8b in Appendix A is quite important to understand the flow of the algorithm FSPDA. Perhaps it would be useful to consider having this description in the main paper at the cost of the some convergence analysis proofs.

Empirical analysis -- while the extreme setting of only one edge being selected at random is interesting, the effect of the convergence of the algorithm on larger random networks would need to be studied. For example, in MNIST, the number of agents = 10, and in ResNet, it is 8. These are reasonably small to assess the overall performance of the algorithm. For instance, what is the relationship between the dynamic nature of the graph topology, the size of the graph, the communication cost incurred on it, and local computation time? Do larger graphs experience bottlenecks that affect eventual convergence?

问题

Empirical analysis is performed on small, complete random graph topologies. What is the relationship between the dynamic nature of the graph topology, the size of the graph, the communication cost incurred on it, and local computation time? Do larger graphs experience bottlenecks that affect eventual convergence?

局限性

NA

最终评判理由

I have read the rebuttal of the authors, including extra empirical results presented. I have also carefully read the discussions and think my score remains justified.

格式问题

NA

作者回复

Thank you for your careful review for our paper and constructive comments. Please find our point-to-point responses below.

The implementation of 8a and 8b in Appendix A is quite important to understand the flow of the algorithm FSPDA. Perhaps it would be useful to consider having this description in the main paper at the cost of the some convergence analysis proofs.

Subject to page limitation, we will try to incorporate the asynchronous implementation in the main text to provide a self-contained discussions in the paper.

Empirical analysis -- while the extreme setting of only one edge being selected at random is interesting, the effect of the convergence of the algorithm on larger random networks would need to be studied. For example, in MNIST, the number of agents = 10, and in ResNet, it is 8. These are reasonably small to assess the overall performance of the algorithm. For instance, what is the relationship between the dynamic nature of the graph topology, the size of the graph, the communication cost incurred on it, and local computation time? Do larger graphs experience bottlenecks that affect eventual convergence?

Empirical analysis is performed on small, complete random graph topologies. What is the relationship between the dynamic nature of the graph topology, the size of the graph, the communication cost incurred on it, and local computation time? Do larger graphs experience bottlenecks that affect eventual convergence?

Thanks for the suggestions. Due to limited time and compute resources, we are only able to extend the experiments under the setting of Fig. 1 to test the effect of graph sizes on convergence, and verifying the linear speedup effects in (19). Instead of complete graphs, we also use ER graphs (with connectivity p=0.5) in these extended experiments.

From our theory, larger graphs actually tend to improve convergence due to the linear speedup effects as shown in (19), albeit such effect may only become dominant as t1t \gg 1. To test it, we fix the batch size of each agent so that the local computation time is equivalent across different graph size. We further fix the communication cost of each edge across n=10,20,40n=10,20,40 by keeping the edge selection probabilities constant. To verify the linear speedup effects, we terminate the FSPDA-SA algorithm after the same number of training samples across agents are used:

Iterationt=1.25e6t=2.5e6t=5e6t=1e7
n=10n=105.01e-52.91e-52.61e-52.58e-5
n=20n=206.74e-52.88e-52.37e-5
n=40n=404.24e-53.58e-5

Table A. Ablation study on the effect of graph size nn. Each run uses 1% coordinate sparsity and random graphs of edge probability 0.0222 (n=20,40)(n=20, 40) or 1-edge random graphs (n=10)(n=10). The table records maxi[n]F(xit)2\max_{i\in[n]} ||\nabla F(\mathbf{x}_i^t) ||^2.

We observe that similar solution stationarity is achieved for nT=1e8nT = 1e8 (i.e., the highlighted entries) regardless of nn. This confirms with the linear speedup observation in (19) as the nTnT factor dominates the error of F(x)2||\nabla F(\mathbf{x})||^2. Note that the solution stationarity with n=40n=40 is slightly higher as expected from (37) that some higher order error may still remain in the bound of F(x)2||\nabla F(\mathbf{x})||^2.

审稿意见
3

This paper studies the decentralized optimization problem over unreliable and time-varying networks. It introduces a new stochastic primal-dual framework (FSPDA) that handles communication randomness through a stochastic augmented Lagrangian. Two algorithms are proposed, FSPDA-SA algorithm that adapts gradient updates to network conditions and FSPDA-STORM algorithm that adds variance reduction for better efficiency. Theoretical results show convergence rates of O(1/T)O(1/\sqrt{T}) for FSPDA-SA algorithm and O(1/T2/3)O(1/T^{2/3}) for FSPDA-STORM algorithm, and experiments demonstrate their practical effectiveness.

优缺点分析

Strengths: The main value in this work lies in the observation that randomness in time varying topolgy can be incorporated in a stochastic augmented Lagrangian formulation, which is an extension based on [Chang et al., 2020].

Weaknesses:

  1. The problem setting addressed in the paper is somewhat limited and less aligned with current research interests. The work could be strengthened by extending the analysis to directed or B-connected time-varying networks, which are more general and practically relevant. As it stands, focusing solely on stochastic optimization over undirected time-varying networks narrows its applicability.

  2. The proposed algorithms, FSPDA-SA and FSPDA-STORM, appear to be straightforward extensions of existing decentralized stochastic methods, incorporating SA and STORM components. This may be perceived as an incremental or relatively minor modification.

  3. The convergence rates derived in the paper align with what is typically expected under the standard assumptions used in the paper and do not represent a significant theoretical advancement.

Overall, the work lacks substantial novelty and does not demonstrate particularly strong contributions relative to the current state of the field.

问题

  1. Could the authors clarify the benefits of formulating an augmented Lagrangian in the context of decentralized optimization? What advantages does this offer over standard formulations?

  2. Are there any unique techniques employed in the convergence analysis for time-varying networks, as compared to the analysis typically used for static network settings?

  3. Could the authors provide a more detailed discussion on how sparsification impacts the overall convergence rate? The experimental results suggest that increased sparsification may lead to slower and less stable convergence.

局限性

Please see above.

最终评判理由

I have read all discussions and still hold concerns on the novelty and technical contributions in the work (please see the discussion with other reviewers for more details) so decided not to raise the score.

格式问题

N/A

作者回复

Thank you for the critical and constructive comments. We wish to emphasize that the FSPDA algorithms work for a general setting of practical interests, and is accompanied with a novel analysis as well as potential to extend further. Please find our responses below.

The problem setting addressed in the paper is somewhat limited and less aligned with current research interests. The work could be strengthened by extending the analysis to directed or B-connected time-varying networks, which are more general and practically relevant. As it stands, focusing solely on stochastic optimization over undirected time-varying networks narrows its applicability. The proposed algorithms, FSPDA-SA and FSPDA-STORM, appear to be straightforward extensions of existing decentralized stochastic methods, incorporating SA and STORM components. This may be perceived as an incremental or relatively minor modification.

Compared to the settings suggested by the reviewer, we believe that our problem setting on random time varying (TV) graphs is not less practical and possibly more relevant to today's challenges. First, to clarify, our setting (also studied in [Lobel & Ozdaglar, 2010]) do not require each of the TV graphs G(ξat),t0{\cal G}(\xi_a^t), t \geq 0 to be connected - in the extreme case agents only need to communicate on a 1-edge random graph at each iteration. The graphs are only required to be connected in expectation (Assumption 3.2), which is similar to the B-connected TV graphs setting suggested by the reviewer. On top of this, we allow for random sparsification for further efficient communication and demonstrates that FSPDA-SA/STORM can be implemented as asynchronous algorithms. To our knowledge, the latter extensions may not be straightforward in the B-connected TV graph settings. We note that algorithms of directed graph requires very different strategies (e.g., push-pull techniques) and it remains an open problem for the analysis on non-convex optimization with PD-based algorithm under TV graph.

In terms of algorithms development, both FSPDA-SA/STORM can be understood as "plug-n-play" derivatives of SA for finding the saddle point(s) of (4), obtained from analyzing the stochastic linear constrained version of the distributed learning problem (1). Together with the stochastic augmented Lagrangian function of the latter, these are novel perspectives offered by our paper and they naturally lead to decentralized algorithms on random TV graphs with random sparsification. In fact, we believe that the FSPDA framework based on (4) can provide a general recipe for designing more powerful algorithms than FSPDA-SA/STORM. The analysis for these algorithms also prompted us to design new proof techniques.

The convergence rates derived in the paper align with what is typically expected under the standard assumptions used in the paper and do not represent a significant theoretical advancement.

While we agree that the rates are as expected, as depicted in Table 1, we emphasize that our paper is the first one to achieve these rates under the settings of TV graphs without relying on the common bounded heterogeneity (BH) assumption, which has caused slow convergence in practice for certain prior works such as DSGD. Although it appears to be unsurprising, this is a crucial step in extending decentralized optimization towards practical scenarios.

Could the authors clarify the benefits of formulating an augmented Lagrangian in the context of decentralized optimization? What advantages does this offer over standard formulations?

Our augmented Lagrangian formulation incorporates a novel stochastic equality constraint to enforce/promote consensus. As demonstrated in the discussions after (4), it enables one to derive algorithms that naturally supports communication on random TV graphs with sparsification. Another advantage is that the approach naturally gives rise (via stochastic gradient descent/ascent) to stable algorithms with strong convergence guarantees, e.g., without the bounded heterogeneity assumption, linear convergence under PL condition with deterministic gradient, etc.

On the other hand, algorithms derived under the FSPDA framework can be seen as TV graph extensions of gradient tracking, EXTRA, etc., see Sec. 2.1 and [Chang et al., 2020]. Note it has been pointed out that the analysis of TV graph GT algorithm is not straight-forward as it involves the communication of two different variables and therefore the non-symmetric product of two mixing matrices [Koloskova et al., 2021]. This brings greater flexibility and stronger guarantees than algorithms from standard approach such as the primal-only DSGD algorithms.

Are there any unique techniques employed in the convergence analysis for time-varying networks, as compared to the analysis typically used for static network settings?

There are several unique techniques deployed in our analysis due to the non-triviality in handling TV graphs. In particular, instead of the natural way (as in [Hong et al., 2017]) of designing a Lyapunov function based on the expected Lagrangian, we view the convergence of FSPDA(-SA/STORM) as that of stochastic approximation scheme on a Lyapunov function that is dependent on the global objective F(x)F(x) and the system error quantities such as consensus error and dual gradient tracking error, see Section 3.1. As shown in Appendix C, we control/introduce the error terms in a step-by-step fashion. The application of Assumption 3.4 is particularly unique in the proof of Lemma C.2. This assumption treats random graphs as stochastic estimation to the graph Laplacian operator AAx\mathbf{A}^\top \mathbf{A} \mathbf{x} and enables us to relate the random graph error to a quantity proportional to the consensus error xK2|| \mathbf{x} ||_{\mathbf{K}}^2.

Could the authors provide a more detailed discussion on how sparsification impacts the overall convergence rate? The experimental results suggest that increased sparsification may lead to slower and less stable convergence.

The reviewer is correct that sparsification will lead to slower convergence and less stable convergence. This is an inevitable tradeoff. Our analysis, however, can shed light on what is the right degree of sparsification to be applied. To set this up, we examine the quantities that are dependent on the degree of sparsification. They are ρmin\rho_{\min} in (13) and σA\sigma_A in (15). In particular, when the degree of sparsification increases which reduces the magnitude R{\bf R}, it is easy to see that ρmin\rho_{\min} decreases and σA2\sigma_A^2 increases.

We focus on FSPDA-SA for illustration purpose. From Theorem 3.5/Theorem C.5 and the discussions that follow, we see that the effects of ρmin,σA\rho_{\min}, \sigma_A vanish as TT \to \infty. In other words, the level of sparsification only affect the performance in the transient stages (when TT is not large). At a closer examination, in (50), we see that the transient terms depending on ρmin,σA\rho_{\min}, \sigma_A get multiplied by L2L^2. From here, we conclude that when LL is large, i.e., when the problem is less smooth (e.g., it involves training an NN with large no of layers), choosing a high sparsification level can be detrimental. We will include a discussion on the above matters in the final version.

We look forward to discussing with the reviewer if any further clarifications are needed during the discussion period.

评论

We thank the reviewer for the continued discussions and appreciate the additional references. We will include them in the revision.

  1. Regarding the connectivity assumption, it is typically hard for distributed optimization algorithm to derive the average consensus in the analysis, especially for time-varying graphs. However, from Assumption 2 it seems that the derivation of average consensus will be trivial and this could diminish the analytical challenge that typically arises in decentralized optimization over time-varying graphs.

We point out that attaining average consensus alone over time varying (TV) undirected graphs is a classical problem that has been extensively studied since 2006 [R1]. Their gossiping condition is a special case of our Assumption 3.2, and a recent treatment focusing on decentralized optimization can be found in [Koloskova et al., 2020]. Notice that to guarantee convergence of average consensus, one also needs to control the second-order moment of the TV gossiping matrices, e.g., the 1-edge activation assumption in [R1], Assumption 3.4 in FSPDA, etc.

The challenge in decentralized optimization on TV graphs, however, lies on showing that the resultant algorithm converges for under desired general conditions. While the analysis extends easily for simple algorithms like DSGD [Koloskova et al., 2020] and for GT based algorithms under strongly convex objectives [Nedic et al., 2017, Li & Lin, 2024], it is noted by [Koloskova et al., 2021] that extending the analysis for GT in general (non-convex) setting is challenging due to the non-symmetric matrix products. We remark that with strongly convex objective, the contraction of iterates towards the unique optimal solution allows to simplify the analysis. One of the FSPDA's innovations is that the "average consensus" step is not explicit, rather the latter is induced as a part of the update direction in the stochastic approximation scheme seeking a saddle point of (4).

[R1] Boyd et al., Randomized Gossip Algorithms, IEEE TIT, 2006.

  1. “On top of this, we allow for random sparsification for further efficient communication” - I appreciate the authors' discussion on the incorporation of random sparsification for communication efficiency. I would like to point out that similar ideas have been explored in [Communication-efficient variance-reduced decentralized stochastic optimization over time-varying directed graphs." IEEE Transactions on Automatic Control 67.12 (2021)], which also addresses communication reduction via sparsification in a time-varying and directed setting. It would strengthen the contribution of this paper if the authors could elaborate on what new theoretical insights or practical behaviors emerge in their current analysis that go beyond existing results.

The mentioned paper indeed uses a similar technique to ours for random sparsification, thanks for letting us know! However, it is noted that they only consider the finite sum & strongly convex objective function setting, which is different from the general stochastic & smooth (possibly non-convex) objective function in FSPDA.

It is important to point out that the generality of our settings is due to the novel modeling of consensus constraint as a stochastic linear equality in (3), (4), where randomness in sparsification & TV graphs are directly embedded into the latter. Plugging in the SA scheme naturally leads to the FSPDA algorithms achieving the desired setting, and the control variate based adaptive variance reduction is easily included, leading to FSPDA-STORM. Without the observations in (3), (4), it may require a number of trial-and-error in constructing an algorithm with similar guarantees as FSPDA. It is unclear whether a similar guarantee achieved by FSPDA in Theorem 3.5 can be provided if one develops an algorithm from the GT-like framework in the mentioned paper, e.g., reference a) in Q3 requires an additional assumption on the instantaneous TV graphs [to be discussed below in the response to Q4].

评论
  1. Regarding the scope of Time-Varying Graph settings, The current work focuses on undirected time-varying graphs. While this is a meaningful setting, I encourage the authors to consider that recent work has tackled the more general and challenging setting of time-varying directed networks. These settings are often closer to real-world decentralized systems (please see papers listed below). Given the maturity of the literature in that area, the exclusion of directed topologies may limit the broader applicability and novelty of the current work. Including a discussion on this point would help clarify the scope and relevance of the contributions. a). "Accelerated distributed stochastic non-convex optimization over time-varying directed networks." TAC, 2024. b). "Accelerated AB/Push–Pull Methods for Distributed Optimization Over Time-Varying Directed Networks." TCNS, 2023. c). "AB/Push-Pull method for distributed optimization in time-varying directed networks." OMS, 2023. d). "Decentralized optimization over time-varying directed graphs with row and column-stochastic matrices." TAC, 2020.

We agree that extending FSPDA for directed and TV graphs is a meaningful direction. However, as noted in the original response, analyzing such setting requires a significantly different set of tools that are more appropriate for a new project/paper.

The suggseted references are relevant, which we will include in the revision. Compared to FSPDA, we should note that except for a) [to be discussed below in the response to Q4], the references b), c), d) only focused on cases with strongly convex objective function. In the latter case, the analysis is arguably easier due to the existence of a unique optimal solution that the algorithms converge to. Such simplification cannot be used in FSPDA as we treat the general smooth (possibly non-convex) objective function setting. Note that most problems in machine learning nowadays cannot be cast as a strongly convex problem.

  1. “ we emphasize that our paper is the first one to achieve these rates under the settings of TV graphs without relying on the common bounded heterogeneity (BH) assumption, which has caused slow convergence in practice for certain prior works such as DSGD.” - Please note that the paper listed above (Accelerated distributed stochastic non-convex optimization over time-varying directed networks.) has established same convergence rate for non-convex problems over time-varying directed networks. I would encourage the authors to more carefully position their claims relative to these recent developments, potentially clarifying what specific aspects of their analysis (e.g., in algorithm design, proof techniques, or assumptions) distinguish their results from prior work.

We have checked the suggested reference (thanks for bringing it up!). However, we notice an important limitation in their Assumption 3, which demands that for all tt, the network is strongly connected with a column-stochastic weight matrix WmtW_m^t. In other words, although the paper considers TV directed graphs, it requires the instantaneous network to be (strongly) connected at each iteration. This will significantly limit the applicability of their results for realistic environment, e.g., it does not support the cases when only 1 edge is active at an iteration.

In contrary, the FSPDA algorithms do not require the instantaneous network to be connected (albeit undirected). As such, we believe that our claim as "the first one to achieve these rates under the settings of TV graphs ..." still holds.

评论

Dear Reviewer H1MF,

As the end of discussion period is approaching, we would like to see if our rebuttal has addressed your concern. We kindly remind the reviewer that you can follow up with new questions to our rebuttal, or raise the rating of our submission if you are satisfied with our response.

Best regards,

Authors of Paper 26853.

评论

I appreciate the authors’ efforts in addressing the concerns. Please see the follow-up comments below:

  1. Regarding the connectivity assumption, it is typically hard for distributed optimization algorithm to derive the average consensus in the analysis, especially for time-varying graphs. However, from Assumption 2 it seems that the derivation of average consensus will be trivial and this could diminish the analytical challenge that typically arises in decentralized optimization over time-varying graphs.

  2. “On top of this, we allow for random sparsification for further efficient communication” - I appreciate the authors' discussion on the incorporation of random sparsification for communication efficiency. I would like to point out that similar ideas have been explored in [Communication-efficient variance-reduced decentralized stochastic optimization over time-varying directed graphs." IEEE Transactions on Automatic Control 67.12 (2021)], which also addresses communication reduction via sparsification in a time-varying and directed setting. It would strengthen the contribution of this paper if the authors could elaborate on what new theoretical insights or practical behaviors emerge in their current analysis that go beyond existing results.

  3. Regarding the scope of Time-Varying Graph settings, The current work focuses on undirected time-varying graphs. While this is a meaningful setting, I encourage the authors to consider that recent work has tackled the more general and challenging setting of time-varying directed networks. These settings are often closer to real-world decentralized systems (please see papers listed below). Given the maturity of the literature in that area, the exclusion of directed topologies may limit the broader applicability and novelty of the current work. Including a discussion on this point would help clarify the scope and relevance of the contributions.

a). "Accelerated distributed stochastic non-convex optimization over time-varying directed networks." IEEE Transactions on Automatic Control (2024).

b). "Accelerated ABAB/Push–Pull Methods for Distributed Optimization Over Time-Varying Directed Networks." IEEE Transactions on Control of Network Systems 11.3 (2023).

c). "AB/Push-Pull method for distributed optimization in time-varying directed networks." Optimization Methods and Software (2023).

d). "Decentralized optimization over time-varying directed graphs with row and column-stochastic matrices." IEEE Transactions on Automatic Control 65.11 (2020).

  1. “ we emphasize that our paper is the first one to achieve these rates under the settings of TV graphs without relying on the common bounded heterogeneity (BH) assumption, which has caused slow convergence in practice for certain prior works such as DSGD.” - Please note that the paper listed above (Accelerated distributed stochastic non-convex optimization over time-varying directed networks.) has established same convergence rate for non-convex problems over time-varying directed networks. I would encourage the authors to more carefully position their claims relative to these recent developments, potentially clarifying what specific aspects of their analysis (e.g., in algorithm design, proof techniques, or assumptions) distinguish their results from prior work.
最终决定

This work proposes an algorithm for decentralized training of neural networks in a time-varying setting. Unlike prior works such as Kovalev et al., which explicitly assume strong convexity, this paper relaxes the assumption to smooth functions. The contribution includes experimental results and convergence bounds, which are of interest, though in my opinion, their generality and relation to the state of the art are limited and not sufficiently contrasted.

The reviewers appreciated the effort to present a rigorous framework for the non convex loss time varying communication matrix setting. However, the formulation is very similar to Kovalev et al., as it also relies on a Lagrangian approach and saddle-point condition. Moreover, the lack of discussion of key related work, e.g., Boyd et al., Randomized Gossip Algorithms, is surprising and weakens the positioning of the contribution. Concerns were raised, in line with reviewer H1MF, regarding the novelty of the work. For a resubmission, addressing these missing references and clarifying the distinct contribution will be necessary, for instance by summarizing and contrasting the differences more precisely (as highlighted by the authors during the rebuttal).

During the rebuttal phase, three reviewers were convinced of the paper’s merits, notably its attempt to provide a framework for a specific problem. However, one reviewer maintained strong concerns about the incomplete literature review. I find this objection persuasive: while the problem is interesting, the related work discussion is insufficiently developed (it's solely done line 29-54 instead of a full section), which prevents acceptance at this stage, given the competitiveness of this specific area.

Thus, I recommend rejection given the substantial rewriting needed.