On Linear Convergence in Smooth Convex-Concave Bilinearly-Coupled Saddle-Point Optimization: Lower Bounds and Optimal Algorithms
摘要
评审与讨论
This paper studies the first-order methods for solving smooth convex-concave saddle-point problems with bilinear coupling i.e. . It establishes the first lower bounds on the number of gradient evaluations and matrix-vector multiplications with and needed for solving these saddle-point problems. Moreover, it develops algorithm which matches this lower bound.
给作者的问题
- What is the number of gradient evaluations required for the algorithm in Kovalev et al 2022b.
- Assumptions 2.6 and 2.7 are not really assumptions right? They follow from assumptions 2.3-2.5.
论据与证据
This paper claims to develop an optimal algorithm for solving saddle point optimization problem with bilinear coupling. However, the paper doesn't provide any experiment to show it's performance and comparison to other existing algorithms in practice.
方法与评估标准
This paper provides no experiments to evaluate the performance of the proposed algorithm. The analysis is entirely theoretical, focusing on deriving lower complexity bounds and developing an optimal algorithm that matches these bounds. No empirical validation is provided to demonstrate its practical performance on real-world problems or benchmark datasets.
理论论述
The paper claims they develop the first optimal algorithm that matches the lower bound. However, they do not provide explicit pseudocode of the algorithm for solving the saddle point problems. Moreover, the iteration complexity of this paper is given in terms of weighted square distance which is completely different from Kovalev et al 2022b. This is an unfair comparison.
实验设计与分析
This paper provides no experiments to evaluate the performance of the proposed algorithm. Therefore, there are no experimental designs or analyses to assess for soundness or validity.
补充材料
I have checked the following parts of the supplementary material:
- Appendix B: the pseudocode
- Appendix C: Table of comparisons
- Appendix E
与现有文献的关系
The saddle-point optimization problem with bilinear coupling is a fundamental problem class that arises in various machine learning applications, including game theory, reinforcement learning, computer vision, robust optimization, and distributed optimization. Developing an optimal algorithm for solving such problems has the potential to significantly benefit the machine learning community.
遗漏的重要参考文献
I think the references discussed in the paper are sufficient.
其他优缺点
Strength:
- Provides lower bound for saddle point optimization problems with bilinear coupling.
Weakness:
- This paper claims they provide the first algorithm which attains the optimal complexity for strongly convex strongly concave. To my understanding, Kovalev et al 2022b already attains the optimal complexity with respect to each of the gradient evaluations (they just didn't mention that explicitly). I think the contribution of this paper is incremental.
- The iteration complexity of this paper is given in terms of weighted square distance which is completely different from Kovalev et al 2022b. This is an unfair comparison.
- The paper provides no experiment to show the performance of proposed algorithm in practice.
- The presentation is poor.
- The main problem under consideration is problem (1). However, authors do not provide any pseudocode to solve this problem class. Only pseudocode is for algorithm 1, which solves problem 14.
- It is hard to follow how problem 14 generalizes problem 1. Authors should add a detailed discussion on this in the Appendix.
其他意见或建议
- Please write down the pseudocode of the algorithm which solves the problem 1.
- Add a discussion on how problem 14 generalizes problem 1.
We thank the reviewer for the time and effort. Unfortunately, the development of the optimal algorithm for solving problem (1), which is one of the key contributions of our paper acknowledged by other reviewers, is missing from the "strengths" list. Moreover, the criticism of our paper is based on claims that are either factually false or highly questionable. We address these below.
Kovalev et al 2022b already attains...Unfortunately, this claim is false and unjustified:- This claim is false. The algorithm of Kovalev et al. (2022b) cannot reach the lower bounds because it does not implement the separation of complexities. Refer to lines 60-72 in Section 1.2 for an explanation. This also clearly follows from Table 1.
- This claim is unjustified. You should support this claim with evidence, including an explanation with references to Kovalev et al. (2022b), such as precise theorems, equations, paragraphs, etc.
This is an unfair comparison.Unfortunately, this statement is false. The exact opposite is true: it is correct to compare with any norm, including the one used by Kovalev et al. (2022b, eq. (14)), due to the norm equivalence theorem. The cost of the transition between the norms is negligible due to linear convergence: it only results in extra additive factors in the complexity. Moreover:- It is standard practice in the field to ignore additive constants in linear complexities. This includes all SOTA papers in Table 1, most of which are A*/Q1.
- On lines 162 and 403, we explicitly mention that we ignore additive constants. This is also acknowledged by Reviewer ydoA.
No experiments.It is a common standard in the field that papers with strong theoretical results are not required to include any experiments, just as strong experimental papers are not required to include any theory. Our paper contains strong theoretical results, which are acknowledged by other reviewers. Hence, the absence of experiments is justified.Presentation is poor.Unfortunately, you have not provided any arguments to support this claim. On the other hand, other reviewers found our paper "well-written" and "clearly written". Regrettably, our only option is to disregard this claim.Authors do not provide pseudocode.In Section 4.3, we provide a clear and comprehensive explanation of how to apply Algorithm 1 with restarting to solve problem (1), including the definitions of functions and operators in eqs. (25) and (26). This is more than enough for anyone with at least some expertise in mathematical optimization to use the algorithm. It is also important to highlight that, according to lines 351-367 in Section 4.2, it is necessary to reorder functions and operators , depending on the values of constants and from Theorem 4.7. This would lead to explicit variants of the algorithm for each order. Hence, a practical implementation would still use Algorithm 1 in combination with separate first-order oracle implementations for functions and operators , which can be easily done in any modern programming language/framework.How problem 14 generalizes problem 1.A brief explanation is available on lines 351-255, Section 4.3. Some explanation is also available in (Lan and Ouyang, 2021; Gidel et al., 2018; Nesterov, 2007). We are strongly convinced that a more detailed explanation is unnecessary because the reduction of convex-concave saddle-point optimization problems to monotone variational inequalities is one of the most basic facts in optimization theory. For a detailed explanation, please refer to books like "Finite-Dimensional Variational Inequalities and Complementarity Problems".
Questions
What is the number...It is given in Table 1 of our paper or Table 1 of Kovalev et al. (2022b).Assumptions 2.6...This is not true. Assumption 2.6 requires , which is not implied by Assumptions 2.3-2.5. Moreover, Assumption 2.6 is the "line" that separates the settings of linear convergence and sublinear convergence. This is clearly mentioned numerous times in the paper, for instance, on lines 160-188 (Section 2.2), and lines 228-262 (Section 3.2). Similarly, it is easy to verify that Assumption 2.7 is not implied by Assumptions 2.3-2.5.
Thank you for your reply.
-
Kovalev et al 2022b haven't mentioned the gradient evaluations separately. But check the terms inside max of equation 25 from Kovalev et al, 2022b. Each of these terms corresponds to , computations and exactly matches what you have.
-
The definition of R^2 contain terms like which in turn depend on . So showing will have another term in the number of gradient evaluations.
-
No, the lack of experiments is not justified. The authors do not provide any pseudocode or discuss how to implement the algorithm in practice, which raises the question of whether the proposed algorithm benefits from the existing ones.
-
I did provide arguments on why I think the presentation is poor. a. There is pseudocode for the main problem (equation 1), b. There is no discussion on why problem (14) generalizes (1).
I will keep my score.
Thank you for the reply. Unfortunately, all four of your arguments are false:
-
Kovalev et al 2022b haven't mentioned the gradient evaluations separately. But check the terms inside max of equation 25 from Kovalev et al, 2022b. Each of these terms corresponds to , computations and exactly matches what you have.
Kovalev et al. (2022b) reported the iteration complexity . The gradients and are computed exactly once at each iteration of their algorithm. Hence, Kovalev et al. (2022b) require computations of and . This does not coincide with and from our paper.
-
The definition of contain terms like which in turn depend on . So showing will have another term in the number of gradient evaluations.
Our algorithm requires computations of the gradient to reach the precision . Hence, to reach the precision , our algorithm requires computations of the gradient . As you can see, these linear convergence rates coincide.
-
No, the lack of experiments is not justified. The authors do not provide any pseudocode or discuss how to implement the algorithm in practice, which raises the question of whether the proposed algorithm benefits from the existing ones.
We do provide a pseudocode for solving problem (1), which is a special instance of problem (14), in Algorithm 1, along with clear and comprehensive instructions on how to implement this algorithm in Section 4.3.
-
I did provide arguments on why I think the presentation is poor. a. There is no pseudocode for the main problem (equation 1), b. There is no discussion on why problem (14) generalizes (1).
Unfortunately, neither argument supports the "poor presentation" claim:
- (a) This is false. Please refer to the information above and our original rebuttal.
- (b) There is no place for such a discussion in a scientific paper aimed at an audience with expertise in optimization, beyond the references that we provide in the paper, which are highlighted in our original rebuttal. In particular, the question of equivalence between problems (1) and (14) is so basic that it is literally asked during the undergraduate optimization course exam at most universities around the world, including ours.
This paper considers deterministic convex-concave minimax optimization problems. In particular, the main focus is on the case where we can obtain linear convergence, as characterized in Assumption 2.6.
- First, the authors establish fine-grained lower bounds by separately counting oracle calls for the gradients and the coupling matrix multiplications, and the results (1) recover [ZHZ'22] for the SCSC (strongly-convex-strongly-concave) case and (2) also cover some other cases (including strongly-convex-concave or bilinear functions).
- Second, the authors propose Algorithm 1, which is based on the idea of solving a more general finite-sum variational inequality. An application of this to minimax optimization problems yield tight convergence upper bounds for the cases described in the lower bound results.
[ZHZ'22] Zhang, J., Hong, M., and Zhang, S. On lower iteration complexity bounds for the convex concave saddle point problems. Mathematical Programming, 2022.
给作者的问题
- Can the authors elaborate on any iteration complexities or number of oracle calls induced in the part of Algorithm 1 in the innermost loop ?
- In the lower bounds, using matrices with 's in the diagonals and 's in the super-diagonals is quite standard (as in [ZHZ'22]), but I wonder what the high-level motivations of the instances in Lemmas G.2 and G.3 (for the parts) were. Could the authors give a bit of an illustration on this, or is this just a magical lower bound?
- I am also curious if previous results by [ZHZ'22] can recover the same results considering separate counts for the gradient and coupling matrix oracles for the SCSC case. (I am aware that [ZHZ'22] deals only with the SCSC case while this paper covers several more cases.)
- This might be a bit out of scope because the proposed problem is already an important, classical problem, but one weakness one could point out (for the upper bound) is that everything applies only when we have the bilinear-coupled structure. Personally, I also have been curious about whether we can construct an intuitive minimax optimization algorithm for SCSC (or maybe more general cases where we can have linear convergence) without the bilinear coupling structure in the objective function. I know that the existence of the coupling matrix is essential for the proposed results, but have the authors ever thought of extending Algorithm 1 to the general convex-concave case by, for instance, replacing the 's with the Hessians , or else?
论据与证据
(The main results are theoretical.) I have not checked the proofs line by line, but the details seem to have no fatal errors.
方法与评估标准
The main results are theoretical.
理论论述
I have not checked the proofs line by line, but I read the appendix, and the details seem to have no fatal errors. (See Strengths & Weaknesses and Questions for details.)
实验设计与分析
The main results are theoretical.
补充材料
There are no separate supplementary materials other than the paper.
与现有文献的关系
The paper can contribute to theoretical guarantees and understandings on convex-concave minimax optimization algorithms and variational inequalities.
遗漏的重要参考文献
I am unaware of any particular closely related work that has not been cited in the paper.
其他优缺点
Strengths
- The paper is well-written (in my opinion), and I really enjoyed reading it.
- The paper suggests both novel lower bounds and matching upper bounds that, altogether, closes the case. The proposed results make solid contributions, especially those for the new tight rates for all of the non-SCSC cases.
- The paper also contains (upper bound) results that consider finite-sum variational inequalities of the form .
Weaknesses
- See Questions for details.
其他意见或建议
- The running title says 'Submission and Formatting Instructions for ICML 2025.' (I have 3 such papers in my review batch, and I don't know why!)
- I think it would have been better if there were more discussions on the lower bound instance constructions, which I think is one of the most interesting parts of the paper.
- (TYPO) Table 1, Footnote (5) symetric → symmetric
伦理审查问题
(None)
We thank the reviewer for the detailed comments, valuable feedback, and high evaluation of our work. Below, we provide our detailed response to the review.
Other Comments Or Suggestions
- Fixed.
- Please refer to the separate paragraph below.
- Fixed.
Questions For Authors
- All functions are just quadratic functions that contain previously computed gradients and operators . This can be shown by analyzing lines 17 and 22 at recursion levels . Hence, line 12 is a simple (possibly constrained) quadratic optimization problem, which does not require any oracle calls.
- Please refer to the separate paragraph below.
- Indeed, the result of Zhang et al. (2022b) recovers the part in the SCSC case. In particular, the proof of Lemma G.3 is inspired by the result of Ibrahim et al. (2020), which is a slightly generalized version of the result by Zhang et al. (2022b). Combining this with the lower bound for smooth and strongly convex optimization by Nesterov (2013), we recover the full result in the SCSC case.
- Some results for general min-max problems beyond SCSC were developed by Kovalev et al. (2022b). However, these results are far from reaching our lower bounds. It is worth highlighting that obtaining accelerated rates is much more difficult for general min-max problems (some results were developed by Alkousa et al. (2020), [2] mentioned by Reviewer 1jnp, or in arXiv:2002.02417, arXiv:2205.05653). Overall, this is indeed an interesting question that we are currently starting to examine.
Lower bounds construction.
Here, we attempt to provide an intuition behind the construction of our lower bounds. We start with the basic example of Nesterov (considering an infinite-dimensional space for simplicity). Consider the following linear system: where \\mathbf{A} = \\begin{bmatrix}1\\\\\\alpha-1&1\\\\&\\alpha-1&1\\\\&&&\\ddots\\end{bmatrix}, and . It has a unique solution . We can construct a minimization problem with the same solution: One can show that after computations of the gradient, no more than coordinates of the current iterate are nonzero (due to the linear span assumption). Hence, the distance to the solution is lower bounded by the remainder of the geometric series \\sum_{i=k+1}^\\infty(1-\\alpha)^i = \\frac{(1-\\alpha)^{k+1}}{\\alpha}. Moreover, the condition number is proportional to , which gives the lower bound of Nesterov.
The minimization problem above has the following min-max reformulation: Here, we have proportional to , and we can apply arguments similar to the ones above to obtain the lower bound . We can also add a regularizer of the form , which, subject to some additional details, leads to the results of Zhang et al. (2022b) and Ibrahim et al. (2020). Our Lemma G.3 can be seen as a finite-dimensional variant of the result by Ibrahim et al. (2020).
Now, we discuss the most challenging case of our lower bounds, which is Lemma G.2. The starting point is the lower bounds of Scaman et al. (2018) for smooth and strongly convex decentralized minimization, which allows for the min-max reformulation of the form (1). It is based on splitting the hard function of Nesterov (see above) into two functions and placing them on the first and the last nodes of the path consisting of nodes (refer to Scaman et al. (2018) for details). To obtain our lower bounds, we need to make the following substantial changes:
- We add dual regularization of the form . This allows us to obtain the desired result for , but it makes the problem much more difficult to work with.
- We replace the -node path with an -node path and attach two -node star-topology networks to its ends, such that Nesterov functions are stored only on the "leaves" of these stars. This step introduces some sort of symmetry, which simplifies finding the solution to the problem.
- We also introduce an extra dual variable, modify matrix , and add an extra regularizer of the form with respect to the new variable to account for nontrivial values of . Additionally, we analyze the spectral properties of matrix in Lemma I.1.
We apply a series of reformulations to the resulting min-max problem and obtain a simple minimization problem similar to the example above but with a different value of . It remains to carefully analyze this value, subject to additional details which we, unfortunately, are unable to discuss here due to the 5000 character limitation.
This paper develops tight lower complexity bounds and matching optimal algorithms for smooth saddle-point problems with bilinear coupling. The work unifies existing results in different regimes (strongly-convex-strongly-concave, bilinear saddle-point, strongly convex with affine constraints) as well as gives new results in the convex-concave setting.
Update after rebuttal
The authors have addressed my concerns regarding the proof of Theorem 3.3 in their rebuttal. I keep my current evaluation of the paper.
给作者的问题
Could you elaborate on the proof of Theorem 3.3 in the cases that or when one of the two strong convexity constants is ?
论据与证据
The claims are clear and supported by convincing evidence.
方法与评估标准
The intuition behind the used methodology is lacking and could be made more clear. The evaluation criteria make sense for the problem at hand.
理论论述
I checked the claims and proofs in all the Appendices except Appendix I. The main issue I have is in the proof of Theorem 3.3 (Appendix G), where the case is proved by assuming that . The authors should provide more explanation why this is justified.
实验设计与分析
There are no experiments in this paper.
补充材料
There is no supplementary material.
与现有文献的关系
The paper studies saddle-point problems which have applications in various fields such as economics, game theory and statistics. A comparison with existing state-of-the-art linearly-converging algorithms is given in Appendix C. Moreover, optimal algorithms and theoretical lower bounds are given, which builds further upon the work of Nesterov.
遗漏的重要参考文献
The related publications are adequately discussed throughout the paper.
其他优缺点
Strengths:
- The paper is clearly written and provides tight theoretical results in the linearly-converging setting.
- The results of the paper are relevant to numerous machine learning applications.
Weaknesses:
- There is no intuition provided for some constants such as the ones in Assumption 2.7 or the Lyapunov function.
- Similarly, the proofs are not very enlightening, being only a series of algebraic manipulations.
其他意见或建议
List of typos and minor errors:
- The page titles are still “Submission and Formatting Instructions for ICML 2025”
- Line 140: would it be possible to write instead of the minimum eigenvalues in the “otherwise” cases, or is this not valid?
- Line 165: Assumption 2.6 is a bit poorly worded, it might give the impression that Assumptions 2.3-2.5 implies the inequality.
- Line 200: Equation (9): missing squares for the distances
- Line 233: fix first set of quotation marks around “hard”
- Line 245: Theorem 3.2: does not hold for any time , in Nesterov’s book there is an upper bound for the iteration number which should translate into an upper bound for
- Line 327: Footnote 5: it is better to specify the proposition/page number instead of the whole book
- Line 345: Theorem 4.5: should also include a reference to Algorithm 1 as well as the problem to be solved
- Line 402: “numbers” should be “number”
- Line 448: capitalisation Polyak-Łojasiewicz (also add en-dash names between instead of hyphen)
- Capitalisation of conference names in references is not consistent (e.g. “International Conference on Machine Learning” vs “international conference on machine learning” or “Advances in Neural Information Processing Systems” vs “Advances in neural information processing systems”
- Line 561: “eigenvalues of a matrix” should be “eigenvalues of a symmetric matrix” since otherwise they might not be real
- Line 566: “argmin” should be “min”
- Line 699: “symetric” should be “symmetric”
- Line 718: “this” should be “these”
- Line 738: Equation (38) does not follow from putting in the values from (37)
- Line 782: A reference to Assumption 2.6 in addition to Assumption 2.5 should be added. Moreover, the implication stated should be explained more thoroughly
- Each time a line number from Algorithm 1 is referenced, it comes out as “algorithm 1 of Algorithm 1” instead of “line x of Algorithm 1”
- Line 1430: “wuch” should be “such”
- Line 1527: should be in the superscript
- Line 2290: Missing an identity matrix in the upper bound
- Line 2430: Lemma K.2: does this hold for ?
- Line 2460: Equation (174): missing superscript on ; the last should be ; should be
- Line 2489: “which is implied by which is implied by”
- Line 2495: “Introduction step” should be “Induction step”
- Line 2495: also have to assume that it holds for or the induction is invalid
- Line 2518: should be
- Line 2596: should be
- Line 2835, 2846, 2850: subscript inside the last product should be instead of (resp. instead of )
- Line 2880: step (c) should be an inequality and should be
- Line 2911: Q(z) should be Q(z^*)
- Line 2930: The Lipschitz constants are missing
- Line 2983: the coefficients 12 could be 6?
- Line 2985: step (e) equality should be an inequality
- Line 3001-3009 could be removed, the statement is obvious since big O does not care about additive constants
- Line 3013: deifnition should be definition
We thank the reviewer for the detailed comments, valueable feedback, and high evaluation of our work. Below, we provide our detailed response to the review.
Question about the proof of Theorem 3.3 in the case or
Thank you for the question! This indeed may need additional explanation. This question is related to cases (i), (ii.b), and (iii.b). For instance, consider case (i) (other cases are similar). It is easy to observe that if the function is -strongly convex with , then it is also -strongly convex or simply convex (see Definition 2.1 and Assumption 2.4). Hence, the class of problems (1) satisfying Assumptions 2.3-2.5 with parameters is contained in the class of problems (1) satisfying Assumptions 2.3-2.5 with parameters . Thus, for case (i), we can choose our hard instance of problem (1) with a -strongly convex function with an arbitrary . In particular, we choose the hard problem instance according to Lemma G.2, which gives the following lower bound:
We can choose and obtain
which is the desired lower bound for case (i).
Intuition behind the numerical constants
- There is no particular intuition behind the actual values of the numerical constants in Assumption 2.7 (4 and 18), except that we chose these constants to simplify our calculations in the proof of Theorem 3.3 and make them less ugly. It is likely that these numerical constants can be reduced, but it would not make much sense because Assumption 2.7 is only used to avoid covering uninteresting corner cases with small, i.e., , condition numbers, as mentioned in the paper.
- The numerical constants in the Lyapunov function in eq. (29) hold little intuition. The values of these constants are mostly driven by the proof and can likely be improved as well, but it would not make much sense as this would only result in logarithmic or additive improvements in complexity.
Intuition behind the proofs
For the intuition behind the construction and the proof of lower bounds, please refer to our response to Reviewer PYLp, who raised a similar question. Unfortunately, we were unable to provide a more detailed intuition behind the convergence proof beyond what we have in Section 4.2 (lines 303-345) due to the 5000 character limit (we really tried). We will include a more detailed explanation in the revised version of the paper.
Typos and minor errors
Thank you for the list of typos and minor errors! We fix them all as follows:
- (1) Fixed.
- (2) Yes, indeed. For instance, in the definition of , if \\lambda_{\\min}(\\mathbf{B}^\\top \\mathbf{B}) > 0, we have \\mathrm{range} \\mathbf{B}^\\top = \\mathcal{X} and fall into the first option.
- (3) Thanks for the suggestion. We have removed the references to Assumptions 2.3-2.5 and added the words "Parameters satisfy the inequality...".
- (4-5) Fixed.
- (6) Indeed, speaking rigorously, we need to mention the transition from the lower bound on the number of iterations in Nesterov's book to the lower bound on , even though it is straightforward. We will add an appropriate comment to the proof in the revised version of the paper.
- (7) Added reference to Lan (2020, proof of Theorem 3.3).
- (8-15) Fixed.
- (16) Indeed, lines 732-740 are worded inaccurately because we cannot choose due to Assumption 2.7. This is also discussed on lines 741-745 ("Note that strictly speaking..."). We will rewrite line 732 and eq. (37) in a more accurate way, i.e., something like "the class of bilinear saddle-point optimization problems falls under Assumptions 2.3-2.5 with parameters..."
- (17) Indeed, the case is trivial. The case implies due to Assumption 2.6, and \\nabla f(x) \\in \\mathrm{range}\\mathbf{B}^\\top = (\\ker \\mathbf{B})^\\perp due to Assumption 2.5 and point #2 from your list.
- (18-21) Fixed.
- (22) Yes, it does. For , it holds due to line 22 (first option). For , we have due to line 22 (second option) and line 17 (second option). Hence, we can prove the desired statement by induction.
- (23-32) Fixed.
- (33) Yes, indeed, they could.
- (34) Fixed.
- (35) Yes, indeed.
- (36) Fixed.
I would like to thank the authors for their reply. The authors have addressed my concerns regarding the proof of Theorem 3.3.
I keep my current evaluation of the paper.
This work studied smooth (strongly)-convex-(strongly)-concave bilinearly-coupled saddle-point problem, and provided lower complexity bounds in terms of the computation time, and achieved the separation of complexities. And they further proposed an optimal algorithm which matches with the lower bound.
给作者的问题
/
论据与证据
The results are supported by convincing evidence, generally I am satisfied with the results. Here are some questions:
- Missing literature. Line 83, you mentioned for strongly-convex-concave case, "To the best of our knowledge, there are no lower complexity bounds that would cover these cases", I think the work [1] should have solved it, check Theorem 2 therein.
- Another missing literature should be [2], whose upper bound result is different from your Table 1.
- Echo on [1], they also extended the oracle class to proximal mapping, which is broader than your setting, which may be an extension that you can consider.
[1] Xie, Guangzeng, et al. "Lower complexity bounds for finite-sum convex-concave minimax optimization problems." ICML 2020.
[2] Wang, Yuanhao, and Jian Li. "Improved algorithms for convex-concave minimax optimization." NeurIPS 2020
方法与评估标准
/
理论论述
/
实验设计与分析
/
补充材料
/
与现有文献的关系
This work advanced the understanding of optimal complexities for bilinear min-max optimization problems.
遗漏的重要参考文献
See above
其他优缺点
/
其他意见或建议
/
We thank the reviewer for the high evaluation of our work and the useful references. We provide our answers to the questions below.
- Thank you for pointing this out. The statement "to the best of our knowledge, there are no lower complexity bounds that would cover these cases" is indeed a bit inaccurate. Our point was that there are no linear lower bounds, i.e., lower bounds for the linear convergence setting. On the other hand, [1, Theorem 2] provides a sublinear lower bound. We will fix the inaccuracy and cite [1] in the revised version of the paper.
- As far as we understand, the main limitation of [2] is that it can achieve the linear SOTA rate in the SCSC regime only for quadratic problems. On the other hand, [2] achieves SOTA rates for general min-max problems without bilinear coupling. We will discuss this in the revised version of the paper.
- Thank you for the suggestion. This is an interesting question to consider for future work. In particular, we think that it is possible to remove the terms and/or from the lower and upper complexity bounds by assuming access to proximal mappings associated with functions and , respectively.
Thank you for the clarification.
A further question, can you clarify whether the dist() in your Equation 9 comes with square? I did not find the detailed formal definition. BTW, Equation 32 should be wrong I think, it should be min, rather than argmin (also questionable on whether to apply square). From the proof in Appendix I, the dist() should be defined in terms of squared norm I think.
Thank you for your reply.
There is a typo in equation (9); the squared distance should be defined using squared distances as follows:
There is indeed also a typo in equation (32); should be replaced with , but no square this time:
The proof in Appendix I indeed uses the squared distances , according to the corrected definitions above. Thank you for pointing out the typos; we have fixed them.
This paper studies bilinearly coupled saddle-point problems with general smooth (non-strongly) convex and . The main contributions of the work include tight upper and lower complexity bounds that establish linear convergence rates. The paper is well written and provides insightful discussion into the nuances that arise in assessing optimal convergence complexity for saddle-point settings. The reviewers were overall in agreement that the paper provides important results for saddle-point optimization, and would therefore be a valuable contribution to the conference.