Robust and Faster Zeroth-Order Minimax Optimization: Complexity and Applications
We propose a robust and faster zeroth-order algorithm to solve nonconvex minimax problems.
摘要
评审与讨论
In this paper, a unified single-loop zeroth-order gradient descent extragradient ascent (ZO-GDEGA) algorithm to solve the nonconvex-concave minimax problem faster and more robustly. The theoretical analysis is provided to guarantee an overall complexity of . The experimental results on the data poisoning attack task and the AUC maximization task are shown to validate the practical performance of the proposed method.
优点
- In this paper, the authors propose a zeroth-order algorithm that achieves lower complexity under the nonconvex-concave condition.
- The first convergence analysis of stochastic zeroth-order algorithm under the nonconvex-concave condition is provided.
- For nonconvex strongly-concave problems, the complexity with respect to the condition number is improved.
缺点
- In section 2 related work, the complexity of first-order minimax algorithms is not discussed. To my understanding, the error of the zeroth-order gradient can be bounded by parameters and that are set as small as . Therefore, the analysis should not change a lot based on the analysis of the first-order counterpart, which probably undermines the novelty and contribution.
- In ref [18] (Huang et al 2022), variance reduction is used to improve the complexity with respect to . The contribution of this paper will be further stronger if this part is also considered.
- The figures in the experiments section is too small and some curves are covered by the legends. I understand the space is limited but maybe some sections could be rearranged to the Appendix.
问题
- Please see weakness [1]. Can you provide more details to explain how the zeroth-order gradient estimator makes the original analysis of the first-order method more challenging?
- Is it possible to further reduce the complexity with respect to for the NC-C problem by integrating variance reduction with the proposed ZO-GDEGA algorithm?
局限性
I did not see any negative societal impact.
Thanks for your insightful thoughts and comments! Below we will clarify the two points in the review.
Response 1.
Related works for first-order methods.
Existing works mainly focus on first-order (FO) methods for solving NC minimax problems. For the NC-SC setting, GDA [1] and AGDA [2] both achieve () gradient complexity in deterministic (sochastic) setting to find an -stationary point. Lin et.al [3] proposed a multi-loop Minimax-PPA algorithm, which further improves the complexity to . For the NC-C setting, GDA [1] and AGDA [2] both achieve a gradient complexity of and for the deterministic and stochastic settings. SAPD+ [4] has reduced the gradient complexity to for the stochastic setting. The above analysis is all measured with the max function . Xu et.al [5] considered the -stationary point in terms of and proposed AGP to achieve a gradient complexity of on this weaker measure, but did not consider the stochastic setting. FO algorithms inspired the design of ZO algorithms. However, existing ZO methods such as [5][6] mainly focus on trivial extensions of existing FO algorithms, while we focus on designing ZO minimax algorithms and directly improving the robustness and convergence rate, which is a non-trivial extension.
In our main paper, we stated the challenges. To address your concerns about our contributions, we clarify them and improve discussion of challenges as follows.
We propose a faster, more robust, and theoretically guaranteed algorithm framework that can solve wider applications.
Stronger robustness. We find that the smoothing parameter bounds have a key impact on the ZO algorithms for black-box minimax problems. Previous research did not focus on this key issue. We designed a novel ZO-GDEGA algorithm to weaken restrictions on the smoothing parameters (see Table 1), which enhances the robustness. Our ZO-GDEGA has also been extended to novel FO methods, which are shown in Algorithms 3 and 4 in the Appendix. Thus, our algorithm is a non-trivial extension of the existing FO algorithms.
Lower complexity. We develop a new and concise analysis framework to prove that our ZO-GDEGA can achieve lower complexity and it is the first ZO algorithm with theoretical guarantees for solving stochastic NC-C problems.
Analysis of Challenges. For minimax problems, the errors and are coupled with each other, which leads to difficult parameter selection. Taking the NC-C setting as an example, the recursion requires and requires . We handle them by constructing reasonable recursions and step sizes. Our Proposition 6 plays a key role and clarifies that the EG structure can result in the coefficient for and for , which allows to choose reasonable step sizes and larger smoothing parameters than existing methods, thus achieving more robust performance.
Even our analysis for the FO counterpart still faces difficulties due to the EG and the proximal operator. In the NC-SC setting, we propose to establish the upper bound on in terms of to construct the recursive relation of , thereby achieving enhanced robustness. In the NC-C setting, we creatively proposed Proposition 6 to handle the proximal operator and EG structure, thereby resulting in the reasonable recursion w.r.t. .
Wider applications. We expand the scope of existing black-box minimax problems by considering more general regularizers rather than being limited to , which can be instantiated as data poisoning attack, AUC maximization, and robust neural network training (see C.3 in the Appendix) problems.
Better experimental results. Our algorithms can find much more effective attack perturbation and perform more robustly.
Response 2. Integrating variance reduction with the proposed ZO-GDEGA can further reduce the complexity.
Taking the NC-SC as an example, we provide an idea. We can use the ZO recursive momentum variance-reduced stochastic gradients as follows:
u_t = \alpha_t \hat\nabla_x f(x_t, y_t; \xi) + (1-\alpha_t)(\hat\nabla_x f(x_t, y_t;\xi) - \hat\nabla_x f(x_{t-1}, y_{t-1}; \xi) + u_{t-1}), $$ to update $x$, as well as $y$, where $\alpha_t \in (0, 1]$. When $\alpha_t = 1$, $u_t$ will degenerate a vanilla ZO stochastic gradients. In theoretical analysis, since STORM's analysis [7] can effectively reduce the complexity with respect to $\epsilon$, we can refer to their analysis and design Lyapunov function $\phi = f(x_t, y_t) + \mathcal{O}(\frac{1}{L^2\eta_x})||u_t - \nabla_x f(x_t,y_t)||^2 + \mathcal{O}(\frac{1}{L^2\eta_y})||v_t - \nabla_y f(x_t,y_t)||^2$ to analyze the complexity, and thus the complexity can be reduced. As for the NC-C setting, we will also think about how to design Lyapunov functions to analyze complexity. We will rearrange tables and figures in our revised version. [1] On gradient descent ascent for nonconvex-concave minimax problems [2] Alternating proximal-gradient steps for (stochastic) nonconvex-concave minimax problems[J] [3] Near-optimal algorithms for minimax optimization [4] Sapd+: An accelerated stochastic method for nonconvex-concave minimax problems. [5] DERIVATIVE-FREE ALTERNATING PROJECTION ALGORITHMS FOR GENERAL NONCONVEX-CONCAVE MINIMAX PROBLEMS [6] Enhanced first and zeroth order variance reduced algorithms for min-max optimization[J]. 2020. [7] Momentum-based variance reduction in non-convex sgdThank you very much for the reply. After reading the rebuttal, I think my concerns are addressed and I have raised my rating to 6.
We appreciate your review and valuable questions that improved our work.
They design a new unified ZO gradient descent extragradient ascent (ZO-GDEGA) algorithm, which reduces the overall complexity to find an ε-stationary point of the function ψ for nonconvex-concave (NC-C) problems. ZO-GDEGA is the first ZO algorithm with complexity guarantees to solve stochastic NC-C problems.
优点
The theoretical section is very thorough and comprehensive.
缺点
(1)The paper lacks detailed link of code implementation. (2)While the paper effectively addresses the NC-C problem from a theoretical perspective, it should also provide more experimental details and demonstrate that the NC-C problem exists within the experimental models.
问题
See in the Weakness part
局限性
None
Response 1. The detailed link of code.
We provide the code of the data poisoning attack experiment on the epsilon_test dataset. We have sent this code sample as an anonymized link to the AC. We will make our code public.
Response 2. The proof of the NC-C problem (5).
Problem (5) is a NC-C problem which can be proved as follows.
For Problem (5) (please see the C.1 in the Appendix for detail), we let and . We prove that the function is general convex as follows. Furthermore, . This Hessian matrix is positively semi-definite. Thus, the function is general convex and is general convex w.r.t. . Thus, Problem (5) can be written as the NC-C formula (1) by setting , , and .
For more experimental details, please refer to the Appendix, where we describe the selection of hyperparameters and more experimental results and analysis.
We appreciate your review and valuable questions that improved our work.
The paper studies zeroth-order methods for nonconvex-(strongly)-concave minimax optimization. The achieved rates improve previous results and tolerate much larger choice of the smoothing parameters. The proposed methods also perform well for some empirical tasks.
优点
Minimax optimization is an important problem that has many applications in machine learning and related areas. In many settings, gradients are hard to estimate or impossible to obtain, which motivates the study of zeroth-order methods. Moreover, nonconvex-concave minimax is itself a class of nonconvex nonsmooth optimization, which is considered as challenge problems in the related literature. The paper proposes new algorithms with improved convergence results compared with previous work for this challenge class of the problem.
缺点
-
I think a discussion on lower bounds can improve the understanding and position of this work. For example, lower bounds on zeroth-order methods justify the dependence on the dimension is inevitable without additional assumptions [1]. This, together with lower bounds for minimax optimization, e.g., [2], provide lower bounds for the considered problem class, which suggest the foundamental limits of this problem and whether the complexity can be further improved.
-
A discussion of the best known upper bounds for first-order methods could also help. As far as I know, the best rate for first-order nonconvex-concave minimax optimization is also [3]. The authors can do some additional literature review and check whether my statement is correct. As the complexity of zeroth-order methods is usually d times that of first-order methods, this suggests the results in this paper match the state-of-the-arts. For nonconvex-strongly-concave minimax optimization, [3] achieves a rate of , which suggests that the upper bounds in this paper can possibly be improved.
-
Have the authors considered to use two-point estimators to construct zeroth-order gradient estiimators? In some cases, two-point estimators give better rates [4]. The paper mentions that for the NC-SC case there is no need to use to update but instead. However, there is involved in the update of . Are there any typos in the statement of the algorithm? Also, I suggest to add dimension in the complexity stated in the abstract. Otherwise, this could be misleading.
I did not have time to carefully check every step in the proof. If the results are correct, I think they make enough contributions to the related literature. Therefore, I will keep a low confidence score.
References
[1] Optimal rates for zero-order convex optimization: The power of two function evaluations. IEEE Transactions on Information Theory, 2015.
[2] The complexity of nonconvex-strongly-concave minimax optimization. UAI, 2021.
[3] SAPD+: An Accelerated Stochastic Method for Nonconvex-Concave Minimax Problems. NeurIPS, 2022.
[4] An optimal algorithm for bandit and zero-order convex optimization with two-point feedback. JMLR, 2017.
问题
See weaknesses.
局限性
See weaknesses.
We would to thank the reviewer for the insightful comments! Below we will clarify the four points in the review.
Response 1. Lower Bounds.
We will add the following discussion about the lower bound to our revised version.
For minimization problems, the lower bounds on ZO methods justify the dependence on the dimension is inevitable without additional assumptions [1]. For minimax problems, some researchers focus on the lower bound of first-order methods. For example, the lower bounds are [2] and [3] for convex-concave and strongly-convex-strongly-concave settings, respectively. For NC-SC minimax problems, there is a lower bound of first-order methods [4][5]. Corresponding to minimization optimization [1], there may be a lower bound for ZO minimax optimization. Unfortunately, to the best of our knowledge, there is no work proving a lower bound for ZO algorithms solving NC-SC minimax problems. On the contrary, our work provides the upper bound for ZO algorithms solving NC minimax problems. Thus, more research is needed.
Response 2. Upper Bounds.
For the NC-C setting, GDA [6], AGDA [7] and GDmax [8, 9] are analyzed to guarantee convergence. To the best of our knowledge, the best complexity for first-order deterministic NC-C optimization is . The works [10, 11] have further extended to the stochastic setting and achieved the complexity matching that of the deterministic setting. For example, SAPD+ [10] achieved the gradient complexity of . As the complexity of ZO methods is usually times that of first-order methods [1], in this sense, it suggests our results match the state-of-the-arts in the deterministic setting.
For the stochastic NC-SC setting, SGDA, SAGDA and SODGA all achieved the gradient complexity of . The works [12][13] reduced the complexity to . Recently, SAPD+ has achievesd a complexity of , which suggests that the upper bounds in this paper can possibly be improved. Unfortunately, there is no similar conclusion as in [1] for NC minimax problems so far. Moreover, existing ZO methods such as [13][14] mainly focus on trivial extensions of existing first-order algorithms, while this paper focuses on designing ZO minimax algorithms and directly improving the robustness and convergence rate, which is a non-trivial extension.
We will add the discussion of upper bounds for first-order methods to our revised version.
Response 3. Other Questions.
To address your concern, we try to analyze our algorithm under two-point estimation, and we find that our algorithm still maintains the advantages of two-point estimation, that is, the second moment of gradient estimate is essentially linear in the dimension . In our revised version, we will clarify this.
Updating does not require the intermediate point , but the sequence is needed as an intermediate point to update . Thus, this is not a typo.
Besides, we will add dimension to the complexity stated in the abstract of the revised version.
Response 4. Our proof is correct.
After our careful inspection, we believe that all our proofs are correct.
[1] Optimal rates for zero-order convex optimization: The power of two function evaluations[J]. IEEE Transactions on Information Theory, 2015, 61(5): 2788-2806.
[2] Lower complexity bounds of first-order methods for convex-concave bilinear saddle-point problems[J]. Mathematical Programming, 2021, 185(1): 1-35.
[3] On lower iteration complexity bounds for the saddle point problems[J]. arXiv preprint.—2018.—https://arxiv. org/pdf, 1812.
[4] The complexity of nonconvex-strongly-concave minimax optimization[C]//Uncertainty in Artificial Intelligence. PMLR, 2021: 482-492.
[5] Complexity lower bounds for nonconvex-strongly-concave min-max optimization[J]. Advances in Neural Information Processing Systems, 2021, 34: 1792-1804.
[6] On gradient descent ascent for nonconvex-concave minimax problems[C]//International Conference on Machine Learning. PMLR, 2020: 6083-6093.
[7] Alternating proximal-gradient steps for (stochastic) nonconvex-concave minimax problems[J]. SIAM Journal on Optimization, 2023, 33(3): 1884-1913.
[8] Minimax optimization: stable limit points of gradient descent ascent are locally optimal (2019). arXiv preprint arXiv:1902.00618.
[9] What is local optimality in nonconvex nonconcave minimax optimization? In International conference on machine learning, pages 4880–4889. PMLR, 2020.
[10] Sapd+: An accelerated stochastic method for nonconvex-concave minimax problems. Advances in Neural Information Processing Systems, 35:21668–21681, 2022.
[11] Adagda: Faster adaptive gradient descent ascent methods for minimax optimization. In International Conference on Artificial Intelligence and Statistics, pages 2365–2389. PMLR, 2023.
[12] Stochastic recursive gradient descent ascent for stochastic nonconvex-strongly-concave minimax problems[J]. Advances in Neural Information Processing Systems, 2020, 33: 20566-20577.
[13] Enhanced first and zeroth order variance reduced algorithms for min-max optimization[J]. 2020.
[14] Zeroth-order alternating randomized gradient projection algorithms for general nonconvex-concave minimax problems[J]. arXiv preprint arXiv:2108.00473, 2021.
Thanks for your detailed response! I will increase my score to 6 because of these detailed discussions and the efforts you took during the rebuttal phase. I have no further comments.
We appreciate your review and valuable questions that improved our work.
This paper proposes zeroth-order method called ZO-GDEGA to find a near-stationary point for nonconvex-concave minimax optimization, with complexity guarantee. The proposed method is also extended to stochastic setting, being the first work on ZO method on stochastic NC-C problem. The method has weaker requirement on ZO gradient estimate, thus also has better dependency on condition number in the special case of NC-SC. Numerical results on data poisoning attack and AUC maximization show the proposed method is comparable and usually slightly better than compared baselines.
优点
-
This work designs a unified single-loop ZO method for NC-C minimax, with better complexity and more robust allowance on ZO gradient estimate. The proposed idea of continuous-time dynamics to assist with updates of dual variable and its related analysis is novel. Overall complexity is proposed, with solid and rigorous proof, under weak assumptions such as not requiring lipschitz continuity, more tolerant ZO gradient, which also results in good complexity in the NC-SC special case.
-
The proposed method can be extended to stochastic setting, with first-ever complexity in this case.
缺点
-
Complexity (deterministic NC-C): The work claims the complexity of the proposed method is a 'reduced' complexity, however to the best of my knowledge, this is not the best-known complexity of ZO method on NC-C minimax. Even for single-loop methods, the existing method in [43] shown by table 1 has a better complexity. Intuitively, only a complexity as good as this existing is near-optimal, because first-order methods on NC-C minimax have complexity. Although assumption-wise, as this work claims, [43] has the extra assumption of 'decreasing regular parameter sequence', and also requires more accurate ZO gradient on the primal variable, but in my opinion, the extra assumption above is not too restrictive, and requiring more accurate ZO gradient would only cost a constant multiple of queries thus would not affect overall complexity. Therefore, the complexity in this work is not good enough, and is in fact worse than certain existing single-loop ZO methods.
-
Complexity (stochastic NC-C): I acknowledge this is the first-ever complexity of ZO method on stochastic NC-C minimax. However, the dependence is not near-optimal, since first-order methods on such problem just have complexity. ** Update: typo fixed, from to .
-
Listed numerical results show the proposed method is similar and generally only slightly better than compared baselines. The difference is not significant in both experiments.
-
Overall, both theoretical complexities and numerical results does not surpass existing methods, thus the contribution may not be strong enough for NeurIPS standard.
问题
-
Please address [Weaknesses 1].
-
Please address [Weaknesses 2].
局限性
N/A
We would like to thank the reviewer for the insightful comments! Below we will clarify the three points in the review.
Response 1. Complexity (deterministic NC-C)
Our ZO-GDEGA achieved the reduced complexity. The key reason for different complexity is different complexity metrics. That is, our method can achieve the complexity of under a more difficult convergence criterion compared with [43]. As we discussed in the tablenote of Table 1, the work [43] achieved an -stationary point of . According to [27, Proposition 4.12] and our Propositions 1 and 2, the -stationary point in terms of is weaker than that of . Thus, [43] found a weaker -stationary point with lower gradient complexity. For a fair comparison, we can convert their complexity in terms of into complexity in terms of , that is, [43] requires the overall complexity of to find an -stationary point of . In this sense, our ZO-GDEGA algorithm reduced the complexity of existing methods.
Assumptions. In addition to satisfying the assumption `decreasing regular parameter sequence', the work [43] requires smaller smoothing parameters (i.e., more accurate ZO gradient) to achieve their gradient complexity. There are two dilemmas. Firstly, the constant multiple of queries may require a lot of experimental verification. Secondly, although satisfactory accuracy ZO gradient can be achieved through constant multiple of queries, smoothing parameters are required to the order of , which is more sensitive than that of our ZO-GDEGA, which means weaker robustness than our methods. In summary, [43] requires the assumption ``decreasing regular parameter sequence'' and has weaker robustness. In the experiment, we also found that the ZO-AGP [43] is more sensitive to smoothing parameters than our ZO-GDEGA. Please see Fig. C.8 in the Appendix for details.
Response 2. Complexity (stochastic NC-C)
To the best of our knowledge, the state-of-the-art complexity of the first-order algorithm is [1][2] for solving stochastic NC-C problems, but there is currently no lower bound for ZO algorithms solving stochastic NC-C problems. One of the contributions of this paper is that our ZO-SGDA firstly achieves the complexity of in the stochastic NC-C setting. As the complexity of ZO methods is usually times that of first-order methods [3], there may be a gap for NC minimax problems. We will bridge this gap by recursive momentum variance reduction technology [4] in our future work.
Response 3. Significant improvements in theory and experiment
In the data poisoning attack experiment, our ZO-GDEGA can reduce the accuracy by about 2%-8% compared with the baseline, which is a significant improvement. In this experiment, the lower the accuracy, the better the algorithm performs. More experimental results and analysis are provided in the Appendix. In summary, our theoretical complexities and numerical results outperform existing methods.
[1] Non-convex min-max optimization: Provable algorithms and applications in machine learning[J]. ArXiv, abs/1810.02060, 2018.
[2] Sapd+: An accelerated stochastic method for nonconvex-concave minimax problems[J]. Advances in Neural Information Processing Systems, 2022, 35: 21668-21681.
[3] Optimal rates for zero-order convex optimization: The power of two function evaluations[J]. IEEE Transactions on Information Theory, 2015, 61(5): 2788-2806.
[4] Momentum-based variance reduction in non-convex sgd[J]. Advances in neural information processing systems, 2019, 32.
I have read the authors' response. I suggest those valuable discussions could be added in appendix to clarify each contribution. While I agree with the theoretical claims in the rebuttal, however each weakness in my original comment still holds. Thus I will keep my score.
Thanks for your comments again. We will add the above valuable discussions in our revised version. We try again to address the weaknesses you mentioned. We further clarified our contributions, and stated our challenges.
Addressed the weakness
Deterministic NC-C. Under the complexity metric , the state-of-the-art gradient complexity is in the deterministic setting [27]. Thus, our ZO-GDEGA achieves the near-optimal gradient complexity in terms of metric . As a comparison, [43] achieves the gradient complexity in terms of metric . In this sense, our ZO-GDEGA is near-optimal and achieved the 'reduced' complexity.
Stochastic NC-C. One of our main contributions is that we first provide the provable complexity framework in the stochastic NC-C setting. Although near-optimal complexity may not (because to the best of our knowledge, no work proves that the complexity can be achieved) be reached, our analysis opens the door to analyzing the ZO NC-C setting.
Our contributions
We propose a faster, more robust, and theoretically guaranteed algorithm framework that can solve wider minimax applications.
Stronger robustness. Developing a ZO minimax algorithm for stronger robustness is challenging, as existing ZO algorithms require a demanding ZO estimate of the gradients, i.e., very small (e.g., [18] [42], [43]) smoothing parameters. We find that the smoothing parameter bounds have a key impact on the ZO algorithms for black-box minimax problems, which is shown in C.1.2 in the Appendix. Previous research did not focus on this key issue. We designed a novel ZO-GDEGA algorithm to weaken restrictions on the smoothing parameters (see Table 1) and offer the first provable guarantee. Our ZO-GDEGA has also been extended to novel FO methods, which are shown in Algorithms 3 and 4 in the Appendix. Thus, our algorithm is a non-trivial extension of the existing algorithms.
Lower complexity. We develop a new and concise analysis framework to prove that our ZO-GDEGA can achieve lower complexity under the same complexity metrics. It is the first ZO algorithm with theoretical guarantees for solving stochastic NC-C problems. Algorithms without theoretical guarantees are unstable and untrustworthy. The stochastic version of ZO-AGP [5] has no provable complexity guarantees. Specifically, in the proof of deterministic ZO-AGP, some parameters are severely coupled, and the stochastic setting will further complicate its analysis, while our proof framework can be easily extended to the stochastic setting.
Wider applications. We expand the scope of existing black-box minimax problems by considering more general regularizers rather than being limited to and , which can be instantiated as data poisoning attack, AUC maximization, and robust neural network training (see C.3 in the Appendix) problems.
Better experimental results. Our algorithms can find much more effective attack perturbation and perform more robustly.
Analysis of Challenges
We propose a novel complexity analysis framework for ZO minimax optimization. For minimax problems, the errors and are coupled with each other, which leads to difficult parameter selection. Taking the NC-C setting as an example, the recursion requires the error and requires the error . We handle them by constructing reasonable recursions and step sizes. Our Proposition 6 plays a key role and clarifies that the EG structure can result in the coefficient for and for , which allows to choose reasonable step sizes and larger smoothing parameters than existing methods, thus achieving more robust performance.
Even our analysis for the FO counterpart still faces difficulties due to the EG and the proximal operator. In the NC-SC setting, we propose to establish the upper bound on in terms of to construct the recursive relation of in Lemma B.9, thereby achieving enhanced robustness. In the NC-C setting, we creatively proposed Proposition 6 to handle the proximal operator and EG structure, thereby resulting in the reasonable recursion w.r.t. .
[5] Derivative-free alternating projection 490 algorithms for general nonconvex-concave minimax problems. 2023.
Thanks for the detailed responses. I agree the proposed methods improves the complexities compared to existing zeroth-order methods, and that is good. My point of weakness, which still holds, was that the proposed complexities are not tight, worse than d * first-order complexities; where existing first-order methods have complexity for deterministic NC-C and complexity for stochastic NC-C. I hope the authors understand this point, and potentially tighten those zeroth-order bounds in future works. With that being said, I carefully read the detailed analysis and additional experiments in the appendix, which made good efforts in making the theory and experiments comprehensive, and the analysis is nontrivial. Thus, I have increased my score from 5 to 6.
We appreciate your review and valuable questions that improved our work.
In this paper, the authors establish a unified framework of zeroth-order optimization for nonconvex-concave minimax optimization problems in both deterministic and stochastic settings. This framework is based on the gradient descent-extragradient ascent algorithm. They claim that their algorithms require weaker assumptions on the zeroth-order estimator, while achieve competitive iteration complexity compared with the existing work. Besides, they provide some numerical experiments to verify the effectiveness of the proposed algorithms.
优点
Main advantages can refer to the Summary. Besides, this paper has a good organization, which makes the readers reading easily. Numerous experiments including abundant baselines and datasets are provided to illustrate the effectiveness of the proposed algorithm.
缺点
In the stochastic setting, the authors make the bounded variance assumption on the zeroth-order estimator in Assumption 3. It seems that the assumption makes the proof very similar to first-order method. It is true that the zeroth-order methods are used when computing derivative are not possible, but does this assumption rules out the most challenging technical part in the proof? This paper has proposed zeroth-order method that is motivated by application scenes when the objective function is not smooth enough to enable one to compute the gradient. But there is no specific examples provided that such method plays a prominent role while the first-order method is not applicable. Since there is a definite trade-off when zeroth-order method are used compared to first-order, the smooth parameter is another factor that should be considered in using zeroth-order method. So I think providing a concrete application that zeroth-order method is inevitable will help convincing the importance of the algorithm in application.
问题
Similar to aforementioned concerns on example, the experiments of poisoning data attack is a clear application that zeroth-order can be used. My question is whether it is possible to use first-order method in some cases? If this is the application that no first-order method can be used, then my previous conern on weakness can be addressed. I don't here mean that the authors should have compare zeroth-order method to first-order method in performance, since it is completely reasonable that first-order method outperforms zeroth-order and even it holds, nothing negative will impact the contribution of the paper.
局限性
This paper is mostly a theoretical work, no negative societal impact may result in.
We would like to thank the reviewer for the insightful thoughts and comments! Below we will clarify the two points in the review.
About Assumption 3.
Firstly, analyzing ZO algorithms has the following two main difficulties compared to analyzing first-order methods. That is, we need to bound two error terms, the error (or ) and (or ). We first bound the first error by Assumption 3. The same assumption also appears in the classic work [1]. In this paper, we mainly focus on solving the second difficulty. We handle the second error of the ZO EG estimates by constructing reasonable recursions. Taking the NC-C setting as an example, our Proposition 6 plays a key role and clarifies that the EG structure can result in the coefficient for and for (please see the inequality B.130 for detail), which allows us to choose larger smoothing parameters (i.e., and ) than existing methods, thus achieving more robust performance.
Secondly, due to the introduction of EG and proximal operator in our ZO-GDEGA, our first-order analysis framework still faces difficulties. That is, the EG and proximal operator of our algorithm make the analysis framework more difficult. Taking the NC-SC setting as an example, we propose to establish the upper bound on in terms of to construct a recursive relation, thereby achieving competitive complexity and enhanced robustness.
In summary, there are still many challenges in our ZO theoretical analysis, except for Assumption 3. To address these difficulties, we provide corresponding solutions in our analysis framework. We will also provide a better upper bound for Assumption 3 in our future work.
The first-order methods can not be used for the poisoning data attack application
Because attackers do not know the internal structure of the model, this application is a black box. Thus, the first-order methods can not access gradient information and are not applicable. We will clarify this in the revised version as well.
[1] Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to Minimax Optimization. The Journal of Machine Learning Research, 2022
Thank you for your time and valuable comments, especially regarding Assumption 3,which is indeed strong and also used in most existing works, such as [1] and [3]. To address your concern, we further improved the variance bound and corresponding theoretical results of our algorithms.
We first derive the following variance bound in the following inequalities (1) and (2) to replace Assumption 3.
and
Furthermore, we use the (1) and (2) instead of Assumption 3 under our analysis framework, and adjust these parameters , , and by adjusting the smoothing parameters , , our ZO-GDEGA still obtains the complexity of for the NC-SC problem. As for the NC-C setting, we can set , , and by setting the smaller smoothing parameters , , our ZO-GDEGA algorithm can still obtain the complexity of .
Thus, these derivation shows that Assumption 3 is a strong assumption. We will add this discussion in our revised version.
[1] Accelerated Zeroth-Order and First-Order Momentum Methods from Mini to Minimax Optimization. The Journal of Machine Learning Research, 2022
[2] Min-Max Optimization without Gradients: Convergence and Applications to Black-Box Evasion and Poisoning Attacks.
[3] Derivative-free alternating projection 490 algorithms for general nonconvex-concave minimax problems. 2023.
Thanks for the reply, after the rebuttal I will keep my score and recommend acceptance for this submission.
We appreciate your review and valuable questions that improved our work.
We would like to thank the reviewers for the detailed reviews. We were asked by the reviewer 4438 to provide code. Our code can be available at the link https://anonymous.4open.science/r/ZO-GDEGA-2F6E
The paper introduces a Zeroth-Order Gradient Descent Extragradient Ascent (ZO-GDEGA) algorithm, which is designed to solve nonconvex-concave (NC-C) minimax optimization problems. The proposed algorithm has iteration complexity of O(epsilon^-6) to find ϵ-stationary point. The algorithm requires weaker conditions on zeroth-order gradient estimations compared to previous methods, offering a more broad/robust theoretical result. It also shows advantages in the nonconvex-strongly concave (NC-SC) case. Experimentally, the paper validates the the performance of ZO-GDEGA through numerical experiments on data poisoning attacks and AUC maximization.
The reviewers found the theoretical analysis rigorous and in line with the claims of the paper. The reviewers also appreciate the relaxed conditions required for applying the theoretical analysis. The experiments are found in line with the message of the paper, demonstrating the relevance of the theory in practical scenarios discussed. As noted by some of the reviewers, the improvements over [43] needs further discussion. I encourage the authors to clarify this improvement further as mentioned in their response (You can use [27, Proposition 4.12] or a a clear version of it with more discussions in arxiv 2002.07919). The reviewers also noted that the paper can improve by including discussions on lower bounds and connections to first-order methods.