PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
5
4
5
4
3.5
置信度
创新性2.0
质量2.8
清晰度3.0
重要性2.5
NeurIPS 2025

Optimal Rates for Generalization of Gradient Descent for Deep ReLU Classification

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29
TL;DR

we established optimal risk bound of $1/(\gamma^2 n)$ for GD with deep ReLU networks

摘要

Recent advances have significantly improved our understanding of the generalization performance of gradient descent (GD) methods in deep neural networks. A natural and fundamental question is whether GD can achieve generalization rates comparable to the minimax optimal rates established in the kernel setting. Existing results either yield suboptimal rates of $O(1/\sqrt{n})$, or focus on networks with smooth activation functions, incurring exponential dependence on network depth $L$. In this work, we establish optimal generalization rates for GD with deep ReLU networks by carefully trading off optimization and generalization errors, achieving only polynomial dependence on depth. Specifically, under the assumption that the data are NTK separable from the margin $\gamma$, we prove an excess risk rate of $\widetilde{O}(L^4 (1 + \gamma L^2) / (n \gamma^2))$, which aligns with the optimal SVM-type rate $\widetilde{O}(1 / (n \gamma^2))$ up to depth-dependent factors. A key technical contribution is our novel control of activation patterns near a reference model, enabling a sharper Rademacher complexity bound for deep ReLU networks trained with gradient descent.
关键词
statistical learning theorygenerlization analysisSGDneural tangent kerneloptimal rates

评审与讨论

审稿意见
5

Improved bounds on the excess risk rate for deep ReLU networks.

优缺点分析

Strengths. Improved bounds wrt those known that are polynomial in depth. Weaknesses. No attempt at an experimental confirmation, but they are on time to present it for the rebuttal.

问题

  • gradient decent -> gradient descent, line 57

局限性

The limitations are the assumptions of their theorems, which are clearly stated.

最终评判理由

I confirm the positive evaluation. The point of saying that their contribution is only theoretical is weak, since to my knowledge no work in NeurIPS is only theoretical. The authors could have missed a mathematical step, that the reviewers could have also missed, so a few experimental confirmations never hurt. To my understanding however the article is strong enough to be accepted.

格式问题

Correct.

作者回复

Thanks for acknowledging the strengths of our work. Our point-to-point response to your comments follows next.

Strengths And Weaknesses Strengths. Improved bounds wrt those known that are polynomial in depth. Weaknesses. No attempt at an experimental confirmation.

  • The improvement over LL is only part of our contribution. The main result is that we are the first to prove that deep ReLU networks trained by gradient descent can achieve optimal rate (1/n1/n). The existing works either obtain a bound of suboptimal rate 1/n1/\sqrt{n} or focus on smooth activations.
  • This is a paper on learning theory. We focus on whether and how deep ReLU networks could achieve optimal generalization error. Hence, we do not take experiments. We believe we have improved the understanding of deep neural networks theoretically.

Q. gradient decent -> gradient descent, line 57

Thank you for pointing it out. We will revise it in the next version.

审稿意见
4

This paper studies the convergence and generalization analysis of deep fully connected neural networks. Specifically, this paper proves that GD achieves a sublinear convergence rate in the order of 1/T1/T. Furthermore, the derived generalization bound exhibits an optimal rate of 1/n1/n with respect to the sample size nn. Additionally, the dependence on LL has been improved from an exponential to a polynomial order by establishing a tighter Rademacher complexity bound.

优缺点分析

Strength:

  1. A tighter and more practically relevant generalization bound.

  2. A thorough comparison with existing works.

Weakness:

Overall, I feel that while this paper shows improvements over prior work, it focuses on a less meaningful direction. The insights gained from this improvement are either unclear or do not align well with what is observed in practice. As a result, I am not convinced that this setting is appropriate for studying the generalization error of deep neural networks.

  1. The assumption that the data is NTK-separable with a margin γ\gamma lacks explanation in practical contexts. Specifically, what kinds of data will succeed or fail in this assumption? In addition, this assumption is extremely strong in the sense that it requires the network to be initialized in such a way that the data, when processed through this predefined model, is already linearly separable.

  2. Compared with prior work, the main difference I observe is that the generalization error bound has been improved from an exponential dependence to a polynomial dependence. However, the key insight remains the same as in previous studies: increasing the number of layers leads to worse generalization. I do not see any fundamentally new insights presented in this paper.

  3. The theoretical insights presented in this paper do not adequately explain the current success and widespread effectiveness of deep learning models that rely on large and deep architectures.

  4. The paper does not include any experiments to validate the theoretical results it presents.

问题

Since the major contribution of this paper is the improvement in the dependence of the generalization bound with respect to LL, I have several questions:

  1. What are the key insights that this paper aims to convey through this improved bound?

  2. If Assumption 3 holds for real-world data, what would be the practical difference between using a deep neural network versus a shallow one? For example, how does LL relate to γ\gamma in this context?

  3. Could you specify lines that reflect "novel control of activation patterns near a reference model"?

局限性

Yes

最终评判理由

I increased the score from 3 to 4 because the author has addressed my concerns regarding the gap between the theoretical results and real applications by providing numerical experiments. However, I still have concerns about Assumption 3, which deviates significantly from real-world applications.

格式问题

N/A

作者回复

Thanks for acknowledging the strengths of our work. Our point-to-point response to your comments follows next.

W2.&Q1. Compared with prior work, the main difference I observe is that the generalization error bound has been improved from an exponential dependence to a polynomial dependence. However, the key insight remains the same as in previous studies: increasing the number of layers leads to worse generalization. I do not see any fundamentally new insights presented in this paper. What are the key insights that this paper aims to convey through this improved bound?

  • The improvement over LL is only part of our contribution. Our key insight is that we are the first to prove that deep ReLU networks trained by gradient descent can achieve optimal rate (1/n1/n). The existing works either obtain a bound of suboptimal rate 1/n1/\sqrt{n} or focus on smooth activations. For more details, please refer to Table 1 of our paper.
  • Despite the success of deep learning in practice, understanding it in theory is often challanging due to its nonconvexity and non-smoothness. We address these problems by developing two novel techniques. Firstly, we derive a tighter Rademacher complexity bound for deep ReLU networks based on the control of activation patterns. Secondly, we prove the O~(L2)\widetilde{O}(L^2) Lipschitzness of deep ReLU networks. These key improvements open the door to further analysis of deep learning, which is beyond the scope of this paper.

W1. The assumption that the data is NTK-separable with a margin γ\gamma lacks explanation in practical contexts. Specifically, what kinds of data will succeed or fail in this assumption? In addition, this assumption is extremely strong in the sense that it requires the network to be initialized in such a way that the data, when processed through this predefined model, is already linearly separable.

We would like to clarify the motivation and interpretation of the NTK-separability assumption, which has been widely used in the literature [2,3,4,5].

  • Role of the NTK separability assumption:

The NTK separability assumption is used in our paper as a tool to analyze F(W)F(\overline{W}) in Theorem 2, which reflects the performance achievable by the network class. This assumption enables a clean characterization of generalization in Theorem 2. Importantly, our main result (Theorem 2) does not require data to be NTK-separable—it holds for any F(W)F(\overline{W}) with bounded complexity. Thus, NTK separability is not a limiting assumption of our theory, but rather a benchmark scenario to make comparisons with existing literature.

  • We believe Assumption 3 is not a strong assumption

NTK separability has been theoretically justified in a range of settings. For instance, [2] analyzes the margin through its dual form in different settings, including linearly separable case and the noisy 22-XOR distribution. [8] also shows that NTK can learn polynomials. Hence, Assumption 3 holds for data that are separable by polynomials. In practice, it can be verified through random features by solving a regularized linear classification problem.

Q2. If Assumption 3 holds for real-world data, what would be the practical difference between using a deep neural network versus a shallow one? For example, how does LL relate to γ\gamma in this context?

This is a good question. In standard NTK theory, the margin γ\gamma does not explicitly depend on LL. However, we conjecture that deeper networks tend to induce larger NTK margins γ\gamma. Our intuition is based on the structure of the NTK and its associated reproducing kernel Hilbert space (RKHS).

  • Let Θ(L)\Theta^{(L)} denote the NTK of an LL-layer ReLU network, and HLH^{\infty}_L its corresponding RKHS. As shown in [6], the NTK satisfies the recursive relation: Θ(L+1)(x,y)=Θ(L)(x,y)Σ˙(L+1)(x,y)+Σ(L+1)(x,y)\Theta^{(L+1)}(x,y) = \Theta^{(L)}(x,y)*\dot{\Sigma}^{(L+1)}(x,y) + \Sigma^{(L+1)}(x,y), where Σ˙(L+1)\dot{\Sigma}^{(L+1)} is some derivative covirance matrix and Σ(L+1)\Sigma^{(L+1)} is the covariance matrix of a Gaussian process. This implies that Θ(L+1)\Theta^{(L+1)} contains Θ(L)\Theta^{(L)} as part of its structure. Consequently, the RKHSs satisfy a nested relationship: HLHL+1H^{\infty}_L \subset H^{\infty}_{L+1}.
  • Consider the finite-width version: HLm={f:WW s.t. f(x)=W,fW(0)(x)/W(0)}.H^m_L = \{f : \exists\, W \in \mathcal{W} \text{ s.t. } f(x) = \langle W, \partial f_{W(0)}(x)/\partial W(0) \rangle \}. As m,HLmHLm\to \infty, H^m_{L}\to H^\infty_{L} (see [7]). Therefore, HLmHL+1mH^m_L \subset H^m_{L+1}, and an (L+1)(L+1)-layer deep network has a larger NTK-margin than an LL-layer network, according to the definition of γ\gamma.

Taken together, these suggest that deeper networks admit richer RKHSs, which can better separate the data and thus lead to a larger margin γ\gamma. Therefore, although γ\gamma is not explicitly parameterized by LL, it is indirectly affected through the expressiveness of the induced RKHS. We emphasize that this argument remains heuristic, and we leave a precise characterization of how γ\gamma scales with LL as an important open question for future work.

W3. The theoretical insights presented in this paper do not adequately explain the current success and widespread effectiveness of deep learning models that rely on large and deep architectures.

Our theoretical framework could not explain the effectiveness of large and deep models adequately, but we provide a partial explanation here.

  • In Theorem 2, we establish a generalization bound of the form O~(L4F(W)n)\widetilde{O}\left(\frac{L^4 F(\overline{W})}{n}\right), where F(W)F(\overline{W}) represents the risk of a well-structured (but fixed) reference model near initialization in the same network class. As long as there exists a deep network with small F(W)F(\overline{W}), gradient descent is guaranteed to find a predictor with low generalization error.
  • Importantly, the quantity F(W)F(\overline{W}) is directly tied to the expressive power of the network class. For more complex data distributions, a shallow network may not be able to achieve a small F(W)F(\overline{W}), while a deep architecture can do so due to its ability to capture hierarchical structures. In particular, existing results (e.g., [1]) show that deep networks can approximate certain function classes with exponentially fewer parameters than shallow ones. Therefore, our theory does capture the advantage of depth: deeper architectures can produce smaller F(W)F(\overline{W}), and Theorem 2 ensures that gradient descent can generalize well in such cases. We believe this provides an explanation for why deep models succeed in practice.
  • In addition, our analysis is conducted in the overparameterized regime, where the network width mm is large. Most of our results require mpoly(L,logn)m \gtrsim \mathrm{poly}(L, \log n) to ensure good optimization behavior (e.g., descent direction, weak convexity). From this perspective, large models also help with optimization—they make gradient-based methods more effective, which partially explains why large-scale networks work well in practice.

Q3. Could you specify lines that reflect "novel control of activation patterns near a reference model"?

  • In Lemma 22, we express the Rademacher complexity via products of sparse matrices, whose norms are tightly controlled using optimization informed estimates. This helps to derive a sharper bound, line 704. Please also refer to remark 2, line 214.
  • Specifically, we first levearage the useful expression fW(xi)fW(0)(xi)=al=1LG^L,0l(xi)(WlWl(0))h0l1(xi).f_W(x_i) - f_{W(0)}(x_i) = \mathbf{a}^\top \sum_{l=1}^L \widehat{G}_{L,0}^l(x_i) (W^l - W^l(0)) h_0^{l-1}(x_i).

Then we control G^L,0l(xi)\widehat{G}_{L,0}^l(x_i) by Theorem 1 and Lemma 15. This leads to a tight bound of Rademacher complexity O~(L2F(W)n)\widetilde{O}\left(L^2\sqrt{\frac{F(\overline{W})}{n}}\right).

  • However, previous works use the bound on the shelf, ignoring how activation patterns would affect the complexity. For example, [4] derive the bound of O~(4LLmF(W)n)\widetilde{O}\left(4^LL\sqrt{\frac{mF(\overline{W})}{n}}\right), which suffers from exponential term and explicit dependence on mm.

W4. The paper does not include any experiments to validate the theoretical results it presents.

This is a paper on learning theory. We focus on whether and how deep ReLU networks could achieve optimal generalization error. Hence, we do not take experiments. We believe we have improved the understanding of deep neural networks theoretically.

[1] Telgarsky et al. Benefits of depth in neural networks.

[2] Ji et al. Polylogarithmic width suffices for gradient descent to achieve arbitrarily small test error with shallow relu networks.

[3] Taheri et al. Generalization and stability of interpolating neural networks with minimal width.

[4] Chen et al. How much over-parameterization is sufficient to learn deep relu networks?

[5] Taheri et al. Sharper guarantees for learning neural networkclassifiers with gradient methods.

[6] Arora et al. On Exact Computation with an Infinitely Wide Neural Net.

[7] Xu et al. Overparametrized multi-layer neural networks: Uniform concentration of neural tangent kernel and convergence of stochastic gradient descent.

[8] Arora et al. Fine-grained analysis of optimization and generalization for overparameterized two-layer neural networks.

评论

Thanks for the rebuttal. I need to point out that I respectfully disagree with the last point. If the theoretical results cannot be justified in real applications, how can we trust and utilize these theoretical insights to genuinely push understanding or guide the design of practical systems? As you mentioned, your goal is to improve the theoretical understanding of deep neural networks (DNNs). However, any results derived under assumptions that significantly deviate from real-world conditions risk being detached from practice. For example, if applying gradient-based methods in real-world applications cannot practically achieve a generalization error that scales on the order of 1/γ2n1/{\gamma^2 n}, then what is the point of establishing such a result? While the theoretical bound might be mathematically elegant, if it cannot be observed under realistic training conditions or architectures, its relevance becomes questionable.

Unless you can justify your assumptions and ensure that the model architecture aligns with real-world applications, it is essential to include numerical experiments to support any theoretical results based on those assumptions.

评论

Thank you for your constructive comments. We agree that numerical experiments are valuable in supporting our theoretical analysis. Accordingly, we have conducted some experimental verifications to support our excess risk analysis.

Our excess risk analysis in Theorem 3 imposes an NTK separability assumption, which has been validated in the literature. For example, the work [1] demonstrates that Assumption 3 holds for a noisy 2-XOR distribution, where the dataset is structured as follows:

(x1,x2,y,,xd)(1d1,0,1),(0,1d1,1),,(1d1,0,1),(0,1d1,1)×1d1,1d1d2.(x_1, x_2, y, \ldots, x_d) \in \\{ (\tfrac{1}{\sqrt{d-1}}, 0, 1), (0, \tfrac{1}{\sqrt{d-1}}, -1), \ldots, (-\tfrac{1}{\sqrt{d-1}}, 0, 1), (0, -\tfrac{1}{\sqrt{d-1}}, -1) \\} \times \\{-\tfrac{1}{\sqrt{d-1}}, \tfrac{1}{\sqrt{d-1}}\\}^{d-2}.

Here, the factor 1d1\frac{1}{\sqrt{d-1}} ensures that x2=1\|x\|_2 = 1, ×\times above denotes the Cartesian product, and the label yy only depends on the first two coordinates of the input xx. As shown in [1], this dataset satisfies Assumption 3 with 1/γ=O(d)1/\gamma=O(d), which implies that our excess risk bound in Theorem 3 becomes O(d2/n)O(d^2/n) for this dataset. We conducted numerical experiments and observed that the test error decays linearly with d2/nd^2/n. The population loss for the test error is computed over all 2d2^d points in the distribution.

  • We train two-layer ReLU networks by gradient descent on noisy 2-XOR data. We fix the width m=128,T=500,η=0.1m=128, T=500,\eta = 0.1.
  • With a fixed dimension d=6d = 6, we vary the sample size nn and obtain the following results
nnd2/nd^2/nTest Error
103.600.4625
123.000.4500
142.570.3625
162.250.3438
182.000.3219
201.800.3063
241.500.2469
281.290.2062
  • With a fixed sample size n=64n=64, we vary the dimension dd and obtain the following results
ddd2/nd^2/nTest Error
70.770.0125
81.000.1313
91.270.2484
101.560.3080
111.890.3365
122.250.4190

In both experiments, we observe that the test error is of the order d2/nd^2/n (approximately 0.15d2/n0.15d^2/n). This shows the consistency between our excess risk bounds in Theorem 3 and experimental results. We will include these numerical experiments in the revised version. We will also conduct more experimental verifications.

[1] Ji, Z., & Telgarsky, M. (2019). Polylogarithmic width suffices for gradient descent to achieve arbitrarily small test error with shallow relu networks.

评论

Thank you for the additional experiments. I have increased my score from 3 to 4.

评论

Dear Reviewer usBV,

Thank you for recognizing the contribution of our paper. We really appreciate your constructive suggestions and comments.

Best regards, Authors

审稿意见
5

The authors establish a convergence rate for Gradient Descent applied on deep neural network with rail activations that is polynomial in the depth of the network LL.

优缺点分析

Strengths

  1. The paper is well written and easy to follow.
  2. The authors present their proof technique and the relation to prior work in a very clear manner.
  3. The result is significant and improves the exponential bound given in previous work.

Weaknesses

  1. My main concern is about the correctness of Lemma 16, which is a central component of the paper. While Lemma 15 is proved only for the data points $x_i$ in the training set, Lemma 16 applies it to points in the covering set $D$, which includes points outside the training set. Why is this use of Lemma 15 justified? Could the authors elaborate on this point? If I’m mistaken, I will revise my score accordingly. Another reason this transition may not be valid is that the network’s weights depends on the training set. Therefore, a bound that holds for the training set may not necessarily apply to points outside it.

  2. Additionally, I am uncertain about the correctness of the proof of Lemma 15. In the inductive argument showing that AlO(Bl)A_l \leq O(B_l) , it is crucial that the constant hidden in the big-OO notation remains uniform across all induction steps. Otherwise, for example, if the constant grows multiplicatively at each step, the resulting bound would be exponential in ll. Could you clarify why this issue does not arise in the proof of Lemma 15?

  3. A minor concern: The authors obtain a bound only in the case where $m \geq d$.

问题

Why is it necessary to assume that R1R \geq 1?

局限性

Yes.

最终评判理由

The paper is clear and well written. The result is significant and improves results presented in previous work. For these reasons, I consider the paper to be above the acceptance threshold.

格式问题

N.A

作者回复

Thanks for acknowledging the strengths of our work. Our point-to-point response to your comments follows next.

W1. My main concern is about the correctness of Lemma 16, which is a central component of the paper. While Lemma 15 is proved only for the data points xix_i in the training set, Lemma 16 applies it to points in the covering set DD, which includes points outside the training set. Why is this use of Lemma 15 justified? Another reason this transition may not be valid is that the network’s weights depend on the training set. Therefore, a bound that holds for the training set may not necessarily apply to points outside it.

We appreciate the opportunity to clarify the connection between Lemma 15 and Lemma 16. We claim that all of our technical lemmas (including Lemma 15) hold for any finite set. In particular, Lemma 15 should be stated as follows:

  • Lemma 15'

Given a set of NN points on the sphere K={x1,,xN}K=\{x^1,\cdots,x^N\}. Suppose mL10log(NL/δ)(logm)4R2m\gtrsim L^{10}\log(NL/\delta)(\log m)^4R^2. Then with probability at least 1δ1-\delta over the randomness of initialization, for any WBR(W(0)),j[N],l[L]W \in \mathcal{B}_R(W(0)),j \in [N],l \in [L], there holds

hl(xj)h0l(xj)2L2Rlogmm,Σl(xj)Σ0l(xj)0L4/3(logm)1/3(mR)2/3.(1)\|h^l(x^j)-h^l_0(x^j)\|_2\lesssim L^2R\sqrt{\frac{\log m}{m}}, \|\Sigma^l(x^j)-\Sigma^l_0(x^j)\|_0\lesssim L^{4/3}(\log m)^{1/3}(mR)^{2/3}. \quad \text{(1)}

As K,NK,N vary, we could get concentration inequalites on different sets.

  • Applying Lemma 15' to the training set

We get the uniform control over nn points x1,,xnx_1,\cdots,x_n (the original Lemma 15): Suppose mL10log(nL/δ)(logm)4R2m\gtrsim L^{10}\log(nL/\delta)(\log m)^4R^2. Then with probability at least 1δ1-\delta over the randomness of initialization, for any WBR(W(0)),i[n],l[L]W \in \mathcal{B}_R(W(0)),i \in [n],l \in [L], there holds

hl(xi)h0l(xi)2L2Rlogmm,Σl(xi)Σ0l(xi)0L4/3(logm)1/3(mR)2/3(2).\|h^l(x_i)-h^l_0(x_i)\|_2\lesssim L^2R\sqrt{\frac{\log m}{m}}, \|\Sigma^l(x_i)-\Sigma^l_0(x_i)\|_0\lesssim L^{4/3}(\log m)^{1/3}(mR)^{2/3} \quad \text{(2)}.
  • Applying Lemma 15' to the covering set DD

Let D={x1,,xD}D= \{x^1,\cdots,x^{|D|}\} to be a 1/(CLm)1/(C^L\sqrt{m})-cover of the unit sphere (CC is a universal constant). Then we have: Suppose mL10log(DL/δ)(logm)4R2m\gtrsim L^{10}\log(|D|L/\delta)(\log m)^4R^2, then with probability at least 1δ1-\delta over the randomness of initialization, for any WBR(W(0)),j[D],l[L]W \in \mathcal{B}_R(W(0)),j \in [|D|],l \in [L], there holds

hl(xj)h0l(xj)2L2Rlogmm,Σl(xj)Σ0l(xj)0L4/3(logm)1/3(mR)2/3.(3)\|h^l(x^j)-h^l_0(x^j)\|_2\lesssim L^2R\sqrt{\frac{\log m}{m}}, \|\Sigma^l(x^j)-\Sigma^l_0(x^j)\|_0\lesssim L^{4/3}(\log m)^{1/3}(mR)^{2/3}.\quad \text{(3)}
  • Uniform control via covering

For any xSd1x\in \mathbb{S}^{d-1}, there exists xjDx^j\in D such that xxj21/(CLm)\|x-x^j\|_2\le 1/(C^L\sqrt{m}). It then follows that hL(x)hL(xj)21m,h0L(x)h0L(xj)21m\|h^L(x)-h^L(x^j)\|_2\le\frac{1}{\sqrt{m}}, \|h^L_0(x)-h^L_0(x^j)\|_2\le\frac{1}{\sqrt{m}}. Hence by triangle inequality we have hL(x)h0L(x)21m+L2Rlogmm\|h^L(x)-h^L_0(x)\|_2\lesssim \frac{1}{\sqrt{m}}+L^2R\sqrt{\frac{\log m}{m}}.

  • On training set dependence
  1. We use (2) to analyze the optimization and estimate the Rademacher complexity. This depends on the training set.
  2. We use (3) to prove Lemma 16, which depends on the covering set DD and is independent of the training set. Lemma 16 demonstrates the O~(L2)\widetilde{O}(L^2)-Lipschitzness of deep ReLU networks near initailization, which is independent of the optimization trajectory. The weights here are not updated via gradient descent; they remain within a neighborhood of the initialization. Lemma 16 is mainly used to derive a bound for GG' in Lemma 1. It is crucial to obtain a generalization bound that is only polynomial in LL (see Appendix C).
  3. As mL10(logm)4R2max(log(nL/δ),log(DL/δ))m\gtrsim L^{10}(\log m)^4R^2\max(\log(nL/\delta),\log(|D|L/\delta)), (2) and (3) hold simutaneously with probability at least 1δ1-\delta. Both of them are important and they contribute to our theory from different aspects.

We will revise Lemma 15 accoardingly and put (2) as a corollary in the next version.

Q. Why is it necessary to assume that R1R\ge1?

The assumption R1R \ge 1 is introduced for notational simplicity and does not affect the generality of our results. In particular, the bound 1m+L2Rlogmm\frac{1}{\sqrt{m}}+L^2R\sqrt{\frac{\log m}{m}} can be written as L2logmmmax(R,1)L^2\sqrt{\frac{\log m}{m}}\max(R,1). Indeed, all of RR in our results can be replaced by max(R,1)\max (R,1) by removing the assumption R1R\ge1.

W2. Additionally, I am uncertain about the correctness of the proof of Lemma 15. In the inductive argument showing that AlO(Bl)A_l\le O(B_l), it is crucial that the constant hidden in the big-OO notation remains uniform across all induction steps. Otherwise, for example, if the constant grows multiplicatively at each step, the resulting bound would be exponential in ll. Could you clarify why this issue does not arise in the proof of Lemma 15?

In our analysis, the constants hidden in the big-OO notation are uniform across all layers, i.e., they do not depend on the layer index ll. This is explicitly ensured by Lemma 12, which provides uniform bounds under overparameterization assumptions for all layers.

  • Specifically, in eq.(32) of our paper, the bound holds uniformly for all l[L]l \in [L]. In Lemma 15, we do not apply a direct recursive argument where each layer builds on the previous one with new constants. Instead, we rely on Lemma 12 to establish uniform control at each layer to finish the induction, assuming the preconditions hold (which are guaranteed by a sufficiently large width mm).
  • More concretely, we assume hl(xi)h0l(xi)2c1L2Rlogmm\|h^l(x_i)-h^l_0(x_i)\|_2\le c_1L^2R\sqrt{\frac{\log m}{m}},Σl(xi)Σ0l(xi)0c2L4/3(logm)1/3(mR)2/3,mc3L10log(nL/δ)(logm)4R2\|\Sigma^l(x_i)-\Sigma^l_0(x_i)\|_0\le c_2L^{4/3}(\log m)^{1/3}(mR)^{2/3},m\ge c_3L^{10}\log(nL/\delta)(\log m)^4R^2 in Lemma 15. Let the universal constant in eq.(32) be c4c_4. Within Lemma 12, let Σ^l(xi)0c5mL2logm,G^ba(xi)2c6Llogmm\|\widehat\Sigma^l(x_i)\|_0\le \frac{c_5m}{L^2\log m},\|\widehat G_b^a(x_i)\|_2\le c_6L\sqrt{\frac{\log m}{m}}. We only need to choose c1,c2,c3c_1,c_2,c_3 to satisfy the following inequalities (which are independent of ll): c12,c4+9c12c2,c3(2c2c5)3,2c6c1c_1\ge 2, c_4+9c_1^2\le c_2, c_3\ge(2c_2c_5)^3, 2c_6\le c_1. Hence, all constants are fixed prior to the induction, and do not accumulate as ll increases. This ensures that the final bounds in Lemma 15 remain polynomial in LL, as claimed.
  • Similar analyses also appeared in other works [1,2], which adopt uniform constants to control inductive arguments across layers.

W3. A minor concern: The authors obtain a bound only in the case where mdm\ge d.

  • We study the generalization performance of deep ReLU networks in the overparameterized regime, which aligns with common practical settings. In this case, the landscape is non-smooth and nonconvex, though the models can generalize well. Our work improves the understanding of gradient descent methods in deep neural networks. The assumption mdm\ge d has been widely adopted in the existing literature [1,2].
  • Moreover, the assumption that the network has sufficiently many neurons is often necessary to ensure expressivity in high dimensions. [3] shows that Ω(dp)\Omega(d^p) neurons are required for NTK to learn a degree-pp polynomial. In the feature learning regime, even for simple data distribution (e.g., Guassian or boolean), [4,5] demonstrate that neural networks need Ω(ds)\Omega(d^s) neurons to learn single-index or multi-index models, where ss denotes the information exponent or leap complexity.
  • We agree that it is an interesting question to study the performance of deep networks in the m<dm<d regime. However, doing so might require additional structural assumptions, such as sparsity in the data distribution, low-dimensional manifold structure, or specific architectural biases. We view this as a promising direction for future research.

[1] Zou et al. Stochastic gradient descent optimizes parameterized deep relu networks.

[2] Chen et al. How much over-parameterization is sufficient to learn deep relu networks?

[3] Ghorbani et al. When Do Neural Networks Outperform Kernel Methods?

[4] Bietti et al. Learning single-index models with shallow neural networks.

[5] Abbe et al. SGD learning on neural networks: leap complexity and saddle-to-saddle dynamics.

评论

Thank you for the clarification. The concerns I raised in my review have been addressed.

Following the rebuttal, I do have one additional question. The authors mention that “the landscape is non-smooth and non-convex.” However, the paper relies on Lemma 1 from Srebro (2010), which was originally proved under the assumption that the predictors are smooth. Could you please clarify why it is valid to apply this lemma in the non-smooth setting considered in the paper?

评论

Thank you for your reply. We are glad to hear that the concerns raised in your previous review have been addressed. We also appreciate your additional comments regarding the smoothness.

To clarify, Lemma 1 from Srebro (2010) only requires the loss function to be smooth; it does not impose any smoothness assumptions on the model itself. The key idea in Srebro (2010) is that the Lipschitz constant of the loss can be bounded by the loss value itself—known as the self-bounding property of smooth losses. This Lipschitz constant plays a crucial role in relating the covering numbers of the loss function class to those of the model class.

By leveraging this self-bounding property, we can bound the Lipschitz constant in terms of the training errors, which then leads to the formulation of Lemma 1 from Srebro (2010). This explains why the training error appears in the bound presented in Srebro (2010). Importantly, this approach depends solely on the smoothness of the loss function, regardless of whether the model itself (e.g., a neural network) is smooth or not.

In our work, we consider a smooth logistic loss. Therefore, Lemma 1 from Srebro (2010) remains applicable in our case. We will clarify this point in the revised version of the paper to avoid this confusion.

评论

Thanks for the clarification regarding the smoothness assumption.

After carefully reading the authors’ responses to all reviewers, I find the improvements in the dependencies on nn and LL to be important contributions. I have therefore raised my score to support the acceptance of this work.

评论

Dear Reviewer iHRn,

Thank you for your constructive suggestions and for recognising our contribution. We truly appreciate your efforts.

best regards Authors

审稿意见
4

This paper investigates the generalization error of LL-th layer ReLU neural networks trained in the NTK regime. On the data that is γ\gamma-separated by the neural tangent kernel, the authors provide the excess risk of order O(L4(1+γL2)/γ2n)O(L^4(1+\gamma L^2)/\gamma^2 n), which is faster than the well-known suboptimal one O(1/n)O(1/\sqrt{n}) and only depends polynomially on the network depth LL. They also provide a global convergence guarantee for the network using gradient descent and the NTK framework. In the analysis, they carefully evaluate the dependence on LL for the convergence analysis and Rademacher complexity bound, which enables them to obtain a sharper bound.

优缺点分析

Strength

  • As the authors mentioned in the paper, much of the existing literature on the generalization error bound of deep neural networks exponentially depends on the network depth. This work mitigates this dependence to the polynomial one, which is a significant refinement.
  • This paper is easy to follow, with detailed explanations of theorems and concise proof overviews.

Weakness

  • The primary concern is that this paper mainly focuses on NTK-separable data, which imposes a substantial restriction on the data structure. It appears that several refinements, such as the polynomial dependence on the network depth, do not require the separability assumption. Therefore, it would be more plausible to compare the results obtained in this work with other NTK-based results that do not focus on the separability condition.

问题

  • How is the initialization scheme defined in Assumption 1 crucial for obtaining the sharper bound? For example, while [Chen et al., 2021] considers the standard Gaussian initialization for the weights, can the same bound be derived for such settings?
  • As the authors also mentioned as a future work, extending the results to stochastic optimization, which is covered by [Chent et al., 2021], will be a significant topic. What is the difficulty of such an extension?
  • Another question is about the extension to other models. For example, how is the ReLU activation essential for obtaining the results, and is it possible to extend the results to other activations? Moreover, I am curious about the applicability of this approach to different models, such as ResNet and CNN.

局限性

The authors properly address the limitations.

最终评判理由

I would like to maintain my original score.

格式问题

N/A

作者回复

Thanks for acknowledging the strengths of our work. Our point-to-point response to your comments follows next.

W1. This paper mainly focuses on NTK-separable data, which imposes a substantial restriction on the data structure.

Our main contribution is that we are the first to prove that deep ReLU networks trained by gradient descent can achieve optimal rate (1/n1/n). The existing works either obtain a bound of suboptimal rate 1/n1/\sqrt{n} or focus on smooth activations. Our results hold in general settings, not limited to NTK-separable data.

  • Specifically, in Theorem 2, we provide generalization bounds in general cases, i.e., O~(L4F(W)n)\widetilde{O}\left(\frac{L^4F(\overline{W})}{n}\right). Hence, as long as there exists F(W)F(\overline{W}) with small error, the performance of the deep network trained with gradient descent is guaranteed.
  • The NTK-separability assumption is only used to analyze F(W)F(\overline{W}) in a concrete case and to better illustrate our main result. This assumption allows us to derive explicit estimates of F(W)F(\overline{W}), rather than being a necessary condition for our general theory. We believe our general generalization bound can be applied to various settings and demonstrate the power of deep learning.
  • The NTK separability assumption has been widely used in the literature [1,2,3,4]. It has been theoretically justified in a range of settings. For instance, [2] analyzes the margin through its dual form in different settings, including linearly separable case and the noisy 22-XOR distribution.

W2. It would be more plausible to compare the results obtained in this work with other NTK-based results that do not focus on the separability condition.

Many NTK-based results assume positive eigenvalues of NTK Gram matrix. [5] points out that NTK separable data is a milder assumption. In general, we need to estimate F(W)F(\overline{W}), or construct a predictor fW(x)f_{\overline{W}}(x) that achieves low training error under the structural data assumption. For example, this is closely related to the Neural Tangent Random Feature (NTRF) model, which has been well studied in the appealing work [3].

Q1. How is the initialization scheme defined in Assumption 1 crucial for obtaining the sharper bound? For example, while [Chen et al., 2021] considers the standard Gaussian initialization for the weights, can the same bound be derived for such settings?

The symmetric initialization assumed in Assumption 1 is a commonly used technique in the theoretical deep learning literature [6], primarily to simplify the analysis.

  • Specifically, this ensures at the beginning, fW(0)(xi)=0,fW(0)(xi)Wl(0)=0,l[L1]f_{W(0)}(x_i)=0, \frac{\partial f_{W(0)}(x_i)}{\partial W^l(0)}=0,l\in[L-1], as illustrated in Lemma 7. This significantly reduces the complexity of bounding the error dynamics.
  • While our current proof uses this symmetric setup, it is not essential. Under standard Gaussian initialization, these quantities can still be shown to be concentrated near zero using standard concentration tools. For example, [3] shows that these quantities can be bounded by poly(log(n/δ))\mathrm{poly}(\log(n/\delta)) terms, which are negligible in our final bounds.

Q2. As the authors also mentioned as a future work, extending the results to stochastic optimization, which is covered by [Chen et al., 2021], will be a significant topic. What is the difficulty of such an extension?

  • While [3] has studied the online SGD setting, we believe that our theoretical framework can be adapted to multi-pass SGD. However, a major difficulty lies in carefully controlling the trade-off between optimization error and generalization gap. For example, how the number of updates TT grows with the dataset size nn.
  • Moreover, analyzing variants of SGD such as noisy SGD, decentralized SGD, or SGD with momentum introduces additional challenges. These algorithms involve extra sources of randomness, communication constraints, or memory terms, which make it harder to track the evolution of the parameters.

Q3. How is the ReLU activation essential for obtaining the results, and is it possible to extend the results to other activations?

  • ReLU is widely used in practice for its efficienct signal propagation and sparsity. Theoretically, it enjoys a variance-preserving property at initialization: Eh0l(xi)22=h0l1(xi)22\mathbb{E}\|h^l_0(x_i)\|_2^2=\|h^{l-1}_0(x_i)\|_2^2. This property makes it possible to control the hidden layers and activation patterns rigorously (Appendix A).
  • For other activations, the core techniques in our paper can still be applied, but additional assumptions may be needed. For example, smoothness, convexity, Lipschitzness or other conditions to ensure signal propagation remains well-behaved.

Q4. The applicability of this approach to different models, such as ResNet and CNN.

Thanks for raising this interesting point. We agree that the generalization analysis in different models is interesting for understanding deep ReLU networks. The core techniques such as controlling the Lipschitz constant and Rademacher complexity can be extended to other architectures. For example, [7] studies the NTK Gram matrix of ResNet and CNN. Building on their insights, it is interesting to explore how our framework can be used in these models, deriving architecture-specific generalization bounds.

[1] Ji et al. Polylogarithmic width suffices for gradient descent to achieve arbitrarily small test error with shallow relu networks.

[2] Taheri et al. Generalization and stability of interpolating neural networks with minimal width.

[3] Chen et al. How much over-parameterization is sufficient to learn deep relu networks?

[4] Taheri et al. Sharper guarantees for learning neural network classifiers with gradient methods.

[5] Nitanda et al. Gradient descent can learn less over-parameterized two-layer neural networks on classification problems.

[6] Xu et al. Overparametrized multi-layer neural networks: Uniform concentration of neural tangent kernel and convergence of stochastic gradient descent.

[7] Du et al. Gradient descent finds global minima of deep neural networks.

评论

I sincerely appreciate the authors' response. The responses adequately address my concern, and then I would like to maintain my current score.

最终决定

This paper proves optimal generalization rates for deep networks in the NTK regime trained on a classification task. A new method for handling the regularity of the outputs w.r.t. to small changes of the weights around initialization, thus replacing an exponential dependence on depth with a polynomial one. The optimal rates of convergence are then proven assuming that the classes are separable with a certain margin w.r.t. the NTK geometry. The paper mixes approaches from NTK analysis and uniform generalization bounds.

We are happy to accept this paper as a poster at NeurIPS, in line with the recommendations of the reviewers.