PaperHub
6.1
/10
Poster4 位审稿人
最低3最高4标准差0.4
4
3
3
3
ICML 2025

Global Convergence and Rich Feature Learning in $L$-Layer Infinite-Width Neural Networks under $\mu$ Parametrization

OpenReviewPDF
提交: 2025-01-24更新: 2025-07-24

摘要

Despite deep neural networks' powerful representation learning capabilities, theoretical understanding of how networks can simultaneously achieve meaningful feature learning and global convergence remains elusive. Existing approaches like the neural tangent kernel (NTK) are limited because features stay close to their initialization in this parametrization, leaving open questions about feature properties during substantial evolution. In this paper, we investigate the training dynamics of infinitely wide, $L$-layer neural networks using the tensor program (TP) framework. Specifically, we show that, when trained with stochastic gradient descent (SGD) under the Maximal Update parametrization ($\mu$P) and mild conditions on the activation function, SGD enables these networks to learn linearly independent features that substantially deviate from their initial values. This rich feature space captures relevant data information and ensures that any convergent point of the training process is a global minimum. Our analysis leverages both the interactions among features across layers and the properties of Gaussian random variables, providing new insights into deep representation learning. We further validate our theoretical findings through experiments on real-world datasets.
关键词
Feature learningNeural networks$\mu$PTensor Program

评审与讨论

审稿意见
4

This paper investigates the training dynamics of infinitely wide neural networks with mup and SGD. They show that these neural networks can learn rich feature spaces and enjoy global convergence, which is better than other mainstream parameterizations such as NTK, MF, and SP. They also validate the theoretical findings through real-world experiments.

给作者的问题

  1. Is the tt in Theorem 4.5 a finite constant w.r.t. width like that in the TP series (e.g. TP4)?

论据与证据

No. The proof of the global convergence property (Corollary 4.6) is too intuitive and seems wrong. Please see the details in "Theoretical Claims".

方法与评估标准

Yes. mup is a popular parameterization in the pretraining of large models. However, its mechanism is rarely explored. This paper tries to prove the advantages of mup over other parameterizations (both in kernel and feature learning regimes).

理论论述

I appreciate the result in Theorem 4.5, which proves that the feature representations evolve while maintaining their diversity and avoiding collapse throughout training.

However, the proof of the global convergence is intuitive and seems wrong. In my opinion, any neural network enjoys the zero training loss and global convergence if it will no longer be updated after some training step.

实验设计与分析

Yes. The paper conducts experiments to verify that mup can perform feature learning (better than NTK and SP) and keep a rich feature space (better than IP). I think the empirical results can support their main results.

补充材料

Yes. I reviewed the Section of Experimental Details and some proofs of the theoretical results.

与现有文献的关系

NA.

遗漏的重要参考文献

NA.

其他优缺点

Strengths:

  1. The paper is written clearly.
  2. The paper tries to study a significant topic in the theory for mup.
  3. As far as I know, the proof idea in this paper is original.

Weakness:

  1. I think the authors need to carefully address my concern about the global convergence.

其他意见或建议

NA.

作者回复

We thank the Reviewer for taking the time to review and give feedback on our manuscript. We appreciate the positive comments regarding the paper's clarity, the topic's significance, the originality of our ideas, and the supportive experimental results. We are particularly grateful for the reviewer's recognition of Theorem 4.5's contribution regarding feature learning under μ\muP.

Regarding the reviewer's main concern about the global convergence result, we believe there is a significant misunderstanding about what our result delivered and why it is non-trivial. The reviewer states (in the comments on "Theoretical Claims" and in Question 1) that

Any neural network enjoys zero training loss and global convergence if it will no longer be updated after some training step.

We respectfully but strongly disagree with this premise as a general statement and wish to clarify this crucial point, as it seems to underlie the reviewer's doubt.


Clarifying the Core Misunderstanding Regarding Global Convergence

While it is true that training algorithms like SGD stop updating when they reach a point where the gradient is zero (a stationary point), this absolutely does not guarantee that this point is a global minimum. In the complex, non-convex loss landscapes of neural networks, such stationary points can very often be suboptimal local minima or saddle points, where the training loss is significantly higher than the lowest possible value (the global minimum) and the model has not achieved the best possible fit to the training data.

The core contribution of our Corollary 4.6 is precisely to demonstrate rigorously that, for the specific setting we study (infinitely wide networks with μ\muP trained using SGD), the optimization dynamics avoid getting trapped in these suboptimal stationary points. Our proof establishes that the training process is guaranteed to converge to a global minimum of the training loss function.

Therefore, proving global convergence is far from trivial; it requires showing that suboptimal stationary points (such as local minima with loss higher than the global minimum and relevant saddle points) are escaped or avoided, leading specifically to the desired global minimum (the lowest possible training loss). This property is not inherent to any network training process that stops; it relies critically on the theoretical properties of μ\muP in the infinite-width limit that we analyze in our paper. Our work provides a theoretical foundation for how μ\muP facilitates finding the global optimum, which is a key aspect of understanding its effectiveness.


Response to Specific Questions:

Q1. In my opinion, any neural network enjoys zero training loss and global convergence if it will no longer be updated after some training step.

A1. As detailed above, this statement reflects the core misunderstanding we wish to clarify. Halting updates only signifies reaching a stationary point (zero gradient), which is not necessarily a global minimum. It could be a suboptimal local minimum or a saddle point with higher loss. Our key result in Corollary 4.6 is the non-trivial proof that, under μ\muP and infinite width, SGD specifically converges to a global minimum of the training loss, overcoming the challenge of potentially getting stuck in other stationary points with higher loss values. This is our specific contribution.


Q2. Is the t in Theorem 4.5 a finite constant w.r.t. width like that in the TP series (e.g., TP4)?

A2. Our current analysis focuses on the behavior of infinitely wide neural networks. Therefore, investigating the dependence of constants like t on finite network widths is beyond the scope of this paper. This remains an interesting open question for exploring finite-width corrections.


We trust this clarification addresses the reviewer's primary concern by highlighting the crucial distinction between halting training updates and achieving mathematically proven global convergence. We believe our theoretical analysis provides a rigorous justification for the observed empirical success of µP and clarifies the optimization dynamics in this important regime.

审稿人评论

Can you prove rigorously that "However, the non-degenerate trajectory ensures that a nonzero error signal would necessitate further parameter updates" in the proof of Corollary 4.6?


My concern has been addressed. I have improved my score from 2 to 4.

作者评论

We appreciate the opportunity to address the new question raised concerning this specific step in the Corollary 4.6 proof: how the non-degenerate trajectory ensures parameter updates follow from a non-zero error signal.

This step "the non-degenerate trajectory ensures that a nonzero error signal would necessitate further parameter updates." establishes a crucial distinction: while neural networks typically have many stationary points in the parameter space (where parameter gradients vanish), our proof shows that under μ\muP with non-degenerate trajectories, convergence can only occur when all error signals χ˚T,i=L(f˚T,yi)\mathring{\chi}_{T, i} = L'(\mathring{f}_T, y_i) are zero. This is significant because zero error signals across all training samples directly imply reaching the global minimum of the training objective (given the convexity of typical loss functions with respect to model outputs).

In other words, we are not merely showing convergence to a stationary point in parameter space, but specifically convergence to a global minimum. This property, stemming directly from the feature independence guaranteed by Theorem 4.5, is critical for global convergence. We provide the detailed derivation below, explicitly demonstrating this link:


More Detailed Derivation: We proceed by contradiction. Suppose at time T, there exists some sample ii with non-zero error signal χ˚T,i0\mathring{\chi}_{T, i} \neq 0, yet the parameters no longer update after time T.

According to our parameter update rule in Equation (3.3) from the main paper, we have:

ZδWL+1_t=ηi[m]χ˚_t1,iZxL_t1(ξi),Z^{\delta W^{L+1}\_t} = -\eta \sum_{i \in [m]} \mathring{\chi}\_{t-1,i} Z^{x^L}\_{t-1}(\xi_i),

where [m]=1,2,,m[m] = \\{1, 2, \dots, m\\} denotes the set of indices for the training samples, and as stated in Section 3, the weights evolve as:

ZWL+1_t=ZWL+1_0+ZδWL+1_1++ZδWL+1_tZ^{W^{L+1}\_t} = Z^{W^{L+1}\_0} + Z^{\delta W^{L+1}\_1} + \cdots + Z^{\delta W^{L+1}\_t}

For the parameters to remain unchanged from time T to T+1, we must have: ZδWT+1L+1=0Z^{\delta W^{L+1}_{T+1}} = 0

Substituting the update rule, this implies:

ηi[m]χ˚_T,iZxL_T(ξi)=0-\eta \sum_{i \in [m]} \mathring{\chi}\_{T,i} Z^{x^L}\_{T}(\xi_i) = 0

Since the learning rate η>0\eta > 0, this simplifies to: i[m]χ˚_T,iZTxL(ξ_i)=0\sum_{i \in [m]} \mathring{\chi}\_{T,i} Z^{x^L}_T(\xi\_i) = 0

However, by Theorem 4.5, we have established that the post-activation features ZxL_T(ξ_i)_i[m]\\{Z^{x^L}\_T(\xi\_i)\\}\_{i \in [m]} are linearly independent at any time T. This linear independence property means that the only way for the equation i[m]χ˚_T,iZxL_T(ξ_i)=0\sum_{i \in [m]} \mathring{\chi}\_{T, i} Z^{x^L}\_T(\xi\_i) = 0 to hold is if χ˚_T,i=0\mathring{\chi}\_{T, i} = 0 for all i[m]i \in [m].

This contradicts our assumption that χ˚_T,i0\mathring{\chi}\_{T,i} \neq 0 for some ii. Therefore, if any error signal is non-zero at time TT, the parameters must continue to update.


We appreciate the reviewer highlighting this key step in the proof of Corollary 4.6. In our revised manuscript, we will include this detailed derivation to strengthen the connection between Theorem 4.5 and Corollary 4.6 while maintaining the natural flow of the paper.

审稿意见
3

The submitted paper analyzes the global convergence of MLPs in feature learning parameterization. By demonstrating that features remain independent during training, they prove global convergence.

给作者的问题

Questions:

  • Can the authors clearly state how they constructed the feature matrix from joint space-time features in Figure 2? I am not sure how this experiment supports the theoretical results.
  • I am unsure about how realistic Assumption 4.1 is for practical settings beyond toy models (MLPs trained on CIFAR; MLPs only achieve around 50% accuracy on CIFAR-10 and may not require feature learning to achieve this performance). Can the authors check the assumption on a few datasets?
  • A recent paper has shown that both SP and NTP can exhibit hyperparameter transfer (and feature learning) [1]. Is it straightforward to show this in the framework introduced?

[1] Scaling Exponents Across Parameterizations and Optimizers https://arxiv.org/abs/2407.05872

论据与证据

Yes. The theorems support the claims.

方法与评估标准

I am not sure if the experiments support the assumptions considered in the paper. I would like the authors to verify these assumptions (see Questions).

理论论述

I have checked the Theorems and arguments presented in the main text but not the Proofs.

实验设计与分析

I am not convinced why the Figure 2 experiments support the results in the paper. I would request the authors to help me understand this result better.

补充材料

I have not reviewed the supplementary material for this paper.

与现有文献的关系

This paper contributed to understanding convergence of feature learning limits of neural networks. Prior literature has mostly focused on the kernel limits.

遗漏的重要参考文献

Relevant references are discussed to the best of my knowledge.

其他优缺点

Strengths:

This work analyzes the global convergence of feature learning parameterizations, which is of significant interest to the community.

Weaknesses:

I am not sure if the assumptions made in this paper generalize to practical settings beyond toy models (see Questions).

其他意见或建议

Suggestions:

  • Can the authors directly measure correlations between different features during training to support their analysis?

Comments:

  • SP and NTP don't learn features at small learning rates only. By comparison, at large learning rates, all parameterizations learn features and can perform equally well or better [1]

[1] Scaling Exponents Across Parameterizations and Optimizers https://arxiv.org/abs/2407.05872

作者回复

We thank the reviewer for their support and constructive feedback. We address each question below:


Q1: Can the authors directly measure correlations between different features during training to support their analysis?

A1: Yes, we have directly measured the correlations between different features during training. Our analysis confirms that features remain largely independent throughout the training process. Specifically, we compute the Gram matrix G of the centered features, where Gij=(h(xi)hˉ)(h(xj)hˉ)G_{ij} = (h(x_i) - \bar{h})^{\top} (h(x_j) - \bar{h}) represents the similarity between features of data points i and j. The set of N centered feature vectors h(xi)hˉ_i=1N\\{h(x_i) - \bar{h}\\}\_{i=1}^N is linearly independent if and only if the minimum eigenvalue λ_min(G) > 0. Our experiments track this minimum eigenvalue throughout training, demonstrating that μ\muP maintains feature diversity while IP shows catastrophic eigenvalue collapse. These results directly validate our theoretical claims. We have additional experimental results here: https://anonymous.4open.science/r/mup_rebuttal-0E0B/Rebuttal.pdf.


Q2. Regarding SP and NTP at large learning rates and feature learning

A2. Thank you for pointing out the paper "Scaling Exponents Across Parameterizations and Optimizers." It is relevant and we will add a citation and brief discussion about it in the revision. This work empirically characterizes when feature learning occurs.

Our work provides a rigorous mathematical framework guaranteeing global convergence during feature learning. The referenced paper shows that SP and NTP can empirically exhibit feature learning at large learning rates but lack theoretical guarantees. The existing NTK analysis framework cannot directly analyze feature learning in SP and NTP at large learning rates, creating a theoretical gap.

While the precise theoretical characterization of SP and NTP at large learning rates remains an open problem in the community, our analysis provides new insights that could potentially be extended to study this challenge in future work.


Q3. Construction of feature matrix from joint space-time features in Figure 2.

A3.In Figure 2, we analyze the combined space-time feature diversity by constructing a feature matrix that simultaneously captures both initial and final representations. This approach provides a comprehensive assessment of how features evolve during training while maintaining linear independence.

For each parameterization scheme, we construct a combined representation matrix as follows:

Let h0(xi)Rnh_0(x_i) \in \mathbb{R}^n represent the feature vector at initialization for input xix_i, and hT(xi)Rnh_T(x_i) \in \mathbb{R}^n represent the feature vector after training completion. For our dataset with NN samples, we form the combined feature matrix HcombinedRn×2NH_{\text{combined}} \in \mathbb{R}^{n \times 2N} by concatenating both initial and final representations:

Hcombined=[h0(x1)h0(x2)h0(xN)hT(x1)hT(x2)hT(xN)]H_{\text{combined}} = \begin{bmatrix} h_0(x_1) & h_0(x_2) & \cdots & h_0(x_N) & h_T(x_1) & h_T(x_2) & \cdots & h_T(x_N) \end{bmatrix}

We then compute the Gram matrix G=HcombinedTHcombinedG = H_{\text{combined}}^T H_{\text{combined}}, where each element GijG_{ij} captures the similarity between combined representations. The minimum eigenvalue of this Gram matrix quantifies the linear independence across both spatial dimensions (different inputs) and temporal dimensions (initialization versus final state).

This combined analysis provides a stronger test of feature diversity than analyzing features in isolation. As demonstrated in Figure 2, μP maintains significantly higher minimum eigenvalues compared to other parameterizations as width increases, confirming its unique capability to preserve feature diversity throughout training while enabling substantial feature learning.

For example, if a network doesn't learn meaningful features (e.g., NTP at large widths) and hT(xi)h0(xi)h_T(x_i) \approx h_0(x_i) for all inputs, the combined matrix would have linearly dependent columns, yielding a minimum eigenvalue near zero. The higher minimum eigenvalues for μP confirm it learns new features that are linearly independent from initialization.


Q4. Realistic nature of Assumption 4.1 for practical settings beyond toy models

A4: Thank you for this question. Assumption 4.1 requires distinct inner product magnitudes between data points. Our empirical verification (sampling 5,000 random triplets from each dataset) shows:

Dataset% Triplets SatisfyingMin Dist Between Inner Products
MNIST100%2.74e-06
CIFAR-10100%3.08e-05
CIFAR-100100%1.51e-05

These results definitively confirm that Assumption 4.1 is not merely a theoretical convenience but reflects geometric properties inherent to real-world datasets. The perfect compliance across standard benchmarks validates that our theoretical framework directly applies to practical settings beyond toy models.

审稿人评论

I thank the authors for their rebuttal. While the rebuttal has clarified my concerns, I am keeping my score because I could not prove several results of the paper in the given time frame, and the score reflects my review confidence. I believe such papers that make fundamental contributions to the field deserve a thorough review in journals with longer review periods.

作者评论

Thank you for acknowledging that our rebuttal addressed your concerns and for sharing the reasoning behind your final assessment. If accepted, we will incorporate these clarifications to enhance the clarity of the manuscript for the ICML audience. We appreciate you recognizing the work's potential significance and thank you for your constructive feedback throughout this process.

审稿意见
3

This paper studies the training of infinite-width LL-layer FFN under the μP\mu P parametrization. The authors establish that features evolve significantly during training while remaining linearly independent, ensuring convergence to a global minimum.

给作者的问题

  1. What is the precise definition of \alpha_k in Lemma C.1? How does it relate to the covariance structure of the random variables involved?
  2. The paper studies the training dynamics of infinite-width feedforward networks (FFNs). Does a finite-width FFN always converge to this infinite-width case? If so, under what conditions does this hold, and what is the convergence rate?
  3. Intuition Behind GOOD Functions: What is the intuitive reasoning behind the definition of GOOD functions? Additionally, why is Condition 4 in Assumption 4.3 necessary?
  4. Gradient Independence Assumption: Does the paper assume the independent gradient assumption, meaning that W and W^\top are independent during training? If so, how does this assumption affect the generality of the results?
  5. Gaussian Process Covariance: Can the authors explicitly write out the specific covariance expressions for the Gaussian processes in Equations (5.3) and (5.4)?
  6. Can the results on independent feature learning and global convergence be extended to the finite-width FFN case? If not, what are the key challenges in making such an extension? If so, how large does the width need to be in your experiments for the results to align with theoretical predictions?
  7. The paper does not seem to impose explicit constraints on the learning rate to ensure convergence, except η=Θ(1)\eta=\Theta(1). Under what conditions on the learning rate does convergence hold?

论据与证据

The theoretical claims in the paper may be correct, as I did not verify all details of the proofs. However, the insights may not generalize beyond the specific setting studied, as the paper makes restrictive assumptions and does not provide comprehensive experimental validation.

Additionally, while the authors claim their results apply to SGD, their theoretical analysis focuses on full-batch GD rather than SGD. In SGD, there is an additional source of randomness from sampling the training data, which may not accounted for in their analysis. As a result, the conclusions drawn for GD may not directly extend to standard SGD.

方法与评估标准

The results seem correct but the paper does not provide comprehensive experiments to support its theoretical results or to verify whether the assumptions are necessary. Without empirical validation, it remains unclear whether the proposed theoretical framework accurately captures real-world training dynamics or if the restrictive assumptions limit its practical applicability.

理论论述

I did not check the proofs in detail, but since they follow results from the Tensor Program framework, the claims are likely correct. However, there are some important concerns.

  1. The authors claim to follow [1], but their setup is actually closer to [2]. In [1], the learning rate is scaled as ηnc\eta n^{-c} uniformly across all layers, whereas [2] allows different layers to use different scaling factors. This difference raises concerns about whether their training dynamics assumptions are fully aligned with prior works.
  2. The paper does not impose explicit constraints on the learning rate for global convergence. This implies that as long as η\eta is constant with respect to width nn , arbitrarily large learning rates (e.g., η=100\eta = 100) could be used theoretically, which is possible in some machine learning models [3]. In practice, learning rates are much smaller ( η1\eta \ll 1), suggesting that there might be missing stability constraints in the theoretical framework.

[1] Tensor Programs IV: Feature learning in infinite-width neural networks. ICML 2021 [2] Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer. NeurIPS 2021 [3] Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability. NeurIPS 2023

实验设计与分析

The experimental results are not comprehensive, and it is unclear what features are being demonstrated in the plots. Based on the authors’ description, I assume the features hh correspond to h2h^2 in the second layer after 10001000 epoch, but this is not explicitly stated. The authors may also consider adding results for x2x^2, as they claim that both pre- and post-activation features remain linearly independent.

Additionally, the paper includes results for Tanh and ReLU, which do not meet their theoretical assumptions, yet the trends in the plots remain similar to those in Figures 1 and 2. This suggests that the assumptions made in the paper may serve primarily for proof convenience rather than being strictly necessary in practice.

补充材料

I reviewed the appendix for experimental details and scanned the proofs, but I did not check every detail thoroughly.

与现有文献的关系

This paper extends the theoretical understanding of infinite-width neural networks by demonstrating that the μP\mu P enables both independent feature learning and global convergence.

遗漏的重要参考文献

The paper omits some recent theoretical works on μP\mu P and NTK for infinite-depth neural networks, which are relevant to its contributions. Notably, the following works should be considered: [1] TP6: Feature Learning in Infinite-Depth Neural Networks [2] Implicit regularization of deep residual networks towards neural ODEs [3] Depthwise Hyperparameter Transfer in Residual Networks: Dynamics and Scaling Limit [4] Global Convergence in Neural ODEs: Impact of Activation Functions

其他优缺点

NA

其他意见或建议

Equation 3.1: should be h1=W1ξh^1 = W^1 \xi, not WξW\xi

作者回复

Thank you for your thorough review. Due to space constraints, our responses are necessarily concise while addressing all key points:


Q1. How does your SGD analysis account for mini-batch randomness when your theoretical focus is on full-batch GD?

A1. Our analysis based on the Tensor Program framework already accounts for mini-batch randomness in SGD. Definition 3.1 models this with the indicator function 1iBt\mathbf{1}\\{i ∈ \mathcal{B}_t\\}, and our derivations incorporate this sampling process. Under SGD, the induced Gaussian processes maintain the same covariance structure as in GD, with the additional randomness from batch selection not affecting the feature independence property central to our global convergence proof.


Q2. Why aren't your experiments more comprehensive in validating theory and testing assumption necessity?

A2. CIFAR-10 provides an ideal balance between complexity and tractability for testing our theoretical claims across multiple parameterizations. In the revised manuscript, we expanded our experimental section with comprehensive evaluations across varying network widths and depths.


Q3. Which layer features do your plots show, h2h_2? Could you add x2 results to validate both pre/post-activation feature independence?

A3. Yes, it is h2h_2. We've analyzed both pre/post-activation features across all layers. Our supplementary analysis [https://anonymous.4open.science/r/mup_rebuttal-0E0B/Rebuttal.pdf] includes complete eigenvalue spectrum plots demonstrating linear independence is maintained throughout training in all layers. These results validate Theorem 4.5 across feature types and network depth. The contrast between μ\muP and IP becomes more significant in deeper layers, where IP shows catastrophic eigenvalue collapse while μ\muP maintains robust feature diversity.


Q4. Why do Tanh and ReLU show similar results despite not meeting your theoretical assumptions?

A4: Tanh satisfies the GOOD function properties in Assumption 4.3. ReLU was included for completeness despite its non-smoothness. As we already discussed in Appendix A.1, extending our analysis to non-smooth activations is a promising future direction. ReLU's similar empirical behavior suggests our theory's core mechanisms may be more general than the specific assumptions in our proofs.


Q5. The authors claim to follow [1], but their setup is actually closer to [2].

A5. Both schemes give the same dynamic. [1] appears to scale learning rate uniformly. But this difference arises because [1] expresses weights as ww_{\ell} with scaling aa_{\ell} in W=nawW^{\ell}=n^{-a_{\ell}}w^{\ell}, while [2] directly uses WW_{\ell}. Such equivalence is justified in Tensor Program V (arXiv:2203.03466).


Q6. No explicit constraints on learning rate for global convergence. Paper only specifies η=Θ(1)\eta = \Theta(1) without clear convergence conditions

A6. We prove that when convergence occurs under μ\muP parameterization, the non-degeneracy of features ensures it will be at a global minimum. We didn't provide explicit convergence conditions because our focus was on characterizing the convergence point, not the convergence process itself. Determining explicit convergence condition of η\eta for μ\muP remains an important open question.


Q7. The paper omits some recent theoretical works on μ\muP and NTK for infinite-depth neural networks.

A7. Thank you for pointing this out. We already discussed Tensor Program 6 in our paper and will discuss the other suggested works in our revision.


Q8. Definition of αk\alpha_k in Lemma C.1

A8. They are arbitrary real numbers representing linear combinations in our independence proof.


Q9. Does a finite-width FFN always converge to this infinite-width case? ... Can the results be extended to the finite-width FFN case? What's the challenge?

A9. Yes - convergence is formally established by Theorem 7.4 in Tensor Program IV. Our empirical results show networks with widths larger than 128 nearly align with infinite-width predictions (Figures 1-4). The main challenge is determining precise convergence rates due to the complexity of tracking higher-order interactions in finite-width networks.


Q10. Can you explicitly write out the covariance expressions for equations (5.3) and (5.4)?

A10. The expressions are provided in Definition 3.1 (Line 187) and Line 212 and restated after Equations (5.3) and (5.4) in Lines 310-311.


Q11. What's the intuition behind GOOD functions, and why is Condition 4 necessary?

A11. GOOD functions maintain feature diversity and ensure non-trivial gradient flow by preventing constant decomposition. Condition 4 helps prevent feature collapse, allowing sigmoid, tanh, and SILU to satisfy our analysis.


Q12. Does your paper assume W and WW^\top are independent during training?

A12. No, we consider standard backpropagation.

审稿人评论

Thank you for your detailed and thoughtful rebuttal.

Given these factors, I will raise my score to a 3 to reflect the strengthened presentation and technical correctness. However, I hesitate to give a strong recommendation for acceptance, as I have not had sufficient time to verify all proofs in detail, and fully digesting the cited works such as Tensor Program V and VI would require more time than was available during the review period. Additionally, while the new experiments are helpful, they still fall short of being comprehensive—for example, broader comparisons and evaluations beyond feature independence would further strengthen the empirical support for your theoretical claims.

作者评论

Thank you for your thorough review and for raising our score after considering our rebuttal – this is very encouraging. We appreciate you acknowledging our responses and have noted your final comments. In the revision, we will incorporate the improvements discussed to strengthen the paper.

审稿意见
3

The paper aims to investigate rich feature learning and convergence to a global minimum via a Maximal Update Parametrization. Networks learn linearly independent features which are different from features at initialization, and due to covariance structure over layers, this implies a convergence to a global minimum under the Maximal Update Parametrization. This holds over the choice of activation function.

给作者的问题

I question if the definition of a good function is somewhat arbitrary (for example, why can't I set r_1 = r_2 = 0 and ensure ϕ\phi is different for each input x?)

I am unsure if the features are empirically shown to be linearly independent.

I am also confused on why minimum eigenvalue is investigated? shouldn't the entire spectrum be looked at in order to measure feature diversity?

论据与证据

I can see the story that the authors tell through the figures, however, I find there to be a lack of experiments. I would appreciate seeing experiments investigate this across different architectures (especially for more conventional networks) and layers. I would also like to verify that these results hold across multiple seeds.

Much of the claims behind the paper relies on theorems, which are given extensive proofs. From what I can understand from looking at the appendix of the main proof claim, the logic appears to hold up. I would appreciate more clear exposition on when the GOOD function definition is used in the proofs.

I am unsure if the features are empirically shown to be linearly independent.

I am also confused on why minimum eigenvalue is investigated? shouldn't the entire spectrum be looked at in order to measure feature diversity?

方法与评估标准

Most of the paper makes a theoretical claim and extensive proofs are provided, so the approach makes sense. I wish more detail was given to explaining the graphs.

理论论述

I looked at the main theorem, and the steps appeared clear and correct.

I question if the definition of a good function is somewhat arbitrary (for example, why can't I set r_1 = r_2 = 0 and ensure ϕ\phi is different for each input x?)

实验设计与分析

I would like more explanation behind why the authors just investigated minimum eigenvalue. I would like to see more experiment over architecture, layers, and seeds of initialization.

I would also appreciate empirical investigations of the linear independence of the features.

补充材料

The proof behind theorem 4.5.

与现有文献的关系

The paper provides a novel investigation of global convergence properties and rich feature learning

遗漏的重要参考文献

I would mention previous works that investigate covariance structure of neural networks and differentiate them from the current study i.e. they don't explicitly study cross-layer interactions

A Rainbow in Deep Network Black Boxes

Structured random receptive fields enable informative sensory encodings

其他优缺点

The paper is easy to read.

其他意见或建议

n/a

作者回复

We thank the reviewer for their thoughtful feedback and valuable suggestions.


Q1. Lack of experiments across different architectures, layers, and seeds.

A1. We focused on MLPs as the conventional building blocks widely used in theoretical studies. While a full investigation across diverse architectures is beyond the scope of this theoretical paper, our core contributions establish the mathematical principles governing feature learning under different parameterizations. To improve experimental breadth, we've expanded our analysis with comprehensive eigenvalue spectrum plots across all three layers for both pre-activation and post-activation features [https://anonymous.4open.science/r/mup_rebuttal-0E0B/Rebuttal.pdf]. These visualizations show μP's advantageous properties becoming more significant in deeper layers.

Layers: While our current plots focus on the second layer, we add results demonstrating the minimum eigenvalue trend across different depths to show that the effect persists. [https://anonymous.4open.science/r/mup_rebuttal-0E0B/Rebuttal.pdf]

Seeds: Robustness across initializations was already considered in our experiments. As detailed in Appendix A (Experimental Details), the presented results are mean values computed over 10 independent trials using different random seeds (seeds 42-51).


Q2. Questions on empirical verification of linear independence and choice of minimum eigenvalue vs. entire spectrum analysis.

A2. Regarding the crucial empirical verification of linear independence among the feature vectors corresponding to different input data points, this is precisely achieved in our work by analyzing the minimum eigenvalue (λmin\lambda_{min}) of the N×NN \times N Gram matrix computed from the centered features. This matrix, capturing the inner products between feature representations, is fundamentally related to kernel methods and the Neural Tangent Kernel (NTK). We compute this Gram matrix GG. If XX is the n×Nn \times N matrix whose columns are the centered feature vectors h(xi)hˉh(x_i) - \bar{h}, then the Gram matrix is G=XXG = X^{\top} X. (Its element Gij=(h(xi)hˉ)(h(xj)hˉ)G_{ij} = (h(x_i) - \bar{h})^{\top} (h(x_j) - \bar{h})represents the similarity between features of data points iiand jj). Crucially, the set of NN centered feature vectors h(xi)hˉ_i=1N\\{h(x_i) - \bar{h}\\}\_{i=1}^N is linearly independent if and only if the minimum eigenvalue λmin(G)>0\lambda{min}(G) > 0. Linear dependence among theseNNvectors occurs precisely when λmin(G)=0\lambda_{min}(G) = 0. Therefore, observing λmin(G)\lambda_{min}(G) provides a definitive and direct verification of whether the features representing different inputs maintain linear independence.

While focusing on the minimum eigenvalue provides the most direct test for linear independence, we have now included comprehensive eigenvalue spectrum analysis [https://anonymous.4open.science/r/mup_rebuttal-0E0B/Rebuttal.pdf] that examines the entire distribution of eigenvalues. These plots reveal that μ\muP maintains significantly higher eigenvalues throughout the entire spectrum compared to other parameterizations. Notably, IP exhibits catastrophic eigenvalue collapse at higher percentiles (e.g., eigenvalues dropping to 10710^{-7} in layer 3), while μ\muP maintains eigenvalue orders of magnitude larger across all percentiles. This full-spectrum evidence further strengthens our claims about feature diversity under μ\muP.


Q3. Definition of GOOD function arbitrary? How about r1=r2=0r_1=r_2=0?:

A3. Regarding Assumption 4.3 / Condition 4: Here, we require that for any real numbers r1r_1 and r2r_2, the function (r1+ϕ(x))(r2+ϕ(x))(r_1 + \phi(x))(r_2 + \phi'(x)) is not almost everywhere constant. It is worth noting that this condition involves "any real number," not the existence of some real numbers satisfying a property.

The GOOD function definition, including this condition, serves as a sufficient condition that guarantees our theoretical results. For an activation function that satisfies our definition, we can show that feature learning maintains linear independence. This condition allows for a broad class of activation functions, including most commonly used ones (e.g., tanh, sigmoid).


Q4. Previous works investigate covariance structure of neural networks and differentiate them from the current study i.e. they don't explicitly study cross-layer interactions.

A4. Thank you for suggesting these valuable references. As recommended, we will add the related works "A Rainbow in Deep Network Black Boxes" and "Structured random receptive fields enable informative sensory encodings" in our revision. We will also highlight, as you noted, that these works don't explicitly study cross-layer interactions, which is one of the key differentiating aspects of our contribution.

最终决定

This paper makes a strong theoretical contribution by proving that infinite-width neural networks trained with the μ\mu-parametrization learn linearly independent features and globally converge to the minimum of the loss. It improves our understanding of feature learning dynamics in contrast to the kernel regime, using the Tensor Program framework to analyze training under SGD.

The reviewers agree that the theoretical insights are novel and technically sound. The authors well addressed concerns raised by reviewers by providing a comprehensive and technically rigorous rebuttal, including detailed derivations of the result, empirical results on full spectrum, and justifications for global convergence.

Overall, this paper offers a meaningful advance in the theoretical understanding of deep learning under the μ\mu-parametrization and is suitable for publication at ICML.