PaperHub
5.8
/10
Rejected4 位审稿人
最低5最高6标准差0.4
6
6
6
5
3.5
置信度
正确性2.8
贡献度2.8
表达2.5
ICLR 2025

$\mu$LO: Compute-Efficient Meta-Generalization of Learned Optimizers

OpenReviewPDF
提交: 2024-09-27更新: 2025-02-05
TL;DR

We derive muP for popular learned optimizer architectures and propose a simple meta-training recipe that significantly improves learned optimizer generalization.

摘要

Learned optimizers (LOs) can significantly reduce the wall-clock training time of neural networks, substantially reducing training costs. However, they can struggle to optimize unseen tasks (meta-generalize), especially when training networks much larger than those seen during meta-training. To address this, we derive the Maximal Update Parametrization ($\mu$P) for two popular learned optimizer architectures and propose a simple meta-training recipe for $\mu$-parameterized LOs ($\mu$LOs). Our empirical evaluation demonstrates that LOs meta-trained with our recipe substantially improve meta-generalization to wider unseen tasks when compared to LOs trained under standard parametrization (e.g., as they are trained in existing work). When applying our $\mu$LOs, each trained for less than 250 GPU-hours, to large-width models we are often able to match or exceed the performance of pre-trained VeLO, the most performant publicly available learned optimizer, meta-trained with 4000 TPU-months of compute. We also empirically observe that learned optimizers trained with our $\mu$LO recipe also exhibit substantially improved meta-generalization to deeper networks ($5\times$ meta-training) and remarkable generalization to much longer training horizons ($25\times$ meta-training).
关键词
Meta LearningLearned OptimizersPre-training

评审与讨论

审稿意见
6

In this paper, the authors aim to improve on the challenging task of improving the meta-generalization of Learned Optimizers (LOs) to longer optimization trajectories and larger base models while maintaining compute efficiency. The authors derive Maximal Update Parametrization for two popular learned optimizer architectures, enabling zero-shot hyperparameter transfer from smaller models to larger ones. A simple training recipe for µ-parameterized LO is proposed which leads to substantial improvement in meta-generalization to wider networks, deeper networks, and longer training horizons with orders of magnitude less computational cost.

优点

• Overall, I think the paper is well-written and addresses a very meaningful problem in the meta-learned optimizer.

• The empirical performance of the proposed algorithm is very good (Figures 4, 5, and 6), and it is particularly impressive that they were able to get such good results using so little compute relative to VeLO (which is extremely expensive …).

缺点

• The proposed method is quite simple, essentially, describing a way to construct a model such that the meta-learned optimizer can generalize to larger-width networks, deeper networks, and longer training trajectories. As far as I am aware the novelty of this paper is relatively limited since most of the findings are originally derived from [1] where Maximal Update Parametrizations was originally proposed. Therefore, from my perception, this appears to be a straightforward application of that work. Could the authors please elaborate on the technical novelty, of this paper? If the technical novelty is more than just an application of Maximal Update Parametrizations, I am happy to increase my score.

• The paper could be made more self-contained. I had to read [1] before really understanding much of the terminology and language being used in this paper.

[1] Yang, Greg, et al. "Tensor programs v: Tuning large neural networks via zero-shot hyperparameter transfer." NeurIPS2021

问题

• From my understanding [1] shows how to initialize the network such that you can scale the width of the network while maintaining optimal hyper-parameters. This paper seems to suggest that using this strategy also allows you to scale the depth of the network and the optimization trajectory. Why is that the case? Can the authors please add some discussion about that and clarify more theoretically why the results are the way they are.

评论

Thank you for taking the time to review our paper. We are pleased to hear oHHy believes our paper is well written, that oHHy thinks it addresses an important problem for learned optimization, and that oHHy finds the empirical performance of our approach to be impressive.

We now address the reviewer's concerns. Each bolded paragraph title refers to one or multiple concerns that have been grouped together. All line numbers refer to the originally submitted manuscript (not the updated version).

Technical contribution beyond applying μ\muP We agree that the change of meta-training LOs in μ\muP is simple, however, the simplicity of our approach is one of its greatest strengths. This small but non-trivial change relative to the existing meta-training framework in learned optimization enables substantial improvements to meta-generalization with no added computational cost to the learned optimizer and no restriction on the problems that can be considered (e.g., any optimizee can be considered in the μ\muLO framework). As such, it represents a net improvement over the existing meta-training framework in learned optimization [4.1,4.2,4.3].

Another contribution of our work is alleviating the limitations of μ\mu-transfer [4.4], itself. As outlined in lines 226-245, μ\mu-transfer increases the complexity of hyperparameter optimization (HPO), while requiring re-tuning for each new task and dataset considered—without the possibility of transferring HPO knowledge from one task to another (e.g. μ\muAdam underperforms in heldout tasks Fig 4. (c),(d),(e)). In contrast, μ\muLOs allows amortizing the cost of future HPO during meta-training at low cost (in contrast to [4.1], which also amortizes HPO) and can benefit from knowledge sharing across multiple tasks during meta-training, unlike μ\mu-transfer.

Therefore, learned optimization helps further alleviate the tuning costs of μ\mu-transfer for new problems, and parameterizing optimizees in μ\muP improves the generalization of LOs from small to large tasks. As such, μ\mu learned optimization is neither a simple application of learned optimization to μ\mu-transfer, nor a simple application of μ\mu-transfer to LOs. It constitute a new meta-training framework for learned optimization that is of interest for amortizing HPO and improving the meta-generalization of LOs.

The paper could be more self-contained with respect to [4.4] We have updated the manuscript by adding a paragraph to section 3.2 that adds more background about μ\muP to the text. If oHHy could provide more details on anything they believe is still missing we would be grateful.

Question regarding generalization to depth and longer unrolls Our observations of stronger generalization to longer unrolls and deeper networks are purely empirical. We have amended the manuscript to emphasize that these findings are only empirical. [4.4] also shows similar empirical findings for longer and deeper versions of the tasks they tune on. Note that our findings for μ\muLOs extend beyond the meta-training tasks to deeper and longer out-of-distribution tasks (e.g. ViT and LM in Figures 5 and 6).


Local References

[4.1] [VeLO: Training Versatile Learned Optimizers by Scaling Up, Metz et al.]

[4.2] [Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves, Metz et al.]

[4.3] [Practical Tradeoffs Between Memory, Compute, and Performance in Learned Optimizers, Metz et al., CoLLAs 2022]

[4.4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

评论

I have read through all the reviewer and author comments, and although I still have some lingering thoughts about the novelty, I think the author/s did a good job improving the paper during the discussion phase. In general, I think this paper would be of use to the community, so I have raised my score, and I suggest acceptance for this paper.

审稿意见
6

This paper proposed new optimizers to overcome the generalizability issue of the existing hand-crafted learned optimizers. It showed extensive experimental results so that the proposed optimizers improve meta-generalization to wider unseen tasks in most cases.

优点

  • Despite some cases where the proposed optimizers do not show the best performance, the “overall” results show that the proposed optimizers show better performance and better stability.

  • It is interesting and informative that it shows a stable performance in the longer horizon.

缺点

• Why don’t you show results on ResNet and Plain ResNet (i.e., without residual connections)? Earlier, it has been shown that the reason why ResNet (with residual connections) works well by loss landscape – ResNet shows a smoother landscape while Plain ResNet shows a bumpy loss landscape in [1]. If the proposed optimizers can show stable performance even in Plain ResNets, I believe this work will be shown to be very effective.

• In a nutshell, the technical addition of this work on top of the standard LP is adding 1/FAN_IN coefficient on the right term (the updated amount for the weight) in (2) if the weight is part of a hidden layer. In that sense, it is unclear how much technical contribution this work would contain. I wonder if the authors’ claim is that it shows a great advancement by adding such a small change.

• The result in Fig 4 is confusing with regard to the goal of the work. The goal of this paper is to develop a better-generalizable optimizer. However, it did not show the best performance on the most “out-of-distribution” tasks (transformer language modeling and ViT image classification, Fig. 4.) (Although the authors argued that their LOs are meta-trained with only 0.004% of VeLO-4000.) Hence, the contribution of this work is diluted. The proposed optimizers do not show the best performance for Transformer LM, in Fig 5 either.

• The legend in Fig 4 needs to be fixed. Either of VeLOM or LOM needs to be dotted line. Also, either of μLOM or μVeLOM needs to be a dotted line (by looking at (c), μVeLOM is a dotted line).

[1]: Hao Li, Zheng Xu, Gavin Taylor, Christoph Studer, Tom Goldstein, “Visualizing the Loss Landscape of Neural Nets,” NeurIPS 2018.

问题

• Why did you select only MLP and Transformers (for language and vision) to show results on?

• I am wondering how the loss landscapes of the approaches on ResNet and Plain ResNet would look.

评论

Thank you for taking the time to review our paper. We are pleased to hear the reviewer believes that overall our proposed optimizers show improved performance and stability and that the generalization to longer training horizons is interesting.

We now address the reviewer's concerns. Each bolded paragraph title refers to one or multiple concerns that have been grouped together. All line numbers refer to the originally submitted manuscript (not the updated version).

Results on ResNets and Plain ResNets The loss landscape is independent of the optimizer. It depends on the data, the loss function, and the optimizee’s (network being optimized) functional form/parameters. It is, nonetheless, interesting to assess the performance of our learned optimizers for plain and residual networks. In section G of the appendix, we report the performance of μ\muLOs and SP LOs on plain and residual networks, finding that our μ\muLOs perform well for both outperforming SP LOs.

The technical contribution is simple While we agree that the change is simple, the strength of our approach, indeed, lies in its simplicity. Changing the optimizee’s parameterization is a small but non-trivial change relative to the existing meta-training framework in learned optimization: it enables substantial improvements to meta-generalization with no added computational cost to the learned optimizer and no restriction on the problems that can be considered. As such, it represents a net improvement over the existing meta-training framework in learned optimization [4.1,4.2,4.3].

Sub-optimal performance relative to VeLO-4000 Fig.4 (d) & (e) As mentioned on L321-323 the tasks in Fig.4 (d) and (e) fall under the meta-training distribution of VeLO-4000. That is, VeLO-4000 was meta-trained on variants of these tasks while μ\muLOM_M and μ\muVeLOM_M were only meta-trained on MLPs. When smaller width versions of the tasks are included in μ\muLO meta-training (subfigures (a) and (b)), we observe that our optimizers outperform VeLO-4000. Note that due to the simplicity of our approach, there is no reason it would not be compatible with VeLO-scale meta-training.

Why study transformers and MLPs We chose to study transformers and MLPs because MLPs allow reporting performance on the meta-training distribution and transformers are the main architecture used in practice today.


Local References

[3.1] [VeLO: Training Versatile Learned Optimizers by Scaling Up, Metz et al.]

[3.2] [Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves, Metz et al.]

[3.3] [Practical Tradeoffs Between Memory, Compute, and Performance in Learned Optimizers, Metz et al., CoLLAs 2022]

[3.4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

评论

Dear reviewer umKW,

Thank you for taking the time to review our paper.

As the ICLR discussion period comes to an end, we would like to offer our assistance in responding to any remaining questions or concerns you may have about our paper.

Best regards,

The authors

评论

Thank the authors for their response. I went through the response, and confirmed my score.

审稿意见
6

The paper investigates the limitations of current learned optimizers (LOs)—small, data-driven neural networks designed to optimize other neural models—in their ability to generalize to longer optimization horizons and larger models than those encountered during meta-training. The authors derive the Maximal Update Parametrization (muP) for two popular LO architectures and propose a simple meta-training recipe for mu-parameterized learned optimizers (muLOs). Empirical evaluations demonstrate that muLOs can match or exceed the performance of VeLO—a state-of-the-art LO trained with 4000 TPU-months of compute—when applied to large-width models.

优点

  1. Effective Identification of Limitations in Learned Optimizers: The authors clearly pinpoint the generalization issues of current LOs, especially their generalization issues with larger models and longer training horizons.
  2. Innovative use of Maximal Update Parametrization: By applying muP to LOs, they leverage theoretical guarantees for zero-shot hyperparameter tuning, going further to transfer optimizers.
  3. Practical Solution to Costly Hyperparameter Optimization: The proposed muLOs offer a way to avoid expensive HPO on each new architecture and task, addressing a significant practical challenge.
  4. Significance in the Field: By enhancing the generalization capabilities of LOs and making them more practical for large-scale applications.

缺点

  1. Lack of Information on Computational Cost: The paper does not provide detailed information about the overall computational resources required for the method, spent on hyperparameter tuning for the mu-based methods. This is crucial for practitioners who may need to pre-train the optimizer themselves and assess the feasibility of implementing the method.
  2. High Pre-Training Computational Cost: According to Line 356, the pre-training of an LO with already tuned hyperparameters amounts to 16 TPU-months. This is a substantial computational expense, prohibiting practitioners from pre-training an LO themselves. This means that a SoTA method would be preferred, and hence the comparison to VeLO seems more critical.
  3. Unclear Method Overhead: There is insufficient information regarding the computational overhead introduced by using the LOs compared to standard optimizers. Specifically, the paper does not quantify how much slower the training process becomes when employing the neural network-based optimizer. For practical applications, understanding this overhead is essential for adjusting compute budgets and training timelines accordingly.
  4. Limited Clarity in Results Presentation: Throughout the paper, the primary metric presented is the training loss. However, details about how this metric is calculated and presented are lacking. For instance, in Figure 2, it is unclear whether the loss is measured exactly at the presented iteration if the error bars represent standard deviation across different seeds, or how the results would differ if showing the best loss achieved so far or using a moving average.
  5. Absence of downstream Evaluation Metrics.
  6. Incomplete Presentation of Experimental Results: In Section 4.1.1, the authors mention collecting data for 1,000 training steps, but the plots presented display only 50% of this range (Figure 3 + appendix). Additionally, information about the number of seeds used and the variability across runs is missing.
  7. Ambiguities in Generalization Evaluation (Figure 4):
  • Convergence Similarities: In Figures 4a and 4b, the loss curves for muAdam and the proposed muLOs converge to similar values. This suggests that the primary benefit is improved early performance (anytime performance), but without considering the method's overhead, it's unclear how significant this benefit is in practical applications.
  • Behavior of muAdam in Figure 4c: The loss trajectory of muAdam has an unexpected shape in Figure 4c without a clear explanation, making it difficult to understand the underlying reasons.
  • Confusing Legend and Color Coding: The legend in Figure 4 uses the same markers for VeLO and LO, and for muVeLO and muLO, making it challenging to distinguish between the different methods.
  1. Terminology Clarification Needed: The paper refers to muAdam, but it is unclear whether this refers to muAdam or muAdamW.
  2. Inconsistent Performance Across Tasks (Figure 5): In Figure 5a, muLO appears to diverge, and across the tasks in Figures 5a and 5b, the ranking between muLO and muVeLO changes significantly. This inconsistency makes it difficult to determine which optimizer performs better overall and under what conditions, complicating the decision-making process for practitioners.
  3. Questionable Claims About Stability: The claim in Line 516 that "muVeLO_M stably decrease training loss over time for each task" seems inaccurate. In Figure 6a, the loss begins to diverge, contradicting this statement. This raises concerns about the reliability of the method and the validity of the conclusions drawn.
  4. High Variance in muVeLO Performance: The figures indicate that muVeLO exhibits a higher standard variance in its performance, while other methods show minimal variance. Understanding why muVeLO has higher variability is important for assessing its reliability and for guiding its practical adoption.

问题

  1. Can you provide detailed information on the computational resources required for your method, specifically the hyperparameter tuning of the mu-based methods?
  2. Given that retraining of both the LO and VeLO is infeasible, how do you justify LO use compared to the SoTA method?
  3. How much additional computational overhead does using the LO introduce compared to standard optimizers/ SoTA, and how does this affect training speed?
  4. Could you clarify how the training loss is calculated and presented throughout the paper, and whether the error bars in Figure 2 represent standard deviation across different seeds?
  5. Could you include any validation metric, to assess the practical impact of your method?
  6. In Figure 3, why do the plots display only 50% of the collected data, and could you provide information about the number of seeds used and variability across runs?
  7. Is muAdam in your experiments referring to muAdam or muAdamW, and could you clarify this terminology?
  8. Given the inconsistent performance between muLO and muVeLO across tasks (as seen in Figures 5a and 5b), can you provide guidance on which optimizer performs better under specific conditions?
  9. How do you reconcile the claim that "muVeLO_M stably decreases training loss over time for each task" (Line 516) with the divergence observed in Figure 6a?
  10. Why does muVeLO exhibit higher variance in its performance compared to other methods, and what are the implications for its practical adoption?
评论

Local References

[2.1] [VeLO: Training Versatile Learned Optimizers by Scaling Up, Metz et al.]

[2.2] [Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves, Metz et al.]

[2.3] [Practical Tradeoffs Between Memory, Compute, and Performance in Learned Optimizers, Metz et al., CoLLAs 2022]

[2.4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

[2.5] [Junjie Yang et al., Learning to Generalize Provably in Learning to Optimize, AISTATS 2023]

[2.6] [Harrison et al. A Closer Look at Learned Optimization: Stability, Robustness, and Inductive Biases, Neurips 2022]

[2.7] [Understanding and correcting pathologies in the training of learned optimizers, ICML 2019]

评论

Thank you for taking the time to review our paper. We are pleased that reviewer s7Qy believes we effectively identify issues with currently learned optimizers, that our use of μ\muP in this context is innovative, that μ\muLOs offer a practical way to reduce HPO costs, and that our work can make LOs “more practical for large-scale applications”.

We now address the reviewer's concerns. Each bolded paragraph title refers to one or multiple concerns that have been grouped together. All line numbers refer to the originally submitted manuscript (not the updated version).

Computational Cost of Meta-training is 16 TPU-months False. The computational cost of meta-training our LOs is provided on line 19 of the abstract, line 356 puts it in the context of VeLO’s meta-training, and line 356 discusses it again. The value on line 356 is given as a percentage, therefore, this corresponds to 0.16 TPU months, less than 5 TPU-days.

Computational cost of μ\muLOs This is addressed in the general reply.

Figure 2 captions missing details Thank you for pointing this out. We have updated the caption accordingly.

Regarding downstream evaluation The main focus of our paper is to study the meta-generalization of μ\muLOs applied to larger-width networks. Therefore, similar to prior published work [2.3], we choose to exclusively study training loss for simplicity. Studying the generalization of the optimizee to downstream tasks is an important orthogonal direction that has been explored multiple times in previous works, as mentioned in L124-127. Specifically, it is well known that the following strategies can improve optimizee generalization: targeting validation loss during meta-training [2.7], using weight decay [2.6], or using flatness-aware regularization [2.5].

“In Section 4.1.1, the authors mention collecting data for 1,000 training steps” This is incorrect, line 375 (section 4.1.1) reports that these plots show 500 steps of training.

Performance relative to μ\muAdam in Figure 4a and 4b and computational cost Indeed, μ\mu Adam (tuned for 500 trials) is a strong baseline. However, our goal is to improve meta-generalization to wider networks, a longstanding problem in learned optimization [2.1]. Therefore, it is most important to compare our approach to previous work in learned optimization, that is to the SP LOs and VeLO, which we handily beat in 4a and 4b. Note that our optimizers outperformμ\muAdam in out-of-distribution tasks 4c, 4d, and 4e.

Behavior of muAdam in Figure 4c μ\muAdam is tuned for the ImageNet MLP width=1024 task (see Table 1). The results in Figure 4c demonstrate that μ\muAdam’s hyperparameters aren't transferable to CIFAR-10 while μ\muLOM_Mand μ\muVeLO’s hyperparameters are. This shows the strength of our μ\muLO framework which can achieve stronger meta-generalization than μ\mu-transfer [2.4].

No mention that μ\muAdam does not use weight decay Thank you for suggesting this. We have updated section 4 to explicitly specify that no weight decay is used with μ\muAdam.

Inconsistent best performer among μ\muLOM_Mand μ\muVeLO This is addressed in the general reply.

μ\muLOM_M diverges in Figure 5a We disagree that μ\muLO diverges in Figure 5a. We acknowledge that μ\muLO’s loss slightly increases after iteration 2000 but it is inconclusive whether it will diverge later on in training. It is important to compare μ\muLOM_M to the LOM_M baseline in this figure. We note that μ\muLOM_M reaches the final loss of value LOM_M after ~ 300 training iterations (16x speedup). Note that the SP baseline VeLOM_M (purple dashed line) does diverge in Figure 5a after the meta-training horizon (red dashed line). This demonstrates the improvements of μ\muLOs over previous work with respect to generalization to deeper networks.

μ\muVeLOM_M begins to diverge in Figure 6a μ\muVeLOM_M does NOT diverge in Figure 6a. Perhaps you are referring to the SP baseline VeLOM_M (purple dashed line) which immediately diverges after the meta-training horizon (red dashed line) in this figure, demonstrating the improvements of μ\muLOs over previous work with respect to generalization to longer training horizons. We stand by our claim that "μ\muVeLOM_M stably decrease training loss over time for each task" in Figure 6.

Variance of μ\muVeLO Answered in the general reply.

评论

Possible divergence in Figure 5a After re-training the models in Figure 5, we noticed that the runs for μ\muLOM_M in Fig. 5a (6 months old) were affected by a bug in the ViT’s μ\mu-parameterization that was fixed a while ago. We have updated the plot accordingly and verified that no other results are affected in the paper. The training curve now shows no sign of divergence. Thank you for asking about this.

μ\muVeLO small loss increase in Figure 6a The VeLO architecture takes as input the number of training steps remaining, thus, requiring the user to specify the total number of training steps (total_steps) a-priori (e.g. as is done for many LR schedules in practice). Therefore, at each step, VeLO’s LSTM is conditioned on an embedding that provides the number of training steps remaining, allowing it to learn a schedule. Previous work analyzing VeLO-4000’s behavior has noted that changing the value of the total_steps hyperparameter leads to variable performance [2.8]. Specifically, they found that increasing the value of total_steps does not always lead to better performance [2.8]. To answer the reviewer’s question about extending training in Figure 6a, we have included training curves for μ\muVeLOM_M on the width 10241024 ViT ImageNet task (same task as in 6a) in section H of the appendix. We show curves for 40,00040,000 steps of training and 25,00025,000 steps of training. We observe that μ\muVeLOM_M can successfully optimize for 40,00040,000 training steps. However, we note this training curve underperforms μ\muVeLOM_M using total_steps=25,000, similar to what was found in previous work for VeLO-4000 [2.8]. We note that this issue is specifically related to the VeLO architecture and not to μ\muLOs

μ\muAdam in Figure 5c μ\muAdam was tuned for 1000 training steps on a width 1024 MLP ImageNet task. Please note that it does not suffer from instability during the tuning window (first 1000 steps) in Figure 5c. While [2.4] shows that, in the cases they study, there is empirical generalization of μ\muAdam’s hyperparameters to longer training unrolls, there is no reason to expect these findings hold for all tasks. We observe that μ\muAdam’s hyperparameters are not transferable beyond the tuning horizon (1000 iterations) to the Cifar10 task in Figure 5c. We have added a clarifying sentence in section 4.1.2 paragraph 4 stating that μ\muAdam suffers from instability outside of its tuning window in figure 5c.

Random seeds used and performance variation across multiple runs All of our plots reporting training loss show the average loss across 5 random seeds. The error bars in these plots report the standard error. Each seed corresponds to a different ordering of training data and a different initialization of the optimizee. We have added a sentence clarifying this in the introductory paragraph of section 4.1 and have made sure the captions of every figure throughout the entire paper information about seeds and error bars if there are any. For the performance variability, we assume the reviewer is referring to the variance in some μ\muVeLOM_M training curves and its performance relative μ\muLOM_M. Please see the general comment for our answer to such concerns.

Thank you for taking the time to thoroughly review our paper. Please let us know if you have any other questions or reservations.


Local references

[2.4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

[2.8] [Is Scaling Learned Optimizers Worth It? Evaluating The Value of VeLO’s 4000 TPU Months]

审稿意见
5

This paper proposes μ\muLO, a new approach, inspired by hyperparameter transfer with the maximal update parameterization (μ\muP), for training learned optimizers. Using μ\muP for a model's initialization and weight updates has been shown to enable transferring high quality hyperparameter settings for the optimizer to networks of greater width, depth and longer training horizons. Similarly, μ\muLO addresses the limited ability of previous learned-optimizers to generalize to wider networks and longer training horizons by deriving the learned optimizer update required by μ\muP. Experiments show that μ\muLO generalizes better than learned optimizers trained with standard parameterizations when applied to networks of larger width, depth, and longer training horizon as was observed for μ\muP in the standard hyperparameter transfer settings.

优点

  • The experimental results are quite compelling and show μ\mu LO addresses the generalization issues of previous incantations of learned optimizers.
  • This is an interesting application of μ\mu P and could potentially reduce the cost of model tuning dramatically. However, the bar is quite high and will require additional work to compare μ\mu LO with newer optimizers like Shampoo and Modula.

缺点

  • While the experimental results are strong, they do not back up the claimed benefit that μ\mu LO addresses the three issues of μ\mu -transfer by avoiding the need for hyperparameter tuning for reach new architecture and task.
  • The experimental results are missing comparisons to VeLO-4000 in Tables 5 and 6. I think another reasonable baseline is per-task tuned Adam and μ\mu Adam in the resource constrained setting to see if μ\mu LO obvious the need for task specific optimizer tuning.
  • The proposed μ\mu LO does not take into account followup work on μ\mu P that extends the original theoretical justification for transferring across width to depth (see Tensor Program VI). Hence, the generalization across depth is more consequential then intended and may be improved with an explicit maximal update parameterization for depth.
  • The legend for Figure 4 does not distinguish between dashed and solid lines.

问题

  • According to Table 1, μ\mu LO_M and μ\mu VeLO_M are trained on ImageNet classification using 3-layer MLP with 3 different width. Is that correct? Does that mean transformer and ViT architectures studied in Figure 4 d and e were not seen during meta-training and are there to show generalization to unseen architectures?
  • Please include VeLO-4000 in Figures 5 and 6.
  • Why is μ\mu LO better than μ\mu VeLO on average despite VeLO being a more powerful model?
  • How does μ\mu LO compare to a per task tuned Adam/ μ\mu Adam?
  • What is the overhead of using the learned optimizer per gradient step relative to Adam?
评论

Thank you for taking the time to review our paper. We are pleased that reviewer i2Y6 believes that μ\muLO addresses generalization issues of previous work and that our approach could reduce the cost of model tuning dramatically.

We now address the reviewer's concerns. Each bolded paragraph title refers to one or multiple concerns that have been grouped together. All line numbers refer to the originally submitted manuscript (not the updated version).

Overclaims in Section 3.3 We believe that the reviewer is referring to Section 3.3 Paragraph 1 and, in particular, to our suggestion that μ\muLOs allow amortizing the tuning cost of multiple architectures, unlike μ\mu-transfer [1.4]. We understand your concern: while the experiments currently demonstrate strong generalization to unseen tasks and architecture families (e.g., as is the case for ViT, LM in Figure 4 (d) and (e)), the performance is not necessarily optimal in the sense that training an LO on small versions of unseen tasks/architectures may perform better than a model attempting to meta-generalize. Our goal for Section 3.3 Paragraph 1 was to state that, unlike μ\mu-transfer, the μ\muLO framework allows meta-generalizing across unseen tasks/architecture families (a significant weakness of μ\mu-transfer). We have revised the text to clarify and tone down those claims in Sec. 3.3. Thank you for bringing this to our attention.

Regarding Figure 5 and 6 baselines We emphasize that these figures study additional advantages of μ\muLO for learned optimization beyond those it was designed to address (width scaling under a certain unroll length). Therefore, these figures only compare with the existing meta-training paradigm in learned optimization. In Figure 6, our focus is to establish the effect of μ\muLO on generalization to longer unrolls. Thus, we directly compare to learned optimizers meta-trained identically (with the same unroll length) but using Standard Parameterization, which is the current paradigm in the learned optimization literature [1.1,1.2,1.3]. VeLO-4000 is meta-trained on a much longer unroll length (20k-200k steps) and as discussed in L321 is not a fair baseline, thus it is not only an unfair comparison but distracts from the question asked by Fig 6. For Figure 5, similarly, our goal is to directly show the unexpected advantage of μ\muLO on generalization to deeper networks over learned optimizers meta-trained under the existing paradigm.

Regarding per-task tuned baselines We did not include the per-tasked re-turned μ\muAdam baseline in our main results because our μ\muLOs do not benefit from meta-training on these tasks and the comparison would not be fair. We do agree that this is an interesting baseline to consider, however. In section F of the appendix, we include results on the ViT w=3072 task where μ\muAdam was tuned for 500 trials in the resource-constrained setting (e.g. on a smaller width version of the task) as you suggested. The results show that the re-tuned μ\muAdam improves over its predecessor, but is outperformed by μ\muLOM_M and μ\muVeLO.

μ\mu Depth Thank you for bringing this up. Depth μ\muP only works for residual networks with a block depth of 1. Since most networks used in practice today (e.g., the transformer) have a block depth > 1 they are incompatible with μ\muDepth. We, therefore, chose not to study meta-generalization under μ\muDepth. The main focus of our paper is to study the meta-generalization of μ\muLOs applied to larger-width networks. However, we noticed that this choice of parameterization is beneficial for meta-generalization to deeper networks and longer unrolls which is very interesting to the learned optimization community, so we also included results for these cases. We have added a brief discussion of \muDepth and our choice not to study \mu depth in section 2, paragraph 3.

Question regarding the meta-training of μ\muLOM_M and μ\muVeLOM_M Yes, the learned optimizers were only meta-trained on the tasks shown in Table 4 for 10001000 steps. This means the tasks seen in Figure 4 (d) and (e) were NOT seen during meta-training and that these subfigures show meta-generalization to new tasks. Moreover, all figures in the paper showing training beyond 10001000 iterations are out-of-distribution with respect to training duration for our μ\muLOM_M and μ\muVeLOM_M.

Why is LO better than VeLO on average despite VeLO being a more powerful model? This question is addressed in the general reply.

Overhead of μ\muLOM_M and μ\muVeLOM_M This is addressed in the general reply.

评论

Thank you for responding to the issues I brought up. However, I do not feel like the main weaknesses I raised were sufficiently addressed due to lack of additional experiments for baselines or incorporation of μ\mu depth. Therefore, I will maintain my score.

评论

As requested by the reviewer we have now run some additional experiments with Depth-μ\muP.

We first want to re-iterate that Tensor Programs VI [1.5] shows theoretically that Depthμ\muP admits hyperparameter transfer in “resnets where each block has only one layer” [1.5] and only in this case. The authors illustrate both theoretically and empirically that “if each block is deeper (such as modern transformers), then [they] find fundamental limitations in all possible infinite-depth limits of such parametrizations”[1.5]. Specifically, in section 9 of [1.5], the authors show an impossibility result for block depth >=2, demonstrating that this case cannot admit hyperparameter transfer under the Depth-μ\muP parameterization they propose. Thus, even when parameterizing μ\muLOs in Depth-μ\muP, any findings for generalization to deeper counterparts of networks with Block depth >=2 (such as modern transformers) will still be more consequential than intended. Therefore, Depth-μ\muP can only help address the weakness that reviewer 12Y6 originally mentioned for the block depth=1 case.

Since Depth-μ\muP is only valid for residual networks with block size=1, we evaluate Depth-μ\muLOs for residual networks with a block depth of 1. Our new experiments incorporate one new baseline, Depth-μ\muAdam, and compare it to μ\muLOM_M paramterized in Depth-μ\muP, henceforth referred to as Depth-μ\muLOM_M.

Depth-μ\muAdam Depth-μ\muAdam is tuned on a with=128 and depth=16 residual MLP ImageNet-32x32x3 classification task. In addition to sweeping the same tunable multipliers used for μ\muAdam, Depth-μ\muAdam also tunes depth_LR_multiplier and depth_branch_multipler by sweeping values in {{24,22,1,22,24}\{2^{−4} , 2^{−2} , 1, 2^2 , 2^4\}}. This results in 12500 total hyperparameter configurations.

Please note that the total GPU-time for Depth-μ\muAdam’s hyperparameter sweep now matches the meta-training time of the Depth-μ\muLOM_M. In the Table below, we report the final loss for with=128,depth=64 (4x tuning) and with=128,depth=128 MLP tasks (8x tuning). We observe that all optimizers following a Depth-μ\mu-parameterization improve the loss as the depth increases. In contrast, SP LO diverges. Comparing within optimizers in Depth-μ\muP, we observe that Depth-μ\muLOM_M outperforms Depth-μ\muAdam. These results are consistent with what was observed throughout our paper for width and they illustrate, as the reviewer suggested, that applying Depth-μ\muP can help improve generalization to deeper networks. However, we note that this parameterization only admits hyperparameter transfer for networks with a block depth=1 [1.5]. While it is outside the scope of our work, we believe that it will be very important for future work to develop parameterizations transferrable for architecutres using a block depth>=2 (e.g., transformers). Thank you for suggesting we explore this, we will incorporate these results in a new section of the appendix.

Generalization to deeper tasks in Depth-μ\muP.

TaskDepth=64 Residual MLPDepth=128 Residual MLP
SP LODivergesDiverges
Depth-μ\muLOM_M5.265.23
Depth-μ\muAdam5.295.27

Local references

[1.5] [Tensor Programs VI: Feature Learning in Infinite Depth Neural Networks]

评论

Local References

[1.1] [VeLO: Training Versatile Learned Optimizers by Scaling Up, Metz et al.]

[1.2] [Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves, Metz et al.]

[1.3] [Practical Tradeoffs Between Memory, Compute, and Performance in Learned Optimizers, Metz et al., CoLLAs 2022]

[1.4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

评论

Here is a list of our changes to the manuscript based on reviewer feedback.

  • We have revised the text to clarify and tone claims in Sec. 3.3 with respect to i2Y6’s feedback.
  • We have added a brief discussion of Depth-μ\muP and our choice not to study Depth-μ\muP in section 2, paragraph 3. This change was made due to i2Y6's feedback.
  • In section F of the appendix, we have added a comparison of the performance of our μ\muLOs to μ\muAdam tuned for 500 trials on a ViT task. This change was made due to i2Y6's feedback.
  • We have added missing details from Figure 2 caption with respect to s7Qy’s feedback.
  • We have added an evaluation of ResNets in section G of the appendix concerning reviewer umKW’s feedback.
  • We have updated the manuscript by adding a paragraph to section 3.2 that adds more background about μ\muP to the text in accordance with oHHy’s feedback.
  • We have minimally updated the abstract, contributions, and section 4 introductory paragraph to emphasize that the findings of improved generalization to deeper networks and more training steps are purely empirical and not theoretically motivated.
评论

General comment: We would like to thank all reviewers for their thoughtful insights and constructive criticism about our paper. We are pleased that reviewers believe we effectively highlight generalization issues with current learned optimizers [s7Qy] and that μ\mu learned optimization (μ\muLO) addresses them [i2Y6,oHHy]. We are particularly pleased that s7Qy thinks the use of μ\muP in this context is innovative and that μ\muLOs offer a practical way to dramatically reduce model tuning cost [i2Y6, s7Qy]. Finally, we are happy that umKW believes our method improves overall performance and oHHy finds the empirical performance to be impressive.

Computational cost of μ\muLOM_Mand μ\muVeLOM_M We chose not to include any such comparison in our paper as the main focus of our work is to study the meta-generalization of μ\muLOs applied to larger-width networks (orthogonal to improving learned optimizer computational cost). As mentioned on L538, the benefits of μ\muLOs come at zero added computational cost relative to learned optimizers from previous work. Nonetheless, we have included results in the table below that compare the computational cost of the learned optimizers in this study to that of AdamW. We refer reviewers to [1] (section B.7) and [3] (sections 4.1,4.2,4.3,4.4) for a more detailed account of the computational cost of μ\muVeLOM_M and μ\muLOM_M, respectively.

Table 1 Computational cost of the learned optimizers in our study and AdamW We report the forward and backward pass time of the optimizee, the optimizer step time, and the total step time per iteration when training ResNet18 on ImageNet. Each number is the median step time over the last 250 training iterations. All timings were measured on an L40S Nvidia GPU.

OptimizerTotal Step TimeOptimizer Step TimeFwBw Time
AdamW0.01990.00220.0177
μ\muLOM_M0.02190.00390.0180
μ\muVeLOM_M0.02530.00760.0177
LOM_M0.02200.00410.0179
VeLOM_M0.02550.00770.0178

μ\muVeLOM_M variance and performance relative to μ\muLO μ\muVeLOM_M uses an LSTM to generate the parameters of a small per-parameter MLP for each tensor in the optimizee. This two-level structure substantially increases its capacity relative to μ\muLOM_M(only a single per-parameter MLP). While μ\muVeLOM_M achieves generalization to wider networks, it appears to suffer from greater variance than μ\muLOM_M, which is not observed in VeLO-4000 (trained on a large distribution of tasks). Moreover, μ\muLOM_M outperforms μ\muVeLOM_M on most tasks, despite its lower capacity. We hypothesize μ\muVeLOM_M may require more extensive meta-training beyond the scope of this research to see it outperform μ\muLOM_M. Our goal here was simply to demonstrate that our parameterization also works for the VeLO architecture.

Figure 4 legend is incorrect We have fixed this in the updated version. Thank you all for pointing this out.


Local References

[1] [VeLO: Training Versatile Learned Optimizers by Scaling Up, Metz et al.]

[2] [Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves, Metz et al.]

[3] [Practical Tradeoffs Between Memory, Compute, and Performance in Learned Optimizers, Metz et al., CoLLAs 2022]

[4] [Yang et al., Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer, Neurips 2021]

评论

Here is a list of our newest changes to the manuscript based on reviewer feedback.

  • We have updated Figure 5a. While responding to reviewer s7Qy, we noticed that the runs for μ\muLOM_M in Fig. 5a (6 months old) were affected by a bug in the ViT’s μ\mu-parameterization that was fixed a while ago. We have updated the plot with corrected runs and verified that no other results are affected in the paper. μ\muLOM_M now consistently decreases the loss throughout training in Fig. 5a.

  • We have added section H to the appendix of our paper, showing training curves of μ\muVeLOM_M for 40,000 training steps.

  • We have made sure that there is a sentence in the captions of every figure stating the number of trials used for each curve and explaining what the error bars designate.

  • We have added a clarifying sentence to section 4.1.2 paragraph 4 to explain the instabilities of μ\muAdam in figure 5c.

评论

Here is a list of our latest changes to the manuscript based on reviewer feedback.

  • We have enhanced the presentation of Figures 2,11,12,13,14 by replacing the legends in each figure with a global color bar and removing redundant x and y labels.
AC 元评审

The paper introduces µ-learned optimizers (µLOs), leveraging Maximal Update Parametrization (µP) to enhance generalization to larger models, deeper networks, and longer training horizons. µLOs achieve competitive or superior performance to state-of-the-art methods like VeLO, with significantly lower computational costs, effectively addressing challenges in hyperparameter tuning. Strenths of the paper include computational efficiency and performance of the approach in the benchmarks studied. However, concerns were raised about the limited novelty of the contribution, as it primarily applies µP to learned optimizers. Also, some key baselines, such as VeLO-4000 and per-task tuned optimizers, were not fully evaluated. Performance inconsistencies on certain out-of-distribution tasks were also pointed out. The main reason for my recommendation to reject is the incomplete experimental comparison.

审稿人讨论附加意见

The reviewers made various changes during the rebuttal, including bug fixes, new experiments, and extensive clarifications in text and visuals. However, novelty and incomplete baselines remained as limitations.

最终决定

Reject