Convergence Analysis of Natural Gradient Descent for Over-parameterized Physics-Informed Neural Networks
We have improved the convergence results of gradient descent and have conducted the convergence analysis of natural gradient descent.
摘要
评审与讨论
This paper considers the convergences of a certain class of PINNs with 2 layers in the overparametrization regime (NTK) and makes two contributions: 1) improves the convergence of GD (conditions for LR) and 2) shows quadratic convergence of natural gradient descent. Specifically for 1) the LR dependency on the smallest eigen value of the gram matrix (fisher information matrix) is removed and for 2) it considers a different stability criterion for the jacobian for the derivation.
给作者的问题
N/A
论据与证据
The claims are justified via theoretical derivations.
方法与评估标准
There was no evaluation as this is a purely theoretical paper
理论论述
Theoretical claims are built on top of previous works and appear correct. I could largely follow the arguments, however, I didn't check for the correctness of derivations rigorously.
实验设计与分析
N/A
补充材料
Did not check
与现有文献的关系
The main focus is on the niche domain of PINNs and the paper is targeted towards such a niche audience. In fact, the paper assumes familiarity of the previous theoretical work and reuses notations (even in the abstract without defining them properly until in the method section). The paper may not be easily accessible for a general audience, even though the topic (convergence of NGD) might be interesting to them.
遗漏的重要参考文献
N/A
其他优缺点
Strengths
- The contributions are interesting and will be useful to the theoretical community, and it may be useful in practice. The proof ideas appear novel.
Weaknesses
- The paper jumps to the derivations straightway, but it might be better to contextualize it first. This would make the paper more accessible to a general audience. also, please define the nations when they were introduced.
其他意见或建议
N/A
Thank you for taking the time to review our article and for your insightful comments. We apologize for the oversight in notations. We agree that adding some context before diving into derivations would improve readability, especially for a broader audience. In the revised version, we will make sure to clarify the definitions of the notations when they are first introduced and add more introductory context.
In this work, for gradient descent, we improve both the requirements for the learning rate and the width . Furthermore, we establish convergence guarantees for natural gradient descent (NGD) in training over-parameterized two-layer PINNs, demonstrating that NGD achieves: (1) an learning rate, and (2) faster convergence compared to gradient descent.
Although this is a purely theoretical paper, we can provide some experiments to justfy our theoretical results. When we are able to revise our paper, we will add the detailed experimental results as a separate section in the manuscript, and the code to reproduce the experiments will be added to the Github.
We conduct experiments on three problems, the 2D Poisson equation with reference solution , the 1D Heat equation with reference solution and the 2D Helmholtz equation with wave number , and the reference solution . All codes are conducted by the Pytorch framework. The configurations used in these examples are listed in Table 1. We report the relative -error of the NGD optimizer, the SGD optimizer, the Adam optimizer, and the L-BFGS optimizer in Table 2. The relative -error is defined as follows: where denotes the predicted solution and represents the reference solution. is the number of interior sampling points, and is the number of boundary sampling points.
Table 1: Configurations of Different Equations
| batch size | hidden layers | hidden neurons | activation function | |||
|---|---|---|---|---|---|---|
| 2D Poisson | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 1D Heat | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 2D Helmholtz | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
Table 2: Relative -error of Different Optimizers
| SGD | Adam | L-BFGS | NGD | |
|---|---|---|---|---|
| 2D Poisson | 1.45e-01 | 5.32e-03 | 3.17e-03 | 1.12e-04 |
| 1D Heat | 5.43e-01 | 6.91e-03 | 4.98e-03 | 3.42e-04 |
| 2D Helmholtz | 8.48e+00 | 1.06e+00 | 3.35e+00 | 6.67e-03 |
In all the experiments,we run the NGD and L-BFGS method for 500 epochs, while the SGD and Adam are trained for 10, 000 epochs. The loss decay during training demonstrating that the NGD method converges significantly faster than other optimization methods.
Table 3 presents the convergence performance of the NGD method with different learning rates on the 2D Poisson equation. The experimental results demonstrate that NGD maintains stable convergence across a wide range of learning rates without significant degradation in final accuracy.
Table 3: Relative -error Comparison Across Different Learning Rates for NGD method
| Learning Rate | 0.5 | 0.1 | 0.05 | 0.01 | 0.005 | 0.001 |
|---|---|---|---|---|---|---|
| Relative -error | 1.18e-03 | 3.24e-04 | 1.87e-04 | 1.12e-04 | 1.22e-04 | 1.68e-04 |
In addition, a comparative analysis of the model performance is performed with progressively increasing network widths. Table 4 presents the variation of -error with respect to network width for 1D Poisson equation with . The results demonstrate that increasing network width leads to accuracy improvements.
Table 4: Relative -error Comparison Across Different Network Width for NGD method
| Width | 20 | 40 | 80 | 160 | 320 | 640 | 1280 | 2560 |
|---|---|---|---|---|---|---|---|---|
| Relative -error | 1.59e-03 | 7.21e-04 | 5.18e-04 | 3.8e-04 | 3.08e-04 | 2.76e-04 | 1.78e-04 | 7.05e-05 |
From an experimental perspective, NGD demonstrates rapid convergence during the training process. Compared to other optimization algorithms, it requires significantly fewer epochs to converge. Furthermore, the experimental results illustrate the strong robustness of the NGD method with respect to hyperparameter selection. Therefore, the empirical findings validate our theoretical conclusions.
The manuscript concerns convergence results for PINNs for shallow neural networks with (or certain smooth) activation functions in the overparametrized setting. Both gradient descent and natural gradient descent are considered. The considered model PDE is a heat equation.
update after rebuttal
I maintain my score. The authors claimed that the natural gradient they are considering is different from Gauss-Newton's method, which is incorrect. I pointed this out, but never received an answer.
给作者的问题
The authors frequently compare to Gao et al. when discussing their results and the improved constants. The cited results in Gao do not contain a dependence on the dimension of the computational domain. Can the authors comment on this? Is this dependence hidden in the constants of Gao et al?
Can the authors comment on the dependence of the results on the norm of the right-hand side ? A complicated right-hand side — for example oscillatory — leads to a complicated solution. It is well known that PINNs can struggle in such situations.
论据与证据
Full proofs for every statement are provided.
方法与评估标准
Not applicable.
理论论述
I checked the proof strategy but I am not familiar enough with the mathematical machinery to certify the correctness of the proofs.
实验设计与分析
Not applicable.
补充材料
Not applicable.
与现有文献的关系
The discussion of the recent literature on convergence results in the NTK regime seems reasonable, although I am no expert in the area. In the broader context, some contextualization may help:
-
The form of natural gradient descent considered here is exactly the classical Gauss-Newton method. Including a short remark reminding readers of this fact should be considered.
-
The analyzed method agrees with energy natural gradients as proposed by Müller and Zeinhofer in the case of linear PDEs. As this is also the setting considered in this paper it can be mentioned.
遗漏的重要参考文献
I am not aware of any.
其他优缺点
The article would greatly benefit from simulations, especially for the natural gradient part. From a practioners point of view, it is important to know to what extent theory and practice meet. Even simple experiments in one spatial dimension would already be interesting (this keeps the scaling with the dimension d under control). Examining equation (23) shows that the cost of NG is manageable: The Jacobians are of size , where is the sample size and is the number of trainable parameters. Consequently, assuming the dominating cost is to compute which should be feasible for in the millions and in the hundreds — which is already quite interesting.
其他意见或建议
My main concern with the article is the lack of simulations.
We sincerely appreciate the reviewer's time and valuable feedback on our manuscript. Let us address your questions point by point.
Q1: Relations to other methods.
A1: We apologize for the insufficient background discussion of Natural Gradient Descent (NGD) in our work. Although NGD shares some similarities with the classical Gauss-Newton method, key differences exist. For instance, in our problem, the Gauss-Newton iteration is given by whereas the NGD iteration follows
Müller and Zeinhofer proposed ENGD for PINNs and empirically demonstrated its ability to yield highly accurate solutions. However, NGD formulation differs from ENGD, and they did not provide theoretical convergence guarantees.
Q2: Lack of simulations.
A2: Thanks for the suggestion. We provide some experiments to validate our theoretical results. In the revised manuscript, we will add a separate section presenting detailed experimental results, and the code to reproduce the experiments will be added to the Github.
The configurations used in these examples are listed in Table 1. We report the relative -error of NGD, SGD, Adam and L-BFGS optimizers in Table 2. is the number of interior sampling points, and is the number of boundary sampling points.
Table 1: Configurations of Different Equations
| batch size | hidden layers | hidden neurons | activation function | |||
|---|---|---|---|---|---|---|
| 2D Poisson | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 1D Heat | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 2D Helmholtz | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
Table 2: Relative -error of Different Optimizers
| SGD | Adam | L-BFGS | NGD | |
|---|---|---|---|---|
| 2D Poisson | 1.45e-01 | 5.32e-03 | 3.17e-03 | 1.12e-04 |
| 1D Heat | 5.43e-01 | 6.91e-03 | 4.98e-03 | 3.42e-04 |
| 2D Helmholtz | 8.48e+00 | 1.06e+00 | 3.35e+00 | 6.67e-03 |
In all the experiments,we run the NGD and L-BFGS method for 500 epochs, while the SGD and Adam are trained for 10, 000 epochs. The loss decay during training demonstrating that the NGD method converges significantly faster than other optimization methods.
Table 3 presents the convergence performance of the NGD method with different learning rates on the 2D Poisson equation. The experimental results demonstrate that NGD maintains stable convergence across a wide range of learning rates without significant degradation in final accuracy.
Table 3: Relative -error Comparison Across Different Learning Rates for NGD method
| Learning Rate | 0.5 | 0.1 | 0.05 | 0.01 | 0.005 | 0.001 |
| Relative -error | 1.18e-03 | 3.24e-04 | 1.87e-04 | 1.12e-04 | 1.22e-04 | 1.68e-04 |
In addition, a comparative analysis of the model performance is performed with progressively increasing network widths. Table 4 presents the variation of -error with respect to network width for 1D Poisson equation with . The results demonstrate that increasing network width leads to accuracy improvements.
Table 4: Relative -error Comparison Across Different Network Width for NGD method
| Width | 20 | 40 | 80 | 160 | 320 | 640 | 1280 | 2560 |
| Relative -error | 1.59e-03 | 7.21e-04 | 5.18e-04 | 3.8e-04 | 3.08e-04 | 2.76e-04 | 1.78e-04 | 7.05e-05 |
Q3: Dependence on .
A3: The dimension-related terms in Gao's results are implicitly hidden in their outcomes. However, our results outperform theirs. For instance, when estimating the initial value , for one of the terms , Gao employed the bounded differences inequality, specifically applying . This leads to their estimate of being at least . In contrast, our analysis, by leveraging Lemma C.2, achieves a better bound of . In their analysis, similar bounding techniques were employed in other parts as well, thus our results concerning the dimension are superior to theirs.
Q4: Dependence on .
A4: From the analysis, we can see that the initial value depends linearly on , and the requirement for the width is . Therefore, a complicated leads to a higher requirement on , making convergence more challenging.
The paper investigates the theoretical convergence of natural gradient descent for overparameterized physics-informed neural networks under NTK regime. The paper extends the previous works on gradient descent to natural gradient descent, with improved bounds. The improvement is based on some inequality techniques adopted and the inherent benefits of natural gradient descent over the vanilla gradient descent. The mathematical proof is overall correct, and the study is meaningful for understanding and improving the training of PINNs.
update after rebuttal
I expect authors can include experiments/implementation details, discussions, and strong motivations of the study, since NGD is not popular in PINNs training. Since my score has already been positive, I will keep it. I think the paper should be accepted.
给作者的问题
Why do you consider natural gradient descent, which is rarely applied for training PINNs? Why not consider LBFGS, which I think is more interesting.
论据与证据
Yes
方法与评估标准
The performance of natural gradient descent for PINNs has not been well investigated. Researchers prefer to use gradient-based methods (e.g., gradient descent, SGD, Adam) or quasi-Newton's methods (e.g., LBFGS) to train PINNs. The natural gradient descent is rarely adopted in PINNs literature. Therefore, I am concerned about the applicability of the optimization algorithm studied in this paper.
理论论述
The proof should be correct.
实验设计与分析
No experiments in the paper. But I think experimental investigations on natural gradient descent for PINNs is necessary. Otherwise, the analysis conducted in the paper does not make sense.
补充材料
Yes. I review the appendix of detailed proofs.
与现有文献的关系
NA
遗漏的重要参考文献
NA
其他优缺点
The motivation for the analysis is not strong, since the algorithm (natural gradient descent) is never adopted for training PINNs. I think experimental investigations on natural gradient descent for PINNs is required to support your motivation and analysis. For example, natural gradient descent is exactly outperforming other methods, which is somewhat supported by your theoretical results. Does the convergence results improve with wider networks? Therefore, I think this paper's motivation and theoretical results are not supported and accompanied by some experimental evidence.
其他意见或建议
NA
We sincerely appreciate the reviewer's time and valuable feedback on our manuscript. Let us address your questions point by point.
Q1: Experimental investigations on NGD for PINNs.
A1: Thanks for the suggestion. We provide some experiments to validate our theoretical results. In the revised manuscript, we will add a separate section presenting detailed experimental results, and the code to reproduce the experiments will be added to the Github.
The configurations used in these examples are listed in Table 1. We report the relative -error of NGD, SGD, Adam and L-BFGS optimizers in Table 2. is the number of interior sampling points, and is the number of boundary sampling points.
Table 1: Configurations of Different Equations
| batch size | hidden layers | hidden neurons | activation function | |||
|---|---|---|---|---|---|---|
| 2D Poisson | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 1D Heat | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 2D Helmholtz | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
Table 2: Relative -error of Different Optimizers
| SGD | Adam | L-BFGS | NGD | |
|---|---|---|---|---|
| 2D Poisson | 1.45e-01 | 5.32e-03 | 3.17e-03 | 1.12e-04 |
| 1D Heat | 5.43e-01 | 6.91e-03 | 4.98e-03 | 3.42e-04 |
| 2D Helmholtz | 8.48e+00 | 1.06e+00 | 3.35e+00 | 6.67e-03 |
In all the experiments,we run the NGD and L-BFGS method for 500 epochs, while the SGD and Adam are trained for 10, 000 epochs. The loss decay during training demonstrating that the NGD method converges significantly faster than other optimization methods.
Table 3 presents the convergence performance of the NGD method with different learning rates on the 2D Poisson equation. The experimental results demonstrate that NGD maintains stable convergence across a wide range of learning rates without significant degradation in final accuracy.
Table 3: Relative -error Comparison Across Different Learning Rates for NGD method
| Learning Rate | 0.5 | 0.1 | 0.05 | 0.01 | 0.005 | 0.001 |
| Relative -error | 1.18e-03 | 3.24e-04 | 1.87e-04 | 1.12e-04 | 1.22e-04 | 1.68e-04 |
In addition, a comparative analysis of the model performance is performed with progressively increasing network widths. Table 4 presents the variation of -error with respect to network width for 1D Poisson equation with . The results demonstrate that increasing network width leads to accuracy improvements .
Table 4: Relative -error Comparison Across Different Network Width for NGD method
| Width | 20 | 40 | 80 | 160 | 320 | 640 | 1280 | 2560 |
| Relative -error | 1.59e-03 | 7.21e-04 | 5.18e-04 | 3.8e-04 | 3.08e-04 | 2.76e-04 | 1.78e-04 | 7.05e-05 |
Q2: The influence of width for convergence results.
A2: Theoretically, the algorithm achieves guaranteed convergence once the network width exceeds a sufficient threshold, with no further improvement in convergence rate from additional width increases. This fact is consistent with established results on gradient descent for PINNs and regression problems. Empirically, wider networks exhibit lower training and test errors.
Q3: Why consider NGD?
A3: From a theoretical perspective, existing convergence analyses of optimization algorithms for PINNs have primarily focused on gradient descent. However, as shown in this paper's improvements on gradient descent, gradient descent imposes relatively stringent requirements on certain hyperparameters (e.g., learning rate) and exhibits slow convergence rates. Therefore, we turned our attention to NGD, which we think may account for curvature information of the function. Theoretically, NGD has more relaxed requirements for the learning rate and achieves faster convergence rates. Experiments also demonstrate that NGD converges more rapidly during training.
As for why L-BFGS was not considered, it is because for non-convex optimization problems like PINNs, L-BFGS is more complex and harder to analyze compared to NGD. Given that recent studies have shown quasi-Newton methods (e.g., SOAP [1]) perform well in optimizing PINNs, analyzing the convergence of such algorithms for PINNs will be an important future direction. Additionally, more efficient implementations or variants of NGD also represent a key area for future research.
[1]: Vyas N, Morwani D, Zhao R, et al. Soap: Improving and stabilizing shampoo using adam[J]. arXiv preprint arXiv:2409.11321, 2024.
Thank you so much for adding experiments comparing Adam, LBFSG, and NGD. The experimental results show the superior performances of NGD, which supports the motivation of this paper to investigate the theoretical convergence of NGD for PINNs. I am just curious about one more thing: how do you deal with the inverse of the Gram matrix? It should be extremely high-dimensional and singular. I don't think the vanilla NGD works well in practice. What kind of practical tricks do you adopt? Thank you.
Thanks for the very timely comment. We can understand your concern about the computing of the Gram matrix. However, for the Gram matrix in NGD, we should note that it differs from the classical Gauss-Newton method. Practical tricks can make computing the inverse of the Gram matrix manageable and stable. We explain it as follows.
In our problem, the classical Gauss-Newton iteration is given by whereas the NGD iteration follows Here is the learning rate, the Jacobian , where is the sample data size and is the neural network parameters, as given by Equation (23) in our paper. The dominating computing cost is the inverse of , which can be manageable and stable for small and large (it is also noticed by Reviewer 4Wzk). In practice, we apply stochastic batch technique to choose small sample size in every iteration, and large network papameters to make invertible (we also use torch.pinverse() in the code). $$
For example, for the 2D Poisson equation in above Table 2, the size in is and , and it is enough to make the computing stable. Much larger is also given in above Table 4, shows accuracy improved when network parameter incresed.
We note that for the classical Gauss-Newton iteration, the Gram matrix is . For larger network parameters , the Gram matrix of Gauss-Newton method will be extremely high-dimensional and singular as you said. However, this does not occur in our NGD method.
Finally, we believe that practical implementation techniques for NGD in PINNs—similar to the classical Newton method's evolution into L-BFGS—require further investigation and will be a focus of our future work.
We would be most grateful for your input if any additional revisions are needed during the remaining time of the discussion period.
The paper investigates the convergence properties of gradient descent (GD) and natural gradient descent (NGD) for training two-layer Physics-Informed Neural Networks (PINNs). The authors improve the learning rate of GD from to , where is the least eigenvalue of the Gram matrix , leading to faster convergence. They also establish the positive definiteness of Gram matrices for various smooth activation functions, such as logistic, softplus, and hyperbolic tangent, applicable to a wide range of PDEs. The paper demonstrates that NGD can achieve a learning rate of , resulting in a convergence rate independent of the Gram matrix, with quadratic convergence for smooth activation functions. By introducing a new recursion formula for GD, the authors reduce the requirements on learning rate and network width, improving convergence results. The study highlights the advantages of NGD over GD, showing faster convergence rates and less stringent network width requirements, making it a promising approach for efficiently training PINNs in scientific computing applications involving PDEs.
给作者的问题
Could the authors propose a method that helps the design of PINNs?
Can the authors conduct experiments to verify the correctness in practice?
论据与证据
Yes.
方法与评估标准
N/A
理论论述
Yes.
实验设计与分析
N/A.
补充材料
Yes, I carefully checked all the supplementary material.
与现有文献的关系
The key contributions of the paper are closely related to several strands of prior research in the broader scientific literature, particularly in the fields of optimization, neural networks, and PINNs.
The paper builds on prior work that demonstrates the convergence of GD for over-parameterized neural networks. Specifically, it extends findings from Du et al. (2018, 2019) and Gao et al. (2023), which show that GD can achieve zero training loss under over-parameterization. The authors improve upon these results by providing a better learning rate and milder requirements on network width.
遗漏的重要参考文献
No specific related work not discussed.
其他优缺点
The current paper is limited to a convergence analysis, based on the improvement proposed in the paper, it may conclude a method to guide the training of PINNs in practice. However, we don't see such a method or guidance.
其他意见或建议
No.
We sincerely thank the reviewer for their valuable time and constructive suggestions on our work.
Q1: Methods to help the PINNs' design.
A1: Thanks for the kind suggestion. The results of the paper motivate us to establish "good" Gram matrix to reduce the strictly learning rate requirement for GD optimizers, and this can be achieved by properly design the loss function, the network architecture, and so on. For example, since the learning rate requirement for GD is related to , existing works [1] have adaptively adjusted the weight in the different loss components of PINNs, to improve the eigenvalue distribution of the Gram matrix , thus we can use a normal learning rate to accelerate PINNs' training and convergence.
[1]Wang S, Yu X, Perdikaris P. When and why PINNs fail to train: A neural tangent kernel perspective[J]. Journal of Computational Physics, 2022, 449: 110768.
Q2: Add experiments to verify theoretical findings.
A2: Thanks for the suggestion. Here, we provide some experiments to validate our theoretical results. In the revised version, we will add the detailed experimental results as a separate section, and the code to reproduce the experiments will be added to the Github.
We conduct experiments on three problems, the 2D Poisson equation with reference solution , the 1D Heat equation with reference solution , and the 2D Helmholtz equation with wave number , and the reference solution . All codes are conducted by the Pytorch framework. The configurations used in these examples are listed in Table 1. We report the relative -error of the NGD optimizer, the SGD optimizer, the Adam optimizer, and the L-BFGS optimizer in Table 2. The relative -error is defined as follows: where denotes the predicted solution and represents the reference solution. is the number of interior sampling points and is the number of boundary sampling points.
Table 1: Configurations of Different Equations
| batch size | hidden layers | hidden neurons | activation function | |||
|---|---|---|---|---|---|---|
| 2D Poisson | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 1D Heat | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
| 2D Helmholtz | 1000 | 200 | 100 | 1 | 128 | tanh(·) |
Table 2: Relative -error of Different Optimizers
| SGD | Adam | L-BFGS | NGD | |
|---|---|---|---|---|
| 2D Poisson | 1.45e-01 | 5.32e-03 | 3.17e-03 | 1.12e-04 |
| 1D Heat | 5.43e-01 | 6.91e-03 | 4.98e-03 | 3.42e-04 |
| 2D Helmholtz | 8.48e+00 | 1.06e+00 | 3.35e+00 | 6.67e-03 |
In all the experiments,we run the NGD and L-BFGS method for 500 epochs, while the SGD and Adam are trained for 10, 000 epochs. The loss decay during training demonstrating that the NGD method converges significantly faster than other optimization methods.
Table 3 presents the convergence performance of the NGD method with different learning rates on the 2D Poisson equation. The experimental results demonstrate that NGD maintains stable convergence across a wide range of learning rates without significant degradation in final accuracy.
Table 3: Relative -error Comparison Across Different Learning Rates for NGD method
| Learning Rate | 0.5 | 0.1 | 0.05 | 0.01 | 0.005 | 0.001 |
|---|---|---|---|---|---|---|
| Relative -error | 1.18e-03 | 3.24e-04 | 1.87e-04 | 1.12e-04 | 1.22e-04 | 1.68e-04 |
In addition, a comparative analysis of the model performance is performed with progressively increasing network widths. Table 4 presents the variation of -error with respect to network width for 1D Poisson equation with . The results demonstrate that increasing network width leads to accuracy improvements.
Table 4: Relative -error Comparison Across Different Network Width for NGD method
| Width | 20 | 40 | 80 | 160 | 320 | 640 | 1280 | 2560 |
|---|---|---|---|---|---|---|---|---|
| Relative -error | 1.59e-03 | 7.21e-04 | 5.18e-04 | 3.8e-04 | 3.08e-04 | 2.76e-04 | 1.78e-04 | 7.05e-05 |
From an experimental perspective, NGD demonstrates rapid convergence during the training process. Compared to other optimization algorithms, it requires significantly fewer epochs to converge. Furthermore, the experimental results illustrate the strong robustness of the NGD method with respect to hyperparameter selection. Therefore, the empirical findings validate our theoretical conclusions.
This paper considers training two-layer PINNs in the NTK regime. The paper extends previous work on gradient descent to natural gradient descent and shows improved convergence bounds. The reviews are generally positive. However, there are several weaknesses identified during the review process:
- Limited novelty: The theoretical contribution is perceived as an incremental extension of existing NTK analysis techniques to NGD within the PINN context, rather than a substantial conceptual advance. The analysis is fairly straightforward given the literature.
- Weak motivation & practical relevance: The paper's motivation is unclear, as NGD is rarely used in practice for training PINNs. The justification for studying this specific optimizer over more prevalent methods (like Adam or L-BFGS) was weak in the original submission, and relying on post-hoc experimental results from the rebuttal doesn't fully address this issue.
- Lack of empirical validation in submission: The paper was initially submitted without any experiments to support its theoretical claims about convergence rates and optimizer performance. This is a major shortcoming for work aiming to provide insights into practical optimization algorithms. Even with the added experiments, Reviewer 4Wzk noted the need for a stronger connection between the empirical results and the specific theoretical claims (e.g., how the experiments directly validate the derived rates or dependencies).
- Unresolved technical clarity: A question was raised regarding the equivalence of the paper's NGD formulation to the Gauss-Newton method in the studied setting. Reviewer 4Wzk noted the authors seemed to misunderstand this relationship and failed to respond to their final clarification, raising concerns about the paper's technical precision and positioning.
If the paper is not accepted, I suggest the authors include strong empirical contributions in the revision.