Spectral Informed Neural Networks
摘要
评审与讨论
This paper proposes spectral-informed neural networks (SINNs), which solve PDEs using spectral information. SINNs uses less memory than PINNs, especially for problems with higher-order derivatives, and it also obtains better accuracy than PINNs across several experimental settings. The authors also provide a theoretical error convergence analysis to show why SINNs outperform PINNs.
给作者的问题
-
Page 4: What is in the definition of ?
-
Page 5: I appreciate the authors addressing why they only explore Fourier basis functions in the paper. However, the references that they cite in point 2 are over 10 years old. Is the approach of transforming Fourier basis functions to other basis functions still an active area of research? If not, this could undermine the authors’ justification for only exploring the Fourier basis.
-
Page 6: In Figure 6a, why does the relative error for SINN approach the relative error for PINN as N increases? This seems to contradict the caption for the figure, which says that “SINNs are robust even with complex spectral structures”, making me skeptical of some of the claims in the paper.
论据与证据
The authors claim that their SINN approach has a lower memory requirement than PINNs, and they clearly show this is true in the paper.
The authors show that SINNs outperform PINNs on several experimental settings, which supports the contributions that they state in the introduction. However I’m concerned that some of the experimental settings in the paper are biased towards the SINN method. For example, the initialization for the 2-D heat and NS equations (see Equations 33 and 37) induce decay in the higher frequency components of the initial condition. If there was no decay, should we still expect SINNs to outperform PINNs?
Moreover, the authors claim to provide an error convergence analysis to show that SINNs are more accurate than PINNs. However, I believe that this claim is not properly justified (see “Theoretical Claims”).
方法与评估标准
I believe that the proposed methods make sense for the problems being solved in the paper. I also appreciate that the authors thoroughly compare their method to other PINN-based approaches. However, I am not sure if this paper is suitable for the application-driven machine learning track. If I recall correctly, a submission to the application-driven ML track is supposed to address a real-world use case, but the experimental settings in the paper are synthetic.
理论论述
In section 3.3, what does it mean to “assume that the capability of MLP is powerful enough”? Is this statement referring to the universal approximation theorem? If so, this statement should be made rigorous.
The result for spectral convergence implicitly relies on the fact that we can find the optimum during optimization. However, this is not a reasonable assumption, since the SINN training problem is non-convex. Perhaps an NTK-style argument could fix some of these issues?
实验设计与分析
The experimental design appears to be sound. The authors’ clearly describe the details for the experiments in Appendix D.
补充材料
I skimmed the entirety of the supplementary material.
In Equation 26, how does the loss function help deal with the aliasing error?
与现有文献的关系
The SINN method proposed in the paper is an interesting approach for solving PDEs with periodic boundary conditions. To the best of my knowledge, there is no existing work that has tried out this exact approach for solving PDEs.
遗漏的重要参考文献
The authors should talk about how their method relates to other methods in the literature that combine spectral methods with PINNs. How does their method fit into the prior work combining these two ideas? Here are some papers that have combined these ideas:
其他优缺点
A weakness of the work, which is also stated by the authors (to their credit), is that SINNs only apply to PDEs with periodic boundary conditions. This limits the utility of SINNs to a special class of PDEs.
其他意见或建议
The font in several figures is too small to read. I would encourage the authors to increase the font size in these figures.
“scaler” should be “scalar” on page 5
“spacial” should be “spatial”
“burgers” should be “Burgers’”
Top of page 6: there appear to be typos in the expression for
“Navier-stokes” should be “Navier-Stokes” on page 6
The y-axes in Figures 4 and 5 should be labeled. Also, what are u and v in the x-axes in Figure 5?
Equation 23 in Appendix B: should be
Thank you for your extensive feedback. We understand that you have several questions about the paper, and we address each of your points in turn below.
Questions 1 and Other Comments Or Suggestions
First and foremost, we apologize for the multiple typos and notational inconsistencies you encountered. We take full responsibility for these and will correct all of them in the final paper. In Figure 5, and represent velocity components in which is the solution of the Navier-Stokes equation. Also, since the notation was confusing, it is better to use and .
Question 2
Transforming Fourier basis functions to other basis functions is a very mature and classical research field. However, only exploring the Fourier basis is not unfair. There are two reasons: 1) it is not practical to explore all basis functions in one paper. 2) exploring one kind of basis is enough. For example, Fourier Neural Operator only explores Fourier basis in Neural Operator, but after FNO, researchers propose Spectral Neural Operator, Wavelet Neural Operator, and so on.
We understand that you are worrying about other basis functions. We conducted a simple example with Chebyshev basis on 1D heat equations:
with the solution . In our experiments, .
The results are:
| PINN | SINN |
|---|---|
Question 3
The in this experiment is the number of terms in the initial condition of Eq.20. To avoid confusion, it is better to use instead of .
This experiment is conducted to verify the robustness of SINNs. As we stated in Section 4.3, 'For most problems, the structure of coefficients in the spectral domain is simpler than the structure of solutions in the physical domain.', low frequencies can represent most cases in the real world. But how about some specific cases in which the structure of coefficients is also complex?
Furthermore, more experiments on testing high-frequency terms are asked by other reviewers and can answer your concern of 'decay in the higher frequency components of the initial condition'. You can find the experiments in the reviewers of zt1d and 34of.
Application-driven ML
When comparing algorithms, precise relative error measurement is essential. Real-world cases are noisy, and full of unknowns, making it hard to isolate and assess algorithm performance. By using synthetic examples, we can better evaluate our algorithm's core capabilities, laying a solid groundwork for real-world applications. For example, because of its inherent mathematical complexity, the 2-D Taylor-Green vortex flow has been established as a representative benchmark in computational fluid dynamics for validating numerical methods, fincluding finite element penalty–projection method (doi: 10.1016/j.jcp.2006.01.019), PINN (doi: 10.1016/j.jcp.2021.110676), and so on.
Furthermore, the periodic boundary condition is not real, but it can be used to understand the real world's features, for example, coherent vortex evolution (doi: 10.1017/S0022112084001750), intrinsic physical dynamics(doi: 10.1017/S0022112095000012), and so on.
Theoretical Claims
'Is this statement ..': Yes, the conclusion is held by the universal approximation theorem, thanks for the question.
'Perhaps an NTK-style ..': The error analysis is based on a very ideal assumption. NTK analysis is based on the loss function, which is constructed by the residual equation rather than the error equation in both PINNs and SINNs. To our best knowledge, there is no research that reveals that the error and the residual have direct relationships. Furthermore, empirically speaking, a smaller loss function may not guarantee a smaller error (see Error Certification (arXiv:2305.10157)). However, Ref(doi:10.1016/j.jcp.2023.112527) proves that: adding some terms to the error yields an upper bound including the residual. Thus, if we can find or construct a connection between the residual (or other loss functions, for example, Astral (arXiv:2406.02645)) and the error, perhaps we can use NTK analysis.
Eq.26
Thanks for being concerned about the aliasing error. How to dealiase the error lies in how to deal with . In a trivial way, one can do inverse FFT back to the physical domain and do those multiplications. Additionally, one can use the 2/3 rule (doi: 10.1175/1520-0469(1971)028<1074:OTEOAI>2.0.CO;2)
Essential References
Thanks, we will give a further discussion on the relevance of those papers.
Thanks for addressing my concerns. I will raise my score accordingly.
The Manuscript describes how to use spatial k-space description of the trial neural network with promising results. The general form of the PDE is assumed to be first order in time with periodic boundary conditions in space. Numerous essential PDEs are considered and it is shown that the naturally sparse k-space representations do provide a significant advantage especially for equations with high oder derivatives and allow non-uniform sampling in k-space for the collocation points. The Authors describe a parametric probability distribution for high sampling rate at lower frequencies diminishing towards the higher frequencies.
给作者的问题
See above.
论据与证据
The claims correspond to the evidence, also the restrictions of the method are stated clearly, like the difficulty for realistic i.e. (complicated) domains important in industrial work.
方法与评估标准
The method does work well for the cases it works well. The challenges are clearly stated by the Authors, for example the experimental finds the for first order derivates the PINNs are more efficient.
理论论述
The theory is sound, in one place when describing continuity equation is k-space the spatial derivative used is gradient when it is supposed to be divergence.
实验设计与分析
Experiments are well performed,
补充材料
The code is given and related data was provided.
与现有文献的关系
The Manuscript as a clear introduction with proper references to the literature positioning the work amongst the physics informed machine dee learning scene.
遗漏的重要参考文献
The relevant references are cited in the Manuscript.
其他优缺点
The Manuscript is well written, and shows very promising results. The basic weakness is, as stated by the Authors, that as a general PDE solver it is restricted to simple periodic geometries. Hence, its impact is restricted although definitely interesting for the industrial community.
However, a discussion and experimentation of defining initial and boundary conditions from experimental measurements could be considered as a choice. This would allow the estimation, for example, he material parameters of a PDE. In this domain and use case, the experimental measurement can be done in particularly designed geometries that fulfill the periodicity expectations of the SINNs and this would not be such a sin anymore, and would improve the value of the method in practical applications.
其他意见或建议
I would have liked to have more discussion on the Burgers equation with low viscosity, where the hight frequency components become essential. How is this scaling?
Thanks for your positive feedback and useful suggestions.
Firstly, we sincerely appreciate your having caught this typo of which should be .
Burgers’ equation with low viscosity
We already conducted an experiment on Burgers’ equation with a fairly low viscosity () which already contrains high-frequency content. But we agree that exploring different viscosities is insightful.
Here we consider a more explicit Burgers' equation:
which contains a high frequency component and a low frequency component.
We set . The results with different are as follows:
| relative error | |
|---|---|
| 20 | |
| 40 | |
| 80 | |
| 160 | |
| 320 |
Helmholtz
Additionally, Reviewer zt1d suggests conducting experiments on Helmholtz equations which only contain a single frequency.
Here we consider a 1D Helmholtz equation:
with the boundary condition
The solution is . The results are as follows:
| PINNs | SINNs | |
|---|---|---|
| 2 | ||
| 4 | ||
| 8 | ||
| 16 | ||
| 32 |
PINNs have arisen as an exciting and promising alternative to classical solution methods for solving partial differential equations. However, PINNs are not without their challenges. One key issue is the cost of automatic differentiation for higher-order derivative PDEs. It is well-known that the cost scales with the dimensionality of spatial variables, which is problematic for higher-order PDEs in high dimensions. To alleviate this issue, the paper proposes Spectral PINNs, which work in spectral space, and reduces applying the differential operator to multiplication, a highly efficient operation on gpu. The authors provide theoretical support to validate their approach, along with experiments to verify its effectiveness in practice.
给作者的问题
Is there any particular reason why you did not compare with STDE?
论据与证据
The paper makes three central claims of contribution:
Claims:
1.) We propose a method that eliminates the spatial derivatives of the network to deal with the high GPU memory consumption of PINNs.
2.) We propose a strategy to approximate the primary features in the spectral domain by learning the low frequency preferentially to handle the difference between SINNs and PINNs.
3.) We provide an error convergence analysis to show that SINNs are more accurate than PINNs. Furthermore, our experiments corroborate that the method can reduce the training time and improve the accuracy simultaneously.
Overall:
The only claim I really have issue with is item 3.) in particular the first part on error analysis. The assumptions made to reach that conclusions is highly idealized, as I discuss more in detail below.
Given that the claims are mostly well supported and the methods does well in the experiments, I'm rating the submission as a weak accept at this point. My main reason for not rating the submission higher is that spectral methods are somewhat limited in the problem complexity they can handle. In particular, they do not handle complex domain geometries well.
方法与评估标准
The proposed methods and evaluation criteria for the paper are appropriate. As the SINNs framework can be combined with most PINN frameworks/strategies, it makes the most sense to compare to the vanilla PINN and PINN frameworks for which SINNs does not apply. This is exactly what the paper does. Moreover, the paper uses standard benchmark PDEs for its evaluation.
Thus, overall I don't have any issues with the baselines for comparison or the experimental design. Though one thing that would have improved the paper is if it had compared with the recently proposed Stochastic Taylor Derivative Estimator (STDE) (Shi et al. 2024) which claims to greatly reduce the cost of differential operators for higher-order PDEs.
理论论述
This is my main issue with the paper. The claims of spectral convergence in section 3.3 is both too informal and oversimplified. I will focus on the PINN case, but analogous statements hold for SINN.
For instance, in the case of a PINN it decompose the error into the sum of two-terms via the triangle inequality.
- Statistical Error:
- Approximation Error: .
It then state assuming the neural network is powerful enough, the approximation error term vanishes, while the statistical term is , as the PINN objective is an empirical estimator. This greatly oversimplifies things. For one the approximation error is likely small, but treating it as exactly zero is somewhat unreasonable. But my bigger issue, is the entire discussion neglects optimization. The PINN objective is a challenging non-convex optimization to solve (and similarly for SINNs), so its highly unlikely the solution found via training corresponds to the globally optimal solution of the empirical risk Thus, in general we should expect achieved in practice to be worse than or .
I'm sympathetic to the authors in that I understand, analyzing the solution error of PINNs is highly non-trivial, with papers entirely focused on this issue. So, I would be satisfied if the authors were more explicit in their assumptions and that they are considering an ideal setting where optimization error is ignored / negligible.
实验设计与分析
Yes, I checked and this seems fine. As I said above, it would have been nice to see a comparison with STDE in Table 1 rather than just Taylor mode automatic differentiation. This is my main criticism.
补充材料
I read the paper's appendix which contained further background on the problem classes, experimental setup, and hyperparameter values.
与现有文献的关系
The paper lies at the intersection of the literature on PINNs and spectral methods for solving PDEs. Its main proposal is to enhance PINNs with techniques from spectral methods for PDEs to reduce the time required to compute derivatives. Given the evaluation in the paper, I believe this approach will enhance the use of PINNs for any PDE problem that possesses structure for which it is appropriate to apply a spectral method. Thus, in terms of the PINNs literature it falls in between papers that propose general strategies to enhance training vs. those who propose problem specific strategies for improving the performance of PINNs.
遗漏的重要参考文献
I cannot think of any essential references the paper missed in its discussion, so the paper's coverage is sufficient.
其他优缺点
Strengths
- The key ideas of the paper are expressed clearly and are easy to follow.
其他意见或建议
I have no other comments or suggestions at this time.
Thank you for your detailed review. We address each of your points below:
Is there any particular reason why you did not compare with STDE?
STDE is indeed a recent work to speed up high-order derivatives in PINNs. However, STDE works by amortizing the cost of computing mixed partials in multivariate problems through random estimations. Thus, for univariate problems, its “stochastic” part isn’t needed and it essentially performs the same operations as Taylor-mode. In our manuscript, to intuitively and concisely demonstrate the results, we choose a univariate function with only one spatial derivative term. Herein, Taylor-mode performs the same as STDE. As for multivariate functions, because STDE is an estimator and still uses univariate Taylor-mode, STDE still scales with the order of derivatives and the dimensionality. However, SINNs remain in a constant number with both the order of derivatives and the dimensionality.
To provide a more convinced conclusion, we conducted a test on the efficiency of -variate problems. If we show SINN is constant with the dimensionality, combining the result that SINNs is constant with the order of derivative, we can claim that SINNs is more efficient than STDE. We consider a Poisson's equation:
with and . The results are as follows:
| Memory (MB) | Iteration rate (ite/s) | |
|---|---|---|
| 2 | 542 | 992 |
| 5 | 550 | 1000 |
| 10 | 550 | 1000 |
| 20 | 790 | 982 |
| 100 | 8736 | 995 |
The memory increases because the input scales with . But, because our operator is only multiplication, the iteration rate is almost constant as expected.
Theoretical Claims
We agree that we are considering an ideal setting that the network can express the truncated Fourier series accurately. And we admit that, in practice, the error we obtained is worse than the theorem results. We will clarify in the paper that the theoretical exponential convergence rate is an asymptotic result assuming sufficient capacity.
Additionally, we don't expect treating the approximation error as exactly zero. is enough in error analysis.
Irregular geometries
Thanks for pointing out the limitation that our experiments were on regular domains. However, we argue that our SINNs can be extended to irregular geometries with additional operations. There are several approaches to deal with this limitation, for example:
-
The moving mesh method can construct a coordinate transformation to map an irregular domain to a regular domain (or map subdomains respectively, for example, doi: 10.1016/j.jcp.2020.109835), and it has already been used in PINNs by PhyGeoNet (Gao et al., 2021). (impressive results in Fig. 5 and 7).
-
One can extend the irregular domain to the regular domain, for example, the fictitious domain method(doi: 10.1137/20M1345153).
-
Even, there is a simpler and more straightforward approach if the irregular region can be expected to be expanded into a regular region: one can just add the irregular geometries as a constraint by inverse Fourier transformation. Mathmatically speaking, for PINNs: we minimize the boundary loss by , where is the output; For SINNs: we minimize the boundary loss by , where is the output.
Here we consider a 2D Helmholtz equation on a plate with a hole using the third approach:
with Dirichlet boundary condition:
where is the ball with the center at the origin (0,0) and a radius of and is derived by the solution .
Here are the results:
| PINN | SINN | |
|---|---|---|
| 1 | ||
| 2 |
I thank the authors for addressing my concerns, and so will raise my score to a 4.
This paper introduces Spectral-Informed Neural Networks (SINNs) as an alternative to standard Physics-Informed Neural Networks (PINNs) for solving PDEs. Instead of computing spatial derivatives through automatic differentiation, SINNs leverage spectral methods, replacing differentiation with simple multiplications in the frequency domain. The authors claim that this approach significantly reduces memory consumption and computational cost while improving accuracy due to spectral methods' exponential convergence properties. The paper presents experimental results demonstrating efficiency gains over PINNs, particularly for higher-order PDEs.
给作者的问题
- The proposed method does not support irregular geometries, which is a major advantage of PINNs. How do you plan to extend it beyond structured domains?
- The computational cost for high-dimensional PDEs could be expensive. Have you tested how well SINNs scale beyond low-dimensional problems?
- Given the reliance on spectral methods, how does the approach handle hyperbolic equations and non-smooth solutions? Have you conducted any experiments to evaluate performance on such cases?
- The convergence analysis does not consider sampling error. How do you justify the claim of spectral convergence without accounting for this factor?
论据与证据
While the proposed method presents an interesting alternative to standard PINNs, some of its key claims lack sufficient evidence:
-
One of the well-known advantages of PINNs is their ability to handle complex geometries flexibly. However, the spectral method used in this paper is inherently tied to structured grids and periodic domains, making it unsuitable for irregular geometries. The paper does not provide any solution or discussion on overcoming this limitation.
-
The spectral approach reduces memory consumption for differentiation but does not address the curse of dimensionality. For high-dimensional PDEs, the cost of maintaining and computing spectral coefficients could be prohibitive. The paper does not discuss how the method scales beyond low-dimensional problems.
方法与评估标准
The paper presents a reasonable experimental setup, but there are significant gaps in the evaluation:
-
The paper evaluates common PDEs but does not test its robustness on problems known to be challenging for spectral methods, such as high-frequency Helmholtz equations or highly nonlinear PDEs. These cases would provide a better understanding of the method’s limitations.
-
The paper does not clearly compute the costs associated with FFT, nor does it discuss how the complex number operations arising from first-order differentiation are handled efficiently.
理论论述
The convergence analysis does not take into account the impact of sampling errors, which could significantly affect the claimed spectral accuracy. In addition, the paper does not provide a theoretical discussion on the significant discrepancy between the experimental results and spectral convergence.
实验设计与分析
The method is evaluated on problems where spectral methods are expected to perform well. There are no experiments demonstrating performance on hyperbolic PDEs, non-smooth solutions, or high-dimensional problems where spectral methods typically struggle. Also, there is no discussion on how computational cost scales with problem dimensionality. Given that spectral methods can become expensive in high dimensions, a cost-benefit analysis comparing against other operator learning methods would be beneficial.
补充材料
Yes, all of them.
与现有文献的关系
The methodology of solving PDEs using machine learning aligns with the broader field of Scientific Machine Learning (SciML), which has shown significant potential in advancing computational efficiency and flexibility across various applications.
遗漏的重要参考文献
The paper does not cite relevant works on neural networks with spectral methods, including:
- Fanaskov et al., 2023 – Spectral Neural Operators (Doklady Mathematics, 2023)
- Choi et al., 2024 – Spectral Operator Learning for Parametric PDEs Without Data Reliance (CMAME, 2024)
其他优缺点
Strengths:
- The idea of leveraging spectral methods in neural networks is well-motivated and could be useful for problems where spectral methods naturally excel.
- The proposed approach effectively reduces memory consumption for automatic differentiation, making it a viable alternative for solving PDEs with high-order derivatives.
Weaknesses:
- The method does not extend well to irregular geometries, significantly limiting its practical applicability.
- Scalability is unclear, especially for high-dimensional problems.
- The method is likely ineffective for hyperbolic PDEs or non-smooth solutions, but no experiments are provided to evaluate these cases.
- The theoretical claims lack rigor, particularly regarding spectral convergence and sampling error analysis.
其他意见或建议
N/A
Thanks for the positive assessment and thoughtful questions.
Challenging Problems
Firstly, we consider a well-defined 1D Helmholtz equation as you suggested:
with the boundary condition
The solution is . The results are as follows:
| PINNs | SINNs | |
|---|---|---|
| 2 | ||
| 4 | ||
| 8 | ||
| 16 | ||
| 32 |
For a 2D Helmholtz equation:
with the condition:
are derived by .
The results are:
| PINNs | SINNs |
|---|---|
Furthermore, we don't think this problem is a challenge for SINNs. Because solving the boundary-value problems are like optimizing:
Thus, changing is merely changing the location of the non-zero element in in SINNs.
Irregular geometries
Due to the limitations, please read it in reviewer PNMU.
High-dimensional problems
Thanks for pointing out this limitation. There is a difference between using SINNs to solve high-dimensional PDEs and low-dimensional PDEs: PINNs can handle high-dimensional problems because their error is dependent on which is the number of sampled points and independent of the dimensionality . As SINNs sample frequencies, if the error is also independent of the dimensionality , SINNs can address the curse of dimensionality (CoD). Fortunately, by utilizing the optimized hyperbolic cross, (Shen et al., 2011) proved that the error is also independent of the dimensionality (Key result is Corollary 8.3). Some examples can be found in (Shen & Yu, 2010). Herein, we will apply high-dim Fourier operators by a sparse matrix to avoid CoD.
And we show the scaling by an ideal example:
with and . Here are the results (memory and rate are for SINN):
| PINN | SINN | Memory (MB) | Iteration rate (ite/s) | |
|---|---|---|---|---|
| 2 | 542 | 992 | ||
| 5 | 550 | 1000 | ||
| 10 | 550 | 1000 | ||
| 20 | 790 | 982 | ||
| 100 | 8736 | 995 |
Theoretical Claims
The sampling error exists in PINNs because the Monte Carlo method is used to estimate the integration. For example, the residual term:
As the integration domain contains infinite points, using finite points to estimate it always has an error, i.e. the sampling error in Monte Carlo methods. However,SINNs don't need Monte Carlo methods because we don't have integrations and our input domain is finite: the frequencies. Herein, our loss function is directly:
With the assumption that the network is powerful enough (i.e. regardless of the sampling method selected, finally, the network learns the total dataset). Then means the output of SINNs exactly equals to the coefficients of the frequencies. But only means the output of PINNs is accurate at the points but still has the sampling error from Monte Carlo methods.
In conclusion, because we truncate the Fourier series to a finite number, SINNs have the spectral error but don't have the sampling error.
Essential References
Thanks for the given relevant papers. However, the frameworks of PINNs and Neural Operators are different. PINNs aim to learn the map from a definition domain to a functional space while NOs learn the map from a functional space to a functional space. Herein, we didn't write the review of Neural Operators in our manuscript. But, indeed, we should write a brief review of neural networks with spectral methods in other fields.
The paper's experiments primarily focus on PDEs that are easily solvable using spectral methods and omit more challenging cases, such as hyperbolic PDEs. Furthermore, the theoretical claims provided address only the sampling error, whereas my original concern pertained specifically to spectral convergence. Additionally, the selected baseline—a plain PINN method from seven or eight years ago—is somewhat outdated, and the paper lacks comparisons with more recent methods specifically developed to address high-frequency or spectral bias issues (e.g., PINNs enhanced by positional encoding). Given these limitations, I will maintain my current score.
Thanks for your further discussion of our paper.
Omit more challenging cases
We apologize for having wrongly thought that solving the Helmholtz equation that you mentioned could address your concern.
Here we choose the classical hyperbolic PDEs: wave equations. We consider the following wave equation from PINNacle (arXiv:2306.08827):
with the solution
The results are as follows:
| PINN | SINN |
|---|---|
You can find other methods from Table 3 Wave1d-C in PINNacle. The best one is which uses NTK loss reweighting.
Hope the results can address your concern. If not, please give us a concrete equation.
The convergence analysis does not take into account the impact of sampling errors, which could significantly affect the claimed spectral accuracy.
As SINNs don't have sampling errors, they will not influence the spectral accuracy theoretically. And we admit that it is hard to discuss the discrepancy between the experimental results and spectral convergence.
More recent methods specifically developed to address high-frequency
Our baselines not only have the vanilla PINNs but also include VSPINN (Hoon Song et al., 2024) which is accepted by NeurIPS2024.
As for the positional encoding, we already have the experiments of Fourier Embedding (FE) which is a representative method in PINNs. As we stated, those methods are also valid for our SINNs, so we only did experiments on SINN combined with FE. We can understand your concern, so we did experiments to compare SINN+FE and PINN+FE with different embedding channels ( means without FE) under the same setting as in Section 4.5:
| 0 | 2 | 4 | 8 | |
|---|---|---|---|---|
| PINN+FE | ||||
| SINN+FE |
We hope the above supplementary can have a positive impact on your evaluation。
The paper uses spectral derivatives in place of derivatives computed through AD in physics informed neural networks (that enforce PDE constraints as soft penalties in the loss) and reformulate the problem in spectral space. The reason is to deal with the high memory/compute cost of higher order derivatives for AD. Experiments demonstrate this advantage clearly. Further, their method retains spectral convergence properties.
Strengths identified:
- Simple idea from classical methods to deal with current bottlenecks in PINN training. All reviewers acknowledged that this can be very useful in certain settings. Experiments demonstrate this clearly.
Weaknesses:
- Spectral derivatives are well-known in classical methods - for the PDEs evaluated it seems spectral methods would work well in the classical sense (with pseudospectral solvers etc) - it's unclear if the experiments were chosen for these specific kind of problems; whereas PINNs are more general and mesh agnostic.
- FFT based methods bring standard issues that reviewers raised - aliasing, conforming to regular geometries and uniform grids. While the authors had good arguments for all (including methods from classical techniques like fictitious domain method for irregular geometries, for example), the experiments were minimal to demonstrate this in the rebuttal.
Overall, the reviewers acknowledged that the contributions were positive and useful. The paper could be made better with more emphasis on harder PDEs (where spectral methods fail in the classical sense due to heavy aliasing etc; solutions with discontinuities visualized), with examples on complex geometries and unstructured grids to show that the FFT based differentiation is still performant and useful compared to standard variants of PINNs. While the authors attempted to answer this during the rebuttal, the experiments were minimal (due to the short response period) with simple examples chosen.