Neural Krylov Iteration for Accelerating Linear System Solving
摘要
评审与讨论
The authors use a neural operator approach to generate subspaces that are used for the acceleration of Krylov subspace methods for several partial differential equations setups.
优点
The authors are interested in a very relevant problem of computational science and engineering of solving linear systems of equations. The attempt of providing a convergence analysis for the proposed method.
缺点
The details of the method remain hidden to me and it seems difficult to grasp the details of the method and many details and even the major workings of the method are unclear.
问题
Why the smallest eigenvalues in line 163 on page 5? I assume that Y_k in Algorithm 1 is the predicted subspace from the neural network, but then explain the derivation used for equation (11)? How is this motivated? How is the method of the authors combined with preconditioning? The details are not given but the choice of the preconditioner is crucial and how would this depend on the learned subspace the method provides?
局限性
The method is limited to linear problems but this is a very broad and general class. If the method was better explained and the details clear and convincing there would be no limitations for these kinds of problems.
We thank the reviewer for the patience to read through our paper. We are pleased to re-introduce our work and respond to your comments as follows. We sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score and your confidence. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Major Workings of Our Method
- We propose a novel method, namely NeurKItt, for accelerating linear systems solving. NeurKItt consists of two modules, one is the subspace prediction module, while the other is the acceleration module.
- The subspace prediction module employs the neural operator for predicting the invariant subspace of given linear systems , as a matrix that contains eigenvectors corresponding to the smallest eigenvalues from .
- The acceleration module uses the predicted subspace for accelerating the linear system solving. It is the Krylov subspace algorithm with predicted subspace involved for iteration.
Subspace Prediction Module
- We use FNO to predict the subspace of the matrix from the given linear system . We give the introduction to FNO in Appendix C.
- The reason why we select the neural operator for subspace prediction is that the mapping from the matrix of linear system to its corresponding invariant subspace can be considered an operator, i.e., a mapping between two Hilbert spaces, thus solving such problems involves finding the corresponding operator.
- We employ the projection loss function for training the FNO and apply thin QR decomposition to the output of the FNO. Projection loss is designed for learning the subspace. The idea behind it is that the predicted subspace should contain the basis vectors in the target subspace. We employ the thin QR decomposition for orthogonalizing the FNO output. We provide the details about thin QR decomposition in Appendix E.
Why the smallest eigenvalues in line 163 on page 5?
- The FNO learns the subspace that associates with the smallest eigenvalues, assuming , where is the eigenvector corresponding to the -th smallest eigenvalue of the given matrix , . We think such subspace is easy to learn because Theorem 5.2 gives proof that if changes in a matrix occur in the subspace corresponding to larger eigenvalues, the subspace associated with the smallest eigenvalues is minimally affected, as long as the changes are smaller than the gap between the smallest and larger eigenvalues. Theorem 5.2 can be found in Section 5.
Acceleration Module
- The acceleration module is the Krylov iteration algorithm with predicted subspace involved, inspired by Krylov recycling algorithms. The detailed pseudo-code about the acceleration module can be found in Appendix A.2. The key of the acceleration module is the Arnoldi relation , which takes the subspace into iteration.
- Here we give an intuitive idea about the acceleration module. Krylov subspace iteration solves the linear system problem by iteratively approximating the invariant subspace of matrix from the linear system . When the subspace information about is given, we don't need to start the approximation from scratch, thus the acceleration module can improve the convergence speed.
I assume that Y_k in Algorithm 1 is the predicted subspace from the neural network, but then explain the derivation used for equation (11)? How is this motivated?
-
is the matrix of predicted subspace from the neural operator. In particular, we obtain and from the thin QR decomposition on , such that and .
-
Equation allows us to find the basis of , which enables us to remove the subspace in Equation (11) by . This approach makes the Krylov subspace iteration more efficient because we only need to search the rest of the invariant subspace for matrix .
-
Equation makes sure the matrix is an upper Hessenberg matrix, which speedups the solving of the linear system in the later iteration procedure.
How is the method of the authors combined with preconditioning? The details are not given but the choice of the preconditioner is crucial and how would this depend on the learned subspace the method provides?
- Supposing the original Arnoldi relation and the preconditioner matrix , the preconditioned Arnoldi relation is . So, for NeurKItt, after the combination, the Arnoldi relation is . Here, is the upper Hessenberg matrices.
- In our experiments, the learned subspace for any preconditioning method is the subspace composed of eigenvectors corresponding to the smallest eigenvalues of the given matrix .
We also provide a theoretical analysis of the acceleration module in Section 5.1, which gives the convergence analysis and explains our choice to predict the subspace rather than the solution. In Section 5.2, we further justify our choice of using the subspace composed of eigenvectors corresponding to the smallest eigenvalues for prediction.
If you are confused about any of the content in our paper, like what kind of obstacle hinders the understanding of our work, please let us know your further concerns, and we will continue to actively respond to your comments.
Thank you for the comments. Regarding your last point on the use of the preconditioner, the main idea of applying a preconditioner P is to cluster the eigenvalues of the preconditioned matrix PA. One typically hopes that the n previously smallest eigenvalues would now be clustered in a tighter region, ideally the perfect preconditioner guarantees only very few distinct eigenvalues What does this mean for your method and the approximated subspace?
Thank you for your question. We address it as follows:
-
We agree with the reviewer about how preconditioners work. However, our method is based on the predicted subspace to improve the convergence speed, which avoids approximating the subspace from scratch for Krylov subspace iteration. Thus, preconditioners that disrupt the invariant subspace, not eigenvalues, will have a negative influence on our method, like ICC and ILU.
-
Despite this potential influence, extensive experiments show that these disruptions do not significantly affect our improvement in practice. In particular, our method accelerates the solving across all preconditioners, especially achieving up to 2.64 time speedup under ICC and 2.6 time speedup under ILU.
-
Besides, optimizations for the approximated subspace could be our future work, which takes the negative influence of different preconditioners on subspace into consideration.
We appreciate the positive feedback and the time you taken to evaluate our work. To improve clarity, we will include these discussions to help readers grasp the key aspects of our study.
The manuscript introduces a novel method, referred to as NeurKItt, which combines neural network techniques with Krylov subspace methods to accelerate the solution of linear systems derived from partial differential equations (PDEs). The core innovation of NeurKItt lies in its capability to predict invariant subspaces associated with the matrices defining these linear systems, thereby significantly reducing the number of iterations required for convergence.
优点
-
Originality: The manuscript introduces a novel approach by integrating neural network techniques with Krylov subspace methods to predict invariant subspaces of linear systems, significantly enhancing the efficiency of these traditional methods.
-
Quality: The manuscript effectively targets the critical issues of computational inefficiency and instability in the context of high-dimensional and poorly conditioned linear systems. The choice of problem is highly relevant to both academic research and practical applications in scientific computing, making the work valuable to a wide audience.
-
Clarity: The manuscript is well-structured, presenting complex concepts in an understandable manner. It effectively communicates the challenges of existing methods and how NeurKItt addresses these issues, providing clear explanations and logical progression from problem statement to solution. However, improvements in typographical accuracy and some clarifications in methodological descriptions could further enhance clarity.
-
Significance: The significance of this work lies in its potential impact on the efficiency of solving large-scale linear systems, which are crucial in various scientific and engineering applications. By reducing computational time and resource consumption, NeurKItt could significantly benefit fields reliant on large-scale computations, making this contribution highly relevant to both academic research and industry applications.
缺点
-
The sentence "Research in using neural networks for accelerating linear system solving" on line 75 appears abruptly terminated and lacks a verb or continuation that would complete the thought.
-
On line 81, the term "precondition" is likely a typographical error. The correct term, given the context, could be "preconditioning".
-
There appears to be a typographical error in Equation (8), where the variable should presumably be .
-
Equation (13) on page 7 starting with an unnecessary 's' before the equal sign seems to be a typo.
-
In the manuscript, Section 4.1 discusses the use of Fourier Neural Operators (FNO) for parametric PDE problems and introduces the subsequent subspace prediction concept. It employs a subspace defined as the span of vectors . The manuscript mentions that is the eigenvector corresponding to the -th smallest eigenvalue, but it does not specify whether these eigenvalues are of the matrix or another related matrix. This omission could lead to confusion about the origin and relevance of these eigenvectors, particularly in how they are integrated into the model and influence the outcome of the subspace prediction.
-
In Section 4.1, the manuscript proposes using invariant subspaces associated with the smallest eigenvalues to train the subspace prediction module. However, an oversight in the theoretical analysis (Section 5.2) arises from applying this concept to potentially non-Hermitian matrices. The section (section 5.2) is based on Hermitian positive definite matrices, whose eigenvalues are all real numbers that can be orderly compared. Non-Hermitian matrices, on the other hand, may exhibit complex eigenvalues, complicating or nullifying the notion of "smallest" eigenvalues due to their lack of a natural order. This discrepancy raises concerns about the applicability of the method to non-Hermitian matrices, which are prevalent in many practical applications.
问题
In reviewing the theoretical analysis provided between lines 237 and 248, I observed a notable similarity in both the phrasing and the mathematical formulations with those presented in reference [23]. Could the authors clarify the extent of originality in these sections? It is crucial for academic integrity to distinguish clearly between direct quotes and original analysis.
局限性
The authors discussed limitations in Section 7 Limitation and Conclusions.
We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score and your confidence. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Weaknesses
Typos mentioned in Weakness 1-4
- Thank you for pointing out these typos. We will fix them in the future version.
The manuscript mentions that is the eigenvector corresponding to the -th smallest eigenvalue, but it does not specify whether these eigenvalues are of the matrix or another related matrix. This omission could lead to confusion about the origin and relevance of these eigenvectors, particularly in how they are integrated into the model and influence the outcome of the subspace prediction.
- Given a linear system , these eigenvalues are derived from the same matrix . Thank you for pointing out this potentially misleading statement. We will fix it in the future version. In particular, we will replace the sentence in lines 164-165 "where is the eigenvector corresponding to the -th smallest eigenvalue, ." with "where is the eigenvector corresponding to the -th smallest eigenvalue of the given matrix , ."
In Section 4.1, the manuscript proposes using invariant subspaces associated with the smallest eigenvalues to train the subspace prediction module. However, an oversight in the theoretical analysis (Section 5.2) arises from applying this concept to potentially non-Hermitian matrices. The section (section 5.2) is based on Hermitian positive definite matrices, whose eigenvalues are all real numbers that can be orderly compared. Non-Hermitian matrices, on the other hand, may exhibit complex eigenvalues, complicating or nullifying the notion of "smallest" eigenvalues due to their lack of a natural order. This discrepancy raises concerns about the applicability of the method to non-Hermitian matrices, which are prevalent in many practical applications.
- Thank you for pointing out this problem. Analyzing perturbations for non-Hermitian matrices involves pseudo-spectral methods, which makes it difficult to investigate. To simplify the analysis, Section 5.2 explores how to select an appropriate subspace in the Hermitian positive definite case. In practice, for Hermitian positive definite matrices, the smallest eigenvalues are compared following the natural order. But for non-Hermitian matrices, the eigenvalues will be sorted by comparing their modulus, such that: where , , is the eigenvalue of a given non-Hermitian matrix . We will add these details to Section 5.2 to avoid the discrepancy.
Questions
In reviewing the theoretical analysis provided between lines 237 and 248, I observed a notable similarity in both the phrasing and the mathematical formulations with those presented in reference [23]. Could the authors clarify the extent of originality in these sections? It is crucial for academic integrity to distinguish clearly between direct quotes and original analysis.
- We are sorry for the misunderstanding caused by our improper citation. We fully adhere to academic ethics, and this issue was due to a writing oversight. Sentences in lines 237-248 provide the definitions and assumptions serving as the preliminaries for Theorem 5.2, which is mentioned in Theorem 5.2. To keep Theorem 5.2, which is cited properly, coherent with the original text, we reused the definitions and assumptions from reference [23], and this might potentially lead to disputes.
- Due to NeurIPS rebuttal restrictions, we are not allowed to submit a revised paper during rebuttal. We will make the quotes clear by adding the sentence " The following definitions and assumptions for Theorem 5.2 are from the reference [23]." at line 238 to indicate that content in lines 237-248 comes from the reference [23].
Dear Reviewer iSRc,
We are writing as the authors of the paper titled "Neural Krylov Iteration for Accelerating Linear System Solving" (ID: 18161). Thanks again for your valuable comments and constructive suggestions, which are of great help to improve the quality of our work. We have followed your suggestions to significantly enhance the quality of our work. As the deadline for the author-reviewer discussion period is approaching (due on Aug 13), we are looking forward to your further comments and/or questions.
We sincerely hope that our rebuttal has properly addressed your concerns. If so, we would deeply appreciate it if you could raise your scores. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Best regards,
Authors
Thanks for the response of the authors.
All of my previous concerns have been addressed in the rebuttal, I think this is a good work and I would like to raise my score from 5 to 6 for this submission.
We appreciate your positive feedback and the efforts you made in reviewing our rebuttal. Thanks for your constructive comments and valuable suggestions. We will incorporate them into our paper in the future version.
This paper proposes a data-driven approach to accelerate solving linear systems. Linear Systems are widespread in scientific computing applications like solving partial differential equations (PDEs), nonlinear systems, etc. so this method has potential for major impact. The proposed method builds upon the idea of neural operators that learn mapping between function spaces.
While most prior methods have attempted to accelerate solving linear systems by predicting a better initial guess, this work instead predicts the matrix invariant subspace. It uses this subspace to accelerate the Krylov Iterations. The method is validated on linear systems originating from PDEs and achieves speedups of around in computation time.
优点
- The presented method is novel, and to the best of my knowledge there doesn't exist a method that uses a similar approach
- Experiments are well designed and use strong baseline methods (GMRES from PETSc). Speedup in computation time and iteration count over such a strong baseline presents a strong case for the proposed method.
缺点
- The Neural Operators need to be trained on each individual problem. So while there is an improved convergence speed, there is an increased overall time when the training time is taken into account. However, this is a pretty common shortcoming across neural operator approaches and hence I haven't used this shortcoming as part of my overall scoring.
问题
- Are there any cases where GMRES (with no-preconditioners) fails to converge while NeurKItt can solve it because it is data-driven?
- Is it necessary to learn the NO for each system? Or is it possible to learn a single network and reuse it?
- (Maybe I missed this) Is it possible to include the training times for the neural networks?
局限性
N/A
We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score and your confidence. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Weaknesses
The Neural Operators need to be trained on each individual problem. So while there is an improved convergence speed, there is an increased overall time when the training time is taken into account. However, this is a pretty common shortcoming across neural operator approaches and hence I haven't used this shortcoming as part of my overall scoring.
- We agree with the reviewer that training the neural network involves additional costs, and it is a shortcoming of any neural operator approach. But considering the benefits of acceleration over millions of linear system solving, such training costs could be negligible due to the time saved. This is one of the reasons why we claim the feasibility of NeurKItt in practice.
Questions
Are there any cases where GMRES (with no-preconditioners) fails to converge while NeurKItt can solve it because it is data-driven?
- There might be the case where GMRES fails but NeurKItt can solve.
- Specifically, NeurKItt consists of the subspace prediction module and acceleration module. The subspace prediction module is data-driven, while the acceleration module uses the predicted subspace to improve the Krylov iteration convergence speed. The case mentioned is due to the acceleration module that replaces the original Arnoldi relation with , which takes the invariant subspace into iteration. It can also be considered as a preconditioned linear system solving, which replaces the original linear system problem with , where . This change directly improves the convergence speed while lowering the condition number of the given matrix , which leads to the case that GMRES fails to but NeurKItt can solve. The data-driven module (Subspace Prediction Module) makes the improvement possible, but it's not the direct answer.
Is it necessary to learn the NO for each system? Or is it possible to learn a single network and reuse it?
- We'd like to give answers given different contexts.
- First, given the case that systems derived from different PDEs, we did not conduct relevant experiments. But we think it is possible to learn a single network and reuse it in the given context, while the network is pre-trained. Recent works like , , and , give experimental proofs that the neural operator can be a pre-train model, which can be reused for different PDEs with only fine-tuning.
- Second, given the case that systems derived from the same PDE but with different scales, like different sizes of the linear systems, we have conducted an experiment to show the potential of NeurKItt in Appendix I. In particular, we train the neural operator on linear systems with a fixed matrix size but predict linear systems with larger sizes, then use the predicted subspace to accelerate the solving for those linear systems with larger sizes. The experimental results show that NeurKItt successfully accelerates the solving, which indicates the potential to learn a single network and reuse it for the linear systems derived from the same PDE but with different scales.
(Maybe I missed this) Is it possible to include the training times for the neural networks?
- Yes, we have reported the training time for each dataset in Appendix H.2. In our experiments, all the training converges within 150 minutes, and keeping training has no additional performance improvement once reaching the convergence.
[1] Yang, Liu, et al. "In-context operator learning with data prompts for differential equation problems." Proceedings of the National Academy of Sciences 120.39 (2023): e2310142120.
[2] Hao, Zhongkai, et al. "DPOT: Auto-regressive denoising operator transformer for large-scale pde pre-training." arXiv preprint arXiv:2403.03542 (2024).
[3] Zhou, Anthony, et al. "Strategies for Pretraining Neural Operators." arXiv preprint arXiv:2406.08473 (2024).
Dear Reviewer XS9H,
We are writing as the authors of the paper titled "Neural Krylov Iteration for Accelerating Linear System Solving" (ID: 18161). Thanks again for your valuable comments and constructive suggestions, which are of great help to improve the quality of our work. We have followed your suggestions to significantly enhance the quality of our work. As the deadline for the author-reviewer discussion period is approaching (due on Aug 13), we are looking forward to your further comments and/or questions.
We sincerely hope that our rebuttal has properly addressed your concerns. If so, we would deeply appreciate it if you could raise your scores. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Best regards,
Authors
I would like to thank the authors for their detailed responses to my questions. Thank you for pointing out the relevant sections in the appendix which clearly answer my remaining questions.
Considering this and the response to Reviewer CSA7 that improves upon the explanation of the method, I have increased the score to 8.
We sincerely appreciate your positive feedback and the time you’ve dedicated to reviewing our rebuttal. Thank you for appreciating the merits of our work. Your constructive comments are invaluable to us.
The paper develops a new method, NeurKItt, for accelerating the solution of non-symmetric linear systems. NeurKItt constructs an approximate invariant subspace of non-symmetric matrix A using the Fourier Neural Operator. This invariant subspace is then used as the initial subspace within GMRES. The paper provides theoretical support for the proposed method and empirical results showing the benefits of NeurKItt.
优点
1.) The paper develops an interesting new method for solving non-symmetric linear systems based on operator learning. In particular, I have not seen the Fourier Neural Operator used in this way before. I believe the idea has the potential to be quite useful and could have valuable implications for other problems.
2.) The authors provide theoretical analysis supporting the method.
3.) The method performs well in practice, which is what is most important. NeurKItt seems to yield significant reductioning in the number of GMRES iterations, and also yields good wall-clock time speed-ups (which is particularly impressive as these matrices are very sparse). So I think the method has the potential to be very useful in practice.
Overall, the paper proposes an exciting new idea that appears to work quite well and has the potential to be quite useful. Moreover, the paper addresses an important hard problem: improving the solutions of non-symmetric linear systems. This setting is quite challenging, as the non-symmetry makes things difficult. Often, methods for accelerating GMRES are highly problem-dependent, so a method that can lead to generic improvements is significant. I'm happy to recommend its acceptance to NeurIPS provided an issue given below in the weaknesses section is properly addressed.
缺点
I'd say the main weakness of the paper is its presentation. In particular, a significant issue is that the paper lacks a precise description of how FNO is implemented to learn the invariant subspace. The appendix only briefly describes FNO but does not explain how the paper applies it. The current discussion in lines 148-151 is vague and unclear. In particular, the output of FNO is the output of an operator on a function, not a subspace. So, how are you getting the predictive subspace from the output of the FNO? I have ideas of how this is done, but rather than guessing, I would like the authors to explain the precise procedure adequately. This should then be added to the paper, ideally in the main body, but at least the supplement with clear pointers in the main paper of where to find the details.
I'm willing to consider raising my score if the authors address this point well. If they don't, I will lower my score, as I cannot recommend acceptance of a paper for which a significant part of the methodology is unclear.
Aside from this, the authors should do a careful spell-check, as there are many typos throughout, which is somewhat distracting.
问题
See weaknesses.
局限性
The authors have adequately addressed the current limitations of the method.
We thank the reviewer for the insightful and valuable comments. We respond to your comments as follows and sincerely hope that our rebuttal could properly address your concerns. If so, we would deeply appreciate it if you could raise your score. If not, please let us know your further concerns, and we will continue actively responding to your comments and improving our work.
Weaknesses
How are you getting the predictive subspace from output of the FNO? I have ideas of how this is done, but rather than guessing, I would like the authors to explain the precise procedure adequately. This should then be added to the paper, ideally in the main body, but at least the supplement with clear pointers in the main paper of where to find the details.
-
Here, we use the 2D Darcy flow problem to demonstrate how the Fourier Neural Operator (FNO) predicts the subspace given the input function.
- Let the input function be from the 2D Darcy flow problem, where is the resolution of the input function, which yields a linear system for numerical solving, where matrix . We aim to predict a subspace , as a matrix , for matrix in the linear system. First, the FNO's lifting layer transforms to a higher-dimensional representation , while the 3rd dimension is channel dimension. After Fourier layers forward, we have the . Because Fourier layers keep the shape unchanged, we apply a transformation to map the to the desired space, , with .
- In practice, transformation is a stack of transformation layers. It first flattens the first and second dimension of , obtaining . Then a fully-connected neural network (FNN) applies the mapping to , obtaining . And another FNN applies the mapping to , obtaining the output . Finally, we apply QR decomposition to for orthogonalizing, obtaining .
-
To address the problems identified in the reviewers' comments, we will make several changes to our paper. First, we will replace lines 148-151 with more precise content to clarify the subspace prediction task. The revisions include the problem setup and the detailed description of how to predict subspace by FNO. Second, we will add the 2D Darcy flow example mentioned above to Appendix C to give an image of how to predict the subspace by the FNO. The sentences quoted below are the revised version of lines 148-151:
-
"Generally, for a linear system derived from the parametric PDE problem, to predict its invariant subspace , the input to FNO is the input function from the given PDE, where . We provide a detailed discussion in Appendix B about how to build a linear system problem from a PDE problem, and what is the input function. Our task is to learn the mapping between two Hilbert space using FNO.
-
For FNO, the lifting transformation first lifts the input to a higher dimensional representation , where is the number of channels. Then we feed the to Fourier layers. After Fourier layers forward, we have from the last Fourier layer, which keeps the same shape as . The FNO's output is the projection of by the transformation . NeurKItt then uses QR decomposition to orthogonalize the matrix , obtaining the predicted subspace . We provide more details about how to predict the subspace given the 2D Darcy flow problem in Appendix C."
-
-
Due to NeurIPS rebuttal restrictions, we cannot submit a revised version during the rebuttal. We will incorporate these modifications into our paper in the future version. If these modifications do not adequately address the issues raised, please let us know your further concerns, and we will continue to actively respond to your comments.
Aside from this, the authors should do a careful spell-check, as there are many typos throughtout, which is somewhat distracting.
- Thank you for bringing this to our attention. We will conduct a thorough spell-check and correct any typos in the future version.
I appreciate the authors' detailed reply to my concerns. This adequately addresses the issues I had with the presentation. I trust the authors to include this in the final revision, as they have promised here. Given this I will raise my score from 7 to 8, and presentation from 1 to 3.
Thank you for your positive feedback and for taking the time to read and respond to our rebuttal. We appreciate your constructive comments and concrete suggestions. We will incorporate them into our paper in the future version.
Keylov subspace method is a standard choice in solving linear system, it suffers from low efficiency when the iterates is less-than-ideal, making it not practical for large-scale sparse linear system. Regarding this difficulty, the submission proposes a neural operator based method, named Neural Keylov Iteration (NeurKItt), which uses a neural operator to predict the invariant subspace of the linear system and in turn to speed up solving the problem. Theoretical analysis is provided, and extensive numerical experiments demonstrate the advantages of the method.
From the reviews, the reviewers reach a consensus that the paper contains significant contribution with two ratings as strong accept. In particular
- Novel idea: solving linear systems via operator learning is new to the literature, and novel contribution to the community. Theoretical guarantees are also provided.
- Strong performance: the method manages to significantly reduce both number of iteration and wall-clock time, implying promising applications.
During the author-reviewer discussions, the authors managed to addressed the reviewers' concerns, and they are easy to incorporate into the final version of the paper. Regarding Reviewer g2gN, though his rating is the lowest, his review and comment are less convincing that the others, implying that he may fail to fully follow the paper. As a result, his rating is not fully taken into my evaluation (but his comments together with Reviewer iSRc's show that the authors need to work on the presentation of the paper). Given the novel idea and strong performance of the method, I recommend for a spotlight.