KoNODE: Koopman-Driven Neural Ordinary Differential Equations with Evolving Parameters for Time Series Analysis
摘要
评审与讨论
Edit: Having read through the rebuttals and the other reviews, I am satisfied with the paper and have increased my score from 3 (weak accept) to 4 (accept). I am grateful to the authors for taking the time to respond, and have left remaining specific thoughts in the Rebuttal Comment.
Original Review:
This paper introduces Koopman-Driven Neural Ordinary Differential Equations (KoNODEs). The motivation is that Neural ODEs given by:
Can benefit from making the parameters time-dependent:
.
One example being that the underlying dynamics might remain unchanged, but the parameters of the dynamics do, for example in engineering the wear and tear of instruments can change the observed dynamics, even though the underlying physics remains unchanged.
An existing solution is ANODE-V2 that models using a neural network and , so the network parameters are given by another ODE.
This work adds further modeling restrictions to the ODE with the aim of modeling more complex behaviors, for longer time series. This is achieved by using Koopman theory, which at a high level says that non-linear dynamics can be achieved by using linear dynamics in a different space. This paper explores how this can be achieved in practice for Neural ODEs, provides an implementation, theoretical analysis and experimental evidence showing modeling Neural ODEs using Koopman operators is an effective way to restrict the dynamics of the parameters to actually improve the model.
给作者的问题
- Is it possible to represent KoNODE as another larger first order ODE:
, ,
If this is the case, it might be easier to write as and as . That way the whole system can be solved together in one ODE solve, which will likely reduce the computation required per solve. This also makes the adjoint sensitivity derivation trivial since the standard adjoint sensitivity can be applied to the extended ODE system.
-
In Figure 4C, why does error increase as the dimension of increases. Is it hard for KoNODE to learn redundancy? Or is it encouraged to use all of its dimensions to learn the underlying Koopman dynamics? The theoretical bound presented says that the error decreases with more dimensions, why is this not the case? Does the model overfit when it has more capacity, i.e. is not large enough, so is negative in Theorem 3.2 so the error bound does not decrease with ?
-
Can you rephrase your argument around Table 4, the argument is that the Euler method can benefit from KoNODEs, with more accurate dynamics. However this does not make sense to me. Firstly Table 4 shows that the Euler method is also better using NODE, so it does not empirically hold. Additionally, RK4 and DOPRI5 should be more accurate than the Euler method during the ODE solve, so should also benefit from improved modeling of the dynamics.
Based on the answers to these questions and the others mentioned I will increase my evaluation.
论据与证据
The claims presented are supported by experimental evidence. However, there are some relevant baselines that would improve the strength of the evidence. For example ODE-RNN or Latent ODE (https://arxiv.org/abs/1907.03907), or Neural Processes (https://arxiv.org/abs/1807.01622) as another method. These are common baselines that would improve the experimental results.
I am also concerned with Figure 3, which shows vanilla NODE not performing on the spiral synthetic dataset. Why is this the case since it does perform in the original paper?
方法与评估标准
The experiments carried out make sense for the application. The datasets are a good mix of synthetic to demonstrate key points and real to show that the method works on real data.
理论论述
The theoretical claims as far as I can tell are correct. I have not checked the proofs in the appendix in significant detail. However, I am confused by Line 254 on the left:
What does it mean for a logarithm to be subscripted? Previously it is given that , so how can be larger than and ? Especially given that is the amount of data, but seems to be the dimension of the Koopman space?
实验设计与分析
The experiments and analyses associated are sound. But as mentioned would be improved with stronger ODE baselines (e.g. ODE-RNN, Latent ODE) and stronger baselines for generalising to new dynamics (e.g. Neural Processes).
补充材料
I read the additional related work, inspiration of KoNODE, and the additional experiments.
与现有文献的关系
The key contributions are framed appropriately in relation to the scientific literature.
遗漏的重要参考文献
There are none as far as I can tell.
其他优缺点
Stengths:
- The idea is robust and theoretically motivated
- The datasets are diverse and relevant
- The experiments support the claims
Weaknesses:
- The writing needs some work, see the Other Comments or Suggestions section of the review for specific cases. The description of KoNODE at points is overly complicated, or unnecessary. For example, at the beginning of page 4, one possible matrix is described in detail, before describing a second matrix with block diagonals, the first version is never used and therefore is not necessary to describe. Section 3.2.2 is another section that overly complicates the description, where the point is to say that the adjoint sensitivity method is used to calculate gradients, which is already described in detail in the original Neural ODE paper
其他意见或建议
This is a list of contained writing changes I would make. They are suggestions so are not required:
- Line 17 Right: "The Ordinary Differential Equations" should be just "Ordinary Differential Equations"
- Line 19 Right: "of ODEs" should be "of the ODE"
- Line 38-39 Right: "encodes the deeper underlying dynamical principles of the evolution (namely deep-level information) inherently" should be "inherently encodes the underlying dynamics of the evolution (namely deep-level information)"
- Line 59 Left: "Based on the insight" should be "Base on this insight"
- Line 80-81 Left: "and ultimately governed by a deeper-leveled, intrinsic linear core - the Koopman linear dynamics" should be "and ultimately governed by the underlying Koopman linear dynamics"
- Line 66 Right: "Applying the Koopman Theory" should be just "Applying Koopman Theory"
- Line 94-96 Right: ", rather than propagating activations through discrete layers as in recurrent or deep networks. Neural ODEs employ numerical solvers" should be ". Rather than propagating activations through discrete layers as in recurrent or deep networks, Neural ODEs employ numerical solvers"
- Line 124 Left: "The Koopman Theory" should be "Koopman Theory"
- Line 124-125 Left "dynamic system analysis" should be "dynamical system analysis"
- Line 141 Left: "commonly finitely approximated" can be just "commonly approximated" since the phrase "finite number" is used later in the sentence
- Line 144 Left "The Koopman operator are" should be "The Koopman operator is" or "The Koopman operators are"
- Line 112 Right: "within the NODE setups" should be "within the NODE setup"
- Line 117-120 Right: "Besides, we assume the trajectories of the intrinsic linear dyanmics are represented as w(t)" should be "We let the intrinsic linear dynamics be represented as w(t)"
- Line 126-144 Right: Describing the three hierarchical levels should refer to the "dynamics" not the "dynamic"
- Line 158 Right: The sentence "We model the h by the neural network due to its flexibility" is not needed, but should be "We model h using a neural network due to its flexibility"
- Line 169 Left: "in replace of the conjugated imaginary eigenvalues of A" should be "that represent the imaginary eigenvalues of A"
- Figure 2 caption: Should refer to the "dynamics" not the "dynamic"
- Figure 2 caption: "over altogether N time points" should be just "over N time points"
- Section 4 description: "just as the red line" and "just as the green line" should be "shown by the red line" and "shown by the green line"
- Line 287 Left: "At last, we get the loss between the predicted trajectory and the true trajectory", should be "Using the true trajectory and the predicted trajectory, we can calculate the loss"
- Line 295-300 Left: "We serve w0 as an encoding" should be "w0 is an encoding" and "Learnable setting when given sequence as input" should be "Learnable setting when given a sequence as input"
- Line 375 Left: "Preformance metric" should be "Performance metric"
- Line 405 Left: "should inherently low-dimensional" should be "should inherently be low-dimensional"
- Lines 433-439 Left: When using quotation marks in Latex they can be properly displayed by changing "S" to ``S''
- Line 420 Right: "ETTh1" should be "ETTH1"
- Line 427 Right: "and the Koopman modeling" should be "and Koopman modeling"
Thank you for your deep understanding of our method and insightful suggestions.
[1. comparison to baselines]
We have included the baselines mentioned, Latent ODE (ODE enc) and Neural Processes, in supplementary_table_reviewers_Snqd_waH1.pdf, https://anonymous.4open.science/r/KoNODE-D8F2/. We didn't compare ODE-RNN due to its official code not supporting prediction tasks.
[2. failure on spiral synthetic dataset]
Lines 283-285 to the right clarified that the spiral dataset we were using is obtained by where , which is more complicated than the one used in vanilla NODE, i.e., . We did successfully reproduce fittings for the spiral proposed in both ours and NODE yet the modified data is more complicated so would demonstrate the limitation of the conventional model.
[3. in line 254]
Sorry for the negligence in the notation. We will clarify in the revision that while the coordinates even for as the true Koopman space has an infinite number of bases \left\\{u_i\right\\}_{i=1}^\infty (though no more than -dimensions due to the data-driven scenario).
[W1. writing]
Thank you for the valuable suggestions. We will modify accordingly (including the minor fixes and the simplifications of the KoNODE description and Section 3.2.2).
[Q1. larger ODE expression]
It is a promising insight that will definitely be easier to write. However, this pipeline sacrifices the consistency of function and does not significantly reduce the computational cost (per solve). The reasons are: (1) In the forward integration, both pipelines should go through function , compute , and another network being either or ; (2) In the adjoint method, both pipelines require the gradient through network (or ) which requires a backpropagation through them; (3) The two networks, and , both require the potential to map from the -space to the -space, hence the sizes of the networks do not differ much.
[Q2. Fig. 4C]
Thank you for your deep understanding of our method. Fig. 4(c) shows a curve dropping in and rising if with a tolerable relative perturbation. Experimentally, more weights are introduced in networks, which may add extra uncertainty and instability during training when is large and [W2 & Q2. selection of Koopman dimension], Response to Reviewer iauo explains the reasons why the practical dimension is not that stable. Theoretically, the error bound in Thm. 3.4, Manuscript is majorly dominated by two terms, one with a factor of in coefficient and the other with the order of , which does indicate the instability when is too large and may infer overfits. On the other hand, the order is negative only when is too small that the first bases fail to successfully model the true observable function , hence this scenario only influences the left-hand side of the graph when is below the theoretical lower bound given in Thm. R.4, [W2 & Q2. selection of Koopman dimension], Response to Reviewer iauo.
[Q3. Tab. 4]
Sorry for the negligence of the phenomenon as the table aimed at showing the more significant improvement from the KoNODE framework using the Euler method. We will rephrase the interpretation in the revision. Regarding the lateral comparison, the integration by RK4 and DOPRI5 is undoubtedly more accurate yet the superiority of Eurler is caused by the different intrinsic objectives under the regularly sampled trajectory, i.e., the network for Euler estimates for a fixed but the others for a mutable . The fixed reduces the complexity of the target function and thus simplifies the task using Euler.
Thank you for taking the time to respond. I have read the other reviews and responses, and am generally happy with the paper now, I will raise my score accordingly. Please see remaining comments below:
Additional Baselines and ODE-RNN: Thank you for running the additional baselines, as well as the additional baselines for Reviewer Snqd. The results are still good for KoNODE, which is reassuring. I'm not particularly convinced by the reason for not including ODE-RNN, the prediction can simply be a function of the hidden state , where follows an ODE-RNN. This should not be too hard to implement. However, that being said I also recognise there is already a large variety of baselines, and another baseline is not necessary.
Spiral Dataset: Thank you for the clarification, my apologies for not noticing this difference. The paper would be improved with clear communication around this point. That is, at the start of Section 5.1, saying something along the lines of "We adapt the fitting task proposed in Chen et al. 2018, by adding a time dependency to the matrix . We add sines and cosines to produce the following new time-dependent matrix: ..."
Line 254: My apologies, but this still is not clear to me. Please can my error in the following reasoning be identified and explained:
- Lines 140-144 Left: In practice the infinite dimensional Koopman space is finitely approximated by modeling a finite number of bases , we have basis functions.
- Line 253 Left: . These are inner products between the function of interest in Theorem 3.2 and the finite Koopman bases . Each of these give scalars, and we have to have finitely many since we are doing a practical implementation. So there are components of since there are basis functions.
- If is dimensional then can only be between and , not and .
- This theorem is about the "Order-m Koopman Operator", so we have basis functions, is the function of interest, and we have data points. Why do we have new basis functions per data point? The true Koopman space has infinite, but this theorem is about the approximation, where there are finitely many?
It is unclear what I am not understanding, and more broadly, as we have discussed, the paper will benefit significantly in the long-term by making the theory clearer.
Larger ODE: This point has been misunderstood, I was really asking about the specific implementation, i.e. the code. It wasn't clear to me if the ODE for was solved first, and then the ODE for , requiring two ODE solves in series and storing the trajectory. I made a mistake and realised the combined ODE could be written as:
,
Without an unnecessary ODE for . Implementing this way requires only one ODE solve and not storing a dense trajectory . Having now looked through the supplementary code I realise this is exactly how KoNODE has been implemented, so I accept I got this wrong. In this case I'm curious what you think would happen if KoNODE is implemented differently:
,
That is, rather than using to generate parameters for , could be used as a concatenated input to which has fixed parameters ? The universal approximation theorem would make this valid, but how do you think it would affect learning, parameter efficiency (no network, not so many evolving s), computation?
Figure 4C: Thank you for your answer. It sounds like it is a case of overfitting am I correct? How would you suggest tuning , are there good values to begin with relative to and the dimension of ?
Table 4: This is not a major concern (it's in the Appendix), and I am still happy with the paper. But I'm still not convinced by this reasoning. No matter the ODE solver, they are predicting for a fixed set . All that should change theoretically is the accuracy of the solve. In practice of course this affects the gradients and therefore the training, because they produce different trajectories. But if gradient descent is run long enough they should reach similar results. Is it because Euler produces "simpler" trajectories, i.e. it will not focus on significantly complicated parts of the trajectory, and therefore despite being a worse solve it is more stable? Maybe training with Euler at first, then changing to a more accurate solver at test time or even part way through training could be better? This is very much a side point, and I know was not the point of the Table (which was to show that KoNODE is robust to the solver), I just think it is curious, and goes against what I've experienced between solvers.
Thank you sincerely for raising the score and for the time spent. Your intelligent feedback has significantly enhanced the depth and clarity of our manuscript.
[1. Additional Baselines and ODE-RNN] Thank you for your thoughtful suggestion. We agree that, theoretically, the ODE-RNN is capable of making predictions. However, we omitted it because the official code raises an error: "Exception: Extrapolation for ODE-RNN not implemented," which prevents a direct comparison. Due to the limited time of the rebuttal, we could not re-implement it from scratch, and we sincerely apologize for this.
[2. Spiral Dataset] Thank you for the helpful suggestion and for pointing this out. We will modify the manuscript accordingly.
[3. in line 254] We apologize for the confusion caused by the notation and thank you for the patient analysis. The first three reasonings are correct. For the fourth one, the basis functions for the true Koopman space is infinite, hence we are able to identify infinite number of . The coordinate in line 254 is the only one in which can be greater than , we will replace it with instead to avoid confusion.
Regarding the questions in reasoning four, we do not have new basis functions per data point. The reason why is that the space the observable function in has a maximal dimension of . As Lemma R.3, Response to Reviewer iauo, defines a possible observable function for any trajectory to model the dataset, these functions span an observable function space with a dimension no more than . Consequently, .
[4. larger ODE] Sorry for the misunderstanding and thank you for the clarification. Your current interpretation of the KoNODE implementation is very precise.
We are very much thrilled by the acute insights you possess. Regarding your proposed alternative formulation (Imp. II), we agree that this is a theoretically valid and feasible variant. This formulation resembles the idea behind ANODEs (https://arxiv.org/abs/1904.01681), where additional variables are appended to the state to enhance expressivity.
While both implementations are in principle functionally equivalent under the universal approximation theorem, their motivation differ. Our design (Imp. I) is motivated by a desire to extracte deep-level dynamic structures through . In contrast, Imp. II treats as a refined input to enhance the model’s capacity, rather than explicitly encoding the system’s intrinsic dynamics. Technically speaking, any output neurons link to in Imp. I while they may disconnect to it and only be activated by using Imp. II.
The following discussed practical factors.
(1) Learning Dynamics. When is concatenated with , its influence on the dynamics must be learned implicitly through a fixed network . Without architectural constraints, this may often lead to learned representations that are dominated by , with playing a minor role, especially when is low-dimensional.
(2) Parameter and Computational Efficiency. Although Imp. II appears simpler, it places the full burden of modeling complex, time-varying interactions on a single network , which may increase training time or convergence difficulty. As highlighted in [1], effectively modeling intricate dependencies from low-dimensional auxiliary inputs (like ) often requires large networks. By separating the responsibilities— modeling latent evolution and focusing on observable dynamics—Imp. I achieves better parameter specialization.
[5. Figure 4C] Yes, "overfitting" is indeed precise. A suggested selection criterion of is increasing it from where is the model size and is the rank of data (described thoroughly in Theorem R.4, Response to Reviewer iauo) until the performance does not improve. An empirical choice is 10 for simple dynamics in 2D space and 100 for high-dimensional ones.
[6. Table 4] We believe the superiority of the Euler solver is caused by the technical reason that the regression w.r.t. low-dimensional inputs (such as ) tends to be more difficult for neural networks to model [1]. Specifically, the Euler method only models but high-order methods RK4 and DOPRI5 require the model of where which rely on more sensitive modeling of , making these models suffer more from the intractability caused by the low-dimension of time. The reason why the correspondence of and relies less on the modeling of is implied by the mapping from state to time in Lemma R.3, Response to Reviewer iauo.
As for the further questions, (1) yes, we agree that the Euler solver may yield more stable results; (2) regarding switching solvers later, we believe this is interesting and worth exploring.
[1] See link: https://link.springer.com/chapter/10.1007/978-3-030-47358-7_27
An architecture based on neural ODE with time-varying parameters is proposed. The idea is to model the dynamics of the parameters with latent linear dynamics, which the authors motivate by referring to the Koopman operators. The superior prediction performance of the proposed method is empirically demonstrated.
Update after rebuttal
Thanks for the clarification, the additional discussion will make the paper more convincing. I keep my originally positive score.
给作者的问题
I do not have major questions that strongly affect my evaluation.
论据与证据
I found two points in the claims. One is about its practical utility, i.e., better performance in prediction. I think this aspect is nicely demonstrated with the empirical results.
The other point of the claims I found is about "understanding" of data or models. For example, the authors claim:
We propose a three-level hierarchical architecture ... that deepens our understanding of system behavior and its governing laws.
By leveraging Koopman operators, our framework could use spectral tools for system analysis ...
Although I do not deny such a possible utility of the method in general, these aspects do not seem supported by concrete empirical observations.
方法与评估标准
The time-series prediction experiments seem valid.
理论论述
A bunch of theoretical claims are presented in Section 3.3 and the appendices. They may be okay, but I do not see how relevant they are to the proposed method. The presented theories are on the estimation error of the Koopman operators, but the proposed method consists of not only the Koopman part but also other components such as the map from to . It would be nice if the authors could elaborate more on the motivation of the theoretical arguments: for example, the overall picture (purpose) of the analyses, key assumptions, and remaining gap to fully analyze the prediction errors of the proposed method, if any.
实验设计与分析
As mentioned in the "Methods And Evaluation Criteria" section, the experiments make sense.
补充材料
I took a quick look at the appendices but did not check any details.
与现有文献的关系
Time-series prediction is widespread in any domain of science and engineering.
遗漏的重要参考文献
N/A
其他优缺点
N/A
其他意见或建议
N/A
Thank you for your insightful comments.
[1. empirical observation supports]
We have added empirical evidence and analysis to support the two quoted claims. We will include concrete empirical observations in the revision. Please refer to [W1. lack sufficient interpretability], Response to Reviewer Snqd for further details, and spectrum_reviewers_Snqd_1pn6.pdf, https://anonymous.4open.science/r/KoNODE-D8F2/ for the visualization.
[2. theoretical claims]
Thank you for the suggestion and we are sorry for the confusion. We will clarify two motivations in Section 3.3 in the revised manuscript, i.e., the presented error provides (1) a theoretical reference for the choice of dimension , and (2) proof for the minority of errors caused by the additional Koopman module compared to conventional ODE models, apart from the potential improvement due to the advanced fitting ability. The theoretical claims are given under the assumption that the parameters indicate the running dynamics, which are intrinsically determined for a given trajectory . Consequently, an accurate model of directly leads to the accurate estimation of . Thm. R.6 below conducts a theoretical comparison of the overall prediction error between the proposed method and a conventional ODE based on the given theories to indicate the superiority of our method.
Definition R.5 (The Static Dynamic Model Condition)
The dynamic satisfies the condition if and only if for the function and any , ,
where is randomly drawn from the ideal trajectory defined in Eq. (1), Theorem R.4, [W2 & Q2. selection of Koopman dimension], Response to Reviewer iauo and is the LSE estimation of a static . The condition holds when the true dynamic behind the trajectory is static which ensures a good fit of the NODE.
Theorem R.6
During the modeling of trajectory , the estimation error upper bound of the proposed model is smaller than that of a conventional ODE method if the observable function does NOT satisfy the static dynamic model condition, i.e., for the bound
there exists,
where is the trajectory predicted by the proposed method while consists of the static parameters in the conventional ODE method.
Proof: We use the notations in Def. R.5. Apparently, Eq. (2) gives a bound of the estimation error where the "sup" exists as is Lipchitz continuous, and the bound is reached for linear mapping .
Note that in NODE and in the proposed Koopman model where satisfies the Koopman model of matrix . Thm. 3.4, Manuscript shows that
with representing the RHS . As the base for observables , Lemma C.12, Appendices implies if and ,
Consequently, if we let
then .
On the other hand, for failed to satisfy the static dynamic model condition at error we have
where and .
As a result, with a probability of , the error bound in Eq. (2) for the proposed method is lower than that of the conventional method.
This paper explores the challenge of modeling time series with NODEs. The authors propose a Koopman-driven framework named KoNODE that hierarchically encodes system dynamics through evolving ODE parameters and Koopman linear operators. Specifically, they introduce a three-level architecture—spanning observed state dynamics, parameter dynamics, and intrinsic Koopman linear dynamics—to disentangle surface-level behaviors from fundamental governing rules. Extensive experiments on synthetic and real-world datasets validate the effectiveness of the proposed approach.
update after rebuttal
All my concerns have now been adequately addressed. Overall, this is a good work. However, considering the limited application scenarios, I will maintain my original score of "weak accept."
给作者的问题
-
Line 178-179, the left half of the page. Could the author clarify why the assumption makes sense?
-
In Section 3.3.2, Theorems 3.3 and 3.4 are formulated in Hilbert space rather than in the observation space of the data. Should and be maintained as characteristic functions to ensure these two theorems valid?
论据与证据
N/A
方法与评估标准
N/A
理论论述
N/A
实验设计与分析
N/A
补充材料
Some of them. C.3
与现有文献的关系
This work is beneficial for the development of AI for Science research.
遗漏的重要参考文献
Pls refer to Weaknesses
其他优缺点
Advantages:
- The proposed KoNODE framework is both theoretically grounded and empirically effective in capturing intrinsic system dynamics.
- The experiments are very detailed and thorough, clearly demonstrating the effectiveness and efficiency of the proposed method.
- The manuscript is well-structured and clearly written, facilitating easy comprehension for readers.
Weakness:
- From my opinion, the first contribution which is claimed as “uncovers the fundamental principles driving system evolution” appears somewhat overstated, as the identified deepest dynamics still lack sufficient interpretability.
- This framework is very generic, but its primary focus on time-evolving ODE parameters is not yet a common paradigm. Could the author compare their settings with other generalizable NODEs [1][2][3]?
Reference:
[1] LEADS: Learning Dynamical Systems that Generalize Across Environments. NeurIPS, 2021.
[2] Generalizing to New Physical Systems via Context-Informed Dynamics Model. ICML, 2022.
[3] Generalizing Graph ODE for Learning Complex System Dynamics across Environments. KDD, 2023.
其他意见或建议
N/A
Thank you for your constructive feedback.
[W1. lack sufficient interpretability]
We apologize for the unclear description of "revealing the fundamental driving forces of system evolution" and will clarify it in the Introduction and Experiment sections in the revision.
In our framework, the deepest dynamics, i.e., Koopman linear dynamics , are obtained through Koopman operators. Fundamental evolution principles are revealed by the spectrum of the operators, which correspond element-wise to matrix and are given by in our framework. Specifically, the real part of the eigenvalues indicates the evolution speed, while the imaginary part corresponds to intrinsic frequencies and periodicity. Analyzing the elements of could help identify the system’s dominant driving modes and frequencies and offer insights into stability, periodicity, or other inherent characteristics, thus gaining a mathematically interpretable understanding of the system's evolution.
To further demonstrate this, we provide a more detailed analysis and interpretation of the learned dynamics in Sec. 5.2 by visualizing the spectrum of approximate Koopman operator for the Oscillator dataset in spectrum_reviewers_Snqd_1pn6.pdf, https://anonymous.4open.science/r/KoNODE-D8F2/. First, the magnitudes of spectrum are approximately one for both two dynamic systems, indicating boundary stability, where the state remains on a stable trajectory without diverging or converging. Second, the dominant spectrum of the nonlinear oscillator (on the left in the figure) has similar imaginary parts (i.e., the frequency components are the same), indicating that the system exhibits clear periodicity. Modes with the same imaginary part have the same periodic components.
[W2. comparison to generalizable NODEs]
We have added the comparison between our framework with other generalizable NODEs, LEADS [1], and CoDA [2] in supplementary_table_reviewers_Snqd_waH1.pdf, https://anonymous.4open.science/r/KoNODE-D8F2/. Due to the lack of publicly available code for [3], we did not compare with it.
[Q1. A≡D]
As stated in lines 174-178 on the left, hence we can assume by modeling instead of . We will clarify the definition of to avoid confusion. The assumption is not only mathematically rigorous but, as addressed in [W1. lack sufficient interpretability], allows us to analyze the system's behavior directly by observing the elements of matrix .
[Q2. Hilbert space]
The theorems are formulated in Hilbert spaces for generality, and this is a very mild assumption. The observation space of the data naturally satisfies this, as we only require the data space to be complete and have an inner product. The two functions are not necessarily characteristic functions. As clarified in line 156 on the right, function h is maintained as the inverse of the characteristic function, while f is the differential function in the NODE framework shown in Eq. (2).
This paper proposes KoNODE, a hierarchical framework that integrates Koopman operators into Neural Ordinary Differential Equations (NODEs) to learn time-evolving parameters.The authors provide theoretical error bounds for the finite-dimensional approximation of the Koopman operator and show how the proposed method improves long-horizon prediction and generalization on both synthetic oscillators and real-world time series tasks.
update after rebuttal
Given the fact, that authors addresed all the raised question and provided a new theorem with the proof, I raise the recommendation to Accept (4).
给作者的问题
- How does KoNODE handle irregularly sampled data?
- Is there a systematic way to select the dimension of the Koopman space m? Can any data-driven rank estimation be integrated?
论据与证据
The paper states that modeling time-evolving ODE parameters via Koopman operators enables a deeper representation of underlying system dynamics, improves long-horizon forecasting, and generalizes better to unseen conditions. It has both theoretical evidence with finite-dimensional Koopman approximation error bound and experimental evidence with superior performance on the synthetic and real world data.
方法与评估标准
The authors conducted several experiments on synthetic and real-world data. They selected a wide range of methods including classic Neural ODEs as well as the models with evolving parameters (ANODEV2), that makes the comparison reasonable.
理论论述
There are several theoretical claims on the error bounds. Proofs, that are provided in the appendix, look reasonable, however, they may be a bit hard to follow, specifically with these equal signs with stars and asterisks, it is better to avoid such “nonlinearities” and be more consequent.
实验设计与分析
I reviewed both the synthetic and real-world experiments in detail, and the overall design and selection of baselines are sound. There are two minor concerns. The first one relates to the potential computational overhead introduced by the added Koopman operator layer. While the appendix does include some runtime analysis, it would be helpful to include a concise overview in the main text to clarify any changes in memory or runtime requirements. The second one is that while the authors provide some insight into Koopman operator dimensionality in synthetic settings, it remains unclear how these dimensional choices generalize to real-world scenarios or whether any systematic criterion (e.g., data-driven rank selection) could be used. Addressing these points would further strengthen the paper’s clarity and applicability. Aside from that, the experimental setup appears valid, with no major issues that would detract from the paper’s conclusions.
补充材料
Yes, I have reviewed the proofs briefly and the ablation study part.
与现有文献的关系
The paper builds on Neural ODE literature (Chen et al. 2018) and prior Koopman approaches (Lusch et al. 2018, among others). It extends ANODEV2 by modeling theta through a linear Koopman system rather than a general ODE, thus adding interpretability and improved long-horizon stability.
遗漏的重要参考文献
n/a
其他优缺点
The paper’s primary strengths lie in its novel integration of Koopman operators with Neural ODEs, supported by comprehensive theoretical analysis and strong empirical results across both synthetic and real-world datasets. Meanwhile, potential weaknesses include limited discussion of computational overhead in the main text (despite some runtime data in the appendix), and a need for deeper examination of how the chosen Koopman dimensionality affects performance in real-world scenarios and whether a systematic selection method could be applied.
其他意见或建议
n/a
Thank you for your valuable feedback.
[W1. changes in memory or runtime]
We will move the runtime analysis from the appendix to the main text to show our advantage in convergence rate and add a discussion on memory complexity. Regarding memory, we clarify that the memory of the proposed framework is compared to of NODE for , , being the dimensions of , , and the hidden layer respectively and the model size for the differential function .
[W2 & Q2. selection of Koopman dimension]
Thank you for the concern about the choice of the Koopman dimension . We provide theoretical lower and upper bounds for . Although other factors may influence the best picks, the bounds provide references for the systematic criterion. Specifically, Thm. R.1 shows that for the rank of trajectory space and Thm. R.4 shows that has a tight lower bound for second-order derivable , where is the dimension of (the model complexity of ) and is the rank of data. Note that the lower bound is commonly small for a network that is strong enough, yet the hyper-parameter may not be static in practice due to (1) the auxiliary dimensions caused by the designed form of the Koopman matrix, (2) the disturbance caused by the multiple choice of the dynamic as , and (3) the attempt to more accurate dynamic regression by multiple sets. The theorems are given as follows.
Theorem R.1 (The Theoretical Upper-Bound of )
Inequality holds if:
(1) The data trajectories are approximately in an -dimensional manifold , i.e.,
where is the projection of to the manifold.
(2) Function satisfies the Lipchitz condition and is differentiable.
Proof: Since admits a bijective chart , the map is bijective. This, together with the uniqueness of (by Picard–Lindelöf), implies that lies on an -dimensional manifold. Hence, .
Definition R.2 (Frontier Manifold)
For a dynamic in Euclidean space, a fixed time and origin , the frontier manifold is the quotient space of the equivalent class decided by the trajectories, i.e., satisfies (1) , and (2) the normal vector at is . Then, we define as the set of evolved with .
Lemma R.3
For , if the trajectory follows the dynamic
a scalar observable function maps it into a Koopman space.
Proof: Note that for , is uniquely identified by the Picard-Lindelöf Theorem. Moreover, all elements in a frontier manifold are on different trajectories, which creates a bijection between and . Let be the mapping from to time under the condition that .
Then, let be a scalar observable function,
where and are constants and is the element-wise inverse of a vector. Therefore,
Hence, lies in the Koopman space.
Theorem R.4 (The Theoretical Lower-Bound of )
To model the trajectory of if it is second-order derivable, the dimension of is ideally where is the rank of data.
Proof: For a differentiable , the dynamic best models the trajectory where
and being the pseudo-inverse of the Jacobian matrix.
Lamma R.3 infers that for , only one dimension of is needed. Thm. R.1 indicates that the gradient lies in a space with the highest rank of , thus the frontier manifold covers a dimension of at least . To ensure a full cover of -space in the initialization, an ideal dimension of is .
[Q1. irregularly sampled data]
NODE-based methods model the system continuously, allowing them to handle irregularly sampled data by interpolating for intermediate time points.
This paper combines Neural ODEs with Koopman operators in a hierarchical approach to model non-stationary dynamics. It provides theoretical error bounds and demonstrates long-term prediction and generalization on various synthetic and real-world examples. All referees agreed that this is a solid and novel contribution, with thorough evaluation and theoretical analysis, and all four recommended acceptance, with which I concur.
Besides what was stated in the reviews and related discussion, the authors may want to take the following additional remarks on board when drafting their revision:
-
One other point that came up and should be clarified in the revision is that Theorem 3.2 is considering the infinite basis case. Also, one referee felt that the presentation of the theory could be further improved.
-
Modeling non-stationary dynamics by assuming parameters to be time-dependent is a quite common approach in time series and dynamical systems modeling. I would like to encourage the authors to more thoroughly check the related lit. (maybe perplexity or just google), especially papers that discuss non-autonomous dynamical systems, and extend the coverage of related lit. which in my mind is a bit thin.
-
Likewise, LEADS and CoDA are somewhat older frameworks. There is much more recent work for hierarchical modeling with time-dependent parameters and testing generalization across domains. I would recommend checking the most recent NeurIPS and ICLR publications on this.