PaperHub
5.0
/10
Rejected4 位审稿人
最低5最高5标准差0.0
5
5
5
5
4.3
置信度
ICLR 2024

Spatio-temporal Twins with A Cache for Modeling Long-term System Dynamics

OpenReviewPDF
提交: 2023-09-20更新: 2024-02-11

摘要

关键词
long-term dynamical systemsspatio-temporal forecasting

评审与讨论

审稿意见
5

This paper proposes a framework called Spatio-Temporal Twins with a Cache (STAC) for modeling long-term dynamics of physical systems. The key ideas are: 1) Using a frequency-enhanced spatial module and an ODE-enhanced temporal module to model spatial and temporal relationships from complementary perspectives; 2) Introducing a cache-based recurrent propagator to store historical feature maps; 3) Optimizing the model with techniques like teacher forcing, Mixup, and adversarial learning. The authors construct a new fire dynamics benchmark and evaluate STAC on 14 datasets, showing superior performance over baselines.

优点

  1. The problem of modeling long-term dynamics is important with many applications. This paper provides a novel perspective by using a cache memory to enhance long-term dependencies.
  2. The motivation is intuitive and reasonable.
  3. This paper is well organized and clearly written.
  4. The new fire dynamics benchmark (FIRE) constructed in this work could facilitate future research in this domain.
  5. Comprehensive experiments on 14 datasets demonstrate the effectiveness and generalizability of the proposed STAC framework.

缺点

  1. Though this paper seems to be promising, I have to say that the novelty seems to be limited. The spatio-temporal twins are actually a two-branch model. Using frequency-based approaches in the spatial domain is nothing new. The temporal module is similar to SimVP v2's [1] but with an ODE solver. The cache memory [2] is also well developed.
  2. I really appreciate the experiments in this paper. However, the ablation study is not satisfying. The authors reported only one metric (RMSE) on only one dataset (Spherical Shallow Water). A more detailed ablation study is needed to figure out why this approach works.
  3. It lacks of complexity comparison. The authors should report the parameters and FLOPs of these baseline models.

[1] Cheng Tan, Zhangyang Gao, Siyuan Li, and Stan Z Li. Simvp: Towards simple yet powerful spatiotemporal predictive learning. arXiv preprint arXiv:2211.12509, 2022.

[2] Lee, Sangmin, et al. "Video prediction recalling long-term motion context via memory alignment learning." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021.

问题

  1. Please discuss the differences between STAC and other similar models.
  2. Could you add a more detailed ablation study? Considering there are many components, a more detailed ablation study can provide more valuable insights.
  3. Please discuss the complexity of these models.

I'm willing to raise my score once these issues have been well solved.

评论

Q2. I really appreciate the experiments in this paper. However, the ablation study is not satisfying. The authors reported only one metric (RMSE) on only one dataset (Spherical Shallow Water). A more detailed ablation study is needed to figure out why this approach works.

A2. Thanks for your comment. Our complete ablation experiments are as follows, verifying that each component of STAC complements each other. We have corrected this in the revised manuscript.

ModelMSE (Fire 10 - 10)SSIM (Fire 10 - 10)MSE/100 (KTH 10 - 20)SSIM (KTH 10 - 20)MSE x 1000 (SWE 20 - 140)SSIM x 1000(SWE 20 - 140)Avg Training Time
STAC0.048792.870.298392.830.82497.68242s / epoch
STAC w/o FNO0.075684.540.302190.792.23579.43239s / epoch
STAC w/o Transformer0.050691.210.476583.721.09394.32192s / epoch
STAC w/o FSM0.076585.430.489382.92.35276.54187 / epoch
STAC (Replacing OTM with ConvLSTM)0.172365.691.204365.421.84583.43321s / epoch
STAC (Replacing OTM with PredRNN)0.392257.98OOMOOMOOMOOM654s / epoch
STAC w/o OTM0.080282.330.564478.432.15379.97209s / epoch
STAC w/o IFTM0.059690.330.732170.431.18391.23237s / epoch
STAC (Replacing CRP with Recall gate)0.099879.490.514580.395.98654.32598s / epoch
STAC w/o CRP0.054390.830.334387.386.32245.65200s / epoch

Q3. It lacks of complexity comparison. The authors should report the parameters and FLOPs of these baseline models.

A3. Thank you for the valuable comments about comparing the complexity of our STAC framework with the baseline model. We have corrected this in the revised manuscript. The reason is that although we have a range of modules, the model depth is relatively small, which can save extensive computation costs.

ModelMemory (MB)FLOPs (G)Params (M)Training time
FNO8.4112.317.27132s / epoch
F-FNO12.313.1211.2176s / epoch
E3D-LSTM2691288.951172s / epoch
MIM2331179.238154s / epoch
PredRNN-V21721117.323.9126s / epoch
SimVP-V242117.246.825s / epoch
LSM10.2114.319.00237 s / epoch
U-NO9232.1136278 s / epoch
STAC57822.8125.498s / epoch

In light of these responses, we hope we have addressed your concerns, and hope you will consider raising your score. If there are any additional notable points of concern that we have not yet addressed, please do not hesitate to share them, and we will promptly attend to those points.

评论

Dear Reviewer UcXH31,

Thank you very much again for the time and effort put into reviewing our paper. We believe that we have addressed all your concerns in our response. We have also followed your suggestion to improve our paper and have added additional experimental analysis. We kindly remind you that we are approaching the end of the discussion period. We would love to know if there is any further concern, additional experiments, suggestions, or feedback, we kindly hope that you can consider increasing the score.

Best regards,

All authors

评论

We are truly grateful for the time you have taken to review our paper and your insightful review. Here we address your comments in the following.

Q1. Though this paper seems to be promising, I have to say that the novelty seems to be limited. The spatio-temporal twins are actually a two-branch model. Using frequency-based approaches in the spatial domain is nothing new. The temporal module is similar to SimVP v2's [1] but with an ODE solver. The cache memory [2] is also well developed.

A1. Thank you for your comment. The novelty of this methodology comes from three parts.

  • New complementary perspective, which explores spatio-temporal dynamics in both discrete and continuous manners. They are both closely related to our target long-term predictions.
  • Information fusion strategy, which includes fine-grained fusion and coarse-grained fusion to tackle feature maps with different granularities.
  • Cache mechanism, which aims to store previous states, thus allowing the system to remember and reuse historical data for future predictions. Compared with E3D-LSTM storing predictions at different timesteps, our cache mechanism stores feature maps with a long interval for long-term spatio-temporal predictions. Moreover, we involve more complicated interaction among current states, short-term states, and long-term states by involving short-term interaction map AmA_m and updated maps QmQ_m. In contrast, E3D-LSTM utilizes a simple recurrent architecture, which achieves much worse performance.

Then, we describe the differences between STAC and the models mentioned in the paper [1,2].

The difference between SimVP-V2 and our STAC:

  • Different objectives. SimVP-V2 focuses on standard video prediction while our STAC focuses on long-term predictions. We also generate a dataset FIRE targetting our problem.
  • Different methodology. SimVP-V2 is entirely based on convolutional neural networks (CNNS) while we introduce complementary modules and cache mechanisms to benefit long-term predictions.
  • Different performance. From the performance comparison, our proposed method performs much better than SimVP-V2 by over 32.18%.

The difference between LMC-Memory and our STAC:

  • Different objectives. LMC-Memory focuses on video prediction while our STAC focuses on dynamics system modeling including temperature, velocity, and pressure. For example, we propose a new FIAR dataset.
  • Different methodology. LMC-Memory uses an external storage device to hold motion contexts while our cache-based propagator focuses on feature maps and incorporates the interaction with short-term prediction with restoring and reusing.
  • Different performance. From the performance comparison, our proposed method performs much better than LMC-Memory by over 3.84%.
评论

Dear Reviewers,

Thank you for your invaluable feedback. As the deadline for the author-reviewer discussion phase is approaching, we hope to make sure that our response sufficiently addressed your concerns regarding the novelty, as well as the revised version of our paper. We hope this could align with your expectations and positively influence the score. Please do not hesitate to let us know if you need any clarification or have additional suggestions.

Best Regards,

Authors

审稿意见
5

This paper addresses the problem of modeling long-term dynamical systems in fields such as fluid dynamics, astrophysics, and earth science. Existing spatio-temporal forecasting approaches based on complex architectures like Transformers have limitations in long-term scenarios due to information loss during semantics exploration and iterative rollouts. To overcome these limitations, the paper proposes a new approach called Spatio-temporal Twins with a Cache (STAC) for long-term system dynamics modeling. STAC comprises a frequency-enhanced spatial module and an ODE-enhanced temporal module that investigates spatio-temporal relationships from complementary perspectives. The information from these twin modules is fused using channel attention to generate informative feature maps. To enhance long-term prediction, a cache-based recurrent propagator is introduced to store and utilize previous feature maps. The paper introduces a new flame flow field benchmark and conducts comprehensive validations across 14 benchmarks. Experimental results demonstrate that STAC outperforms other methods in long-term spatio-temporal prediction and partial differential equation-solving challenges. The contributions of the paper include the construction of a fire dynamics benchmark, the incorporation of cache memory concept into long-term system modeling, the proposal of a novel framework, and extensive experiments showcasing the effectiveness of STAC.

优点

The strengths are as follows:

  1. Effective modeling of long-term dynamical systems: The proposed STAC approach overcomes challenges in long-term forecasting by capturing spatio-temporal relationships and leveraging historical information.

  2. Integration of cache memory: By incorporating a cache-based recurrent propagator, the model effectively stores and reuses informative feature maps, enhancing the accuracy of long-term predictions.

  3. Comprehensive experimental validation: The paper includes extensive experiments on various benchmarks, demonstrating the superior performance of STAC in long-term spatio-temporal prediction and partial differential equation-solving challenges.

  4. Information fusion: STAC combines complementary perspectives through twin modules, using channel attention to generate feature maps with rich semantics, leading to more informative predictions.

  5. Effective optimization strategies: The paper employs teacher forcing, adversarial learning, and mixup techniques to stabilize the learning process and improve the accuracy of iterative updating.

缺点

My main concern about this paper is several potential drawbacks:

  1. Lack of truly innovative contributions: While the paper introduces several components and techniques, such as FSM, OTM, IFTM, CRP, Fourier-based Spectral Filters, teacher forcing, adversarial learning, and the new FIRE dataset, only CRP and IFTM can be considered as relatively novel contributions. The other techniques mentioned are already known and used in existing methods, which may limit the originality and novelty of the proposed approach.

  2. Limited explanation for the CRP technique: The paper mentions the use of a cache-based recurrent propagator (CRP) to prevent forgetting previous events and enhance long sequence prediction. However, it does not provide a clear explanation of the key parameter "α\alpha" and whether it is a learnable parameter. Additionally, CRP's similarity to traditional RNNs raises questions about its parallelization capabilities and potential limitations.

  3. Artificial handling and limited interpretability of IFTM: The separation of temporal and spatial processing, as well as the channel-independent merging in IFTM, appears to be a forced transformation without much interpretability. The lack of learnable factors and reliance on manual processing may hinder the scalability and extensibility of the method.

  4. Potential loss of spatial information in FSM: FSM applies different treatments to the same data and forcibly merges them, potentially leading to a loss of spatial information. Additionally, the direct fully connected mapping of the segmented data raises concerns about the preservation of spatial relationships and the possibility of information loss.

问题

My questions and concerns about this paper are listed in the Weakness part. I will raise my rating if the author can address my concerns with reasonable evidence.

评论

Q4 . Potential loss of spatial information in FSM: FSM applies different treatments to the same data and forcibly merges them, potentially leading to a loss of spatial information. Additionally, the direct fully connected mapping of the segmented data raises concerns about the preservation of spatial relationships and the possibility of information loss.

A4. Thanks for your comment. The information fusion of different sources is very common in recent studies [1,2]. Since we have provided extensive learnable parameters in both parts, their fusion is adaptive and can enhance semantics learning. Moreover, we conduct ablation studies as below to show that removing either part decreases the performance.

Here, we utilize fully connected mapping to make our spatial information more adaptive and more related to our target. Since we utilize an optimization process, our added parameters would be adaptively learned to provide more representation of learning capacity rather than information loss.

Ablation study, We have added four model variants as follows:

  • STAC w/o FNO, which removes FNO and only uses Transformer in our FSM module.
  • STAC w/o Transformer, which removes the Transformer module and only uses FNO in our FSM module.
  • STAC w/o FSM, which removes FSM.
  • STAC w/o FC, which removes the Fully Connected Layer.
ModelMSE (Fire 10 - 10)SSIM (Fire 10 - 10)MSE/100 (KTH 10 - 20)SSIM (KTH 10 - 20)MSE x 1000 (SWE 20 - 140)SSIM x 1000(SWE 20 - 140)Avg Training Time
STAC0.048792.870.298392.830.82497.68242s / epoch
STAC w/o FNO0.075684.540.302190.792.23579.43239s / epoch
STAC w/o Transformer0.050691.210.476583.721.09394.32192s / epoch
STAC w/o FSM0.076585.430.489382.902.35276.54187s / epoch
STAC w/o FC0.079480.850.492181.842.44874.32179s / epoch

Reference

[1] Peng, Zhiliang, Wei Huang, Shanzhi Gu, Lingxi Xie, Yaowei Wang, Jianbin Jiao, and Qixiang Ye. "Conformer: Local features coupling global representations for visual recognition." In Proceedings of the IEEE/CVF international conference on computer vision, pp. 367-376. 2021.

[2] Zhang, Jingyi, Jiaxing Huang, Zhipeng Luo, Gongjie Zhang, Xiaoqin Zhang, and Shijian Lu. "DA-DETR: Domain Adaptive Detection Transformer With Information Fusion." In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 23787-23798. 2023.

In light of these responses, we hope we have addressed your concerns, and hope you will consider raising your score. If there are any additional notable points of concern that we have not yet addressed, please do not hesitate to share them, and we will promptly attend to those points.

评论

Dear Reviewer zrZ831,

Thank you very much again for the time and effort put into reviewing our paper. We believe that we have addressed all your concerns in our response. We have also followed your suggestion to improve our paper and have added additional experimental analysis. We kindly remind you that we are approaching the end of the discussion period. We would love to know if there is any further concern, additional experiments, suggestions, or feedback, we kindly hope that you can consider increasing the score.

Best regards,

All authors

评论

Q3. Artificial handling and limited interpretability of IFTM: The separation of temporal and spatial processing, as well as the channel-independent merging in IFTM, appears to be a forced transformation without much interpretability. The lack of learnable factors and reliance on manual processing may hinder the scalability and extensibility of the method.

A3. Thank you for your comment. We answer it point-by-point:

  1. Why separation. Compared with temporal signals, spatial signals include information from frequency domains, which cannot be considered simultaneously. Therefore, we separate the temporal and spatial processing.

  2. Why transformation. We do the transformation to align the dimensions of features from different sources, which is a popular strategy like padding.

  3. Why deterministic operation. Our Conv2d and FFN modules consist of extensive parameters, which allow our model with strong extensibility. Therefore, to increase efficiency, we utilize the deterministic operation for alignment. From the experiments, we can observe that our STAC performs better than baselines consistency across all datasets, which validate our scalability and extensibility.

BackboneSTAC MSESTAC MAELSM MSELSM MAEU-NO MSEU-NO MAEFNO MSEFNO MAEF-FNO MSEF-FNO MAEE3D-LSTM MSEE3D-LSTM MAEMIM MSEMIM MAEPredRNN-V2 MSEPredRNN-V2 MAESimVP-V2 MSESimVP-V2 MAE
Turbulence0.51230.53450.64120.75530.56540.60930.65670.77890.81240.98761.12341.45670.83210.94721.01631.09871.27651.4321
ERA51.98651.77913.98312.31323.46122.29312.85342.29838.98537.343173.09522.98543.35673.22362.27312.64533.08433.0743
SEVIR1.97311.40542.98312.44312.20311.56323.08331.883110.98315.44324.17022.55633.98422.00123.90141.97572.93711.7743
Fire0.54930.72171.28311.09321.56430.98530.99851.04322.74121.65571.09210.87311.87431.53240.77890.68631.77431.0321
KTH28.832124.221639.983138.443235.873234.432233.198329.742131.874129.875386.174385.556356.594254.842651.151250.645740.842143.2931
评论

Q2 . Limited explanation for the CRP technique: The paper mentions the use of a cache-based recurrent propagator (CRP) to prevent forgetting previous events and enhance long sequence prediction. However, it does not provide a clear explanation of the key parameter "α" and whether it is a learnable parameter. Additionally, CRP's similarity to traditional RNNs raises questions about its parallelization capabilities and potential limitations.

A2. Thanks for your comment. We solve your concerns as follows:

  1. Description of parameter "α": α is a fixed hyperparameter set to 0.20.2.
  2. Parallelization strategy: The computational cost of our method mainly depends on the generation of feature maps, which is suitable for parallelization. Moreover, the complexity comparison of the proposed method is shown below. From the results, we can observe that the efficiency of our method is competitive.
ModelMemory (MB)FLOPs (G)Params (M)Training time
FNO8.4112.317.27132s / epoch
F-FNO12.313.1211.2176s / epoch
E3D-LSTM2691288.951172s / epoch
MIM2331179.238154s / epoch
PredRNN-V21721117.323.9126s / epoch
SimVP-V242117.246.825s / epoch
LSM10.2114.319.00237 s / epoch
U-NO9232.1136278 s / epoch
STAC57822.8125.498s / epoch
  1. When it comes to superlong prediction tasks, our cache-based design could suffer from low efficiency, which would be studied in our future works.
评论

We are truly grateful for the time you have taken to review our paper and your insightful review. Here we address your comments in the following.

Q1 . Lack of truly innovative contributions: While the paper introduces several components and techniques, such as FSM, OTM, IFTM, CRP, Fourier-based Spectral Filters, teacher forcing, adversarial learning, and the new FIRE dataset, only CRP and IFTM can be considered as relatively novel contributions. The other techniques mentioned are already known and used in existing methods, which may limit the originality and novelty of the proposed approach.

A1. Thanks for your comment. The novelty of this methodology comes from three parts:

  • New complementary perspective, which explores spatio-temporal dynamics in both discrete and continuous manners. They are both closely related to our target long-term predictions.
  • Information fusion strategy, which includes fine-grained fusion and coarse-grained fusion to tackle feature maps with different granularities.
  • Cache mechanism, which aims to store previous states, thus allowing the system to remember and reuse historical data for future predictions. Compared with E3D-LSTM storing predictions at different timesteps, our cache mechanism stores feature maps with a long interval for long-term spatio-temporal predictions. Moreover, we involve more complicated interaction among current states, short-term states, and long-term states by involving short-term interaction map AmA_m and updated maps QmQ_m. In contrast, E3D-LSTM utilizes a simple recurrent architecture, which achieves much worse performance.

Then, the motivation for introducing these modules for dynamical system modeling is closely related to real-world data as follows:

  • Why FSM? FSM includes FNO and Transform to explore comprehensive semantics from complementary views. On the one hand, learning from the frequency domain is crucial for our problem. For example, ERA5 datasets containing comprehensive atmospheric and ocean information need to be analyzed in the frequency domain to discern complex climate patterns. Therefore, we design FNO in the FSM module to efficiently capture these complex data in the frequency domain. On the other hand, mining Long-distance dependencies is the key to our long-term prediction tasks. Especially, in KTH datasets and FIRE datasets these dependencies are the key for accurate forecasting. With its self-attention mechanism, the Transformer effectively captures these long-distance dependencies and continuous actions in video sequences. Therefore, we design the Transformer in the FSM module to accurately identify and correlate distant frames in the video.

  • Why OTM? Still, Long distance dependencies are crucial for our long-term prediction tasks. Our ODE-based module can learn these naturally without increasing the depth, i.e., infinite depth. Moreover, Turbulence data contains complex dynamic information such as fluid velocity, pressure, and temperature. Therefore, we incorporate multi-scale convolution to fully understand the physical dynamic system.

  • Why IFTM? We design IFTM for information fusion from previous modules for comprehensive semantic learning. Moreover, our real datasets (e.g., SEVIR and Fire datasets) contain storm details and rapidly changing local extreme events such as flame and temperature fluctuations. Therefore, we generate informative feature maps with both coarse-grained and fine-grained semantics.

The above modules complement each other. We find from experiments that if a certain module is removed, the results will drop significantly, so we propose a unified model. Besides, this work also the contribution of new dataset and new benchmark. In the future, we will fully open-source the data sets of all working conditions and provide the code for reading and processing.

评论

Dear Reviewers,

Thank you for your invaluable feedback. As the deadline for the author-reviewer discussion phase is approaching, we hope to make sure that our response sufficiently addressed your concerns regarding the contributions, as well as the revised version of our paper. We hope this could align with your expectations and positively influence the score. Please do not hesitate to let us know if you need any clarification or have additional suggestions.

Best Regards,

Authors

审稿意见
5

This paper investigates the problem of modeling long-term dynamical systems, which are essential for understanding fluid dynamics, astrophysics, earth science, etc. The authors propose a new approach called STAC, which contains a discrete frequency-enhanced spatial module and an ODE-enhanced temporal module to capture spatial-temporal relationships of the observational data and employs a cache-based recurrent propagator to ensure the long-term prediction ability of the framework. They also utilize teacher forcing and semi-supervised adversarial learning to stabilize the learning process and enhance the reality of predicted trajectories, respectively. Moreover, the paper constructs a new benchmark (FIRE) to model fire dynamics for dynamics forecasting, which potentially benefits the research community. Extensive experiments on complex dynamics modeling, extreme local events sensing, and video prediction tasks demonstrate the superior performance of the proposed framework compared to other SOTA methods.

优点

  1. This paper tackles an important research problem, complex dynamical system modeling, which benefits our understanding of fluid dynamics, astrophysics, earth science, etc.
  2. This paper provides a well-prepared benchmark, FIRE, to facilitate the research in this field and benefit the community.
  3. This paper proposes to consider spatial-temporal correlations in observational data during prediction by utilizing vision Transformer, Fourier neural operator, and neural ODEs, and incorporating cache memory concept into long-term system modeling.
  4. The authors conduct extensive experiments to verify the performance of the dynamical modeling of the proposed methods from multiple perspectives.

缺点

  1. The design of the whole framework is complicated. Although the author explains the reason why they design each module, it still lacks straightforward motivation. Do such challenges really exist in the real data? This straightforward utilization of existing techniques makes the paper novelty seem incremental.
  2. The pictures in Figures 3, and 4 do not seem to show a significant improvement of STAC compared to other SOTA methods in terms of visualization.
  3. In the part of the ablation study, some designs, for example, TF/M, CA, and SSAL, only contribute slightly improvement. However, SSAL may make the training of the framework become unstable. Others may increase the time complexity of the framework, which the authors do not report.
  4. Some notations in the paper are confusing. For example, in Section 4.3, the notation definitions of input, feature map, and output are hard to match the subsequent statement.

After Response

I have read the response and found it addresses most of my concerns. However, I still think the straightforward utilization of existing techniques makes the paper's novelty seem incremental. Moreover, I also have concerns regarding the claimed advantage of using NODE, i.e., modeling long-distance dependencies, even with a learnable t.

问题

  1. Can authors provide their motivation for such complicated module design through data?
  2. Can authors provide the standard deviation of their experimental results?
  3. The authors can conduct more persuasive experiments to address my concerns mentioned in the Weaknesses part.

伦理问题详情

Not applicable.

评论

Q3. In the part of the ablation study, some designs, for example, TF/M, CA, and SSAL, only contribute slightly improvement. However, SSAL may make the training of the framework become unstable. Others may increase the time complexity of the framework, which the authors do not report.

A3. Thanks for your comment. We have added more ablation studies on more datasets and the results are shown below. From the results, we can observe that the performance gain of our full model is between [1.27%, 14.5%] and over 5% in most cases, validating the effectiveness of every component. In addition, we include the variance of our model and STAC w/o SSAL. The results are shown below. From the results, we can observe the model is still stable comparably. Finally, we have completed experiments to verify the effectiveness of the proposed components.

ModelMSE (Fire 10 - 10) (±std)SSIM (Fire 10 - 10) (±std)MSE/100 (KTH 10 - 20) (±std)SSIM (KTH 10 - 20) (±std)MSE x 1000 (SWE 20 - 140) (±std)SSIM x 1000(SWE 20 - 140) (±std)Avg Training Time
STAC0.0487 ± 0.000392.87 ± 0.69060.2983 ± 0.001992.83 ± 0.54810.824 ± 0.00497.68 ± 0.6655242s / epoch
STAC w/o FNO0.0756 ± 0.000484.54 ± 0.62870.3021 ± 0.001990.79 ± 0.5362.235 ± 0.010879.43 ± 0.5412239s / epoch
STAC w/o Transformer0.0506 ± 0.000391.21 ± 0.67830.4765 ± 0.003183.72 ± 0.49431.093 ± 0.005394.32 ± 0.6426192s / epoch
STAC w/o FSM0.0765 ± 0.000585.43 ± 0.63530.4893 ± 0.003182.9 ± 0.48942.352 ± 0.011376.54 ± 0.5215187s / epoch
STAC (Replacing OTM with ConvLSTM)0.1723 ± 0.00165.69 ± 0.48851.2043 ± 0.007765.42 ± 0.38621.845 ± 0.008983.43 ± 0.5684321s / epoch
STAC (Replacing OTM with PredRNN)0.3922 ± 0.002357.98 ± 0.4312OOMOOMOOMOOM654s / epoch
STAC w/o OTM0.0802 ± 0.000582.33 ± 0.61230.5644 ± 0.003678.43 ± 0.4632.153 ± 0.010479.97 ± 0.5448209s / epoch
STAC w/o IFTM0.0596 ± 0.000490.33 ± 0.67180.7321 ± 0.004770.43 ± 0.41581.183 ± 0.005791.23 ± 0.6216237s / epoch
STAC (Replacing CRP with Recall gate)0.0998 ± 0.000679.49 ± 0.59110.5145 ± 0.003380.39 ± 0.47465.986 ± 0.028854.32 ± 0.3701598s / epoch
STAC w/o CRP0.0543 ± 0.000390.83 ± 0.67550.3343 ± 0.002187.38 ± 0.51596.322 ± 0.030445.65 ± 0.311200s / epoch
STAC w/o SSAL0.0632 ± 0.000486.93 ± 0.64650.3281 ± 0.002187.89 ± 0.51891.772 ± 0.008583.98 ± 0.5722229s / epoch
评论

We are truly grateful for the time you have taken to review our paper and your insightful review. Here we address your comments in the following.

Q1. The design of the whole framework is complicated. Although the author explains the reason why they design each module, it still lacks straightforward motivation. Do such challenges really exist in the real data? This straightforward utilization of existing techniques makes the paper novelty seem incremental.

A1. Thank you for your comment. The motivation for introducing these modules for dynamical system modeling is closely related to real-world data as follows:

  • Why FSM? FSM includes FNO and Transform to explore comprehensive semantics from complementary views. On the one hand, learning from the frequency domain is crucial for our problem. For example, ERA5 datasets containing comprehensive atmospheric and ocean information need to be analyzed in the frequency domain to discern complex climate patterns. Therefore, we design FNO in the FSM module to efficiently capture these complex data in the frequency domain. On the other hand, mining Long-distance dependencies is the key to our long-term prediction tasks. Especially, in KTH datasets and FIRE datasets these dependencies are the key for accurate forecasting. With its self-attention mechanism, the Transformer effectively captures these long-distance dependencies and continuous actions in video sequences. Therefore, we design the Transformer in the FSM module to accurately identify and correlate distant frames in the video.

  • Why OTM? Still, long distance dependencies are crucial for our long-term prediction tasks. Our ODE-based module can learn these naturally without increasing the depth, i.e., infinite depth. Moreover, Turbulence data contains complex dynamic information such as fluid velocity, pressure, and temperature. Therefore, we incorporate multi-scale convolution to fully understand the physical dynamic system.

  • Why IFTM? We design IFTM for information fusion from previous modules for comprehensive semantic learning. Moreover, our real datasets (e.g., SEVIR and Fire datasets) contain storm details and rapidly changing local extreme events such as flame and temperature fluctuations. Therefore, we generate informative feature maps with both coarse-grained and fine-grained semantics.

  • Why Cache? Still, Long-term memories are crucial for our long-term prediction tasks, especially in Fire datasets. The forgetting phenomenon is common in current methods, resulting in inferior performance. Therefore, we design a cache-based mechanism to store and reuse historical feature maps.

  • Why Optimization Objectives? Long-term predictions could be unreliable. Therefore, we introduce semi-supervised adversarial learning to improve the long-term predictive capacity of the model without the ground truth.

In summary, all these modules can benefit from modeling long-term dependencies for accurate long-term predictions from different perspectives (e.g., dependency mining, infinite depth, information fusion, memory reusing, and reliability).

Q2. The pictures in Figures 3, and 4 do not seem to show a significant improvement of STAC compared to other SOTA methods in terms of visualization.

A2. Thanks for your comment. We have revised the figure for better visualization. In Fig. 3, we use the red box to mark the local difference phenomenon and enlarge the area. in Fig. 4, we introduce a professional cartopy library to highlight the information on land, longitude, and latitude on the ERA5 dataset, which can better distinguish visual differences. We have updated the visualization results in Figure.3 and Figure.4 of the manuscript.

评论

Q4. Some notations in the paper are confusing. For example, in Section 4.3, the notation definitions of input, feature map, and output are hard to match the subsequent statement.

A4. Thanks for your comment. We have introduced a notation table as below to make the paper clear.

SymbolExplanation
TTTotal length of the time series
TobsT_{obs}Time interval of observed states
u(t)u(t)State at time step t
Ω\OmegaSpatial coordinates
ttTime coordinate
xxSpatial position
RC×H×WR^{C\times H \times W}Tensor representation of the state
I^in\hat{I}_{in}Input tensor of the time series
F0F_0Initial feature mapping
I^m\hat{I}_mInput for the time interval [tm1,tm][t_{m-1}, t_m]
XmX_mOutput feature mapping for the time interval [tm1,tm][t_{m-1}, t_m]
QmQ_mUpdate mapping for the time interval [tm1,tm][t_{m-1}, t_m]
AmA_mAuxiliary variables for the time interval [tm1,tm][t_{m-1}, t_m]
WmW_mWeights for the time interval [tm1,tm][t_{m-1}, t_m]
RRSize of the cache
α\alphaBalance parameter
ϕ()\phi(\cdot)Computes interactions between current and historical feature mappings
MMThe most recent R historical feature mappings stored in the cache
f()f(\cdot)Function for updating the hidden state
g()g(\cdot)Function for computing the update mapping

Q5. Can authors provide the standard deviation of their experimental results?

A5. Thanks for your comment. We have added the standard deviation (SD) to the main table as shown in the table below.

BackboneSTAC MSE ± SDSTAC MAE ± SDFNO MSE ± SDFNO MAE ± SDF-FNO MSE ± SDF-FNO MAE ± SDE3D-LSTM MSE ± SDE3D-LSTM MAE ± SDMIM MSE ± SDMIM MAE ± SDPredRNN-V2 MSE ± SDPredRNN- MAE ± SDSimVP-V2 MSE ± SDSimVP-V2 MAE ± SD
Turbulence0.5123 ± 0.050.5345 ± 0.040.6567 ± 0.060.7789 ± 0.050.8124 ± 0.070.9876 ± 0.061.1234 ± 0.081.4567 ± 0.070.8321 ± 0.050.9472 ± 0.061.0163 ± 0.091.0987 ± 0.081.2765 ± 0.101.4321 ± 0.09
ERA51.9865 ± 0.151.7791 ± 0.122.8534 ± 0.202.2983 ± 0.188.9853 ± 0.257.34317 ± 0.223.0952 ± 0.192.9854 ± 0.183.3567 ± 0.203.2236 ± 0.192.2731 ± 0.172.6453 ± 0.163.0843 ± 0.213.0743 ± 0.20
SEVIR1.9731 ± 0.121.4054 ± 0.103.0833 ± 0.151.8831 ± 0.1410.9831 ± 0.305.4432 ± 0.254.1702 ± 0.222.5563 ± 0.203.9842 ± 0.182.0012 ± 0.173.9014 ± 0.191.9757 ± 0.182.9371 ± 0.201.7743 ± 0.15
Fire0.5493 ± 0.030.7217 ± 0.020.9985 ± 0.041.0432 ± 0.032.7412 ± 0.061.6557 ± 0.051.0921 ± 0.040.8731 ± 0.031.8743 ± 0.051.5324 ± 0.040.7789 ± 0.030.6863 ± 0.021.7743 ± 0.051.0321 ± 0.04
KTH28.8321 ± 0.8024.2216 ± 0.7033.1983 ± 0.9029.7421 ± 0.8531.8741 ± 0.9529.8753 ± 0.9086.1743 ± 1.1085.5563 ± 1.0556.5942 ± 0.9554.8426 ± 0.9051.1512 ± 0.9550.6457 ± 0.9040.8421 ± 0.8543.2931 ± 0.80

In light of these responses, we hope we have addressed your concerns, and hope you will consider raising your score. If there are any additional notable points of concern that we have not yet addressed, please do not hesitate to share them, and we will promptly attend to those points.

评论

Dear Reviewer CD63,

Thank you very much again for the time and effort put into reviewing our paper. We believe that we have addressed all your concerns in our response. We have also followed your suggestion to improve our paper and have added additional experimental analysis. We kindly remind you that we are approaching the end of the discussion period. We would love to know if there is any further concern, additional experiments, suggestions, or feedback, we kindly hope that you can consider increasing the score.

Best regards,

All authors

评论

Dear Reviewers,

Thank you for your invaluable feedback. As the deadline for the author-reviewer discussion phase is approaching, we hope to make sure that our response sufficiently addressed your concerns regarding the motivation and details, as well as the revised version of our paper. We hope this could align with your expectations and positively influence the score. Please do not hesitate to let us know if you need any clarification or have additional suggestions.

Best Regards,

Authors

审稿意见
5

This paper focuses on the long-term spatiotemporal dynamics modeling. They propose the STAC by combing the advanced spatial and temporal modeling backbones and presenting a cache-based recurrent propagator to store the previous feature maps to avoid information loss. Besides, the authors propose a compound training loss to optimize STAC. Experimentally, STAC shows favorable performance in a wide range of benchmarks, including the newly generated flame flow field benchmark.

优点

  1. This paper presents the STAC model to tackle the problem in long-term dynamic prediction, which is technologically reasonable.

  2. The authors experiment on a wide range of benchmarks to demonstrate the effectiveness of STAC.

  3. This paper is clearly presented and well-written.

缺点

  1. About the novelty.

Generally, I think the technology design is reasonable. However, in my opinion, I think it is insufficient in novelty. STAC just combines a series of advanced models, including FNO, Vision Transformer, Neural ODE and a similar recall mechanism proposed by E3D-LSTM. The proposed training strategy is also in a combination style. For me, it is hard to find the novel part in this model.

Note that I am not attempt to enforce the authors to build a completely new model or block. I just think they fail in illustrating their advantages beyond other models. For example, they should consider the following questions:

  • Why should they combine vision transformer and FNO? FNet [1] has shown that the feedforward layer can perform like FFT. Why not just only use Transformer or FNO?

  • Why can Neural ODE capture the continuous dynamics? I know that Neural ODE can achieves the adaptive depth or adaptive temporal interval. But according to the equation and code, I think the usage here is equivalent to a simple rk4 algorithm. It is hard to claim that they learn the continuous dynamic feature. Besides, They don’t present the necessity in using Neural ODE.

  • Are the experimental datasets temporally irregular? According to the paper, I think the input sequences are equally collected along the temporal dimension.

  • About the cache-based design. I think it is necessary to demonstrate its advancement over the temporal recall gate in E3D-LSTM.

[1] FNet: Mixing Tokens with Fourier Transforms, ACL 2022.

  1. About the experiment.

(1) In addition to the performance, they should compare the efficiency with other baselines, including running time, GPU memory and parameter size.

(2) In the current version, they only compare STAC with video prediction baselines. How about the advanced neural operators, such as LSM [2], U-NO [3]?

(3) Are all the baselines trained by the same loss as STAC? This point is essential to ensure a fair comparison.

(4) More detailed ablations are expected. They should also conduct the following experiments:

  • Removing FNO or Transformer in FSM.

  • Replacing OTM with ConvLSTM or PredRNN.

  • Replacing the CRP with the recall gate in E3D-LSTM.

[2] Solving High-Dimensional PDEs with Latent Spectral Models, ICML 2023

[3] U-NO: U-shaped Neural Operators, TMLR 2023

问题

All the questions are listed above, including novelty, experiment design. Here are several serious problems that should be clarified:

(1) The acutal usage of OTM is inconsistent to their expection.

(2) Are all the baselines trained by the same loss?

(3) More neural operator baselines are expected.

(4) Demonstrate the novelty of STAC.

I think it is favorable that the authors experiment on extensive benchmarks. But I have some serious concerns. if the authors reply my questions properly, I am willing to raise my score.

评论

Q7. In the current version, they only compare STAC with video prediction baselines. How about the advanced neural operators, such as LSM [2], U-NO [3]?

A7. Thanks for your comment. We have added LSM and U-NO for a more comprehensive evaluation. The experimental results show that the STAC model outperforms LSM and U-NO on different datasets, which validates the superiority of the proposed STAC.

ModelSTACSTACLSMLSMU-NOU-NO
DatasetMSEMAEMSEMAEMSEMAE
Turbulence0.51230.53450.64120.75530.56540.6093
ERA51.98651.77913.98312.31323.46122.2931
SEVIR1.97311.40542.98312.44312.20311.5632
Fire0.54930.72171.28311.09321.56430.9853
KTH28.832124.221639.983138.443235.873234.4322

Q8. Are all the baselines trained by the same loss as STAC? This point is essential to ensure a fair comparison.

A8. Thanks for your comment. The loss function of the baseline follows the original paper, which is MSE loss. Considering fairness, we train STAC with only MSE loss. The experimental results are shown in the table below. It can be seen that with only MSE loss, our STAC can outperform the baselines, and adding our designed loss would further increase the performance.

DatasetFireFireERA5ERA5
ModelMSESSIMMSESSIM
STAC0.549392.31921.986587.4531
STAC(only MSE Loss)0.643290.87522.112985.4853
PredRNN-V20.778985.97432.273184.9853
SimVP-V21.774377.86543.084377.3212
FNO0.998582.22182.853479.3834
LSM0.965483.94822.983179.4393
评论

Q9. More detailed ablations are expected. They should also conduct the following experiments:

Removing FNO or Transformer in FSM.

Replacing OTM with ConvLSTM or PredRNN.

Replacing the CRP with the recall gate in E3D-LSTM.

A9. Thanks for your comment. We have added five model variants as follows:

  • STAC w/o FNO, which removes FNO and only uses Transformer in our FSM module.
  • STAC w/o Transformer, which removes the Transformer module and only uses FNO in our FSM module.
  • STAC (With ConvLSTM), which replaces the OTM module with ConvLSTM.
  • STAC (With PredRNN), which replaces the OTM module with PredRNN.
  • STAC (With Recall gate), which replaces the CRP with the Recall gate.

Experiments show that removing FNO and Transformer from the STAC model reduces performance especially Transformer has a significant impact. Replacing OTM with ConvLSTM or PredRNN significantly degrades performance, especially PredRNN replacement leads to memory overflow. Replacing CRP with Recall gate also shows the importance of CRP, whose removal leads to performance loss.

ModelMSE (Fire)SSIM (Fire)MSE/100 (KTH)SSIM (KTH)
STAC0.048792.870.298392.83
STAC w/o FNO0.075684.540.302190.79
STAC w/o Transformer0.050691.210.476583.72
STAC (With ConvLSTM)0.172365.691.204365.42
STAC (With PredRNN)0.392257.98OOMOOM
STAC (With Recall gate)0.099879.490.514580.39

In light of these responses, we hope we have addressed your concerns, and hope you will consider raising your score. If there are any additional notable points of concern that we have not yet addressed, please do not hesitate to share them, and we will promptly attend to those points.

评论

I appreciate the authors' effort in clarifying their novelty and providing sufficient ablations and baselines. They made a great effort obviously.

Most of my concerns are resolved. However, the questions about continuous modeling still remain (Q3 and Q4).

(1) I don't think the usage in STAC is Neural ODE. It is just a rk4 algorithm. If you do want to claim that your model takes benefits from Neural ODE, you need to use the learnable temporal steps tt.

(2) Would you please clarify this comment "which can allow the capture of long-range correlations naturally without increasing the depth" in your rebuttal? What kind of long-range correlations do you mean?

Overall, I think the ODE claim in this paper is not rigorous, which obviously requires a huge revision to the paper. Thus, I would like to raise my score to 5 but cannot give an acceptance recommendation.

评论

Q4. Are the experimental datasets temporally irregular? According to the paper, I think the input sequences are equally collected along the temporal dimension.

A4. Thanks for your comment. The dataset we use is temporally regular. The complete dataset statistics are as follows:

Table: Dataset statistics. N_tr and N_te denote the number of instances in the training and test sets. The lengths of the input and prediction sequences are I_l and O_l, respectively.

DatasetN_trN_te(C, H, W)I_lO_lInterval
Turbulence50001000(3, 300, 300)50501 second
ERA5 (Global)100002000(1, 1440, 720)10101 day
ERA5 (Local)50001000(2, 200, 200)881 hour
KTH1087174086(1, 128, 128)10201 step
SEVIR4158500(1, 384, 384)13125 mins
Fire60001500(2, 32, 480)503501 second
SWE2000200(1, 128, 256)201401 second
dynamics system60001200(3, 128, 128)221 second

Q5. About the cache-based design. I think it is necessary to demonstrate its advancement over the temporal recall gate in E3D-LSTM.

A5. Thanks for your comment. We have added a model variant STAC w/ E3D, which replaces our cache-based design with the temporal recall gate in E3D-LSTM. The compared results are shown below. From the results, we can observe that STAC performs the best in all cases. The reason is that our module not only stores feature maps with a long interval but also involves short-term interaction maps into spatio-temporal predictions.

Iuput-Output length50 - 40050 - 40020 - 14020 - 14010 - 4010 - 40
ModelMSE (Fire)SSIM (Fire)MSE x 100 (SWE)SSIM (SWE)PSNR (KTH)SSIM (KTH)
E3D-LSTM1.298375.87210.229876.331230.593187.4221
STAC0.663188.87310.082497.676233.983192.1176
STAC w/ E3D0.998783.47300.122287.28331.398488.3984

Q6. In addition to the performance, they should compare the efficiency with other baselines, including running time, GPU memory and parameter size.

A6. Thanks for your comment. We have added the following table as follows. From the comparison, we can observe that our efficiency is competitive. The reason is that although we have a range of modules, the model depth is relatively small, which can save extensive computation costs.

ModelMemory (MB)FLOPs (G)Params (M)Training time
FNO8.4112.317.27132s / epoch
F-FNO12.313.1211.2176s / epoch
E3D-LSTM2691288.951172s / epoch
MIM2331179.238154s / epoch
PredRNN-V21721117.323.9126s / epoch
SimVP-V242117.246.825s / epoch
LSM10.2114.319.00237 s / epoch
U-NO9232.1136278 s / epoch
STAC57822.8125.498s / epoch

Q7. In the current version, they only compare STAC with video prediction baselines. How about the advanced neural operators, such as LSM [2], U-NO [3]?

A7. Thanks for your comment. We have added LSM and U-NO for a more comprehensive evaluation. The experimental results show that the STAC model outperforms LSM and U-NO on different datasets, which validates the superiority of the proposed STAC.

ModelSTACSTACLSMLSMU-NOU-NO
DatasetMSEMAEMSEMAEMSEMAE
Turbulence0.51230.53450.64120.75530.56540.6093
ERA51.98651.77913.98312.31323.46122.2931
SEVIR1.97311.40542.98312.44312.20311.5632
Fire0.54930.72171.28311.09321.56430.9853
KTH28.832124.221639.983138.443235.873234.4322
评论

We are truly grateful for the time you have taken to review our paper and your insightful review. Here we address your comments in the following.

Q1. However, in my opinion, I think it is insufficient in novelty. STAC just combines a series of advanced models, including FNO, Vision Transformer, Neural ODE, and a similar recall mechanism proposed by E3D-LSTM. The proposed training strategy is also in a combination style. For me, it is hard to find the novel part in this model.

A1. Thanks for your comment. The novelty of this methodology comes from three parts.

  • New complementary perspective, which explores spatio-temporal dynamics in both discrete and continuous manners. They are both closely related to our target long-term predictions.
  • Information fusion strategy, which includes fine-grained fusion and coarse-grained fusion to tackle feature maps with different granularities.
  • Cache mechanism, which aims to store previous states, thus allowing the system to remember and reuse historical data for future predictions. Compared with E3D-LSTM storing predictions at different timesteps, our cache mechanism stores feature maps with a long interval for long-term spatiotemporal predictions. Moreover, we involve more complicated interaction among current states, short-term states, and long-term states by involving short-term interaction map AmA_m and updated maps QmQ_m. In contrast, E3D-LSTM utilizes a simple recurrent architecture, which achieves much worse performance.

Besides, this work also the contribution of new dataset and new benchmark. In the future, we will fully open-source the data sets of all working conditions and provide the code for reading and processing.

Q2. Why should they combine vision transformer and FNO? FNet has shown that the feedforward layer can perform like FFT. Why not just only use Transformer or FNO?

A2. Thanks for your comment. We have added four model variants as follows:

  • STAC w/o FNO, which removes FNO and only uses Transformer in our FSM module.
  • STAC w/o Transformer, which removes the Transformer module and only uses FNO in our FSM module.
  • STAC w/o FNO&Transformer +Fnet, which removes the FNO and Transformer modules and replaces them with Fnet.
  • Fnet, which only uses Fnet.

The compared performance is shown below. From the results, we can find that both mechanisms have a crucial effect on performance. The reason is that Transformer focuses on capturing the correlation between pixels, while FNO learns function mapping in the frequency domain. Combining them can provide a comprehensive for semantics learning. In addition, incorporating Fnet into STAC and the predictions alone cannot achieve superior performance as well.

ModelMSE (Fire 10 - 10)SSIM (Fire 10 - 10)MSE/100 (KTH 10 - 20)SSIM (KTH 10 -20)
STAC0.048792.870.298392.83
STAC w/o FNO0.075684.540.302190.79
STAC w/o Transformer0.050691.210.476583.72
STAC w/o FNO&Transformer +Fnet0.128179.480.732578.23
Fnet0.764145.4324.543256.41

Q3. Why can Neural ODE capture the continuous dynamics? I know that Neural ODE can achieve the adaptive depth or adaptive temporal interval. However, I think the usage here is equivalent to a simple rk4 algorithm. It is hard to claim that they learn the continuous dynamic feature. Besides, They don’t present the necessity in using Neural ODE.

A3. Thanks for your comment. Our approach utilizes Neural ODE to explore spatiotemporal relationships in a continuous way, which can allow the capture of long-range correlations naturally without increasing the depth. This point is important for effective long-term predictions. We have added a model invariant STAC w/o OTM, which removes the ODE module. The performance is shown below. From the results, we can observe our ODE module is necessary for accurate long-term prediction.

ModelMSE (Fire 10 - 10)SSIM (Fire 10 - 10)MSE/100 (KTH 10 - 20)SSIM (KTH 10 - 20)MSE x 1000 (SWE 20 - 140)
STAC w/o OTM0.080282.330.564478.432.153
STAC0.048792.870.298392.830.824
评论

Thanks for your feedback and we are happy to resolve your further concerns as follows:

Q1. I don't think the usage in STAC is Neural ODE. It is just a rk4 algorithm. If you do want to claim that your model takes benefits from Neural ODE, you need to use the learnable temporal steps tt.

A1. Thanks for your comment. Clearly, the earliest neural ODE [1] also focuses on the fixed depth, but introduces the continuous propagation with benefits. Moreover, [2] also utilizes the fixed step tt with performance improvement, which shows the benefit of adopting neural ODE with fixed steps. Moreover, we would include learnable tt in our future work.

Would you please clarify this comment "which can allow the capture of long-range correlations naturally without increasing the depth" in your rebuttal? What kind of long-range correlations do you mean?

A2. Thanks for your comment. Here, the long-range correlations mean the relationships of pixels far away in tensor Iˉin\bar{I}^{in}, which require a large receptive field to capture. Here, neural ODE utilizes continuous propagation to get rid of stacking a range of CNN blocks.

Thanks again for appreciating our work and for your constructive suggestions. Please let us know if you have further questions.

[1] Chen et al. Neural Ordinary Differential Equations.

[2] Jin et al., Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs

评论

Thanks for your feedback and we are happy to resolve your further concerns as follows:

Q1. I don't think the usage in STAC is Neural ODE. It is just a rk4 algorithm. If you do want to claim that your model takes benefits from Neural ODE, you need to use the learnable temporal steps tt.

A1. Thanks for your comment. Here, we have tried the learnable tt with compared performance as follows. The results show that the learnable tt can slightly improve the performance compared to the fixed tt with more flexibility. Thanks for your suggestion. We have added the results into our manuscript. (A smaller mse means better model performance and a larger ssim means better model performance.)

Besides, the earliest neural ODE [1] also focuses on the fixed depth, but introduces the continuous propagation with benefits. Moreover, [2] also utilizes the fixed step tt with performance improvement, which shows the benefit of adopting neural ODE in both settings.

ModelMSE (Fire 10 - 10)SSIM (Fire 10 - 10)MSE/100 (KTH 10 - 20)SSIM (KTH 10 - 20)MSE x 1000 (SWE 20 - 140)SSIM x 1000(SWE 20 - 140)Avg Training Time
STAC0.048792.870.298392.830.82497.68242s / epoch
STAC (learnable tt)0.046493.020.287393.110.81498.53257s / epoch

Q2. Would you please clarify this comment "which can allow the capture of long-range correlations naturally without increasing the depth" in your rebuttal? What kind of long-range correlations do you mean?

A2. Thanks for your comment. Here, the long-range correlations mean the relationships of pixels far away in tensor Iˉin\bar{I}^{in}, which require a large receptive field to capture. Here, neural ODE utilizes continuous propagation to get rid of stacking a range of CNN blocks.

Thanks again for appreciating our work and for your constructive suggestions. Please let us know if you have further questions.

[1] Chen et al. Neural Ordinary Differential Equations.

[2] Jin et al., Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs

评论

Dear Reviewer rchA,

Thank you very much again for the time and effort put into reviewing our paper. We believe that we have addressed all your concerns in our response. We have also followed your suggestion to improve our paper and have added additional experimental analysis. We kindly remind you that we are approaching the end of the discussion period. We would love to know if there is any further concern, additional experiments, suggestions, or feedback, we kindly hope that you can consider increasing the score.

Best regards,

All authors

评论

Dear reviewers,

We sincerely appreciate your valuable feedback. As the deadline for the author-reviewer discussion phase is approaching, we would like to check if our response adequately addressed your concerns regarding our neural ODE and learnable tt. If there are any remaining issues or if you require further clarification, please feel free to inform us.

Best Regards,

Authors.

评论

Dear reviewer,

Thanks for your response and try the learnable t. But for me, the description of "which can allow the capture of long-range correlations naturally without increasing the depth" is still not well-supported. Even though you can "increase" the depth in favorable efficiency with NODE, the reception field may not increase as you want.

Considering the novelty and soundness of this paper, I would like to keep my original score.

评论

Thanks for your feedback and we are happy to resolve your further concerns as follows:

Q1. The description of "which can allow the capture of long-range correlations naturally without increasing the depth" is still not well-supported.

A1. Thanks for your comment. Here, we revised our statement into "which can allow the capture of long-range correlations naturally with robustness" since Neural ODE has the property of robustness with increased propagation depths, as stated in Sec. 5.4 of [1]. This is one of the most important properties of Neural ODE compared with discrete methods. We have revised the manuscript accordingly.

[1] Multivariate Time Series Forecasting with Dynamic Graph Neural ODEs

Q2. About the soundness.

A2. Thanks for your comment. We have added extensive clarification and experiments to address all the concerns. Moreover, the other reviewers all give a "3" for soundness. If there are any remaining issues or if you require further clarification, please feel free to inform us.

Q3. About the novelty.

A3. Thanks for your comment. As mentioned earlier, our method consists of three parts, i.e., complementary perspective, information fusion strategy and cache mechanism, which are novel compared with existing works. This work also the contribution of new dataset and new benchmark in advancing the community. And we are committed to making our complete code and the dataset publicly available.

As the deadline for the author-reviewer discussion phase is approaching, we hope to make sure that our response sufficiently addressed your concerns regarding the revised version. We hope this could align with your expectations and positively influence the score. Please do not hesitate to let us know if you need any clarification or have additional suggestions.

评论

Dear Reviewers,

We thank you for your careful reviews and constructive suggestions. We acknowledge the positive comments such as "technologically reasonable” (Reviewer rchA), "clearly presented and well-written" (Reviewer rchA), "a wide range of benchmarks" (Reviewer rchA), "an important research problem,” (Reviewer CD63), "a well-prepared benchmark” (Reviewer CD63), "extensive experiments” (Reviewer CD63), "effective modeling” (Reviewer zrZ8), “Comprehensive experimental validation” (Reviewer zrZ8), “Effective optimization strategies” (Reviewer zrZ8), "novel perspective" (Reviewer UcXH), "intuitive and reasonable" (Reviewer UcXH), "well organized and clearly written" (Reviewer UcXH), "new fire dynamics benchmark" (Reviewer UcXH) and "comprehensive experiments" (Reviewer UcXH). We also believe that our new benchmark FIRE would contribute to the community more.

We have made extensive revisions based on these valuable comments, which we briefly summarize below for your convenience, and have accordingly revised our article with major changes highlighted in blue.

  • We have clarified our novelty in four points including new benchmark FIRE, new complementary perepsectives for long-term prediction, information fusion strategy and new cache mechanism with differences stated.

  • We have introduced detailed motivation of each component related to real data to show these compotents typically focus on long-term dependency mining for our long-term prediction tasks.

  • We have added more sufficient ablation studies to show the effectiveness and reliability of each component.

  • We have included more competing baselines including LSM and U-NO to demonstrate the superiority of our approach.

  • We have included more detailed efficiency comparisons between our method and various baselines and observed that our method shows competitive efficiency.

  • We have proofread the manuscript to correct some typos and mistakes.

We have also responded to your questions point by point. Once again, we greatly appreciate your effort and time spent revising our manuscript. In summary, our paper introduces a new benchmark dataset, extensive performance comparison with 14 benchmarks, and a practical approach to achieve superior long-term predictive performance, which we believe can greatly contribute to the community. Please let us know if you have any follow-up questions. We will be happy to answer them.

Best regards,

the Authors

评论

Dear Reviewers,

Thanks for your time and commitment to the ICLR 2024 review process.

As we approach the conclusion of the author-reviewer discussion period (Wednesday, Nov 22nd, AOE), I kindly urge those who haven't engaged with the authors' dedicated rebuttal to please take a moment to review their response and share your feedback, regardless of whether it alters your opinion of the paper.

Your feedback is essential to a thorough assessment of the submission.

Best regards,

AC

AC 元评审

This paper proposes a composite solution to modeling long-term dynamics, which combines Vision Transformer, NODE, and Cache mechanism, and more. Authors also provided lengthy rebuttal with valuable experimental results, which make the paper strong in the empirical aspect, and were acknowledged by reviewers with increased scores. After rebuttal, reviewers still held crucial concerns towards this paper. Most reviewers felt that the novelty is relatively slim, given that most of the components are inherited from previous works. The introduction of NODE and its actual role in the final solution is not clear. The cache design is incremental compared with previous works. The overall solution is way more complicated than the previous counterparts. In a nutshell, the paper needs to undergo a major revision.

为何不给更高分

After rebuttal, unanimous rejection. Novelty is not pronounced enough.

为何不给更低分

N/A

最终决定

Reject