PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
5
5
4
4
3.8
置信度
创新性2.5
质量2.8
清晰度3.3
重要性3.0
NeurIPS 2025

MoFo: Empowering Long-term Time Series Forecasting with Periodic Pattern Modeling

OpenReviewPDF
提交: 2025-04-14更新: 2025-10-29

摘要

关键词
Time Series; Deep Learning; Position Information Encoding

评审与讨论

审稿意见
5

The paper proposes the MoFo framework for long-term time series forecasting, which models periodicity as the correlation of period-aligned time steps and the trend of period-offset time steps. By designing a periodic structural patch and a periodic-aware modulator, the framework captures periodic patterns effectively. Experiments demonstrate that it outperforms 17 baselines, achieving both a 14x improvement in memory efficiency and a 10x faster training speed.

优缺点分析

Strengths

  1. Modeling periodicity as the correlation of period-aligned time steps and the trend of period-offset time steps provides a novel modeling perspective for long-term time series forecasting.
  2. MoFo demonstrates competitive predictive performance compared to 17 state-of-the-art baselines on multiple popular benchmark datasets, validating the effectiveness of the proposed method.
  3. The model's computational complexity is related to the period length rather than the input sequence length, achieving up to a 14×\times improvement in memory efficiency and a 10×\times speedup in training.

Weaknesses

  1. Presentation Issues
    • Grammatical errors. For example, there is a subject-verb agreement issue on line 182 regarding singular and plural forms.
    • The names of all designed modules should be in lowercase.
  2. Clarifications Needed
    • The description of the patch-based structured input is unclear, especially the padding strategy. Using data from the last period for padding may introduce potential information redundancy —could this affect model generalization?
    • How to determine the cycle length? What does formula H mean in Eq.10?
  3. Baselines
    • How many layers of Transformer did the author use? Did the author perform a sensitivity analysis on this parameter?
    • A hyperparameter study on the period length is suggested, for example, using double the default period length, to evaluate its impact on performance.

问题

The authors can refer to the weakness listed above.

局限性

The authors can refer to the weakness listed above.

最终评判理由

The paper proposes the MoFo framework for long-term time series forecasting, which models periodicity as the correlation of period-aligned time steps and the trend of period-offset time steps. By designing a periodic structural patch and a periodic-aware modulator, the framework captures periodic patterns effectively. Experiments demonstrate that it outperforms 17 baselines, achieving both a 14x improvement in memory efficiency and a 10x faster training speed. The author's response has largely addressed my concerns. Considering the significance of MoFo for long-term time series forecasting, I have raised my score to 5.

格式问题

The authors can refer to the weakness listed above.

作者回复

Dear Reviewer Q3nE,

Thank you for your kind words. Your feedback and guidance have been instrumental in improving our paper, and we sincerely appreciate your support and constructive suggestions. Next, we will respond to your concerns point by point.

W1. Presentation Typos

We sincerely appreciate your constructive feedback and will thoroughly revise and proof-read the manuscript in the next version to address every point you raised.

W2.1. Padding Strategy

Our padding strategy is already described in detail in Section 3.1, lines 109–119 and formalized in Equation (1) of the manuscript. In fact, Given an input sequence X=[x1,,xT]\mathbf{X} = [\mathbf{x}_1, …, \mathbf{x}_T]^\top, we start from the current time step xT\mathbf{x}_T and move past to delineate periods of length PP. Any prefix that does not form a complete period is left-padded with the leftmost elements of the nearest full period on its right; this yields a padded input length of T/PP\lceil T / P\rceil * P and exactly T/P\lceil T / P\rceil periods.

To evaluate the impact of our strategy, we introduce a variant "+ zero padding" that pads the value with zeros and conpare on ETTm1 and ETTm2 datasets. The results are shown below and the best-performing metric for each experiment is bolded.

DatasetMethodMoFo+ zeros padding
Prediction WindowMSE, MAEMSE, MAE
ETTm1960.286, 0.3350.292, 0.345
1920.320, 0.3630.328, 0.372
3360.347, 0.3820.348, 0.386
7200.388, 0.4110.401, 0.435
ETTm2960.155, 0.2400.156, 0.247
1920.211, 0.2830.215, 0.293
3360.258, 0.3140.261, 0.318
7200.342, 0.3680.345, 0.372

Our padding strategy consistently outperforms zero padding, demonstrating that re-using the immediately preceding historical pattern enhances generalization rather than introducing redundancy.

W2.2.1. Determine the cycle length

We determine the period length by calculating the distance between the peaks of the autocorrelation coefficients of the time series data on the training set. The specific method for this calculating can be found in Appendix A.2.

W2.2.2. What is H\mathcal{H} ?

We apologize for omitting the precise definition. In our notation, H\mathcal{H} denotes the Heaviside step function satisfying

H(s)=0 if s<0;H(s)=1 if s>0.\mathcal{H}(s) = 0 \text{ if } s < 0; \mathcal{H}(s) = 1 \text{ if } s > 0.

We will add this clarification to the next version.

W3.1. Transformer layers in MoFo

MoFo attains best performance with only one Transformer layer. The corresponding ablation is provided in Section 4.6 and Figure 5 of the manuscript, where we systematically vary the number of Transformer blocks. Across all settings, stacking additional layers yields negligible to zero gains, confirming that a single layer already saturates the model’s capacity.

W3.2. Period length of MoFo

Following your suggestion, we designed two variants—"+ Half" and "+ Double"—to investigate MoFo’s sensitivity to the chosen the half and the double period length. As shown the below Table, the results on three diverse datasets with different period reveal that a well-calibrated period is essential: an inappropriate choice disrupts the model’s ability to capture the intrinsic periodicity of the time series, whereas the original setting preserves this structure and retains optimal performance.

DatasetMethodMoFo+ Half+ Double
Prediction WindowMSE, MAEMSE, MAEMSE, MAE
ETTm1960.289, 0.3360.302, 0.3460.295, 0.359
1920.326, 0.3670.335, 0.3790.339, 0.388
3360.361, 0.3910.374, 0.4040.372, 0.408
7200.405, 0.4190.414, 0.4250.419, 0.429
Weather960.141, 0.1860.152, 0.1990.157, 0.206
1920.186, 0.2300.193, 0.2390.194, 0.234
3360.233, 0.2720.241, 0.2790.244, 0.284
7200.312, 0.3310.348, 0.3630.342, 0.355
Traffic960.362, 0.2470.376, 0.2580.370, 0.252
1920.379, 0.2540.380, 0.2620.388, 0.275
3360.390, 0.2580.406, 0.2860.395, 0.283
7200.424, 0.2810.434, 0.2970.438, 0.302
评论

Thanks for the your reply. This has largely addressed my concerns. Considering the importance of MoFo for long-term time series forecasting, I have raised my score to 5.

评论

Dear Reviewer Q3nE,

We sincerely value your positive feedback and the improvement in our score. Your insightful guidance has been immensely helpful, and we will carefully revise our work to incorporate all of your suggestions.

With heartfelt thanks,

Authors.

审稿意见
5

This paper introduces MoFo, a Transformer-based approach for long-term sequence forecasting. The method employs a periodicity-structured patching strategy along with a periodicity-aware modulator, inspired by theoretical mathematical insights. Extensive experimental results demonstrate that MoFo achieves highly competitive performance and superior efficiency.

优缺点分析

Strength:

S1. The paper has a strong and clearly articulated motivation. The authors have effectively identified and explained the gap in existing work on time series forecasting.

S2. The proposed method is novel, incorporating a periodicity-aware modulator and an input representation strategy that emphasize the modeling of temporal periodicity. And the approach is driven by theoretical mathematical insights.

S3. Extensive experiments validate the model’s effectiveness, showcasing significant performance improvements and approximately 15× efficiency gains.

S4. The paper is well-written and well-structured, and The mathematical formulas are presented in a clear and explicit manner.

Weakness:

W1. Some time series forecasting benchmarks need to be supplemented. For example, how does the model compare to FITS and SparseTF models?

W2. Why does the time complexity of MoFo remain stable as the input time series becomes longer?

W3. What does MACs mean in Table 3? What is the purpose of this indicator?

问题

Please kindly see the Weaknesses above.

局限性

YES

最终评判理由

Since authors have targeted and addressed my questions in rebuttal I keep the score conservatively.

格式问题

NO

作者回复

We appreciate your guidance—your suggestions have helped us produce a more precise, readable, and broadly accessible manuscript.

W1. More Baselines

Thank you for your valuable suggestion. The FITS [5] model is already included in Table 4 of Appendix E.3. To further address your concern, we have conducted additional experiments comparing FITS [5] and SparseTF [6] on the ETTm1 and ETTm2 datasets; the results are reported below. For clarity, the best performance in each experiment is highlighted in bold. Across all evaluations, MoFo consistently achieves the most competitive results, demonstrating superior performance.

DatasetMethodMoFoFITSSparseTSF
Prediction WindowMSE, MAEMSE, MAEMSE, MAE
ETTm1960.286, 0.3350.303, 0.3450.316, 0.355
1920.320, 0.3630.337, 0.3650.348, 0.376
3360.347, 0.3820.368, 0.3840.373, 0.387
7200.388, 0.4110.420, 0.4130.434, 0.422
ETTm2960.155, 0.2400.165, 0.2540.166, 0.256
1920.211, 0.2830.219, 0.2910.220, 0.292
3360.258, 0.3140.272, 0.3260.273, 0.327
7200.342, 0.3680.359, 0.3810.361, 0.381

Furthermore, as requested by Reviewer 7GM5, we have added comparisons with the TimesNet [1], WITRAN [2], PGN[3], and TimeKAN[4] models. Please refer to our response to W1 for details.

W2. Why does the time complexity of MoFo remain stable as the input time series becomes longer?

In MoFo, due to the unique patch strategy, the number of patches is equal to the period length and thus remains constant. Each patch contains exactly T/P\lceil T / P\rceil elements, where TT is the input sequence length and PP is the period length. Since the time complexity depends solely on the number of patches (i.e., the period length), MoFo is not sensitive to increases in the input sequence length.

W3. What is MACs

MACs (Multiply–Accumulate operations) quantify the total number of fused multiply-add operations executed during a single forward pass. This metric serves as a hardware-agnostic proxy for computational cost: fewer MACs generally translate to lower latency and reduced energy consumption on both CPUs and accelerators. By reporting MACs alongside accuracy, we provide a clear, hardware-independent view of the efficiency–performance trade-off offered by each model.

Ref:

[1] Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023). Timesnet: Temporal 2d-variation modeling for general time series analysis. In The Eleventh International Conference on Learning Representations.

[2] Jia, Y., Lin, Y., Hao, X., Lin, Y., Guo, S., & Wan, H. (2023). WITRAN: Water-wave Information Transmission and Recurrent Acceleration Network for Long-range Time Series Forecasting. In Thirty-seventh Annual Conference on Neural Information Processing Systems.

[3] Jia, Y., Lin, Y., Yu, J., Wang, S., Liu, T., & Wan, H. (2024). PGN: The RNN's New Successor is Effective for Long-Range Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.

[4] Huang, S., Zhao, Z., Li, C., & Bai, L. (2025). TimeKAN: KAN-based Frequency Decomposition Learning Architecture for Long-term Time Series Forecasting. In The Thirteenth International Conference on Learning Representations.

[5] Xu, Z., Zeng, A., & Xu, Q. (2023). FITS: Modeling time series with 10k10 k parameters. arXiv preprint arXiv:2307.03756.

[6] Lin, S., Lin, W., Wu, W., Chen, H., & Yang, J. (2024, July). SparseTSF: Modeling Long-term Time Series Forecasting with* 1k* Parameters. In International Conference on Machine Learning (pp. 30211-30226). PMLR.

评论

Thank you for your detailed response and the additional experimental results. Regarding the additional experiments and analysis you provided, I appreciate the efforts of the authors in the rebuttal phase, and as a result, I conservatively keep the rating. I will keep tracking the discussion with 7gm5, who provide actionable discussions.

评论

Dear Reviewer XTBU,

Thank you for your insightful feedback during the review phase, and we appreciate your support of our submission and your attention during the discussion phase between us and Reviewer 7gm5.

In response to Reviewer 7gm5's actionable discussions on two points: (1) the specific impact of model layers, and (2) evaluation on larger-scale datasets, we have conducted corresponding experiments and have addressed Reviewer 7gm5's concerns, as detailed in our subsequent response to Reviewer 7gm5 at [https://openreview.net/forum?id=sbvLts2HqR&noteId=fAooVNcIEf].

Thank you once again for your valuable feedback. Your thorough review has been truly encouraging for us.

Best regards,

The Authors of Submission 2224

审稿意见
4

This paper proposes MoFo, a Transformer-based forecasting method that models time series in a two-dimensional structure by leveraging their periodic patterns. The authors conduct correlation-based experiments in an attempt to validate the effectiveness of the proposed approach.

优缺点分析

Strengths:

S1: The method for handling periodicity in the paper is reasonable and demonstrates a certain degree of novelty.

S2: The figures and tables throughout the paper are clear, and the presentation is well-structured.

Weaknesses:

W1: The paper lacks comparisons with important baseline methods, especially since the proposed method also relies on 2D modeling. Therefore, it should be compared with previous classic SOTA methods that also use 2D modeling, such as TimesNet[1] (CNN-based), WITRAN[2] (RNN-based), and TPGN[3] (PGN-based, where PGN is a new paradigm based on RNN). Moreover, in the related work section, these key methods are not addressed, and other emerging approaches, such as TimeKAN[4] (KAN-based, where KAN is a new paradigm based on Linear), are also omitted.

[1] Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023). Timesnet: Temporal 2d-variation modeling for general time series analysis. In The Eleventh International Conference on Learning Representations.

[2] Jia, Y., Lin, Y., Hao, X., Lin, Y., Guo, S., & Wan, H. (2023). WITRAN: Water-wave Information Transmission and Recurrent Acceleration Network for Long-range Time Series Forecasting. In Thirty-seventh Annual Conference on Neural Information Processing Systems.

[3] Jia, Y., Lin, Y., Yu, J., Wang, S., Liu, T., & Wan, H. (2024). PGN: The RNN's New Successor is Effective for Long-Range Time Series Forecasting. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.

[4] Huang, S., Zhao, Z., Li, C., & Bai, L. (2025). TimeKAN: KAN-based Frequency Decomposition Learning Architecture for Long-term Time Series Forecasting. In The Thirteenth International Conference on Learning Representations.

W2: The experimental setup contains clear flaws. Although the authors claim to follow configurations from previous studies, the actual design lacks fairness and rigor.

(1) Improper handling of historical look-back length

(1.1) Different look-back lengths may alter the data splitting process, thereby affecting model training. This represents a lack of control over experimental variables.

(1.2) Different models exhibit varying sensitivity to look-back lengths, so treating it uniformly as a hyperparameter is clearly inappropriate. A more fair and rigorous approach would be to predefine several common look-back lengths and treat them as separate experimental groups, enabling a more systematic comparison of how each method performs under different historical horizons.

(2) Limited hyperparameter search space The TFB paper cited by the authors only expands to 8 parameter configurations. Taking Transformers as an example, even commonly used parameters like d_model, e_layer, and d_layer are barely covered—let alone other hyperparameters. Such a narrow search space makes it difficult to draw convincing conclusions.

(3) Why an extensive hyperparameter search is necessary

(3.1) Even with identical random seeds and hyperparameter values, experimental results may still vary across hardware platforms.

(3.2) Different models have different levels of sensitivity to hyperparameters. When the experimental platform changes, directly reusing published settings cannot guarantee that baseline models operate at their optimal capacity, nor that results are consistent with those reported originally.

Therefore, to ensure a fair comparison, it is essential to perform a fresh hyperparameter search for all models within a unified and sufficiently broad search space by selecting optimal configurations on the validation set and then evaluating and reporting results on the test set. While models with more hyperparameters may lead to a larger search space, this process is critical for ensuring fairness and rigor. Only after eliminating platform differences and hyperparameter biases, and ensuring that all models perform at their best, can a trustworthy and meaningful comparison be made.

Furthermore, the authors do not compare their approach with the baselines mentioned in W1. Consequently, the current experiments are insufficient to convincingly establish the effectiveness of the proposed method.

W3: The efficiency experiments likewise omit comparisons with the baselines listed in W1, as well as with FITS and DLinear. In addition, the paper does not specify how the hyperparameters were chosen for these tests. It must be emphasized that efficiency studies are intended to measure the resource consumption inherent to each model architecture; hence, all experimental variables should be controlled to ensure a fair comparison. In particular, shared hyperparameters should be kept identical so as to eliminate confounding effects, such as one method being run with very small settings while another is evaluated with much larger ones.

问题

Is MoFo still effective on time series data without clear periodicity? Has the method been explored on such datasets, for example, the Human Connectome Project (HCP), which involves non-periodic time series? I would like to see more discussion on the applicability and performance of the proposed method in non-periodic scenarios.

局限性

The authors mentioned limitations regarding computational complexity.

最终评判理由

The authors have addressed my concern. Therefore, I am raising my score to 4.

格式问题

It seems that the NeurIPS 2025 submission guidelines explicitly state that tables should not contain vertical rules, yet vertical rules are included in Table 2 and Table 4 of the paper.

作者回复

Thank you very much for your time and effort; they are crucial for improving the quality of our manuscript.


W1. More baselines

We have already compared TimesNet in Table 4. Next, we added comparative experiments with PGN, TimeKAN, and WITRAN. The performance results are shown in the table below.

DatasetrMethodMoFoTimeKANPGNTimesNetWITRAN
Prediction WindowMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAE
ETTh1960.360, 0.3890.369, 0.3960.402, 0.4070.389, 0.4120.660, 0.627
1920.397, 0.4130.402, 0.4170.446, 0.4330.440, 0.4430.844, 0.723
3360.407, 0.4240.419, 0.4300.494, 0.4560.523, 0.4870.967, 0.766
7200.447, 0.4540.442, 0.4630.490, 0.4780.521, 0.4950.965, 0.768
ETTh2960.273, 0.3340.279, 0.3430.323, 0.3580.334, 0.3700.601, 1.285
1920.327, 0.3730.326, 0.3800.426, 0.4200.404, 0.4134.193, 1.611
3360.361, 0.4050.368, 0.4110.452, 0.4570.389, 0.4355.036, 1.819
7200.379, 0.4250.419, 0.4430.438, 0.4490.434, 0.4484.556, 1.817

Fairness statement. Although the TFB-benchmark setting we adopted may differ from your expectations, we would like to clarify that all models are run under the same experimental environment and unified configuration to ensure basic fairness in evaluation. We chose to conduct experiments on the latest TFB mainly because it has been extensively validated and is academically recognized—the benchmark was nominated for the Best Paper Award at VLDB 2024 and has been cited over 62 times on Google Scholar.


W2. Historical look-back length

Please allow us to clarify TFB’s approach regarding historical window selection. According to the TFB GitHub guidelines, multiple look-back window lengths are preset in the experiments, covering typical settings commonly used in time series forecasting tasks. For each model, tests are conducted independently on all candidate historical window lengths, and the best-performing result is ultimately reported.

We believe this approach does not introduce fairness issues; on the contrary, it is fairer to models with varying sensitivities. Forcing all models to use exactly the same look-back length might disadvantage those that are particularly sensitive to this parameter, causing them to be unfairly regarded as “underperforming”. In contrast, as long as the device conditions and the amount of historical data are consistent, allowing each model to flexibly choose its optimal historical window according to its structural characteristics both ensures a fairer evaluation of predictive performance and better reflects the flexible practices found in real-world applications.

Despite having a different perspective, in response to your suggestion, we conducted experiments using several commonly adopted predefined look-back window lengths. Specifically, for forecasting with a future window length of 720, we used {96, 336, 512} as the look-back windows. Due to time constraints, partial experimental results are shown below.

MetricMoFoTimeKANFITSPGNTimesNetWITRANDLinear
Look-back WindowMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAE
ETTh1960.447, 0.4540.463, 0.4700.547, 0.5180.490, 0.4780.521, 0.4950.965, 0.7680.515, 0.511
3360.459, 0.4690.509, 0.4900.475, 0.4870.522, 0.5150.542, 0.5190.976, 0.7710.471, 0.493
5120.443, 0.4630.442, 0.4630.435, 0.4580.692, 0.5980.560, 0.5311.034, 0.8020.464, 0.488
ETTh2960.416, 0.4420.438, 0.4510.439, 0.4520.438, 0.4500.434, 0.4484.556, 1.8170.650, 0.571
3360.393, 0.4280.420, 0.4490.397, 0.4310.466, 0.4740.472, 0.4804.272, 1.7450.704, 0.597
5120.379, 0.4250.408, 0.4430.382, 0.4250.440, 0.4650.480, 0.4684.668, 1.7870.786, 0.623

W3. Hyperparameter search

(1) According to the TFB guidelines, the authors follow the hyperparameters specified in the original papers. And they further performed additional hyperparameter searches on 16 configuration sets to achieve a more comprehensive evaluation, which covers the key hyperparameters sufficiently.

(2) TFB has fully open-sourced the necessary environment configuration, including detailed version information, along with a fixed running protocol, enabling users to reproduce the same performance.

(3) Please refer to W1.


W4. Complexity

Our efficiency comparison, as stated in the caption of Table 3, is conducted under the condition that each model achieves its optimal predictive performance—a standard and widely accepted practice in academic research. Since the ultimate goal of time series forecasting is to enhance prediction accuracy, efficiency comparisons that consider performance are both meaningful and practical. For instance, if all models are configured with extreme parameter settings, like compressing the feature dimension to 1, computational efficiency might seem significantly improved, but prediction performance would suffer greatly. Therefore, a fair and informative efficiency evaluation must be conducted at comparable performance levels to ensure both validity and practical significance.

In response to your suggestion, we used the ETTh1 dataset as an example and optimized this configuration for the best performance, while keeping certain hyperparameters such as batch size and learning rate fixed. We report the efficiency for fixed batch sizes (specifically 4 and 128), learning rate 1e-4 and look-back window 512, as well as for each model under its best-performing setting. Please refer to the table below for details (K: Kilo-10310^3, M: Million-10610^6). We observe that our model achieves significant performance improvements with only a slight increase in complexity.

ETTh1ModelMSEMAEMemory# Para# MACs# FLOPs
bs=4MoFo0.3720.399542 MB33.0 K2.5 M2.9 M
FITS0.3810.406540 MB24,0 K0.7 M0.7 M
DLinear0.3870.415532 MB98,5 K2.8 M2.8 M
bs=128MoFo0.3800.390636 MB22.4 K42.3 M50.6 M
FITS0.4920.478584 MB24,0 K21.4 M21.4 M
DLinear0.3890.418558 MB98.5 K88.5 M88.1 M

Q1.

The Human Connectome Project (HCP) is a large-scale dataset with over 1TB of data, requiring mandatory institutional approval for access. Due to time constraints, we were unable to use and study it during the rebuttal period. To address your concerns, we opted to utilize the Influenza-Like Illness (ILI) dataset for evaluation instead. The ILI dataset, published by the US Centers for Disease Control and Prevention (CDC) from 2002 to 2021, comprises health-related data that record the weekly proportion of ILI patients relative to the total number of visits, without any obvious seasonality. We predict future values for four windows: LL\in {24,36,48,60}.

To emphasize our modeling of non-periodic time series, we follow the approach of TimesNet by computing the main pseudo-period as the reciprocal of the dominant frequency obtained via Fast Fourier Transform (FFT) on each input time series. Although MoFo was designed with periodic patterns in mind, the results below demonstrate that it still achieves remarkably high accuracy on non-periodic data, highlighting its strong capability in capturing complex temporal dynamics even in the absence of clear periodicity.

MoFoTimeKANFITSPGNTimesNetWITRAN
LLMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAE
242.113, 0.9272.176, 0.9282.182, 1.0022.255, 0.9362.131, 0.9583.486, 1.218
361.952, 0.9242.166, 0.9932.330, 1.0512.132, 0.9402.612, 0.9745.129, 1.596
481.714, 0.8242.011, 0.9282.761, 1.1842.182, 0.9441.916, 0.8973.862, 1.332
601.800, 0.9062.010, 0.9672.929, 1.2172.169, 0.9401.995, 0.9054.812, 1.543
评论

Thank you for your response. I noticed that additional experiments have been provided, which seem to demonstrate the effectiveness of the MoFo method from a different perspective. In addition, supplementary experiments on non-periodic data have been included to support the method’s broader applicability. My main concerns have been addressed, and I am therefore willing to adjust my score to a positive rating.

However, the following points would benefit from further clarification:

(1) The specific impact of model layers on performance;

(2) Regarding efficiency, it would be beneficial to include experiments on larger-scale datasets to allow for a more comprehensive evaluation.

评论

We sincerely appreciate your positive response. Moving forward, we will conduct additional experiments to ensure your satisfaction. Once again, we deeply thank you for improving the evaluation to a positive score.

评论

Thank you very much for your response. As the rebuttal period is nearing its end, we have done our utmost to address your concerns and ensure your satisfaction.

1. Model Layers

1.1. Model Layer of MoFo

MoFo attains best performance with only one Transformer layer. The corresponding ablation is provided in Section 4.6 and Figure 5 of the submission, where we systematically vary the number of Transformer blocks. In all configurations, adding more layers does not provide additional performance gains.

1.2. Model Layer of Baselines

To further address your concerns regarding the baseline model layer parameters, we conducted an investigation using DUET, TimeKAN, and TimesNet. Due to time constraints, we were only able to perform hyperparameter experiments on a single dataset, ETTh1, with prediction lengths of 96, 192, and 336. The optimal layer choices for each prediction length are listed in parentheses. The experimental results are as follows:

Model Layer1234
DUETMSE, MAEMSE, MAEMSE, MAEMSE, MAE
960.352, 0.3840.358, 0.3870.371, 0.4000.368, 0.396
1920.398, 0.4090.402, 0.4130.410, 0.4180.408, 0.418
3360.414, 0.4260.413, 0.4300.413, 0.4320.417, 0.436
Model Layer1234
TimeKANMSE, MAEMSE, MAEMSE, MAEMSE, MAE
960.375, 0.4020.369, 0.3960.378, 0.4020.376, 0.401
1920.415, 0.4200.402, 0.4170.412, 0.4270.416, 0.427
3360.434, 0.4370.419, 0.4300.434, 0.4360.435, 0.434
Model Layer1234
TimesNetMSE, MAEMSE, MAEMSE, MAEMSE, MAE
960.412, 0.4290.389, 0.4120.403, 0.4210.418, 0.430
1920.485, 0.4700.440,0.4430.475, 0.4620.470, 0.462
3360.530, 0.4930.523, 0.4870.533, 0.4910.531, 0.492

2. Larger-scale Experiments

To further address concerns about efficiency and performance on large-scale datasets, we conducted experiments on two large-scale time series datasets: XXLTraffic [1] and XTraffic [2]. The long-period dataset XXLTraffic includes 12 years of traffic flow data from 2012 to 2024, while the large-scale dataset XTraffic contains traffic data recorded since 2023 across 16,972 feature channels (nodes), showcasing a significantly larger number of channels.

Due to time constraints, we compared the models you are concerned with (TimeKAN, DLinear, and the current SOTA time series model DUET). The input length was uniformly set to 512, and the output length to 720. The performance results and training efficiency comparisons on two datasets are as follows:

XXLTrafficMSEMAEMemoryTraining Time
MoFo (Ours)0.0660.1451,486 MB149.1 s/epoch
DUET0.0720.1585,728 MB296.7 s/epoch
TimeKAN0.0800.17922,272 MB1008.6 s/epoch
DLinear0.0740.163792 MB84.4 s/epoch
XTrafficMSEMAEMemoryEpoch Time
MoFo (Ours)0.0480.09032,182 MB765.1 s/epoch
DUET--Out-of-Memory-
TimeKAN--Out-of-Memory-
DLinear0.0620.10714,548 MB552.15 s/epoch

In larger-scale scenarios, MoFo demonstrates the ability to achieve the highest performance while maintaining competitive efficiency. Specifically, for XTraffic dataset with a large number of nodes (channels), TimeKAN and DUBT are too complex to run. Compared to DLinear, MoFo achieves a 22.58% performance improvement on the large-scale dataset. Furthermore, for the long time-span dataset, MoFo achieves competitive performance while maintaining high efficiency.

Ref:

[1] Yin D, Xue H, Prabowo A, et al. Xxltraffic: Expanding and extremely long traffic dataset for ultra-dynamic forecasting challenges[J]. arXiv e-prints, 2024: arXiv: 2406.12693.

[2] Gou X, Li Z, Lan T, et al. Xtraffic: A dataset where traffic meets incidents with explainability and more[J]. arXiv preprint arXiv:2407.11477, 2024.

评论

I have no further concerns. I suggest that the authors include the above points from the rebuttal in the revised paper for clarity. I am raising my score to 4.

评论

Thank you very much for your time and feedback. We promise to incorporate our discussion into the new version. Once again, we sincerely appreciate your encouraging evaluation.

审稿意见
4

The paper introduces MoFo, a Transformer-based long-term time-series forecasting model comprising

  1. Period-Structured Patching (PSP): reshapes an input length-TT series into PP rows of phase-aligned "patches", reducing the number of tokens processed;
  2. Period-Aware Modulator (PAM): multiplies attention logits by a learnable sigmoid mask of phase distances, theoretically shown to approximate a hard phase gate. The authors claim significant memory and runtime savings, as well as competitive accuracy, on eight periodic benchmarks, and provide a theoretical bound on the mask approximation error.

优缺点分析

Strengths S1) PSP and PAM are easy to implement and interpret; the paper is mostly well written. S2) On datasets with a single, stable period, MoFo yields substantial token reduction and lower memory/FLOPs. S3) The paper provides a concise proof that the sigmoid mask converges to a binary phase filter.

Weaknesses W1) The proposed method relies on the assumption of a period length PP. However, real-world time series often exhibit complex, time-varying periodic patterns. For example, in Figure 1(a), the "Successive Patching" clearly shows multiple peaks representing mixed periodicities. Why is it reasonable to treat these as belonging to the same period? When dealing with multiple samples from different time periods, how can we ensure accurate and meaningful period selection? When data faces mixed periodicities and time-varying periods (e.g., summer vs. winter electricity consumption patterns), can MoFo effectively handle such complex periodic variations?

问题

How does accuracy change if the supplied PP is perturbed by ±5–15 % or if multiple periods (e.g., 24 h & 7 d) coexist?

局限性

Yes

最终评判理由

The authors have addressed most of my other concerns. I suggest that the authors provide detailed explanations from the rebuttal in the revised paper for clarity. I have raised my score to 4.

格式问题

NA

作者回复

We sincerely thank you for your precious time and insightful comments. We will now address your concerns point by point below.


W1.

W1.1. Confusion about period

In Figure 1(a), the “Successive Patching” does not represent the coexistence of mixed periods; rather, it includes multiple peaks, which are dynamic components within a single period and together constitute a complete period.

We cannot simply assume that a single peak corresponds to one complete period. For instance, in traffic data, there are distinct morning and evening peak periods; however, the periodicity of traffic is typically considered to be one full day rather than 12 hours. To accurately determine the period length, we employ the autocorrelation function (ACF). As shown in Figure 1(b), the autocorrelation coefficients of the sequence in Figure 1(a) indicate that a truly complete period often contains multiple peaks.

W1.2. Accurate period selection

As stated in Appendix Section A.2 of the manuscript, we utilize three steps to accurately determine periodicity.

Firstly, we identify potential period lengths of repeating patterns within the sequence using the autocorrelation function (ACF). The ACF is a widely used tool in the field of time series analysis for examining periodic patterns [1-3].

Secondly, we verify the existence of these periods through statistical significance testing, such as utilizing a 95% confidence interval, to eliminate any noise interference.

Finally, we also conduct hyperparameter experiments to assess the sensitivity of model performance with respect to different period lengths.

W1.3. Time-varying periods

The electricity dataset (ETT) used in our experiments aligns perfectly with your requirements. The ETT dataset records electricity consumption from two counties in China over a continuous period from July 2016 to July 2018, thus encompassing both winter and summer patterns. However, it is worth noting that the periodic lengths for the two seasons do not differ significantly. For your convenience, we have included the relevant performance results below.

DatasetMethodMoFoDUETDLinear
Prediction WindowMSE, MAEMSE, MAEMSE, MAE
ETTh1960.360, 0.3890.352, 0.3840.379, 0.403
1920.397, 0.4130.398, 0.4090.408, 0.419
3360.407, 0.4240.414, 0.4260.440, 0.440
7200.447, 0.4540.429, 0.4550.471, 0.493
ETTm1960.286, 0.3350.279, 0.3330.300, 0.345
1920.320, 0.3630.320, 0.3580.336, 0.366
3360.347, 0.3820.348, 0.3770.367, 0.386
7200.388, 0.4110.405, 0.4080.419, 0.416

Furthermore, to address your concerns more comprehensively, we have created a synthetic dataset (Mixed-ETT) by systematically combining ETTh1 (with a period of 24) and ETTm1 (with a period of 96). We divided each dataset into four equal parts along the time axis. Then, we alternately sampled each segment to construct the new dataset Mixed-ETT, represented as {ETTh1[:14\frac{1}{4}], ETTm1[14\frac{1}{4}:12\frac{1}{2}], ETTh1[12\frac{1}{2}:34\frac{3}{4}], ETTm1[34\frac{3}{4}:1]}. Consequently, the Mixed-ETT dataset displays a mixed periodic length of (24, 96, 24, 96). We compare the performance of state-of-the-art models, DUET and DLinear, as shown below.

DatasetMethodMoFoDuetDLinear
Prediction WindowMSE, MAEMSE, MAEMSE, MAE
Mixed-ETT960.178, 0.2230.186, 0.2290.194, 0.232
1920.185, 0.2320.184, 0.2410.190, 0.247
3360.180, 0.2370.187, 0.2470.196, 0.256
7200.191, 0.2530.218, 0.2780.219, 0.282

Even under time-varying periods, MoFo retains the competitive performance, underscoring the soundness of its architectural design and its robust capability to capture dynamic temporal patterns.


Q1.

Q1.1. Performance under perturbed period setting

To resolve any remaining concerns, we introduced four variants:

  • "+5 %" and "+15 %": the period setting is increased by 5 % and 15 %, respectively.
  • "–5 %" and "–15 %": the period setting is decreased by 5 % and 15 %, respectively.

We evaluated these perturbations on three datasets with distinct period lengths; the quantitative results are reported below.

DatasetMethodMoFo+ 5%+ 15%- 5%- 15%
Prediction WindowMSE, MAEMSE, MAEMSE, MAEMSE, MAEMSE, MAE
ETTm1960.286, 0.3350.292, 0.3350.297, 0.3380.291, 0.3370.297, 0.341
1920.320, 0.3630.329, 0.3680.336, 0.3700.333, 0.3690.344, 0.386
3360.347, 0.3820.357, 0.3880.372, 0.3910.354, 0.3850.376, 0.402
7200.388, 0.4110.406, 0.4200.403, 0.4160.402, 0.4150.411, 0.420
Weather960.141, 0.1860.143, 0.1920.155, 0.2060.157, 0.2210.146, 0.188
1920.186, 0.2300.196, 0.2360.202, 0.2520.199, 0.2400.201, 0.242
3360.233, 0.2720.238, 0.2770.245, 0.2860.235, 0.2800.243, 0.286
7200.312, 0.3310.335, 0.3580.323, 0.3580.332, 0.3520.335, 0.355
Traffic960.362, 0.2470.388, 0.2660.395, 0.2650.383, 0.2550.382, 0.265
1920.379, 0.2540.397, 0.2750.408, 0.2870.398, 0.2840.409, 0.288
3360.390, 0.2580.395, 0.2600.405, 0.2750.395, 0.2660.417, 0.294
7200.424, 0.2810.424, 0.2810.443, 0.2980.437, 0.2830.443, 0.303

Q1.2. Performance under multiple periods setting

We utilized the traffic dataset, which includes the multiple time intervals you mentioned: 1 day and 7 days. In practice, when dealing with multi-period data, we typically set the shorter period as our baseline configuration. Our experimental results were very impressive, as shown in Table 2 in the manuscript.

To address your concern regarding the performance when integrating multiple MoFo models to handle different periods, we introduced three models. One original model (MoFo-1) excels in sequences with a period length of 24 hours (1 day), another (MoFo-7) is designed for a weekly period of 7 days, and the final ensemble model (Mix-MoFo) learns and predicts in parallel using the two MoFo models, combining their outputs to produce the final prediction. Below are the results from our experiments on the Traffic dataset.

The experimental results indicate that using the shorter period as the configuration allows MoFo to capture more precise periodic patterns, leading to enhanced performance.

MoFo-1Mix-MoFoMoFo-7
Prediction WindowMSE, MAEMSE, MAEMSE, MAE
Traffic960.362, 0.2470.365, 0.2540.381, 0.257
1920.379, 0.2540.380, 0.2550.394, 0.262
3360.390, 0.2580.395, 0.2630.409, 0.272
7200.424, 0.2810.430, 0.2850.451, 0.289

Experimental results demonstrate that our straightforward MoFo implementation achieved strong performance. When dealing with multi-period time series, we set the smallest period length as our baseline configuration. The “Mix-MoFo” variant added assumptions regarding additional periods, which increased the risk of overfitting and consequently led to a decline in performance.

Ref:

[1] Wu H, Xu J, Wang J, et al. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting[J]. Advances in neural information processing systems, 2021, 34: 22419-22430.

[2] Lin S, Lin W, Hu X, et al. Cyclenet: Enhancing time series forecasting through modeling periodic patterns[J]. Advances in Neural Information Processing Systems, 2024, 37: 106315-106345.

[3] Lin S, Lin W, Wu W, et al. SparseTSF: Modeling Long-term Time Series Forecasting with* 1k* Parameters[C]//International Conference on Machine Learning. PMLR, 2024: 30211-30226.

评论

Thank you for your response. I have read the response and most of my concerns have been addressed. I have improved the score. I would like to suggest adding more detailed explanations and experimental results from the rebuttal in the revised paper.

评论

Hi Reviewer EWUL,

The authors have submitted their rebuttal. Do you have any further questions or comments based on their response?

Best regards, The AC

评论

Dear Reviewer EWUL,

Thank you very much for your thorough review. If you have any further questions or suggestions, we would be delighted to discuss them with you.

To make our response clearer, we have summarized it as follows:

(1) We would like to clarify that the selection of the period length is determined using the autocorrelation function with Bartlett's Test under the Null Hypothesis, a widely adopted method for periodic analysis in time series data. Further details can be found in Appendix A.2 of the paper.

(2) We designed experiments to evaluate the model's performance in dynamic-, mixed-, and multi-period scenarios. The highly competitive results demonstrate the outstanding performance of our model under these complex conditions.

Additionally, we are happy to share the latest progress of our experiments.

In our response to Reviewer 7GM5, we utilized a 12-year-long time series dataset, XXLTraffic, which potentially contains complex periodic scenarios over such an extended time span. On this dataset, our model also achieved competitive performance. For convenience, we have reproduced the performance comparison below.

XXLTrafficMoFo (Ours)DUETTimeKANDLinear
MSE0.0660.0720.0800.074
MAE0.1450.1580.1790.163

Thank you again for your professional and valuable review! Your feedback can significantly enhance the quality of our paper, and we sincerely appreciate it.

Sincerely,

The authors of the paper.

评论

Dear Reviewer EWUL,

We are deeply grateful for your thoughtful and constructive feedback, as well as the time and effort you have dedicated to reviewing our manuscript. Your insightful suggestions have been invaluable in guiding us to enhance the quality of our research.

Once again, we sincerely thank you for your support and expertise, and we truly appreciate your valuable contribution to this work.

With warm regards,

The Authors of Submission 2224

最终决定

This paper proposes a modulated method for period modeling in long-term time series forecasting. All reviewers acknowledged the strong motivation, clear presentation, concise model design, and (after rebuttal) comprehensive evaluation. Following the rebuttal, two reviewers increased their ratings to positive scores, and the overall consensus was in favor of acceptance.

I strongly encourage the authors to revise the paper in line with the reviewers’ comments and to release their code to the community to enhance reproducibility.

Recommendation: Accept as poster.