PaperHub
7.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
5
5
4
4.0
置信度
创新性2.8
质量3.0
清晰度3.3
重要性3.0
NeurIPS 2025

SEMPO: Lightweight Foundation Models for Time Series Forecasting

OpenReviewPDF
提交: 2025-05-10更新: 2025-10-29
TL;DR

SEMPO is a lightweight time series foundation model with energy-aware spectral decomposition and mixture-of-prompts enabled Transformer for effective and efficient zero-shot and few-shot forecasting.

摘要

关键词
time seriesfoundation modelforecastinglightweight

评审与讨论

审稿意见
5

The paper proposes a lightweight foundation model for time series forecasting, SEMPO, which includes an Energy-Aware Spectral Decomposition (EASD) module and a Mixture-of-Prompts enabled Transformer (MoPFormer) module. The former leverages high- and low-energy frequency signals to improve data utilization, while the latter achieves parameter-efficient adaptation through a lightweight prompt-based expert system. Experiments show that SEMPO, with 6.5 million parameters and pre-trained on 83 million data samples, outperforms mainstream models in terms of prediction accuracy under zero-shot and few-shot scenarios, demonstrating its efficiency and generalization capability.

优缺点分析

Strengths

  1. Proposes a lightweight SEMPO model that achieves zero-/few-shot prediction advantages with only 6.5 million parameters and 83 million pre-training data, offering high resource efficiency and practical value.
  2. The experiments cover multiple datasets and baseline models, including ablation studies and visualization analyses, providing thorough validation.
  3. The paper is well-structured, with clear formulas, standardized figures, precise technical descriptions, and smooth logical flow.

References

  1. It is recommended to include full-shot experiments to evaluate model performance with the full dataset, thereby validating the benefits of pre-training.
  2. A comparison with frequency debiasing models such as Fedformer [1] should be added to highlight the superiority of the EASD module.
  3. The description of the two-stage training paradigm in the main text is too brief; more details should be provided.
  4. The EASD module uses a masking mechanism for data augmentation, which actually increases the amount of training data. Therefore, further verification is needed to confirm whether the claimed reduction in pre-training cost is valid.
  5. The zero-shot experiment in Table 1 and the efficiency analysis in Figure 6 do not include the lightweight model TTM. It is suggested to add its results to fully illustrate the performance gradient of the lightweight methods. Additionally, since new versions of Chronos and TimesFM are available, it is advisable to update the comparisons to the latest state-of-the-art models to ensure the conclusions remain current with field advancements.
  6. The visualization of ZLecZ_{Lec} in Figure 2 appears to contain an error; the complementary relationship between ZLecZ_{Lec} and ZHecZ_{Hec} should be corrected.
  7. In Figure 4(b), the low-energy feature response of Moirai seems to reflect prediction inaccuracies rather than being completely neglected. The authors are encouraged to provide a more reasonable explanation for this phenomenon.

Weaknesses

[1] Fredformer: Frequency Debiased Transformer for Time Series Forecasting, KDD 2024

问题

The authors can refer to the weakness listed above.

局限性

The authors can refer to the weakness listed above.

最终评判理由

The paper proposes a lightweight foundation model for time series forecasting, SEMPO, which includes an Energy-Aware Spectral Decomposition (EASD) module and a Mixture-of-Prompts enabled Transformer (MoPFormer) module. The former leverages high- and low-energy frequency signals to improve data utilization, while the latter achieves parameter-efficient adaptation through a lightweight prompt-based expert system. Experiments show that SEMPO, with 6.5 million parameters and pre-trained on 83 million data samples, outperforms mainstream models in terms of prediction accuracy under zero-shot and few-shot scenarios, demonstrating its efficiency and generalization capability. The author's response has largely addressed my concerns. Considering the significance of SEMPO for lightweight time series forecasting foundation models, I have raised my score to 5.

格式问题

The authors can refer to the weakness listed above.

作者回复

Thank you very much for the constructive comments. We are grateful for the positive comments on the technical contribution, sound experiments and well-structured presentation. Please see our detailed response below.

Weakness #1 Include full-shot experiments to evaluate model performance.

Please see our detailed response to Reviewer Y38u’s Weakness #1, where we added the full-shot experiments on four representative datasets of different scales from the TSLib benchmark.

Weakness #2 Compare with frequency debiasing model Fredformer.

Thank you for highlighting the debiasing model Fredformer. Following your suggestion, we added a comparison between SEMPO and Fredformer under 10% few-shot and full-shot settings, as summarized in Table A1 below. SEMPO demonstrates strong capabilities, consistently outperforming Fredformer under both settings in both MSE and MAE. While Fredformer shifts attention toward mid- and high-frequency components, it overlooks energy-level bias, as low-energy signals can occur across all frequency bands. In contrast, the EASD module in SEMPO suppresses dominant high-energy components to better capture subtle yet informative low-energy patterns, thereby reducing information loss and improving the forecasting generalization.

Table A1. SEMPO vs. Fredformer with 10% and 100% training data. Results are averaged over horizons of {96,192,336,720}.
DatasetMetricsETTh1ETTh2ETTTm1ETTm2WeatherElectricityTraffic
SEMPO (10%)MSE0.4170.3390.3610.2560.2280.1630.417
MAE0.4360.3870.3860.3150.2660.2610.288
Fredformer (10%)MSE0.5920.3930.4730.2790.2290.1870.428
MAE0.5270.4250.4580.3320.2790.2790.299
SEMPO (100%)MSE0.4190.3400.3560.2500.2240.1720.385
MAE0.4320.3860.3830.3130.2610.2700.261
Fredformer (100%)MSE0.4240.3480.3500.2560.2250.1610.394
MAE0.4350.3900.3790.3170.2650.2560.273

Weakness #3 Overly brief description of the two-stage training paradigm in the main text.

We have explained the details of two-stage training paradigm in Lines 882-902 of Appendix B.2. For clarity, we will move this to the main text in the final version.

Weakness #4 The claimed reduction in pre-training cost needs further verification.

The pre-training data reduction we argue in the paper is mainly focused on the original data size, not considering the data size resulting from data augmentation techniques. However, even we take into account of the data size increment due to the data augmentation, our data reduction argument still holds. Specifically, random masking is a widely adopted strategy in recent time series foundation models (FMs), such as TimesFM [Ref1] and Moment [Ref2], meaning that their pre-training data size is even larger than the one, TimesFM (100B) and Moment (1.13B), we reported in Table 1 if we consider the augmented samples as part of the pre-training data size.

Unlike these FMs that merely apply masking on original samples, SEMPO duplicates each sample NmN_m times across both energy branches, yielding a training size of 83M×Nm×283M \times N_m \times 2. Yet, despite this expansion, the total pre-training data of SEMPO remains substantially smaller than the billion-scale data required by other FMs such as TimesFM (100B) and Moment (1.13B). More importantly, SEMPO requires only 83M time points as original pre-training data, making it well-suited for deployment in resource-constrained environments. The cost introduced by duplication is already reflected in the reported memory usage and computational overhead. Therefore, our claim of reduced pre-training cost remains valid. We will add these discussions in the final version.

Weakness #5 Lack of comparison to TTM in zero-shot and efficiency comparisons, and results of new versions of Chronos and TimesFM.

To clarify, we have provided zero-shot comparison results with the lightweight MLP-based FM TTM in Table 9 in Appendix D.6, as explained in Lines 264-267, since Table 1 focuses on zero-shot results against Transformer-based FMs.

Following your suggestion, we have additionally included an efficiency comparison with TTM on ETTh1 and ETTm2, summarized in Table A2 below. While SEMPO is less competitive in parameter size and inference time due to its more complex Transformer-based architecture, it uses only 83M pre-training data (vs. TTM's 1B) and still achieves lower MSE on both datasets. Moreover, SEMPO outperforms TTM in 31 out of 50 cases across five diverse datasets (see Table 9 in Appendix D.6), with particularly substantial advantages on large-scale datasets such as Traffic and Electricity, demonstrating stronger knowledge transfer capability and superior sample efficiency.

Table A2. Efficiency comparison with TTM on ETTh1 and ETTm2.
DatasetsModelsInference Time (Seconds)Parameters (M)MSE
ETTh1TTM3.210.411
SEMPO226.50.410
ETTm2TTM5.210.288
SEMPO676.50.286

To address your concern regarding the updated FMs, we added zero-shot results for the latest Chronos-Bolt and TimesFM (500M), summarized in Table A3 below for an up-to-date comparison. Across five datasets, Chronos-Bolt achieves the best performance in 16 cases, SEMPO in 12, and TimesFM (500M) in 5. We can observe that SEMPO significantly outperforms TimesFM (500M). It should be noted that, even with the small version of Chronos-Bolt (i.e., Chronos-BoltS_S), it is trained on nearly 100B time points with 48M parameters. In contrast, SEMPO is trained on only 83M points with just 6.5M parameters, yet it still ranks in second in terms of the overall performance. This compelling data efficiency–performance balance highlights SEMPO’s promise as an effective lightweight FM for resource-constrained environments, to which the other FMs such as Chronos-Bolt and TimesFM are infeasible.

Table A3. Zero-shot comparison on five datasets with Chronos-Bolt and TimesFM (500M).
DatasetHorizonSEMPOChronos-BoltS_SChronos-BoltL_LTimesFM (500M)
MSEMAEMSEMAEMSEMAEMSEMAE
ETTh11920.4090.4260.4510.4150.4350.4120.4280.409
3360.4170.4330.4970.4370.4760.4330.4510.424
7200.4320.4540.4990.4560.4800.4490.4630.447
ETTh21920.3340.3840.3510.3710.3550.3710.3700.385
3360.3550.4030.3960.4050.4000.4030.4160.420
7200.3950.4350.4130.4290.4140.4220.4430.448
ETTm11920.4840.4550.3820.3660.3810.3660.3880.376
3360.5060.4690.4330.3960.4320.3930.4260.402
7200.5570.4980.5300.4470.5250.4390.4910.444
ETTm21920.2520.3230.2390.2910.2470.2950.2680.307
3360.3050.3540.3060.3340.3140.3360.3270.347
7200.3910.4040.4130.3980.4150.3950.4210.407
Weather1920.2160.2690.2180.2500.2220.258--
3360.2670.3040.2750.2900.2820.298--
7200.3360.3500.3540.3380.3660.348--
1st^{st} Count12165

Weakness #6 The visualization of ZLecZ_{Lec} in Figure 2 appears to contain an error.

We have carefully checked Figure 2, and confirm that ZLecZ_{Lec} and ZHecZ_{Hec} are indeed complementary, with ZLec=ZZHecZ_{Lec}=Z-Z_{Hec}.

Weakness #7 In Figure 4(b), the low-energy feature response of Moirai seems to reflect prediction inaccuracies rather than being completely neglected.

Yes, your observation is correct. As shown by the spectral discrepancy between the prediction and ground truth, MoiraiL_L exhibits non-negligible responses in low-energy regions, in contrast to ChronosS_S, which largely neglects low-energy components. However, these responses in MoiraiL_L do not consistently align with the spectral peaks of the ground truth. This indicates that while MoiraiL_L is able to partially capture low-energy signals, it still predominantly attends to high-energy components, resulting in less accurate learning of low-energy features, such as the important positional information. We will clarify this in the final version.

References:

  • [Ref1] A Decoder-only Foundation Model for Time-Series Forecasting, ICML 2024

  • [Ref2] MOMENT: A Family of Open Time-series Foundation Models, ICML 2024

评论

Thanks for the your reply. This has largely addressed my concerns. Considering the significance of SEMPO for lightweight time series forecasting foundation models, I have raised my score to 5.

评论

Dear Reviewer EgfG,

Thank you very much for your appreciation of SEMPO and for raising your recommendation score. We're very pleased that our response has largely addressed your concerns. Please kindly let us know if there are any further questions. Thanks again!

Best regards,

Authors

审稿意见
5

The paper introduces SEMPO, a lightweight foundation model for general-purpose time series forecasting that achieves strong performance while using fewer parameters and less pre-training data than existing large-scale models. SEMPO integrates two key features: (1) an Energy-Aware Spectral Decomposition (EASD) module that transforms time series into the frequency domain and models both high- and low-energy frequency signals, improving utilization of pre-training data; and (2) a Mixture-of-Prompts-enabled Transformer (MoPFormer) that uses dataset-specific prompts to capture heterogeneous temporal patterns via a lightweight routing mechanism. SEMPO achieves zero-shot and few-shot forecasting on various datasets using less parameters and pre-training data points. Experiments show that it outperforms several LLM-based models.

优缺点分析

Strengths:

S1. SEMPO achieves state-of-the-art results in zero-shot and few-shot time series forecasting across 16 datasets, despite using only 6.5M parameters and a modest 83M time points for pre-training, which are substantially smaller than competing models that use hundreds of millions of parameters and billions of data points.

S2. The proposed EASD module effectively captures both high- and low-energy frequency components of time series.

S3. The paper provides extensive experiments on benchmark time series datasets covering various domains. It includes zero-shot, few-shot, ablation, and efficiency analyses that can demonstrate the superiority of SEMPO.


Weaknesses:

W1. Some claims are not sufficiently justified. One notable example is the claim that existing time series FMs suffer from energy bias and fail to effectively handle low-energy frequency components is primarily supported by a case study. A more rigorous justification for this limitation by other models would be necessary to convincingly support this claim.

W2. The presentation of SEMPO’s data efficiency is unclear. While the model is positioned as requiring less pre-training data, the paper does not clearly quantify the significance of this reduction compared to baselines. It is unclear how many data points are utilized in other models. Also, the epochs for pre-training play an important role in its real data utilization.

W3. There is a lack of clarity regarding the relationship between the pre-training and inference stages. There’s a gap between training and inference stage. It is better to illustrate the process of this two stages in details.

W4. The analysis of low-energy components in the experiments is weak. The paper includes only a qualitative case study to demonstrate SEMPO’s effectiveness in modeling such signals. However, a more comprehensive quantitative evaluation is preferred to be provided.

问题

Q1. The claimed advantage of using less pre-training data is not convincingly demonstrated. The paper does not clearly specify how the number of datasets or total data volume used for pre-training compares to other models. Additionally, the number of pre-training epochs significantly affects the effective data utilization and should be reported to ensure a fair and meaningful comparison.

Q2. MoE architectures typically rely on sparse routing to activate only a subset of expert networks, while maintaining diversity among the experts is essential for generalization. It remains unclear how SEMPO addresses these challenges, particularly how it ensures both routing efficiency and functional diversity in the proposed prompt-based expert system.

Q3. Please respond to the other concerns raised in W1-W4.

局限性

Yes

最终评判理由

The authors have sucessfully addressed my concerns. I would adjust my score accordingly

格式问题

There are no formatting concerns in this paper.

作者回复

Thank you very much for the constructive and positive comments on our technical contribution and empirical justification. Please see our response to your comments one by one below.

Weakness #1 Lack of sufficient justification for the energy bias issue.

Thank you very much for the suggestion. We’d like to clarify that the claim regarding energy bias has been widely justified in the task-specific forecasting domain in the literature [Refs1-3], and we explained the basis of this claim in Lines 145-148. This phenomenon reflects a learning bias in Transformers, where the self-attention tends to prioritize low-energy components over low-energy ones. Such limitation has been rigorously supported by both empirical and theoretical studies. For example, Fredformer [Ref1] and FilterNet [Ref2] provide comprehensive empirical evidence showing that the dominance of low-frequency components, due to their substantial spectral energy, can lead to frequency bias, as Transformer-based forecasting models tend to disproportionately attend to these signals during learning; Amplifier [Ref3] complements this by offering a theoretical justification that frequency components with low spectral energy contribute minimally to the training loss and receive slower parameter updates compared to high-energy components.

Building upon these previous findings, and given that most existing foundation models (FMs) [Refs4-6] are built upon Transformer architectures, we empirically investigate whether energy bias persists in general-purpose forecasting domain. Our experiments on ETTh1 and Electricity datasets confirm its presence in foundation models, such as ChronosS_S [Ref6] and MoiraiL_L [Ref5], as illustrated in Figures 1, 4. To further address your concern regarding this claim, we have added additional quantitative results in our response to Weakness #4, showing that other FMs like TimesFM and Timer also struggle to effectively capture low-energy frequency components.

Together, prior studies and our quantitative and qualitative results provide a rigorous justification for our claim that existing FMs exhibit energy bias.

Weakness #2 and Question #1 Unclear presentation of SEMPO’s data efficiency, such as pre-training data size and epochs.

We’d like to clarify that both the model size and pre-training data size of SEMPO and all baseline FMs have been reported in Table 1 in main text and Table 6 in Appendix C.2 (in the (,)(\cdot,\cdot) values just below the names of the models). SEMPO achieves strong zero-shot forecasting performance with just 83M time points, underscoring its remarkable data efficiency compared to baselines trained on data with hundreds of millions of time points to billions (or even hundreds of billions) of time points.

Additionally, the number of pre-training epochs have been provided in Lines 1010-1012 in Appendix C.3: in SEMPO, the energy-aware pre-training is conducted for 10 epochs, followed by 20 epochs of MoP tuning. To enable a straightforward comparison, Table A1 summarizes the pre-training epochs and data sizes of SEMPO and other four strong FMs. Notably, SEMPO requires only 83M time points, with 30 epochs of training, achieving a substantial reduction in training data and demonstrating superior data efficiency compared to baseline FMs trained on billions of time points.

Table. A1. Pre-training epochs of SEMPO and four representative FMs.
ModelsSEMPOTTMTimerChronosMoment
Data size83M1B28B84B1.13B
Epoch3020102002

To further validate SEMPO’s data efficiency, we extended the pre-training by 5 additional epochs in either the pre-training or the tuning stage, with the results presented in Table A2 below. TTM with 20 training epochs is used as baseline. SEMPO reaches optimal performance within 30 total epochs under 83M time points, indicating that SEMPO has effectively converged and fully utilized the pre-training data.

Table. A2. Zero-shot performance with increased pre-training epochs, averaged over horizons of {96,192,336,720}.
DatasetsETTh2ETTm2ElectricityTraffic
MetricsMSEMAEMSEMAEMSEMAEMSEMAE
SEMPO (10,20)0.3420.3910.2860.3420.1970.2950.4670.344
SEMPO (15,20)0.3480.3960.2880.3400.2060.3040.4970.363
SEMPO (10,25)0.3490.3980.2870.3400.2020.3000.4830.353
TTM (20)0.3570.3960.2880.3300.2150.3030.5590.365

Weakness #3 Unclear relation between pre-training and inference stages.

We’d like to clarify that we’ve provided the relationships/differences for the energy-aware pre-training stage, the MoP Tuning stage, and the inference stage in Figure 2 (Details) in Appendix B.
As mentioned in Lines 867–881 of Appendix B.1, the energy-aware pre-training does not include the MoP module, which is introduced and used only in the MoP tuning.

As also discussed in Lines 886–902 of Appendix B.2, after pre-training, the model enters the fine-tuning phase, where the model is trained on the target dataset using the same joint optimization strategy as in the MoP tuning. In this stage, only the parameters in the mixture of prompts and prediction heads are updated, while the Transformer backbone remains frozen. During zero-shot inference, the reconstruction decoder and head are removed, and there are no parameter updates in all modules. We will clarify these relationships and differences in the final version.

Weakness #4 Weak analysis of low-energy components.

We appreciate your constructive suggestion. We’d like to clarify that our analysis of low-energy components is not limited to the qualitative case study in Figure 4. In fact, we’ve also investigated SEMPO’s capability to model low-energy signals across diverse prediction lengths, with the results shown in Figure 7 in Appendix D.1.

In terms of quantitative analysis, we conducted ablation studies comparing SEMPO with its variant A.1 Multi-band Masking, which removes the energy-based partitioning mechanism (i.e., a variant that also ignores the energy bias, as what is done in existing FMs). The results across all datasets and forecasting horizons in Table 3 demonstrate the superior performance of SEMPO over this variant, in which modeling low-energy components is the only difference between SEMPO and this variant. The same observation can be found in the full ablation study results in Table 13 in Appendix D.6.

While these results provide meaningful evidence, we agree that more quantitative evaluations would strengthen the argument. In response, we justified the importance of low-energy modeling from another perspective, i.e., whether including such a component in the pre-training could improve the performance of existing FMs. The results of SEMPO and three strong FMs under two settings: with and without low-energy modeling are given in Table A3 below. Note that as the energy distribution varies across time windows and the distinction between high- and low-energy components is inherently relative, applying a unified threshold for statistical evaluation across the entire dataset is infeasible. We therefore randomly selected five batch samples from Traffic and ETTm1 and report their averaged results in Table A3.

Table A3. Impact of modeling low-energy components (LEC) on MAE across different FMs.
DatasetsLEC SettingChronosL_LTimesFMTimerSEMPO
Trafficw/o LEC0.381-0.3910.304
with LEC0.341-0.3570.256
Improvement10.5%-8.7%15.8%
ETTm1w/o LEC0.5020.4500.4910.472
with LEC0.4550.3940.4340.409
Improvement9.4%12.4%9.6%13.3%

Two key observations can be found in Table A3. 1) Adding the low-energy modeling consistently results in very significantly decreased MAE across all FMs, highlighting the positive role these components play in enhancing the prediction accuracy; 2) SEMPO achieves the largest performance gains from incorporating low-energy components, with a 13.3% improvement on ETTm1 and 15.8% on Traffic, indicating its superior ability to effectively mitigate energy bias and model low-energy signals. We will add this quantitative evaluation to the final version.

Question #2 How it ensures routing efficiency and functional diversity?

Unlike traditional MoE that requires sparse activation over large expert networks to ensure routing efficiency and reduce computation, SEMPO adopts lightweight, learnable prompt vectors as experts. Based on these lightweight experts, SEMPO applies soft gating via a simple linear-softmax function over input tokens, enabling efficient expert selection without the need for sparse activation. These prompts are prepended to the key-value pairs in self-attention, allowing all tokens to share the same Transformer backbone. This avoids multi-branch execution, resulting in minimal computational overhead only, and well supports parameter-efficient adaptation.

To ensure functional diversity, each prompt-based expert is tailored to a specific small dataset, independently learned, and assigned input-adaptive weights via soft gating. A reparameterization strategy further transforms each expert into structured key-value pairs through a two-layer MLP, enhancing its expressiveness and promoting diverse functional roles across the expert set.

References:

  • [Ref1] Fredformer: Frequency Debiased Transformer for Time Series Forecasting, KDD 2024
  • [Ref2] FilterNet: Harnessing Frequency Filters for Time Series Forecasting, NeurIPS 2024.
  • [Ref3] Amplifier: Bringing Attention to Neglected Low-Energy Components in Time Series Forecasting, AAAI 2025.
  • [Ref4] Timer: Generative Pre-trained Transformers Are Large Time Series Models, ICML 2024
  • [Ref5] Unified training of universal time series forecasting transformers, ICML 2024
  • [Ref6] Chronos: Learning the language of time series, TMLR 2024
评论

The authors have addressed my concerns. I will adjust my score accordingly.

评论

Dear Reviewer uD87,

We sincerely appreciate your consideration of our responses and the adjustment to your recommendation score. We are very pleased that our reply has addressed your concerns. If you have any further questions or concerns, please feel free to let us know.

Best regards,

Authors

审稿意见
5

This paper tackles the time series forecasting task by utilising a lightweight foundation model, involving 2 key modules, one decomposes the time series in the spectral domain by thresholding the frequencies and masking the frequencies randomly; and one mixture of prompts block. It is claimed that with such modules the model improves the time series forecasting performance in MSE, MAE or sMAPE.

优缺点分析

  • Strength

    The paper is well written and easy to follow. The approach is clearly described, and the experimental setup and the results are well illustrated.

    The proposed approach is to the best of my knowledge novel and sound.

  • Weakness

    It seems the results in the experiments are from only one run. Since the masks are generated randomly, it would be more convincing if multiple runs are conducted and the standard deviations are reported.

    It is claimed in the checklist 7, but I failed to see a discussion about the statistical test in the paper.

问题

  • What do Hec and Lec mean? What would you expect, if the energy-wise spectral partitioning leads to more than 2 partitions?

  • In figure 4, what is "input", and what is the point of plotting "input" in the figure?

  • In figure 8a in Appendix, could you explain the peak at N_m=5? Does it make sense to show the cases with N_m > 6?

局限性

Although claimed in the checklist, I failed to find a proper discussion of limitations in the conclusion.

最终评判理由

My questions are mostly answered.

格式问题

I don't see any formatting issue.

作者回复

Thank you very much for the constructive comments. We are grateful for the positive comments on our methodological innovation, paper clarity and empirical justification. Please see our response to your comments one by one below.

Weakness #1 It seems the results in the experiments are from only one run. Since the masks are generated randomly, it would be more convincing if multiple runs are conducted and the standard deviations are reported

All experimental results are obtained by averaging over three runs with different random seeds. In deep time series forecasting, performance variance across seeds is typically small, and following prior work [Refs1-3], we report only the mean values. To address your concern regarding random sampling, we additionally report the standard deviation of the 5% few-shot experiments to demonstrate the stability of our method in Table A1 below.

Table. A1. Statistical results (mean ± std) of SEMPO and TTM under the 5% few-shot setting. The horizon length H is all set to 336.
ModelsSEMPOTTM
MetricsMSEMAEMSEMAE
ETTh10.425 ±\pm 0.00060.439 ±\pm 0.00000.400 ±\pm 0.00100.421 ±\pm 0.0006
ETTh20.355 ±\pm 0.00050.402 ±\pm 0.00000.383 ±\pm 0.00120.415 ±\pm 0.0008
ETTm10.376 ±\pm 0.00120.393 ±\pm 0.00050.389 ±\pm 0.00050.396 ±\pm 0.0008
ETTm20.274 ±\pm 0.00000.328 ±\pm 0.00000.319 ±\pm 0.00080.347 ±\pm 0.0000
Weather0.249 ±\pm 0.00050.285 ±\pm 0.00000.255 ±\pm 0.00000.292 ±\pm 0.0000
Electricity0.168 ±\pm 0.00000.267 ±\pm 0.00000.185 ±\pm 0.00050.282 ±\pm 0.0005
Traffic0.421 ±\pm 0.00080.292 ±\pm 0.00000.446 ±\pm 0.00000.315 ±\pm 0.0000

The results show that both SEMPO and TTM exhibit stable performance across different seeds. Furthermore, SEMPO demonstrates its superiority by outperforming TTM in both stability and predictive accuracy in 12 out of 14 cases.

Question #1 What do Hec and Lec mean? What would you expect, if the energy-wise spectral partitioning leads to more than 2 partitions?

Hec refers to high-energy components, and Lec refers to low-energy components. If energy-wise spectral partitioning yields more than two groups, it effectively creates multiple patches along the energy axis. This is similar to the frequency patching technique. The latter offers finer granularity to capture energy patterns but introduces significant computational overhead during masking.

Question #2 In figure 4, what is "input", and what is the point of plotting "input" in the figure?

In Figure 4, "input" refers to the spectrum of the lookback window X1:LX_{1:L}. In time series forecasting, it is a common practice to include the lookback window in both time- and frequency-domain visualizations to present a complete forecasting sample, see similar examples in [Ref1, Ref4]. More importantly, it provides essential contextual information that helps assess whether the model's predicted spectrum preserves the spectral structure of the input and whether the model tends to emphasize certain frequency components.

Question #3 In figure 8a in Appendix, could you explain the peak at Nm=5N_m=5? Does it make sense to show the cases with Nm>6N_m>6?

In principle, increasing NmN_m provides the model with more complementary frequency perspectives via multiple random samplings, facilitating the learning of diverse spectral features. However, the spike observed at Nm=5N_m=5 is likely caused by stochastic noise rather than a stable trend. Such fluctuations may arise as redundant or conflicting frequency views introduce noise or disrupt convergence, ultimately leading to performance degradation.

Unfortunately, it does not make sense to show cases with Nm>6N_m>6. This is because, as shown in Figure 8(a), performance does not consistently improve beyond Nm=4N_m=4, and it may even degrade due to increased randomness in masking and potential over-fragmentation of frequency views. Meanwhile, a larger NmN_m imposes considerable memory and computational overhead, as frequency masking is applied separately to NmN_m duplicated inputs in both high- and low-energy branches. Considering the trade-off between performance and efficiency, we choose not to explore Nm>6N_m>6.

Limitation Although claimed in the checklist, I failed to find a proper discussion of limitations in the conclusion.

We briefly touched on this aspect in Lines 360-362, where we discussed future directions that reflect the current limitations of SEMPO. Specifically, one limitation is that the channel-independent mechanism may hinder the modeling of critical multivariate interactions. Another is that supporting only point forecasting reduces the SEMPO’s flexibility in capturing uncertainty through probabilistic forecasting. These limitations may constrain the SEMPO’s ability to capture richer time series characteristics.

We will add all the above empirical results and clarifications in the final version and hope we have addressed your concerns.

References:

  • [Ref1] Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts, ICLR 2025

  • [Ref2] Tiny Time Mixers (TTMs): Fast Pre-trained Models for Enhanced Zero/Few-shot Forecasting of Multivariate Time Series, NeurIPS 2024

  • [Ref3] Timer: Generative Pre-trained Transformers Are Large Time Series Models, ICML 2024

  • [Ref4] Fredformer: Frequency Debiased Transformer for Time Series Forecasting, KDD 2024

评论

Thanks to the authors for the answer.

Regarding weakness 1, is it stated anywhere in the paper averaging over three runs? Because I failed to find it.

And thanks for the table A1. One follow-up question, is it conducted with the same setting as table 2 in the paper? Why the average values are different from table 2?

评论

Dear Reviewer 3Jvu,

We greatly appreciate your continuous valuable feedback. Please find our response to your remaining concerns below.

Follow-up question #1 Regarding weakness 1, is it stated anywhere in the paper averaging over three runs? Because I failed to find it.

This information was indeed not stated in the paper. To address your concern, we will clearly clarify this point in the final version.

Follow-up question #2 And thanks for the table A1. One follow-up question, is it conducted with the same setting as table 2 in the paper? Why the average values are different from table 2?

The settings for the two set of results are different. Table 2 reports results averaged over four forecasting horizons H96,192,336,720H \in \\{96,192,336,720\\}, while Table A1 presents results averaged over different random seeds under a single horizon H=336H=336. The values in Table A1 are consistent with those in the full version of Table 2, namely Table 11 in Appendix D.6, at the same horizon.

We sincerely hope the above replies help address your follow-up questions. If you have any further questions or concerns, please feel free to let us know.

Best regards,

Authors

审稿意见
4

The paper proposes SEMPO, a lightweight foundation model designed for general time series forecasting that significantly reduces both model size and pre-training data requirements, while still achieving state-of-the-art performance in zero-shot and few-shot scenarios.

优缺点分析

Strength

  1. The paper is well-structured, with clear organization and a logical flow that makes it easy to follow.

  2. Contribution: The paper explores an efficient way to address time series forecasting through Energy-aware Spectral Decomposition and Mixture-of-Prompts Transformer. Extensive experiment results demonstrate that SEMPO is able to outperform baselines with only 6.5M parameters which is pre-trained on 83 million time points

Weakness

  1. The core contribution of SEMPO is to present a lightweight yet effective foundation model for time series forecasting. However, the experimental evaluation primarily focuses on zero-shot and few-shot settings, where many of the baseline methods were not originally designed to perform. This may limit the fairness and comprehensiveness of the comparison. It would strengthen the empirical analysis if the authors could additionally include in-distribution evaluations—similar to what was done in Time-MoE—to demonstrate that SEMPO also achieves competitive or superior performance in standard supervised settings, despite its reduced model size.

  2. In lines 51–53, the authors state: “Specifically, modeling full energy spectrum of temporal signals from pre-training data is essential for learning a comprehensive knowledge base of temporal patterns to underpin strong transferability across different datasets.” This is an important claim, but it lacks empirical evidence or citations to support it.

  3. While the proposed Mixture-of-Prompts (MoP) mechanism contributes to model adaptability, the overall technical novelty of SEMPO is limited. The use of FFT for spectral decomposition is a well-established technique in time series analysis, and the design of the model architecture—particularly the encoder-decoder structure and masking strategies—follows relatively standard practices. As a result, the contribution of the model as a whole may not be sufficient for a top-tier venue.

  4. There are several formatting issues in the current draft, such as broken links to corresponding figures and missing or improperly linked citations. These issues hinder readability and should be addressed to improve the overall presentation quality.

问题

In lines 126–130, the paper introduces the concepts of DtuneD_tune and DtestD_test in the context of few-shot and zero-shot evaluation. However, it is unclear whether these datasets are drawn from entirely new domains or from domains already included in the pre-training set TtrainT_train. Please clarify this distinction to help assess the generalization capabilities of the proposed model more accurately.

In Figure 1 (right), SEMPO (Zero) and SEMPO (Few) are marked separately, while other models such as PatchTST and Time-LLM are not differentiated by evaluation setting. This inconsistency may lead to confusion when interpreting the results. It would be clearer to present two separate figures: one for zero-shot comparisons and another for few-shot comparisons, ensuring a fair and consistent visual comparison across models.

局限性

yes

最终评判理由

I would like to raise my score as my concerns are mostly addressed, although I am still have reservation on novelty.

格式问题

NA

作者回复

We sincerely appreciate your constructive and positive comments on the clear presentation and technical contribution. Please see our response to your comments one by one below.

Weakness #1 Lack of in-distribution evaluations similar to Time-MoE

Thank you very much for the insightful suggestion. To ensure fairness and enhance the comprehensiveness of our evaluation, we have added in-distribution experiments under the standard supervised setting adopted by Time-MoE. Specifically, we compare SEMPO against LLM-based time series foundation models (FMs) (S2IP-LLM and GPT4TS) as well as strong task-specific models (iTransformer, DLinear, PatchTST and TimesNet). Results on four datasets of different scales are presented in Table A1 below:

Table A1. In-domain forecasting results on four datasets. 
DatasetHorizonSEMPOGPT4TSS2PLLMiTransformerDLinearPatchTSTTimesNet
MSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETTh2960.2730.3340.2850.3420.2780.3400.2970.3480.3020.3680.2740.3370.3510.399
1920.3330.3760.3540.3890.2460.3850.3710.4030.4040.4330.3410.3820.3940.429
3360.3550.4000.3730.4070.3670.4060.4040.4280.5110.4980.3290.3840.4150.443
7200.3990.4350.4060.4410.4000.4360.4240.4440.8150.6400.3790.4220.4770.481
ETTm2960.1600.2510.1730.2620.1650.2570.1750.2660.1640.2550.1660.2560.2330.305
1920.2210.2940.2290.3010.2220.2990.2420.3120.2240.3040.2230.2960.2650.328
3360.2730.3280.2860.3410.2770.3300.2820.3400.2770.3390.2740.3290.3790.392
7200.3490.3800.3780.4010.3630.3900.3780.3980.3710.4010.3620.3850.3900.407
Weather960.1430.1930.1620.2120.1450.1950.1590.2080.1700.2300.1490.1980.1930.244
1920.1890.2340.2040.2480.1900.2350.2000.2480.2120.2670.1940.2410.3200.329
3360.2400.2800.2540.2860.2430.2800.2530.2890.2570.3050.2450.2820.3630.366
7200.3250.3380.3260.3370.3120.3260.3210.3380.3180.3560.3140.3340.4400.404
Traffic960.3550.2460.3880.2820.3790.2740.3630.2650.4110.2940.3600.2490.6110.323
1920.3730.2530.4070.2900.3970.2820.3850.2730.4210.2980.3790.2560.6090.327
3360.3840.2600.4120.2940.4070.2890.3960.2770.4310.3040.3920.2640.6160.335
7200.4270.2860.4500.3120.4400.3010.4450.3120.4680.3250.4320.2860.6560.349
1st^{st} Count25030050

The results show that SEMPO consistently outperforms six advanced time series models, demonstrating that its pre-training framework effectively equips the model with strong inductive biases for downstream adaptation, despite its lightweight design. These results will be included in the final version to address this concern.

Weakness #2 Lack empirical evidence or citations to support claim on ''modeling full energy spectrum of temporal signals from pre-training data...''

Low-energy frequency components represent subtle yet informative short-term variations, such as periodicities over short durations, which serve as good indicators for forecasting. Prior studies [Refs1-3] have revealed that attending to these low-frequency components with substantial spectral energy and mitigating this energy bias are critical for enhancing performance in task-specific forecasting domain. In particular, Fredformer [Ref1] and FilterNet [Ref2] provide compelling empirical evidence that the dominance of low-frequency components with substantial spectral energy can create a bottleneck to full-spectrum information utilization, ultimately hindering forecasting performance.

Building upon this, multi-source time series pre-training can be viewed as an aggregation of learning processes across multiple individual datasets. If energy bias occurs within each dataset, it can compound during pre-training, leading to irreversible information loss. Therefore, we extend these prior findings to general-purpose forecasting domain, arguing that modeling the full energy spectrum is also essential for learning comprehensive forecasting patterns from a large collection of pre-training datasets to underpin strong generalization across datasets.

To further address your concern regarding empirical evidence, we conducted experiments comparing SEMPO with three strong FMs (ChronosL_L, TimesFM, and Timer) under two settings: with and without low-energy components, to highlight the indispensable role of low-energy pattern modeling, please see our response to Reviewer uD87’s Weakness #4 for the detailed results and discussion. We will include the references and empirical evidence in the final version.

Weakness #3 The use of FFT and the encoder-decoder architecture with masking strategies appears standard and lacks novelty

FFT-based spectral decomposition is indeed widely adopted in time series analysis, such as in [Refs1-4]. However, it is important to clarify that these prior works primarily address the learning bias between high- and low-frequency components, a phenomenon characterized by the prioritization of low-frequency features at the expense of high-frequency features. In contrast, our work goes a step further by empirically revealing that existing time series FMs [Refs5-6] are generally biased toward high-energy components, while neglecting low-energy components that may distribute across mid-, high-, or even low-frequency bands (see Figures 1, 4, 7). The coarse-grained frequency bias correction is insufficient in this context, as it can lead to significant information loss of low-energy frequency information. To address this, we propose the novel EASD module, which performs adaptive spectral partitioning and dual-branch spectral masking to model low-energy yet informative frequency signals that are ignored in current FMs.

Additionally, most existing FMs adopt Transformer-based architectures, such as Moirai [Ref5] and Chronos [Ref6]. Leveraging a well-established encoder-decoder framework and following standard design practices does not diminish the novelty of our work in the two proposed modules. Instead, it enables us to clearly identify the above critical limitation in current Transformer-based FMs, namely energy bias. Building upon this mature architecture, we further propose the novel MoPFormer module, which enables parameter-efficient model adaptation across different datasets and domains. Together, these two innovative modules enable SEMPO to substantially reduce both the pre-training data scale and model size, while achieving strong generalization.

Weakness #4 Several formatting issues in the current draft

Thank you for pointing this out. The broken links may be due to the separation between the main text and appendix. We will address the formatting issues, including figure links and citations, in the final version to ensure preferred readability.

Question #1 Clarify whether DtuneD_{tune} and DtestD_{test} are drawn from new domains or overlap with the pre-training set TtrainT_{train}

Following standard few-shot and zero-shot settings, DtuneD_{tune} and DtestD_{test} are drawn from entirely new domains that are not presented in TtrainT_{train}. We will clarify this point in our final version.

Question #2 Fair and consistent visual comparison in Figure 1 (right)

We would like to clarify that Time-LLM, GPT4TS, and PatchTST are used only in the few-shot setting, while other FMs are evaluated in the zero-shot setting (see Table 1 and Table 2). Accordingly, we distinguish SEMPO (Zero) and SEMPO (Few) for fair comparison in each setting.

To address your concern regarding consistency, we will revise the figure and split it into two subplots for zero- and few-shot settings respectively in the final version.

References:

  • [Ref1] Fredformer: Frequency Debiased Transformer for Time Series Forecasting, KDD 2024

  • [Ref2] CATCH: Channel-aware Multivariate Time Series Anomaly Detection via Frequency Patching, ICLR 2025

  • [Ref3] Towards a General Time Series Forecasting Model with Unified Representation and Adaptive Transfer, ICML 2025

  • [Ref4] FilterNet: Harnessing Frequency Filters for Time Series Forecasting, NeurIPS 2024

  • [Ref5] Unified training of universal time series forecasting transformers, ICML 2024

  • [Ref6] Chronos: Learning the language of time series, TMLR 2024

评论

Dear Reviewer Y38u,

Thank you very much for going through our replies. May we check whether our responses have addressed your concerns? If you have any further questions or concerns, please feel free to let us know.

Best regards,

Authors

评论

Dear Reviewer Y38u,

As the End of the author-reviewer discussion period approaches, we would like to kindly check if our responses have addressed your concerns. Your feedback is truly valuable to us, and we would be more than happy to engage in further discussion and paper improvements. Thank you again for your time and effort in helping us improve our paper!

Best regards,

Authors

最终决定

This paper proposes a new lightweight time series foundation model with strong experimental results across 16 datasets in 0 and few short forecasting tasks. The reviewers mostly found the proposed approach to be novel and the experiments to be good although (prior to the author/reviewer discussion completing) not thorough enough. However, after the author/reviewer discussion, the bulk of reviewer concerns were addressed, and the reviewers unanimously favored acceptance for this paper (3 accept, 1 borderline accept). I am recommending acceptance for this paper as a result.