PaperHub
7.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
4
5
5
3.5
置信度
创新性2.8
质量3.3
清晰度2.8
重要性3.0
NeurIPS 2025

Synthetic Series-Symbol Data Generation for Time Series Foundation Models

OpenReviewPDF
提交: 2025-05-01更新: 2025-10-29
TL;DR

Based on the complex dynamics perspective, this paper proposes a bimodal time series generation mechanism to mitigate the current data shortage problem in time series analysis.

摘要

关键词
Time Series AnalysisFoundation ModelData Scarcity

评审与讨论

审稿意见
5

This paper introduces a novel mechanism for generating unlimited synthetic time series data paired with their underlying symbolic (mathematical) expressions—based on complex dynamical systems theory—to overcome real-world data scarcity and imbalance. Leveraging this dual-modality dataset, the authors pretrain SymTime, a foundation model that fuses masked time-series and symbolic-language modeling with contrastive learning, achieving state-of-the-art performance across five major time-series analysis tasks.

优缺点分析

Strengths

  • I very much like the careful design of the S² generator and the theoretical background provided around Takens’ theorem.
  • The idea of synthetically generating time-series seems with their proposed approach seems to be highly effective.
  • I like the broad experimentation carried out by the authors, and the consistent gains that their method seems to achieve.
  • The ablations performed by the authors support the strength of each of the changes introduced.

Weaknesses

  • If I understood correctly, there is no comparison against real-data pre-training with the same architecture. This would have been a valuable addition to the paper.
  • There are no standard deviation results in the tables presented.

问题

  • Would the synthetic data generation process be able to reproduce phenomena observed in real-world domain specific time-series datasets, e.g. volatility clustering in finance?

局限性

Yes

最终评判理由

The authors addressed my concerns and added new experiments. My initial review was positive and I raise my confidence now.

格式问题

None

作者回复

We sincerely thank you for taking the time to review our paper and your constructive comments. We appreciate your recognition of our mechanism. Although deep learning technology has made great progress in time series analysis, we still hope to return to the essence of time series and introduce the complex dynamical system modeling perspective in machine learning to solve the current problem like data imbalance.


We will respond to your questions one by one below:

W1: If I understood correctly, there is no comparison against real-data pre-training with the same architecture. This would have been a valuable addition to the paper.

Thank you for the question. The paper's core is the Taken-theorem-based S2S^2 generator, which produces unlimited time-series and symbolic data. SymTime is designed to learn both modalities jointly, leveraging symbolic representations to boost time series modeling—this benefit is verified in our ablation. However, real-world data lack symbolic expressions, so we can not pre-train SymTime with only time series data through Equation (6).

But in order to conduct fair and effective comparative experiments, we adopt the configuration of the ablation experiment in Section 4.4, randomly and uniformly selecting 1/61/6 (we use 50B time series points data in S2S^2 for pre-training) of real time series points from the Time300B dataset (Time-MoE), pre-training the model using Equation (2) as the target (only using mask time series modeling, not considering symbolic expressions), and fine-tuning on the ETTh1 and ETTh2 datasets. The experimental results (mean (std)) are shown below:

TypeSymTimeReal-Dataw/o Symbol
MetricMSEMAEMSEMAEMSEMAE
ETTh10.442(0.008)0.448(0.006)0.446(0.012)0.453(0.007)0.450(0.0010)0.454(0.006)
ETTh20.382(0.007)0.425(0.005)0.389(0.009)0.430(0.005)0.401(0.009)0.432(0.006)

We conduct multiple experiments by switching random seeds, with the results for each experiment being the average of forecasting lengths {96, 192, 336, 720}. The results show that pre-trained SymTime on real data outperforms models pre-trained solely on time series data from the S2S^2 dataset. However, it underperforms models pre-trained using both series and symbolic data. This result further demonstrates the relationship between symbolic expression data and time series, and that this pairing effectively enhances SymTime's performance. We did not intentionally exclude samples from the ETT dataset when sampling from Time300B, so our masked time series modeling may include our test samples. However, the performance of SymTime pre-trained on the synthetic S2S^2 dataset is not inferior to that of real data.

We will add these experimental results to the ablation experiments in the main text (Section 4.4) and provide a detailed description of our experimental setup.

W2: There are no standard deviation results in the tables presented.

Following TimesNet's benchmark across five tasks, we omitted SDs for brevity but reported their ranges in each table header (see Tables 17–26 in Appendix). For example, Table 17 notes SD ≤ 0.5 % for long-term forecasting.

We will highlight this note more prominently in the main text.

Q: Would the synthetic data generation process be able to reproduce phenomena observed in real-world domain specific time-series datasets, e.g. volatility clustering in finance?

Thank you for the insightful question and examples. The diversity of our S2S^2 generator stems from (1) the unrestricted variety of symbolic systems f()f(\cdot) and (2) the diversity of sampled series XX. Hence, S2S^2 can in principle generate any real-world domain. Section 4.1 empirically confirms this.

Financial series, as highly chaotic systems, exhibit strong heteroskedasticity and volatility clustering, typically modeled by ARCH or GARCH. Searching S2S^2, we find an f()f(\cdot) that reproduces this behavior:

y=xexp ⁣(sin2(x1)2),y = x \cdot \exp\!\left(\frac{\sin^{2}(x-1)}{2}\right),

with xN(0,1)x \sim \mathcal N(0,1) or any stationary input (in S2S^2 we can generate this input sampling series). The resulting yy shows clear volatility clustering akin to ARCH(1)/GARCH(1,1): large shocks are followed by calm periods.

The effect can be observed with the short Python code below.

import numpy as np
from matplotlib import pyplot as plt

np.random.seed(42)
x = np.random.normal(0, 1, 251)

fig, ax = plt.subplots(2, 1, figsize=(12, 8))
y = x * np.exp((np.sin((x - 1) ** 2)) / 2)

ax[0].plot(y, color="royalblue")
ax[0].set_title("Financial Time Series with Volatility Clustering")
ax[1].plot(y[:-1] ** 2, "tomato")
ax[1].set_title("Conditional Variance (Shows Clustering)")

By observing this visualization, we can clearly see the small fluctuations after the large fluctuations in the time domain. This is well reflected in the real-time changes in the variance of the series below, where large and small variances appear alternately. Furthermore, real financial time series may contain trend information, and the complex system generated by our S2S^2 dataset is more complex than the above formula. It may contain expected trend information (derived from the autoregressive model in our sampled series). See the visualization in Section B.2 for specific possible representations.

The series is particularly similar to the simulated ARCH(1) model:

rt=σtt1εt,r_t = \sigma_{t|t-1} \varepsilon_t, σtt1=ω+αrt12,\sigma_{t|t-1} = \omega + \alpha r^2_{t-1},

when ω=0.01\omega=0.01 and α=0.9\alpha=0.9.

To prevent explosion, we omitted autoregressive operators such as yt=ϕ1yt1+ϕ2yt2y_t = \phi_1y_{t-1} + \phi_2y_{t - 2}. (In order to prevent excessively large values from affecting the standardization during pre-training, we mentioned in line 132 that excessively large sample values will be discarded.) But your example shows volatility clustering depends on past values, so we will reintroduce autoregression under tight constraints to yield richer series.

We believe this is a compelling and valuable case and add it to our paper to further demonstrate the validity of S2S^2.

评论

Thank you for the additional experiments and explanations provided by the authors. I will take them into account in my final score.

评论

Dear Reviewer,

Thank you very much for your positive recognition of our work. We are sincerely grateful for the valuable time and effort you have dedicated to reviewing our paper and considering our rebuttal materials. Your constructive feedback throughout the review process has been instrumental, and we appreciate your acknowledgment of the additional experiments and explanations provided.

We are delighted that you found merit in our contributions. Rest assured, we will carefully consider all points discussed during this review cycle as we work to further refine and improve the paper. We hope that your recognition of our perspectives on complex system modeling will further advance the field of time series analysis.

Thank you again for your support and insightful comments.

Sincerely, The Authors

审稿意见
4

The article focuses on improving model performance through data augmentation and the use of foundation models. Specifically, they propose a series-symbol data generation mechanism that alleviates issues of scarcity and imbalance in data. Additionally, they design a time series foundation model consisting of two encoders, where one encoder processes time series data and the other encoder processes symbol inputs. These two encoders are trained using contrastive learning to enhance the representation ability of time series. The approach has achieved good results in multiple time series tasks. Experiments verify the ability of data augmentation in terms of stability, predictability, frequency domain analysis, and trendiness. The comparative experiments validate the performance of the foundation model across multiple downstream tasks.

优缺点分析

S1: The author proposes a data generation mechanism that addresses issues of data scarcity and imbalance, and experimental results show that it can alleviate these problems. S2: The pre-trained model demonstrates strong capabilities in multiple time-series tasks.

W1: Lack of implementation details for the symbol data generation mechanism. W2: No explanation was given regarding the complexity issue of the symbol data generation mechanism when dealing with long series, as well as whether the symbol selection strategy needs to be adapted based on specific tasks.

问题

  1. Does the symbol data generation mechanism have a complexity issue when dealing with long or non-stationary time series data?
  2. Is the symbol generation mechanism required to be adjusted according to downstream tasks? Please provide detailed information on how symbols are selected.
  3. How were the 6-layer Transformer encoder for handling time series data and the 6-layer DistilBERT for handling symbol data chosen? Does layer depth have any impact on performance?

局限性

Yes

最终评判理由

These supplementary materials and the responses from the authors to me and other reviewers basically answered my questions. I think it is sufficient to rating this work, and I will maintain my scores.

格式问题

N/A

作者回复

Thank you for your recognition of the data generation mechanism, model performance, and experimental results in this paper. Below I will answer your questions one by one.

W1: Lack of implementation details for the symbol data generation mechanism.

Thank you for your valuable feedback. We will open a new chapter in the appendix to further explain each step and hyperparameter settings of the S2S^2 data generation mechanism in detail, and we have responded to your questions in detail in Q1-Q3.

We will open source our S2S^2 data generation code in the future.

W2: No explanation was given regarding the complexity issue of the symbol data generation mechanism when dealing with long series.

We define the specific symbol explanation and its as follows. Then, we use the divide-and-conquer approach to analyze the complexity of our S2S^2 data generation mechanism.

symbolexplanation
LLThe length of time series
MMNumber of input channels
NNNumber of output channels
kkTotal number of mixed distributions used
ppAutoregressive model order in ARMA model
qqMoving average model order in ARMA model
PPProbability of choosing a sampling method
bbNumber of binary operators used to construct symbolic expressions
uuThe number of unary operators used to construct symbolic expressions
  1. Symbolic Expression Generation: We construct symbolic expressions using a tree structure as a medium. When we have bb binary operators, we further insert (b+1)(b + 1) leaf nodes (the process from (a) to (b) in Figure 3 in our paper). Therefore, after inserting uu unary operators (Figure 3 (c)), the total number of nodes in the tree is n=2b+u+1n = 2b + u + 1. Because there are many ways to construct a tree, we consider the time complexity of constructing a balanced tree. Therefore, for NN symbols constructed, the specific complexity of this process is O(N×nlogn)O(N\times n\mathrm{log}n).

  2. Sampling series generation: When we want to generate a sampling time series with MM channels, each channel has a probability of PP to be sampled using a mixture distribution and a probability of (1P)(1-P) to be sampled using an ARMA model. When the sampling length of the series is LL, the complexity of generating kk mixture distribution and ARMA (pp, qq) series is O(kL)O(kL) and O(L(p+q))O(L(p+q)). Therefore, the time complexity of this process can be quantified as O(ML×[Pk+(1P)(p+q)])O \left ( ML \times [Pk + (1-P)(p+q)] \right ).

  3. Sampling through symbolic expressions and series: We simplify the specific operational details of this process and only consider the time complexity of operations with variables. For a series of length L, we have NN symbolic expressions to be sampled, and each symbol has an average of M+12\frac{M+1}{2} variables (Each symbolic expression may contain any number of variables from 1 to M, so here we take M+12=(1+2++M)M\frac{M+1}{2}=\frac{(1+2+\cdots+M)}{M} as the average probability). Then the process can be quantified as O(NM+12L)O(N \cdot \frac{M+1}{2}\cdot L).

To sum up, since other variables that affect the S2S^2 sampling process are usually small, it can be intuitively understood that the time complexity of the entire sampling process is proportional to the length LL.

In Q1 below, we further answer why this paper does not sample long time series.

W2: whether the symbol selection strategy needs to be adapted based on specific tasks.

We do not select any symbols. Instead, to learn the richest representation of time series, we use all symbols to generate the S2S^2 dataset for unified pre-training of SymTime. See Q2 for details.

Q1: Does the symbol data generation mechanism have a complexity issue when dealing with long or non-stationary time series data?

Through W2, we find that the time complexity of data generation is indeed related to the length LL. The specific explanation for avoiding sampling long and non-stationary series is as follows:

  1. Out of domain: As described in line 132, the symbolic expression f()f(\cdot) has a domain restriction. The longer the series, the more likely it is that the value will fall outside the domain, and the probability of sampling failure increases. When we fail to sample, we will resample.
  2. Stationarity: Non-stationary series often exhibit strong autoregressive properties, which can easily lead to numerical inflation and sampling failure (line 132). Furthermore, stationary series have bounded prediction errors and are better characterized; non-stationary series have errors that tend to diverge and are poorly characterized. Therefore, to ensure sampling success and forecast performance, we prioritize generating stationary series.
  3. More channels: Since we use mask time series modeling for representation learning, the benefits of long series on our model learning are not significant. We are more likely to generate data with more channels (with M=6,N=12M=6, N=12) to help the model learn the feature between multivariate time series channels.

In summary, in order to more effectively generate data with better predictability, we adopted the strategy described in the paper in the S2S^2 data generation mechanism.

Q2: Is the symbol generation mechanism required to be adjusted according to downstream tasks? Please provide detailed information on how symbols are selected.

As described in Experiment 4.1, our goal is to generate as much data as possible through an unrestricted data generation mechanism to cover the broadest possible representation of time series, thereby training a time series foundation model (Definition 1 in Line 54). Using this dataset, we uniformly pre-train SymTime, enabling the model to learn the most comprehensive representation of time series data. To this end, we do not specifically tailor the symbolic expressions we use to specific downstream tasks or datasets. This approach demonstrates the generalization capabilities of SymTime as a foundation model.

Q3-1: How were the 6-layer Transformer encoder for handling time series data and the 6-layer DistilBERT for handling symbol data chosen?

  1. Time Series Encoder: We drew on the frameworks of time series foundation models such as Moirai, Timer, and Moment, as well as model architectures in the computer vision such as ViT and ALBEF. Ultimately, we chose to use a 6-layer Transformer encoder to learn the basic representation of time series data.
  2. Symbolic Encoder: This module aims to extract the representation of symbolic expressions during contrastive learning and enable the time series encoder to learn the pairing relationship between series and symbols. Considering computational complexity, we decide to chose a smaller, but more linguistically robust, pre-trained LLM as the symbolic encoder. Therefore, the two most commonly used 6-layer Transformer models are gpt2-small and DistilBERT. DistilBERT was chosen due to its superior performance and the compatibility of its encoder architecture with the time series encoder.

Q3-2: Does layer depth have any impact on performance?

Based on the ideas in Q3-1, we chose to determine the model architecture of SymTime. To further verify the impact of the number of model Transformer layers on downstream task performance, we set different parameters for the time series encoder and constructed the following two control groups:

namelayersdmodeld_{\mathrm{model}}dffd_{\mathrm{ff}}
SymTime65122048
A67683072
B33861536

We keep the symbol encoder unchanged and pre-train the control group using the pre-training configuration used in C.3 of the original paper (since the parameters of the model in A are further expanded, we reduce the batch size from 128 to 96, while keeping the other configurations unchanged). Then, we fine-tune the model on the four ETT datasets. The average forecasting results for the lengths {96, 192, 336, 720} are shown below:

DatasetSymTimeAB
MetricsMSEMAEMSEMAEMSEMAE
ETTm10.3710.3900.3700.3980.3780.395
ETTm20.2740.3210.2780.3230.2790.324
ETTh10.4300.4360.4350.4390.4340.439
ETTh20.3650.4020.3690.4050.3750.410
Average0.3600.3870.3630.3910.3670.392

It is clear from this that SymTime's original model parameter configuration achieves optimal results on the ETT dataset.

However, when the number of parameters changes, the forecasting results will decrease to a certain extent, but the decrease is not significant. Therefore, as long as the model can meet the representation learning requirements during the pre-training phase, the performance improvement of the model mainly comes from our pre-training. A model that is too large will, to a certain extent, bring greater computational complexity.

评论

Dear reviewer,

The authors have provided a response. Do you have any additional questions for them? If so, please ask them now. If not, please summarize your conclusions after reading their response.

Best,

AC

评论

Thank you for providing detailed symbol explanation and additional experiments. These supplementary materials and your responses to other reviewers basically answered my questions. I think it is sufficient to rating your work, and I will maintain my scores.

评论

Dear Reviewer,

Thank you for acknowledging our supplementary materials and experimental additions. We sincerely appreciate your confirmation that the symbol explanations and extended results addressed your concerns, and we are grateful for your commitment to maintaining the evaluation score.

Your focus on methodological clarity and empirical rigor has significantly strengthened our paper. We will further add the time complexity analysis of the S2 data generation algorithm to our paper.

Respectfully,

The Authors

审稿意见
5

The paper introduces a novel approach to address data scarcity and imbalance in time series analysis. The authors propose a series-symbol data generation pipeline. Inspired by complex dynamic system theories, the authors generate high-quality univariate time series data paired with corresponding symbolic expressions, enabling the creation of diverse time series datasets. The authors present SymTime, a pre-trained encoder-only time series foundation model, leveraging contrastive learning and masked modeling. It shows notable scalability and generalization performance.

优缺点分析

Quality: The submission is technically sound, with a strong theoretical foundation and extensive experimental validation. SymTime’s architecture is well-motivated, leveraging contrastive learning and masked modeling on numerical and symbolic tokens. About this architecture option, I wonder if the author has ever tried generative modeling on a decoder-only Transformer like TimesFM and Timer. The benchmarks include five major TSA tasks (forecasting, classification, imputation, anomaly detection). Although it is comprehensive, I'd like to raise my score if the model can be evaluated more public leaderboard, such as GIFT-Eval and FEV.

Clarity: The mathematical formulation (e.g., tree-based symbolic sampling, ARMA processes) is well-explained. The paper is generally well-written and organized, but there are some typos that could be improved (e.g., "prmutation" -> "permutation"; "pretrain" -> "pre-train").

Significance: The paper provides new insights for (1) pre-training generalizable time series foundation models, (2) inspiring new benchmarks for time-series pre-training. As symbolically generated time-series data can effectively replace or augment real-world data for pre-training, it can be beneficial for the community if the author releases the data generation framework.

Originality: Different from previous works, which are pre-trained solely on synthetic data or partially synthetic data. The proposed symbolic framework can provide aligned synthetic data and symbolic formulation. In addition to enabling contrastive learning, it is a promising direction is to take symbolic formulation as the aligned textual description for multimodal time series models, which can help current models better understand the underlying dynamics of time series. In terms of the model design, the contributions of SymTime mainly focus on pre-training tasks with contrastive learning. The structure is similar to previous works (such as moment, moirai).

问题

  1. Under the proposed symbolic system, what is the maximum scale of synthetic data that can be constructed?
  2. Are there any suggestions to overcome the diminishing marginal returns when pre-training on an ever-larger scale of synthetic data?
  3. What is the range of different statistical metrics in Figure 4? The generated data has relatively low diversity in some metrics (PE, FFT Mean, Seasonality). How does pre-training based on metrics of different degrees of diversity affect the performance of downstream tasks?
  4. Can the proposed method generate multivariate time series? It may also provide insights about how to define how "multivariate" a set of time series is.

局限性

Please discuss whether omitted operators (e.g., exponent and higher power) limit the generation of certain dynamics.

最终评判理由

Based on the additional empirical results provided, I am confident that the framework has the potential to offer valuable insights for the development of time series foundation models. Therefore, I will maintain my positive evaluation. Additionally, I am pleased to see the release of this framework.

格式问题

No

作者回复

We sincerely thank you for your recognition of the novelty of the data generation mechanism and the effectiveness of the SymTime model. Below we will answer your valuable questions one by one.

Quality: Although it is comprehensive, I'd like to raise my score if the model can be evaluated more public leaderboard, such as GIFT-Eval and FEV.

Thank you for your recognition of the comprehensiveness of our experiments. We are happy to verify SymTime on the GIFT-Eval benchmark. We will split the training sets according to the requirements and fine-tune SymTime.

However, due to the lots of sub-datasets included in this benchmark and time constraints, we have not yet completed fine-tuning. Once we have completed it, we will immediately inform you of the results and upload the online leaderboard. The results will also be added to Section A.1 in our paper.

Clarity: The paper is generally well-written and organized, but there are some typos that could be improved.

Thank you for your recognition of the clarity of our algorithm writing. We already have corrected the above errors and further checked the full paper.

Significance: As symbolically generated time-series data can effectively replace or augment real-world data for pre-training, it can be beneficial for the community if the author releases the data generation framework.

We sincerely appreciate your recognition of our S2S^2 data generation mechanism. Deep learning have made significant progress in time series analysis, but we continue to strive to address existing challenges by incorporating the insights of complex dynamical systems into the essence of time series.

We promise to open source the code for our S2S^2 data generation mechanism in the future.

Q1: Under the proposed symbolic system, what is the maximum scale of synthetic data that can be constructed?

Theoretically, the S2S^2 data generation mechanism can generate unlimited sequence-symbol pairings. This diversity is primarily reflected in two aspects:

  1. It can generate an unrestricted variety of symbolic expressions f()f(\cdot): Due to the existence of random constants and a variety of symbols and operators, the symbols we can generate are infinite, and the time series generated by different symbols are also diverse.
  2. It can generate a variety of sampled sequences XX: Since we switch between different random seeds and specify different initializations at each sampling, XX generated from a mixture of uniform and normal distributions and randomly parameterized ARMA(p, q) models is also diverse. This also makes the data we can generate infinitely.

Q2: Are there any suggestions to overcome the diminishing marginal returns when pre-training on an ever-larger scale of synthetic data?

This is a very valuable question for large-scale pre-training. Currently, many large-scale foundation models trained on real datasets face the problem of diminishing marginal returns. However, in the experiments in Section 4.3 that verified the impact of pre-training dataset size on SymTime performance, we find that as we continuously increase the pre-training dataset, the model's performance continue to improve randomly. This phenomenon has little impact on SymTime. Therefore, we conduct a detailed analysis, and the specific reasons are as follows:

When the S2S^2 dataset is generated, its representation is more "random," meaning that most adjacent samples may not come from the same distribution. As our data volume increases, our S2S^2 data will appear evenly across the representation space of time series data. However, real time series datasets often consist of multiple sub-datasets, each of which comes from the same distribution and has the same generation mechanism. Therefore, as the amount of real data increases, it may only be able to fill a certain position in the representation space (a specific position and range within this distribution).

Therefore, for other synthetic datasets, <u>we believe that we should make the data more random when generating it</u>. Rather than restricting the possible distribution of the data, we should make it more diverse. This approach we believe can alleviate this problem.

Q3-1: What is the range of different statistical metrics in Figure 4? The generated data has relatively low diversity in some metrics (PE, FFT Mean, Seasonality).

This is a very interesting question. Radviz visualizes high-D data in 2D using a spring force balance. After normalization, sample features act as "pulls" from anchor points evenly spaced around a circle. The sample's final position is determined by the equilibrium of all these pulls. Consequently, the 2D coordinates themselves lack inherent physical meaning, as they result from the combined influence of all features. This method is useful for observing clustering in reduced dimensions. The observed lack of diversity in indicators like PE, FFT Mean and Seasonality within the Radviz plot is an artifact of this balancing process across all indicators, not an indication that those specific features lack diversity in the actual data.

Q3-2: How does pre-training based on metrics of different degrees of diversity affect the performance of downstream tasks?

To quantitatively explore this issue, we selecte 25B of time series data from the 50B S2S^2 dataset, which ranked high for Seasonality (S), Forecastability (F), and Permutation Entropy (PE) respectively. So the subdataset can reflect the seasonality, predictability, and permutation complexity of the data, respectively. We pre-train SymTime using the configuration in C.3 and fine-tuned it on four ETT datasets for downstream tasks. The experimental results are shown below:

DatasetFSPESymTime
MetricsMSEMAEMSEMAEMSEMAEMSEMAE
ETTm10.3740.3930.3800.3950.3850.3890.3780.393
ETTm20.2830.3290.2780.3250.2900.3310.2780.325
ETTh10.4330.4380.4450.4430.4430.4400.4340.438
ETTh20.3860.4110.3820.4100.4080.4190.3700.408
avg0.3690.3930.3710.3950.3820.3950.3650.391

We believe this is a very interesting experiment. The results show that metrics with varying degrees of diversity do have a certain impact on the performance of downstream tasks. We can preliminarily conclude that Forecasting and Forecasting efficiency metrics are more effective in promoting model representation learning than PE.

Q4: Can the proposed method generate multivariate time series? It may also provide insights about how to define how "multivariate" a set of time series is.

Our S2S^2 data generation mechanism can generate a multivariate time series Y=f(X)RN×LY = f(X)\in \mathbb R^{N\times L} by constructing a multivariate complex system f()f(\cdot) and a multivariate sampling series XRM×LX\in \mathbb R^{M\times L}, where MM and NN are the input and output channels respectively and LL is the length of time series.

For example, we can generate a multivariate symbol and series as follows:

f(x1,x2,x3)=[f1(x1,x2,x3)f2(x1,x2,x3)f3(x1,x2,x3)]f(x_1, x_2, x_3) = \begin{bmatrix}f_1(x_1, x_2, x_3) \\ f_2(x_1, x_2, x_3) \\ f_3(x_1, x_2, x_3) \end{bmatrix} X=[x1x2x3]X = \begin{bmatrix} x_1 \\ x_2 \\ x_3 \end{bmatrix}

We simultaneously obtain time series data for the three input and three output channels using Y=f(X)Y=f(X). This allows us to combine sampling series from different channels through symbolic operations. Therefore, our data generation mechanism is capable of generating multivariate time series with channel correlation. Specifically, we present the generated symbolic expressions and time series data in Section B.1 on page 21, including multivariate time series data, from which we can observe the correlation between channels.

Limitations: Please discuss whether omitted operators (e.g., exponent and higher power) limit the generation of certain dynamics.

Thank you for this valuable comment. In the current S2S^2, omitting certain operators does restrict the creation of some dynamic data. Earlier versions included a richer set of numerical operations—differentiation, integration, exponential functions with various bases, and higher-order powers. Extensive experiments showed that several of these operators severely impeded the data-generation algorithm; the main reasons are detailed below.

  1. Value explosion: To maintain quality, we cap large magnitudes (l. 132). Integration, exponential and high-order powers readily cause overflow, so they were dropped; exp alone is retained for diversity.
  2. Numerical differentiation: Symbolic differentiation introduces truncation/round-off trade-offs. Combined with reciprocal and absolute-value operators, functions such as x|x| or sin(1x)\mathrm{sin}(\frac{1}{x}) because non-smooth or high-frequency vibration at x=0x=0, breaking differentiation.
  3. Numerical integration: Randomly built expression trees often yield integrands with singularities (e.g., 011xdx\int_{0}^{1}\frac{1}{\sqrt{x}}\mathrm{d}x ) or force costly oscillatory integrals (e.g., 0100sin(100x)dx\int_{0}^{100}\mathrm{sin}(100x)\mathrm{d}x); interval selection is non-trivial.
  4. Symbolic cost: Differentiation and integration are slow and can trigger exponential memory growth. Many elementary functions lack closed-form antiderivatives (e.g., ex2dx\int \mathrm e ^{-x^2}\mathrm d x).

Considering factors like numerical stability, symbolic complexity, computational efficiency, and sampling success rate, we selectively omitted some symbolic operations.

Nevertheless, the S2S^2 data generation framework proposed is essentially a complete theory. It already incorporates the vast majority of symbolic operations, and new operators or user-defined symbolic operations can be easily added. The omission of some operations due to the above factors does not undermine the validity of this framework. We consider this a valuable point and will add it in the paper.

评论

Thanks to the authors' responses, which have effectively addressed the majority of my concerns. Based on the additional empirical results provided, I am confident that the framework has the potential to offer valuable insights for the development of time series foundation models. Therefore, I will maintain my positive evaluation.

评论

Dear Reviewer,

Thank you for your constructive feedback and continued support throughout the review process. We deeply appreciate your recognition of our framework’s potential to contribute to time-series foundation models and your commitment to maintaining a positive assessment. Your insights have been invaluable in strengthening both our paper and our perspective on the broader implications of this work.

We are sincerely grateful for the time and expertise you dedicated to evaluating our revisions.

Sincerely,

The Authors

审稿意见
5

This paper introduces a synthetic data generation method for training time series foundation models. At the core of the synthetic data generation method is a composition of symbolic expressions that gets randomly generated through sampling (sampling input/output dimensions, operators, affine transformations, etc.). The input to these symbolic expressions are time series sampled from mixed distributions and ARMA processes. The transformation of the symbolic expression then forms the synthetic time series. The paper also introduces SymTime, a time series foundation model which uses patched encodings, cross-entropy loss, and additional contrastive learning and momentum distillation terms.

The manuscript compares the generated time series dataset features with existing datasets from Monash to evaluate whether the generated time series are realistic. The authors then evaluate their models on forecasting, imputation, classification, and anomaly detection tasks. Finally, the paper conducts ablations on the pre-training objectives and analyses the obtained time series and symbol representations.

优缺点分析

The paper introduces an interesting idea for synthetic time series generation based on sampling of symbolic expressions. This offers a flexible way of generating synthetic time series. As the authors state in the paper, pretraining time series models requires synthetic data due to a relative scarcity (compared the text data) and therefore novel ideas for synthetic data generation are needed. Therefore, the explored topic in the paper has significance and is original. Another strength of the paper is that the thorough evaluation of the model on several downstream tasks (forecasting, imputation, classification, anomaly detection) along with ablations and analysis of the generated synthetic time series and SymTime representations.

One weakness of the paper is the clarity of writing. I found several parts of the paper a bit confusing and even after reading several times I am still not sure how they work. While I understood the data generation process after some re-reads, I still do not understand the contrastive learning part of SymTime. For example, how are positive and negative pairs sampled? I understand that the authors covered a lot of ground and had to accommodate this in the space constraints, but I believe a clearer explanation of the core concepts would strengthen the paper.

Another major weakness is the specific setup of the forecasting results and the claims that are made on SymTime’s performance in forecasting. The paper for the long-range forecasting dataset, a fixed lookback window of 96 is used for most comparing models. This makes the conclusion that SymTime outperforms the other baseline models on this dataset invalid because it only does so under this (artificial) conditions that make the baseline methods perform worse. For example, the MSE/MAE windows for PatchTST are much lower in its original paper on these datasets, if the proposed lookback window of 336 or 512 time steps is used. Thus, SymTime does not reduce the absolute error on these datasets compared to existing methods. I have a similar criticism on the short-term datasets. Here again, while the results for SymTime are good, the OWA for the state-of-the-art method is also lower (see Gasthaus et al. (2019), "Spline Quantile Function RNNs") than the error of the methods presented here.

Additionally, while the long-term forecasting benchmark has been used extensively in time series forecasting, the value of this benchmark to still measure progress in this field is questionable. This has been explored in recent position papers (https://arxiv.org/abs/2502.14045) and in the NeurIPS 2024 time series workshop (https://neurips.cc/virtual/2024/workshop/84712#wse-detail-108471, see Christoph Bergmeir's talk). I think it is important for the time series forecasting filed to move towards more extensive evaluation which is proposed in several recent paper (Gifteval: https://arxiv.org/abs/2410.10393; FEV (from the Chronos paper): https://openreview.net/forum?id=gerNCVqqtR).

I understand and appreciate that this work also considers other tasks (like classification and anomaly detection, where I have less insight on the SOTA). Therefore, the paper provides additional experimental evidence. However, I would ask the authors to at least run some of the baseline methods with their recommended settings (specifically PatchTST with 336/512 context length) make sure their comparison captures the error that SOTA models are capable of and therefore an accurate reflection of the SOTA.

问题

How are positive/negative pairs selected in the contrastive learning step?

局限性

The authors appropriately address limitations.

最终评判理由

The authors addressed a major weakness in their evaluation of long-term forecasting benchmark, along with other clarifications and improvements as evident by the discussions with the other reviewers. Hence, I raise my score.

格式问题

n/a

作者回复

We sincerely appreciate your valuable feedback and your acknowledgement of the thoroughness of our experiments on downstream tasks. We will address your questions one by one.

W1: For example, how are positive and negative pairs sampled?

Thank you for your question. In our S2S^2 data generation mechanism, we generate a three-tuple pair of data: (sample sequence, generated sequence, symbolic expression). Specifically, we first construct the symbolic expression f()f(\cdot). Then, we construct the sample series XX. Finally, by inputting the sample series into the symbolic expression, we obtain the generated series Y=f(X)Y=f(X). In this process, we believe that the generated YY will have a natural correspondence with the symbol f()f(\cdot) that generates it.

Therefore, for a certain time series YY, we believe that the series and the symbolic expression f()f(\cdot) that generates it are mutually positive samples, and that other symbolic expressions are mutually negative samples. Correspondingly, for a certain symbol f()f(\cdot), the time series YY generated by it is mutually positive, and other irrelevant symbols are mutually negative samples.

In the abstract and line 153, we repeatedly mention that we generate paired data of sequences and symbols, and the pairing of the two makes them mutually positive samples. We will make this clear in the main paper.

W2: The paper for the long-range forecasting dataset, a fixed lookback window of 96 is used for most comparing models. This makes the conclusion that SymTime outperforms the other baseline models on this dataset invalid because it only does so under this (artificial) conditions that make the baseline methods perform worse.

However, I would ask the authors to at least run some of the baseline methods with their recommended settings (specifically PatchTST with 336/512 context length) make sure their comparison captures the error that SOTA models are capable of and therefore an accurate reflection of the SOTA.

Thank you for your valuable feedback. As you mentioned, for long-term forecasting, the vast majority of models use a look-back window of 96. Therefore, to ensure fairness, we adjusted the look-back length of all models to 96 in our initial comparison.

However, We are pleased to accept your suggestion. To this end, we adjust the look-back window to 336 and 512, re-run the experiments, and compared them with a series of current baseline models. Below we show the average results for four lengths {96, 192, 336, 720} on 8 datasets:

336SymTimePatchTSTTimeMixerTimesNetAutoformerDLineariTransformer
MetricsMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETTm10.3540.3810.3520.3820.3680.3920.4210.4230.6180.5390.3570.3790.3680.395
ETTm20.2600.3170.2580.3150.2620.3180.2820.3340.4000.4200.2910.3530.2720.329
ETTh10.4250.4360.4190.4320.4300.4370.4850.4800.5800.5390.4250.4400.4500.457
ETTh20.3520.3930.3310.3790.3960.4250.4090.4400.6630.6040.4900.4760.3900.416
weather0.2380.2730.2580.2920.2350.2730.2500.2860.4410.4500.2450.2980.2380.272
ECL0.1650.2670.1650.2940.1690.2600.1970.2970.2360.3460.1700.2690.1630.257
Traffic0.3930.2660.3960.2680.4110.2710.6150.3310.6760.4130.4650.3200.3850.273
Exchange0.3700.4080.3850.4200.4150.4380.5480.5321.0530.8070.4480.4620.3920.427
Average0.31980.34270.32040.34780.33570.35190.40090.39030.58320.51490.36140.37470.33230.3533
512SymTimePatchTSTTimeMixerTimesNetAutoformerDLineariTransformer
MetricsMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETTm10.3560.3800.3520.3820.3710.3920.4250.4300.5560.5180.3580.3800.3670.397
ETTm20.2650.3200.2560.3170.2630.3230.2940.3440.3710.4160.2750.3400.2730.331
ETTh10.4140.4320.4130.4340.4290.4440.4810.4860.6270.5790.4180.4380.4460.460
ETTh20.3650.4050.3570.4090.3730.4100.3970.4320.6870.6090.4990.4780.3880.417
weather0.2340.2730.2450.2840.2310.2710.2510.2880.4890.4860.2410.2920.2490.280
ECL0.1630.2670.1690.2690.1770.2740.2010.3020.3530.3930.1670.2670.1620.257
Traffic0.3950.2680.3990.2720.4100.2700.6240.3340.7050.4350.4330.3050.3830.273
Exchange0.3840.4120.3980.4230.5170.4970.7180.6080.9440.7680.5000.4940.4270.467
Average0.32210.34460.32370.34870.34640.36010.42390.40300.59140.52530.36140.37440.33680.3603

The results show that increasing the look-back does improve forecasting performance for most models, likely because longer series contain more historical information. SymTime, PatchTST, and TimeMixer show particularly significant improvements, especially for datasets with larger numbers of channels, such as Weather, ECL, and Traffic. We summarize the possible reasons for the SymTime performance improvement:

  1. SymTime is Transformer encoder framework, and has also undergone large-scale pre-training.
  2. The longer look-back window allows us to construct more tokens, which is closer to the average number of tokens in our pre-training (144), thus better leveraging our pre-training performance.

The detailed test results and experimental configuration of this experiment will be added to the paper in Section A.1 (long-term forecasting benchmark).

W3: Additionally, while the long-term forecasting benchmark has been used extensively in time series forecasting, the value of this benchmark to still measure progress in this field is questionable. This has been explored in recent position papers and in the NeurIPS 2024 time series workshop. I think it is important for the time series forecasting filed to move towards more extensive evaluation which is proposed in several recent paper.

We appreciate your insights on time series forecasting benchmarks. Our paper uses the benchmark from TimesNet, which evaluates models across 5 different TSA tasks to assess multi-task generalization capabilities. This benchmark is widely adopted by models like GPT4TS and Peri-midFormer. The goal of SymTime is to demonstrate strong applicability in a variety of tasks.

While, we recognize newer zero-shot forecasting benchmarks like GIFT-Eval and FEV (used by mary models such as Sundial and Chronos), these primarily target zero-shot forecasting scenarios which differ from SymTime's objectives. Therefore, we prioritized the TimesNet benchmark.

To further validate SymTime, we follow GIFT-Eval's training splits and performed fine-tuning. Testing across all GIFT-Eval sub-datasets is ongoing due to time constraints. We commit to reporting full results upon completion, submitting them to the online leaderboard, and will include comparative analyses with other benchmarks in our paper's future work section.

Another key aspect of this paper lies in the S2S^2 paradigm, which offers the advantage of unrestricted data generation. We can construct unlimited high-quality time series datasets for model pre-training. When the data representations we generate are similar to those in real-world test scenarios, we believe that models pre-trained with S2S^2 will also achieve excellent zero-shot performance on real-world data.

Q: How are positive/negative pairs selected in the contrastive learning step?

In W1, we point out that the time series data generated by the same symbol f()f(\cdot) will be positive samples with the symbol expression, and negative samples with other symbols.

评论

Thank you for response and your added experiments. The increased lookback window results do not indicate significant improvements over PatchTST in the long-term forecasting task. I think this is a more realistic reflection of a real-world scenario as there would be no reason to truncate the lookback window. Thus, I would advocate to update the main results in the paper (including the main table/figures) to reflect the findings from the increased lookback experiment.

Overall, this paper has strong empirical results in other downstream tasks and long-term forecasting is only one component. Even if SymTime "only" matches a competing method like PatchTST, this would be enough for acceptance given the only paper.

I will consider raising my score if the authors commit to update the main figures/tables in the main text to reflect these findings at longer lookback horizons, as opposed to only updating section A.1.

评论

Dear Reviewer,

Thank you for your thoughtful feedback and for recognizing the empirical strengths of our work. We sincerely appreciate your time and expertise in evaluating our rebuttal and providing constructive suggestions.

We fully agree with your observation that the extended lookback window results offer a more realistic reflection of long-term forecasting performance. We will update the main figures and tables in the main text to reflect these findings over the longer review period. We promise to add the experimental configurations of 336 and 512 lengths, as well as the experimental results and conclusions in the main text. We will also add this part of the complete experimental results to the appendix of the paper and draw a result tables similar to Table 17 (long-term time series forecasting results) in the paper to show this part of the results.

We are grateful for your support and believe these revisions will strengthen the paper's transparency and rigor. Your guidance is invaluable to our work.

Sincerely,

The Authors

评论

Dear Authors,

Thank you for your responds and your commitment to strengthen the empirical results. I will raise my score as my points have been addressed.

评论

Dear Reviewer,

Thank you for your support and valuable feedback throughout the review process. We sincerely appreciate your recognition of our revisions and your decision to raise the score—this means a great deal to us. We are grateful for the time and expertise you dedicated to improving our paper. Your insights have undoubtedly strengthened our work.

Sincerely,

The Authors

最终决定

This paper studies synthetic data generation for time series. They use symbolic expressions to develop a data generator, generate a dataset, then show that models pre-trained on synthetic data can rival models pre-trained on real data when finetuned and tested on real data. The paper was received very well by reviewers, who appreciated the approach is innovative and interesting and that it seems to work well and will hopefully inspire future works. The negatives appear to largely be clarifications that can be addressed in the writing, as noted by the reviewers who were receptive to the authors' responses. One key remaining drawback identified during the discussion period was that the performance evaluations are limited to the TimesNet datasets. These datasets have been so overstudied that improvements on them mean very little, especially the forecasting datasets. So extending the evaluations to include others (e.g., GIFT-Eval) will boost the impact of this work.