PaperHub
5.2
/10
Rejected5 位审稿人
最低5最高6标准差0.4
5
5
5
6
5
3.6
置信度
正确性2.6
贡献度1.8
表达2.0
ICLR 2025

UTSD: Unified Time Series Diffusion Model

OpenReviewPDF
提交: 2024-09-20更新: 2025-02-05
TL;DR

In this paper, a Unified Time Series Diffusion (UTSD) model is established for the first time to model the joint probability distribution among multiple domains, utilizing the powerful probability distribution modeling ability of Diffusion.

摘要

关键词
time series anlysisfoundation modelcross-domandiffusition model

评审与讨论

审稿意见
5

This work introduces a conditional diffusion model for MTS prediction task. Different from previous non-Diffusion based methods, it employs a complicated denoised process to model the latent discrepancies in both temporal and channels (domains as described in this work). Different from the diffusion-based methods, it is flexible since it can be easily scalable for N time steps input and M time steps output scenario thanks to its well-designed interaction between conditional embeddings and denoised outputs. In general, the proposed method achieves promising results in various MTS prediction tasks such as prediction and cross-domain prediction.

优点

  1. The motivation of designing conditional diffusion model for flexible N time steps input and M time steps output is clear and reasonable, the model architecture is also well-designed.
  2. Experimental results are soundness and promising.

缺点

  1. The dimension transform is not clearly presented, which may make readers confused. For example, given N time steps input lookback, how to obtain a M time steps output prediction? Although it might be discussed in the manuscript, however, it is still not so easy to follow.
  2. As claimed by the authors in line 201-206 that “existing conditional diffusion models often use simple NN layer to capture … demonstrates low accuracy”. However, according to Sec. 3, as far as I understand that the authors use CNN or Transformer based modules to transform features along the time embedding dimension of input series. For example, in denoising net of Fig. 2, the dimension of condition h\overline{h} is PL×PdP_L \times P_d while the dimension of denoised series before decoder block-a is also PL×PdP_L \times P_d. The time dimension transformation happens at the very last block in denoising net, decoder block-a. Therefore, it should still be a simple NN layer conceptually, which is conflicted with the authors claim.
  3. How do the authors obtain a so called pretrained models? It is not clear in the manuscript.
  4. The authors mentioned many times in the manuscript on down stream tasks of MTS, it might be unsuitable since they only concern MTS prediction task, although including traditional prediction and cross-domain prediction, but it still belongs to MTS prediction task, not involving other down stream MTS tasks, such as anomaly detection or impainting.
  5. In Fig. 4, the authors show results of some short-term periodic series. How about the results for long-term periodic series and short-term non-periodic series? These should be crucial for practical use.

问题

See the weakness above.

评论

Q5: In Fig. 4, the authors show results of some short-term periodic series. How about the results for long-term periodic series and short-term non-periodic series? These should be crucial for practical use.

A5: Thank you for your insightful feedback. We introduce more visualisation results in the appendix section of the revised version, including the ‘long-term cyclic series’ and ‘short-term acyclic series’ as your suggestion, please refer to the revision. Figure 10 (Page. 24 in Revision) shows real extremely long multi-periodic series sampled from the ECL and Traffic datasets, and in addition, Figure 11 (Page. 25 in Revision) shows real short period non-periodic sequences sampled from the ETT dataset. Specifically, UTSD also shows good performance in extremely long sequence prediction, which verifies that UTSD has the ability to capture extremely long term dependencies, which is crucial for practical applications. In addition, UTSD also shows satisfactory prediction results in short-term non-periodic sequences.

评论

Q4: The authors mentioned many times in the manuscript on down stream tasks of MTS, it might be unsuitable since they only concern MTS prediction task, although including traditional prediction and cross-domain prediction, but it still belongs to MTS prediction task, not involving other down stream MTS tasks, such as anomaly detection or impainting.

A4: Thank you for your valuable comments. Due to the page limit, the results of the multitasking experiments could not be displayed in the main draft, so we put the experiments of the other tasks inside the appendix. The following table demonstrates imputation experiments in the full-sample training-from-scratch setting. The average MSE is reduced by 17.0%, 20.1% and 38.7% compared to the existing GPT4TS, TimesNet and PatchTST. Surprisingly, the proposed UTSD shows better results on the ECL dataset characterized by multi-periodic patterns, which conforms to the multi-scale representation mechanism designed in our Condition-Denoising component.

DatasetsMask RatiosUTSD (MSE)UTSD (MAE)TimeLLM (MSE)TimeLLM (MAE)GPT4TS (MSE)GPT4TS (MAE)TimesNet (MSE)TimesNet (MAE)LLMTime (MSE)LLMTime (MAE)
ETTm112.5%0.0190.0770.0170.0850.0230.1010.0410.1300.0960.229
25.0%0.0220.0920.0220.0960.0250.1040.0470.1390.1000.234
37.5%0.0280.1100.0290.1110.0290.1110.0490.1430.1330.271
50.0%0.0350.1170.0400.1280.0360.1240.0550.1510.1860.323
Avg.0.0260.0990.0280.1050.0270.1070.0470.1400.1200.253
ETTm212.5%0.0180.0790.0170.0760.0180.0800.0260.0940.1080.239
25.0%0.0190.0820.0200.0800.0200.0850.0280.0990.1640.294
37.5%0.0210.0850.0220.0870.0230.0910.0300.1040.2370.356
50.0%0.0240.0940.0250.0950.0260.0980.0340.1100.3230.421
Avg.0.0200.0850.0210.0840.0220.0880.0290.1020.2080.327
ETTh112.5%0.0400.1370.0430.1400.0570.1590.0930.2010.1260.263
25.0%0.0530.1550.0540.1560.0690.1780.1070.2170.1690.304
37.5%0.0700.1750.0720.1800.0840.1960.1200.2300.2200.347
50.0%0.0930.2020.1070.2160.1020.2150.1410.2480.2930.402
Avg.0.0640.1670.0690.1730.0780.1870.1150.2240.2020.329
ETTh212.5%0.0400.1240.0390.1250.0400.1300.0570.1520.1870.319
25.0%0.0430.1310.0440.1350.0460.1410.0610.1580.2790.390
37.5%0.0490.1430.0510.1470.0520.1510.0670.1660.4000.465
50.0%0.0530.1550.0590.1580.0600.1620.0730.1740.6020.572
Avg.0.0470.1380.0480.1410.0490.1460.0650.1630.3670.436
Electricity12.5%0.0430.1290.0800.1940.0850.2020.0550.1600.1960.321
25.0%0.0490.1420.0870.2030.0890.2060.0650.1750.2070.332
37.5%0.0560.1510.0940.2110.0940.2130.0760.1890.2190.344
50.0%0.0650.1650.1010.2200.1000.2210.0910.2080.2350.357
Avg.0.0530.1470.0900.2070.0920.2100.0720.1830.2140.339
Weather12.5%0.0240.0400.0260.0490.0250.0450.0290.0490.0570.141
25.0%0.0260.0430.0280.0520.0290.0520.0310.0530.0650.155
37.5%0.0300.0470.0330.0600.0310.0580.0810.1800.0580.121
50.0%0.0330.0520.0370.0650.0340.0620.0380.0630.1020.207
Avg.0.0280.0460.0310.0560.0300.0540.0600.1440.0760.171

Indeed we have never stopped exploring the potential of UTSD, and we believe that the introduction of multitasking scenarios helps to fully demonstrate the cross-domain generalisation capabilities of diffusion models. An encouraging piece of idea is to integrate the three tasks of prediction, interpolation, and anomaly detection using a masking mechanism, we use the masking information to make only the unmasked part visible to the Condition Net and reconstruct the masked part using the Denoising Net, which shows that UTSD has the potential to be a multitasking general purpose time series model.

评论

Q2: As claimed by the authors in line 201-206 that “existing conditional diffusion models often use simple NN layer to capture … demonstrates low accuracy”. However, according to Sec. 3, as far as I understand that the authors use CNN or Transformer based modules to transform features along the time embedding dimension of input series. For example, in denoising net of Fig. 2, the dimension of condition h\overline{h} is PL×PdP_L\times P_d while the dimension of denoised series before decoder block-a is also PL×PdP_L\times P_d. The time dimension transformation happens at the very last block in denoising net, decoder block-a. Therefore, it should still be a simple NN layer conceptually, which is conflicted with the authors claim.

A2: We appreciate the reviewers for carefully reviewing our article and asking very valuable questions and apologize for the confusion in our original manuscript. We have clarified the discussion of prior works as follows:

  1. The main innovative contribution of UTSD is the establishment of a unified temporal diffusion model as well as the diffusion model architecture which support of the pre-trained and efficient fine-tuned. In the fine-tuning paradigm of UTSD, the pre-trained condition net captures generic fluctuation patterns from observed sequences as conditional information, whereas the denoising network connects the unified representation space with the proprietary representation space using a fine-tuning-supporting adapter, subsequently, utilizes specific fluctuation patterns from noises in the target domain to reconstruct sequence. Compared with the existing condition diffusion model, UTSD demonstrates excellent cross-domain generalisation capabilities, which is attributed to fact that condition net captures conditional embeddings at multiple scales to enhance denoising net modelling of inductive bias from conditional probability distributions in different domains.
  2. In Figure 2, to minimise the complexity and to help the reader understand our simple intention, we deliberately unified the shape of the tensor at both ends of the Adapter Block to be (PL×Pd)(P_L\times P_d). Indeed, haRPL×Pdh_a\in\mathbb{R}^{P_L\times P_d} and haRPH×Pd\overline{h_a}\in\mathbb{R}^{P_H\times P_d} are defined in the code implementation, and of the entire feature map in the Denoising Net, are (PH×Pd),(PH×Pd2),(2PH×Pd4),(4PH×Pd8)(P_H\times P_d),(P_H\times\frac{P_d}2),(2P_H\times\frac{P_d}4),(4P_H\times\frac{P_d}8). Therefore, the time dimension transformation happens in the all adapter-blocks, and the hybrid structure containing the entire Condition Net and Adapter acts as a substitute for the ‘simple NN layer’. The ablation experiments show that the degradation of the model performance in the case of the w/o condition net and w/o adapter validates our view.

We have highlighted this clarification in blue in the revised manuscript. Thank you for your insightful comments, sincerely.

Q3: How to obtain a pretrained models?

A3: We pretrain our UTSD in the mixed dataset with hybrid-domains, which follows the experimental setup of UniTime[1]. Concretely, we provide the detailed experimental configuration about the batch size and number of training steps for pretrain datasets and several benchamrk in Table 7 (Page. 17 in Revision).

[1] Xu Liu, Junfeng Hu et al. Unitime: A language-empowered unified model for cross-domain time series forecasting. In Proceedings of the ACM Web Conference 2024.

评论

Thank you for your thoughtful review and valuable feedback. We appreciate recognition of our innovative work in condition-denoising architecture with flexible observation and prediction timesteps. We response your concerns as following.

Q1: The dimension transform is not clearly presented, which may make readers confused. For example, given N time steps input lookback, how to obtain a M time steps output prediction? Although it might be discussed in the manuscript, however, it is still not so easy to follow.

A1: Thank you for pointing out the need for clarity regarding the dimension transformation. First, UTSD processes the observation sequence XL+1:00RB×d×LX_{-L+1:0}^0\in\mathbb{R}^{B\times d\times L} and forecast sequence Y1:H0RB×d×HY_{1:H}^0\in\mathbb{R}^{B\times d\times H} with the same channel-independence operation and patch-tokenizer operation to obtain XembRBd×PL×PdX_{emb}\in\mathbb{R}^{B\cdot d\times P_{L}\times P_{d}} and YembRBd×PH×PdY_{emb}\in\mathbb{R}^{B\cdot d\times P_{H}\times P_{d}} where PdP_d is the length of the patch, respectively, and we adopt a padding-free and overlap-free approach to perform the patch-tokenizer operation to ensure that PL=L/PdP_L=L/P_d and PH=H/PdP_H=H/P_d. Specifically, we convert the observation sequence and forecast sequence into PLP_L and PHP_H tokens, respectively, by setting the hyperparameter PdP_d, where the dimension of the token is fixed to PdP_d.

Subsequently, the Condition Net accepts XembRBd×PL×PdX_{emb}\in\mathbb{R}^{B\cdot d\times P_{L}\times P_{d}} as input and outputs a set of multiscale condition variables (hmR4PL×Pd8h_m\in\mathbb{R}^{4\cdot P_L\times\frac{P_d}8},hcR2PL×Pd4h_c\in\mathbb{R}^{2\cdot P_L\times\frac{P_d}4},hbRPL×Pd2h_b\in\mathbb{R}^{P_L\times\frac{P_d}2},haRPL×Pdh_a\in\mathbb{R}^{P_L\times P_d}). Correspondingly, the Denoising Net accepts a set of multiscale condition variables (hmR4PH×Pd8\overline{h_m}\in\mathbb{R}^{4\cdot P_{H}\times\frac{P_{d}}{8}},hcR2PH×Pd4\overline{h_c}\in\mathbb{R}^{2\cdot P_{H}\times\frac{P_{d}}{4}},hbRPH×Pd2\overline{h_b}\in\mathbb{R}^{P_{H}\times\frac{P_{d}}{2}},haRPH×Pd\overline{h_a}\in\mathbb{R}^{P_{H}\times P_{d}}) as input and outputs YembRBd×PH×PdY_{emb}\in\mathbb{R}^{B\cdot d\times P_{H}\times P_{d}}, which is straightened directly as the prediction result Y1:H0RB×d×HY_{1:H}^{0}\in\mathbb{R}^{B\times d\times H}.

The most innovative and critical design is that the innovative adapter module is designed to align the channel dimensions (the number of tokens PLP_L and PHP_H) of the two tensors (hm,hc,hb,ha)(h_m,h_c,h_b,h_a) and (hm,hc,hb,ha)(\overline{h_m},\overline{h_c},\overline{h_b},\overline{h_a}) in the observation space and the generation space. Specifically, the 1 ×\times 1 Conv1D in Adapter Block-a accepts haRPL×Pdh_a\in\mathbb{R}^{P_L\times P_d} as input, keeping the feature dimension (PdP_d) unchanged, and converts the number of input channels from (PLP_L) to (PHP_H). This can be expressed as Equation ha=AdapterBlocka(ha)\overline{h_a}=AdapterBlock_a(h_a), where haRPL×Pdh_a\in\mathbb{R}^{P_L\times P_d} and haRPH×Pd\overline{h_a}\in\mathbb{R}^{P_H\times P_d}.

Furthermore, a question that may be of interest to reviewers is how to allow Condition Net and Denoising Net to be able to handle an arbitrary number of tokens (e.g. PLP_L and PHP_H)? Indeed, this is achieved through the formula X=(Conv1D(XT))TX=(Conv1D(X^T))^T, where XTRPd×PLX^T\in\mathbb{R}^{P_d\times P_L} denotes the transpose of XRPL×PdX\in\mathbb{R}^{P_L\times P_d}. Thus, it is only necessary to ensure that the number of input and output channels of Conv1D is fixed to PdP_d. On the other hand, Attention is not sensitive to PLP_L or PHP_H when capturing global dependencies, and does not change the shape of the input tensor, so UTSD is able to accept any timesteps as an observation sequence, and can also output a prediction sequence with any timesteps.

Detailed implementation details will be updated in the revised manuscript, making it easier for readers to follow the transformation steps.

评论

Dear Reviewer pg2F,

Thank you very much for devoting time on reviewing our manuscript. As the author-reviewer discussion process is ending soon, we are wondering whether our responses have well addressed your concerns. If you have any further questions regarding our manuscript, please let us know and we are glad here to provide further discussion and clarification to improve the quality of this manuscript.

Thank you for your contributions,

Authors of Paper 2173

评论

Dear Reviewer pg2F,

Thank you very much for your time and valuable comments on our manuscript. It is heartening to know that a modified version is ready with the help of valuable comments from reviewers, and we would like to know if our modified version has solved your problem. If you have any further questions about our manuscript, please let us know and we are happy to provide further discussion here. Our improvements are centered around the following suggestions:

  1. Confusion of dimension transformations: We have fixed the confusing part of Figure 2 and highlighted the Tokenizer operation in L214. Detailed implementation details will be updated in the revised manuscript, making it easier for readers to follow the transformation steps.
  2. Implementation details of pre-trained hybrid dataset: We provide the detailed experimental configuration about the batch size and number of training steps for pretrain datasets and several benchamrk in revision Table 7, Page 17.
  3. Supplementary imputation task: The UTSD model is intended to be a versatile tool for various time-series tasks, so we added the imputation experiments on all six benchmarks. Detailed experimental results are presented in Table 11, Page 21 and the experimental analyses were in the revision L1076-1137
  4. More visualization experiments: We introduce more visualisation results in the revision appendix section (Figure 10,11 Page 24,25), including the long-term periodic series and short-term non-periodic series as reviewer's suggested.

Sincerely thank you for your contribution and responsibility.

Author of paper 2173

评论

Dear Reviewer pg2f:

Thank you once again for your valuable and constructive review. As the author-reviewer discussion period is coming to a close, we kindly request your feedback on our rebuttal provided above. Please let us know if our responses have sufficiently addressed your concerns. If there are any remaining issues, we would be happy to engage in further discussion, as we believe that an in-depth exchange will help strengthen our paper.

Best regards!

Authors

审稿意见
5

This paper proposes a time series diffusion model for time series forecasting in multi-domain settings. The proposed model consists of the condition net, adapter, and denoising net. Different from prior autoregressive models, a diffusion denoising process is applied to capture the mixture distribution of cross-domain data and generate predictions directly. Experimental results show the effectiveness of the proposed method on probabilistic time series forecasting tasks.

优点

According to the authors' claim, this is the first work to incorporate diffusion models into time series foundation models.

缺点

  • The paper is generally not well-written. There are many typos and formatting issues.

  • Example of typos: adaptor in the abstract

  • Example of formatting issues: Text size is not consistent in the abstract, and citep is not used.

  • There are false statements when criticizing prior works to motivate the proposed method. For example, while UniTime [Liu et al., 2024a] and Moirai [Woo et al., 2024] are non-autoregressive, authors claim that they "rely on an auto-regressive mechanism" in L085-088.

  • Different from the statement around key contributions, the idea of denoising with condition has already been studied in many prior works, including [Fan et al.], [Shen et al., 2024], [Yuan and Qiao, 2024].

[Fan et al.] MG-TSD: Multi-Granularity Time Series Diffusion Models with Guided Learning Process. In ICLR, 2024.

  • L114-116: "each iteration of the denoising stage causes an accumulation of errors in the latent space, which are further amplified during the alignment of the latent space to the actual sequence space." -> Is there any evidence on this statement?

  • Not enough justification on the effectiveness of the proposed components. The ablation study in Table 5 is insufficient to explain why they should work.

  • L177-179: Any supporting theory or experiments on the inherent drawbacks stated here?

  • While it is natural to consider large-scale datasets for "unified" time series diffusion model, the datasets used in this paper is limited to small single-task datasets.

  • Average performance should not be counted when ranking methods.

  • Please distinguish citet and citep.

  • Zero-shot forecasting is done in an unusual way; while the source and target datasets are similar ETT variations in this work, they usually have a significant gap in the standard zero-shot forecasting settings.

  • Table 3 caption does not make sense.

  • In Table 5, it is not clear what happens if condition net is missing, which implies that no input is given.

问题

Please address concerns above.


I appreciate the response from authors, they addressed some of my concerns. I raised my rating accordingly.

评论

Q10: Zero-shot forecasting is done in an unusual way; while the source and target datasets are similar ETT variations in this work, they usually have a significant gap in the standard zero-shot forecasting settings.

A10: The ETTh1,h2,m1,m2 datasets presented in Table 3 are multimodal data collected from multiple power transformer stations with load and oil temperatures under different external factors such as season, weather, temperature and sampling rate. The diverse and complex scene modes ensure the validity of the ETT datasets, so we adopt the same experimental setup as the existing work in Table 3, and the experimental results demonstrate the good zero-sample generalisation capability of UTSD.

In addition, to further validate the zero-sample performance of UTSD, we also introduced the new cross-domain scenarios RiverFlow[1]\toExchange and Sunspot[2]\toWeather and verified that UTSD still shows good zero-sample prediction capability. Note that the source and target domains therein have completely different domain background information. The experimental results are shown in the following table. We choice RiverFlow[1] and Sunspot[2] as source datasets and Exchange and Weather as target datasets, which ensures that the source and target datasets have completely different domain contexts, sampling rates, period information, and other statistical characteristics. Where RiverFlow reports river flow consisting of more than 23741 time steps, Sunspot contains a single long-day time series representing solar flare observations recorded every 4 seconds starting in 2019.

DatasetsMetricsUTSD(MSE)UTSD(MAE)TimeLLM(MSE)TimeLLM(MAE)LLMTime(MSE)LLMTime(MAE)GPT4TS(MSE)GPT4TS(MAE)PatchTST(MSE)PatchTST(MAE)
RiverFlow\toExchange960.0900.2200.1310.2710.2780.3420.0980.2320.1380.286
1920.1980.3310.2280.3590.4350.4380.1970.3340.6680.660
3360.3560.4350.4000.4770.5330.5220.3570.4440.8980.760
7200.8790.7111.0990.8551.0920.8461.0320.8221.6831.057
Avg.0.3810.4240.4640.4910.5850.5370.4210.4580.9420.765
Sunspot\toWeather960.1810.2350.1980.2530.2000.2680.1860.2420.6690.621
1920.2260.2690.2410.2830.2420.2980.2350.2790.7000.632
3360.2750.3010.2830.3100.2820.3230.2800.3080.7040.633
7200.3260.3410.3330.3430.3290.3530.3520.3580.7460.651
Avg.0.2540.2860.2640.2970.2630.3100.2630.2970.7050.634

[1] Dai, et al. "Changes in Continental Freshwater Discharge from 1948 to 2004. " Journal of Climate (2009).

[2] Nie, Y., et al. SKIPPD: A SKy Images and Photovoltaic Power Generation Dataset for short-term solar forecasting. Solar Energy, 255, 171-179 (2023).

Q11: In Table 5, it is not clear what happens if condition net is missing, which implies that no input is given.

A11: Indeed, as we stated in L522-523, 'In w/o ConditionNet, the observation sequences are directly input into the denoising model as prompt information'. Specifically, with ConditionNet removed, the network structure in Fig.2 contains only DenoisingNet, which accepts only the Prompt Embedding PembP_{emb} captured from the observation sequences as condition variables PembP_{emb} is utilised in the two sub-modules shown in Fig.3 (a) Transformer1D-writer,reader to compute the cross-attention score.

评论

Q7: Not enough justification on the effectiveness of the proposed components. The ablation study in Table 5 is insufficient to explain why they should work.

A7: Thank you for your valuable feedback. To asses the effiectiveness of our proposed components, Table 5 implements the ablation experiments on all four datasets for Adaptor, Classifier-free, and ConditionNet. The average degradation of model performance after removing the three components exceeds 8.7%, 13.2%, and 24.3%, respectively. The most representative result is that when the proposed Condition Net is removed, the performance of the model on datasets ETTh1, ETTm1, Electricity, Weather decreases by more than 25.3%, 19.1%, 27.9% and 25.2% respectively. The results shows the effectiveness of the proposed components.

For convenience, we show ablation experiments on three novel components or mechanisms in the following table, where the data are taken exclusively from Table 5 of the manuscript (page. 10).

AblationETTh1 (MSE)ETTh1 (MAE)ETTm1 (MSE)ETTm1 (MAE)ECL (MSE)ECL (MAE)Weather (MSE)Weather (MAE)
w/o Adapter0.3470.4120.3440.4120.1690.2780.2040.254
\downarrow3.9%7.6%11.7%15.7%13.4%13.9%0.9%3.25%
w/o Classifier-free0.3710.4430.3390.4080.1700.2850.2240.279
\downarrow11.1%15.7%10.1%14.6%14.1%16.8%10.9%13.4%
w/o ConditionNet0.4120.4870.3610.4310.1890.3160.2510.311
\downarrow23.4%27.2%17.2%21.1%26.9%29.5%24.3%26.4%

Q8: L177-179: Any supporting theory or experiments on the inherent drawbacks stated here?

A8: Thank you for your insightful comment. Existing classification guidance mechanisms have inherent drawbacks, such as reliance on the quality of trained classifiers, which affects the accuracy of category-specific generation. Additionally, gradient-based updates can introduce adversarial effects, causing generated images to mislead classifiers with subtle, imperceptible details. These theoretical and experimental insights are supported by Ho's work on classifier-free diffusion guidance [3]. Furthermore, extensive ablation demonstrates the advantages of the Classifier-free strategy utilised by UTSD. when the proposed Condition Net is removed, the performance of the model on datasets ETTh1, ETTm1, Electricity, Weather decreases by more than 13.3%, 12.2%, 15.4% and 12.0% respectively.

AblationETTh1 (MSE)ETTh1 (MAE)ETTm1 (MSE)ETTm1 (MAE)ECL (MSE)ECL (MAE)Weather (MSE)Weather (MAE)
w/o Classifier-free0.3710.4430.3390.4080.1700.2850.2240.279
\downarrow11.1%15.7%10.1%14.6%14.1%16.8%10.9%13.4%

[3] Ho, Jonathan. “Classifier-Free Diffusion Guidance.” Neural Information Processing Systems (2023).

Q9: While it is natural to consider large-scale datasets for "unified" time series diffusion model, the datasets used in this paper is limited to small single-task datasets.

A9: Thank you for your valuable feedback. In the real world, unlike the ‘foundation model’, which is directly pre-trained on large-scale datasets, the ‘unified model’ has the advantage of being able to utilise multiple small datasets from different domains for joint pre-training. This means that the Unified Model is more adaptable to specific domains with many small and fragmented datasets. In addition, the cross-domain results in Table 2 show that UTSD has good cross-domain modelling capability and model scalability, which verifies that ‘the unified model has a wide range of application scenarios’.

As we have discussed in the related work, Timer, Moirai and Moment use a huge corpus (more than 10,000 million timesteps) in pretrain stage. In contrast, UTSD is only pre-trained on a mixed dataset (only 27 million timesteps) obtained by merging 7 small datasets, while the experimental results on the right hand side of Table 2 are obtained by pre-training UTSD from scratch on a specific single small dataset (0.12-1.52 million timesteps). Compared to Moirai, which uses a huge corpus pre-training in the temporal space, UTSD shows excellent data utilisation and can learn sufficiently data patterns from a small number of corpus for inference on downstream datasets, achieving better results on the prediction task.

评论

Q6: L114-116: "each iteration of the denoising stage causes an accumulation of errors in the latent space, which are further amplified during the alignment of the latent space to the actual sequence space." -> Is there any evidence on this statement?

A6: Thank you for your constructive comment. Recent work by Li et al. [2] supports our view, proposing a Timestep-aware Quantization (TaQ) method to effectively eliminate accumulated quantization errors in the multi-step denoising process.

The recent Q-DM [2] confirms our view and proposes a develop a Timestep-aware Quantization (TaQ) method to effectively eliminate the accumulated quantization error caused by the multi-step denoising process.

To further validate our perspective, we conducted a controlled experiment with the Latent-UTSD model, focusing on iterative denoising in the hidden representation space. The model trains the self-encoder end-to-end using a mixed multi-domain synthetic dataset, then reconstructs target sequence embeddings directly in the hidden space. As shown in the following table, Latent-UTSD exhibits relatively poor performance, supporting our claim experimentally. For more details, please see Section C in the appendix. The comprehensive performance of Latent-UTSD across four datasets is summarized in the table below.

MdoelsMetricsETTh1(96)ETTh1(192)ETTh1(336)ETTh1(720)ETTh1(Avg)ETTm1(96)ETTm1(192)ETTm1(336)ETTm1(720)ETTm1(Avg)
UTSDMSE0.2740.2900.3830.3870.3340.2990.3040.3120.3170.308
UTSDMAE0.3010.3390.4240.4280.3830.3330.3580.3650.3680.356
Latent-UTSDMSE0.3230.3420.4500.4540.3920.3520.3650.3650.3700.363
Latent-UTSDMAE0.3480.3860.4930.4920.4300.3900.4180.4350.4430.422
MdoelsMetricsTraffic(96)Traffic(192)Traffic(336)Traffic(720)Traffic(Avg)Weather(96)Weather(192)Weather(336)Weather(720)Weather(Avg)
UTSDMSE0.2840.2930.3080.3190.3010.1330.1840.2070.2640.202
UTSDMAE0.2030.2110.2150.2230.2130.1950.2370.2580.3130.246
Latent-UTSDMSE0.3140.3160.3410.3430.3290.1540.2110.2380.3130.229
Latent-UTSDMAE0.2230.2360.2420.2420.2350.2260.2710.2970.3590.288

These results highlight differences in inductive bias focus during training stages. The encoder may lose critical information during the projection into hidden space, leading the model to reconstruct 'incomplete hidden representations.' Consequently, the noise reduction model struggles to capture all pattern distributions, causing error accumulation in latent space inference.

[2] Li, Yanjing et al. “Q-DM: An Efficient Low-bit Quantized Diffusion Model.” Neural Information Processing Systems (2023).

评论

A5: To evaluate the quality of generated time series, we refer to the metrices (i.e., Context-FID, Correlational, Discriminative, and Predictive) from DiffusionTS[1] and perform the comparison experiments. The results are shown in the following tables. Specifically, the generation metrics Context-FID Score, Correlational Score, Discriminative Score and Predictive Score is reduced by 18.2%, 16.2%, 18.5% and 17.8% on ETTh dataset compared to DiffusionTS. Besides, the average improvement of all indicators of UTSD on ETTh and Energy datasets is 17.6% and 8.2%, respectively.

DatasetMetrics (scores)LengthUTSDDiffusion-TSTimeGANTimeVAEDiffwaveDiffTimeCot-GANImprove(%)
ETTh1Context-FID640.522±.0310.631±.0581.130±.1020.827±.1461.543±.1531.279±.0833.008±.27717.2
1280.633±.0290.787±.0621.553±.1691.062±.1342.354±.1702.554±.3182.639±.42719.5
2560.347±.0100.423±.0385.872±.2080.826±.0932.899±.2893.524±.8304.075±.89417.9
ETTh1Correlational640.070±.0020.082±.0050.483±.0190.067±.0060.186±.0080.094±.0100.271±.00714.5
1280.072±.0020.088±.0050.188±.0060.054±.0070.203±.0060.113±.0120.176±.00618.4
2560.054±.0030.064±.0070.522±.0130.046±.0070.199±.0030.135±.0060.222±.01015.8
ETTh1Discriminative640.087±.0170.106±.0480.227±.0780.171±.1420.254±.0740.150±.0030.296±.34818.1
1280.120±.0230.144±.0600.188±.0740.154±.0870.274±.0470.176±.0150.451±.08016.8
2560.048±.0110.060±.0300.442±.0560.178±.0760.304±.0680.243±.0050.461±.01020.6
ETTh1Predictive640.098±.0030.116±.0000.132±.0080.118±.0040.133±.0080.118±.0040.135±.00315.6
1280.087±.0030.110±.0030.153±.0140.113±.0050.129±.0030.120±.0080.126±.00120.8
2560.090±.0060.109±.0130.220±.0080.110±.0270.132±.0010.118±.0030.129±.00017.1
EnergyContext-FID640.136±.0140.135±.0171.230±.0702.662±.0872.697±.4180.762±.1571.824±.144-0.5
1280.084±.0150.087±.0192.535±.3723.125±.1065.552±.5281.344±.1311.822±.2713.6
2560.122±.0190.126±.0245.032±.8313.768±.9985.572±.5844.735±.7292.533±.4673.4
EnergyCorrelational640.618±.0120.672±.0353.668±.1061.653±.2086.847±.0831.281±.2183.319±.0628.1
1280.426±.0310.451±.0794.790±.1161.820±.3296.663±.1121.376±.2013.713±.0555.4
2560.341±.0390.361±.0924.487±.2141.279±.1145.690±.1021.800±.1383.739±.0895.6
EnergyDiscriminative640.066±.0050.078±.0210.498±.0010.499±.0000.497±.0040.328±.0310.499±.00115.0
1280.127±.0380.143±.0750.499±.0010.499±.0000.499±.0010.396±.0240.499±.00111.4
2560.252±.0470.290±.1230.499±.0000.499±.0000.499±.0000.437±.0950.498±.00413.2
EnergyPredictive640.225±.0010.249±.0000.291±.0030.302±.0010.252±.0010.252±.0000.262±.0029.6
1280.221±.0010.247±.0010.303±.0020.318±.0000.252±.0000.251±.0000.269±.00210.7
2560.214±.0010.245±.0010.351±.0040.353±.0030.251±.0000.251±.0000.275±.00412.8

[1] Xinyu Yuan and Yan Qiao. Diffusion-ts: Interpretable diffusion for general time series generation. In ICLR, 2024.

评论

Q5: Different from the statement around key contributions, the idea of denoising with condition has already been studied in many prior works, including [Fan et al.], [Shen et al., 2024], [Yuan and Qiao, 2024]. [Fan et al.] MG-TSD: Multi-Granularity Time Series Diffusion Models with Guided Learning Process. In ICLR, 2024.

A5: Thank you for your insightful comments. We appreciate the opportunity to clarify the uniqueness of our work compared to existing studies like MG-TSD and Dish-TS. While prior works often focus on domain-specific models tailored to individual datasets, limiting their generalization, our approach introduces a unified time series diffusion model designed to excel across diverse and unseen domains. Here's a summary of our contributions and innovations:

  1. Unified Time Series Diffusion (UTSD) Model: For the first time, we model multi-domain probability distributions using diffusion's robust capabilities. Unlike autoregressive models, our approach employs a diffusion denoising process to handle cross-domain data, generating predictions directly through conditional sampling.
  2. Superior Generalization: The pre-trained UTSD surpasses existing foundation models across various data domains, demonstrating exceptional zero-shot generalization. It matches domain-specific models when trained from scratch, validating its potential as a foundational time series model.
  3. Flexible Architecture: UTSD supports arbitrary input and output time steps, thanks to a patch-transpose operation and an adapter module. This allows for seamless alignment of multiscale representations, enabling the generation of forecasting sequences without a dedicated forecasting head.
  4. Efficient Fine-Tuning: Our model's architecture supports efficient 'plug-and-play' fine-tuning. The pre-trained Condition Net captures generalized patterns, while the Denoising Net reconstructs target domain sequences. The Adapter bridges the unified and proprietary representation spaces, minimizing computational overheads.
  5. Multi-Scale Feature Capture: UTSD extracts features at different scales across various domains, capturing both trend information and complex multi-periodic patterns. This contrasts with traditional models that focus on single-scale variables.

We believe these innovations distinguish our work from existing models and contribute significantly to the field of time series analysis.

In addition, extensive experimental results have shown that our UTSD achieves state-of-the-art integrated performance compared to existing conditional diffusion models. We validate the predictive performance of the model on including the most widespreadly utilized Electricity and Traffic datasets. And compared with the baseline model DiffusionTS on Solar Forecasting and MoJoCo Imputation, the proposed UTSD achieves the best performance on almost all benchmarks.

DatasetsMetricsUTSD across datasets (MSE)UTSD across datasets (MAE)UTSD from scratch (MSE)UTSD from scratch (MAE)Diffusion-TS (MSE)Diffusion-TS (MAE)
ETTh2960.3210.3620.2410.3010.4410.417
1920.4170.4250.2750.3750.5060.473
3360.4260.4370.3020.3720.5960.554
7200.4730.4740.3230.3860.6240.561
Avg.0.4090.4250.2850.3580.5550.506
ETTm2960.1950.2890.1910.2840.3190.371
1920.2410.3220.2210.3060.4150.421
3360.2860.3450.2350.3140.4310.439
7200.3710.4040.2830.3530.4530.477
Avg.0.2730.3400.2330.3140.4490.481
Electricity960.1830.2980.1280.2210.2460.282
1920.1920.3020.1470.2400.2610.292
3360.2030.2980.1490.2440.3220.335
7200.2300.3330.1720.2720.4300.407
Avg.0.2020.3080.1490.2440.3580.375
Traffic960.3090.2140.2840.2030.4350.299
1920.3200.2400.2930.2110.4450.309
3360.3280.2410.3080.2150.4490.320
7200.3310.2600.3190.2230.4600.339
Avg.0.3220.2390.3010.2130.4270.297
Solar ForecasintgUTSDDiffusion-TSSSSDCSDITLAETransMAFGP-copula
MSE2.83e2±0.51e13.75e2±3.6e15.03e2±1.06e19.0e2±6.1e16.8e2±7.5e19.30E+029.8e2±5.2e1
MuJoCo ImputationUTSDDiffusion-TSSSSDCSDINRTSINAOMILatent-ODENeuralCDEODE-RNNRNN GRU-D
70% Mask0.25(2)0.37(3)0.59(8)0.24(3)0.631.4638.359.8611.34
80% Mask0.37(2)0.43(3)1.00(5)0.61(10)1.222.322.9510.7112.0914.21
90% Mask0.48(5)0.73(12)1.90(3)4.84(2)4.064.423.613.5216.4719.68
评论

Thank you for your thoughtful review and valuable feedback. We appreciate recognition of our innovative work in incorporating diffusion models into time series foundation models. We response your concerns as following.

Q1-Q3: Typos and formatting issues.

A: We appreciate your feedback regarding the typos and formatting issues. We have corrected the spelling error in the abstract, ensured consistent text size throughout, and properly formatted all citations. We have carefully reviewed the entire manuscript to enhance the overall writing quality. Thank you for your valuable input.

Q4: There are false statements when criticizing prior works to motivate the proposed method. For example, while UniTime [Liu et al., 2024a] and Moirai [Woo et al., 2024] are non-autoregressive, authors claim that they "rely on an auto-regressive mechanism" in L085-088.

A4: We apologize for the confusion in our original manuscript. We have clarified the discussion of prior works as follows:

Specifically, the existing multi-domain generalization methods (e.g., Timer) rely on autoregressive mechanisms to establish connections between observed and predicted sequences. These methods are susceptible to error accumulation in long sequence prediction and often struggle to capture cross-domain relationships. In contrast, other models, such as UniTime and Moirai, endeavor to directly establish the projection between observation and forecast sequences. However, sequences from different domains differ significantly, and the projection relationships learned during training are not flexible enough to be directly applied to unseen domains. This limits the zero-shot and cross-domain generalization capabilities of the model.

We have highlighted this clarification in blue in the revised manuscript. Thank you for your insightful comments.

评论

Dear Reviewer eVfc,

Thank you very much for devoting time on reviewing our manuscript. As the author-reviewer discussion process is ending soon, we are wondering whether our responses have well addressed your concerns. If you have any further questions regarding our manuscript, please let us know and we are glad here to provide further discussion and clarification to improve the quality of this manuscript.

Thank you for your contributions,

Authors of Paper 2173

评论

Thank you for your response. While I feel that the response is somewhat verbose, I see that authors addressed some of my concerns. Below I leave additional comments:

  1. The revision around L88-95 still does not read well, maybe it has some grammatical issue.

9-1. Could you elaborate on the difference between foundation models and the proposed unified model? Specifically, I think "being able to utilise multiple (small) datasets from different domains for joint pre-training" sounds like a shared idea with foundation models, and I don't understand why you specifically mentioned small datasets here? Also, if you claim foundation models and your unified model are different, could you justify how experimental comparisons between them (e.g., in Table 2) are fair?

9-2. For Table 2, I understand your experiment is based on UniTime. However, I cross-checked the UniTime paper and found that the list of datasets is not matched. Why is Exchange missing in Table 2 while described in Section 4, and why is Illness never appeared?

9-3. As you are submitting your work to an ML venue, I personally recommend borrowing experimental settings from papers published in ML venues. For example, TSDiff published in NeurIPS 2023 experimented 8 benchmarks, providing a strong baseline. Note that I am not asking to conduct a new experiment during this review process, as it might not be feasible.

评论

Q4: As you are submitting your work to an ML venue, I personally recommend borrowing experimental settings from papers published in ML venues. For example, TSDiff published in NeurIPS 2023 experimented 8 benchmarks, providing a strong baseline. Note that I am not asking to conduct a new experiment during this review process, as it might not be feasible.

A4: Thank you for your valuable suggestion. We have evaluated the forecasting performance of UTSD on the benchmarks related to TSDiff (Electricity, Traffic, Exchange), with the results presented in Table 2 (page 8). However, due to the differences in experimental settings (TSDiff uses a prediction length of 24, whereas the commonly used prediction lengths for long-term forecasting tasks are {96, 192, 336, 720}), we did not include a direct comparison with TSDiff on long-term time series forecasting to maintain consistency with the majority of baselines.

Additionally, based on your suggestion, we conduct comparison experiments between UTSD and TSDiff on the Solar, UberTLC, M4, and Wikipedia datasets. The results of these experiments are presented in the table below.

DatasetsContextPredictionUTSDTSDiff-QTSDiff-MSTSDiff-CondCSDITFTTransformerDeepStateMQ-CNNDeepAR
Solar336240.290±0.0070.358±0.0200.391±0.0030.338±0.0140.352±0.0050.417±0.0230.419±0.0080.379±0.0020.790±0.0630.389±0.001
UberTLC336240.175±0.0020.172±0.0050.183±0.0070.172±0.0080.206±0.0020.193±0.0060.192±0.0040.288±0.0870.436±0.0200.161±0.002
M4312480.036±0.0010.036±0.0010.045±0.0000.039±0.0060.040±0.0030.039±0.0010.040±0.0140.041±0.0020.046±0.0030.052±0.006
Wikipedia360300.186±0.0010.221±0.0010.257±0.0010.218±0.0100.289±0.0170.229±0.0060.214±0.0010.318±0.0190.220±0.0010.231±0.008

The experimental results demonstrate that UTSD achieves the best performance on three out of four datasets (Solar, M4, and Wikipedia), while TSDiff-Q slightly outperforms UTSD on the UberTLC dataset. This highlights the effectiveness and robustness of UTSD across diverse benchmarks.

We sincerely appreciate your valuable feedback, which has been instrumental in improving the quality of our manuscript. If you have any additional concerns, we would be more than happy to address them.

评论

Thank you for your positive response and your willingness to provide constructive feedback! We believe that addressing these questions will greatly help readers gain a deeper and more comprehensive understanding of our work. We sincerely appreciate your thoughtful and insightful comments. Below, we provide detailed responses to address your concerns:

Q1: The revision around L88-95 still does not read well, maybe it has some grammatical issue.

A1: Thank you for pointing this out. We have carefully reviewed lines 88–95 and made further revisions to resolve any grammatical issues and enhance the overall readability. We kindly invite you to review the updated manuscript.

Q2: Could you elaborate on the difference between foundation models and the proposed unified model? Specifically, I think "being able to utilise multiple (small) datasets from different domains for joint pre-training" sounds like a shared idea with foundation models, and I don't understand why you specifically mentioned small datasets here? Also, if you claim foundation models and your unified model are different, could you justify how experimental comparisons between them (e.g., in Table 2) are fair?

A2: Thank you for your thoughtful question. We appreciate the opportunity to elaborate on the distinction between foundation models and our proposed unified model, as well as to clarify the reasoning behind specifically mentioning small datasets and the fairness of our comparisons.

1. Difference Between Foundation Models and the Unified Model: Foundation models are typically pre-trained on extremely large-scale datasets, often comprising billions of examples from diverse domains. These models rely on such massive datasets to generalize effectively across tasks and domains. In contrast, our unified model is designed to jointly utilize multiple smaller datasets from different domains for pre-training without relying on a single, massive dataset. This approach allows us to efficiently leverage domain-specific information from smaller datasets while maintaining the flexibility to generalize across domains. In this sense, the unified model emphasizes the integration of smaller, diverse datasets in a principled way, rather than relying on the sheer scale of data, which is a hallmark of foundation models.

2. Why Mention Small Datasets: The emphasis on small datasets highlights the practical advantage of our approach. In many real-world scenarios, obtaining a massive, high-quality dataset akin to those used for training foundation models is infeasible. Our model addresses this limitation by effectively combining and learning from smaller, more accessible datasets, which may not individually cover the breadth or size of a typical foundation model's training data.

3. Fairness of Experimental Comparisons: We acknowledge that foundation models and our unified model have different design philosophies and data requirements, which may raise concerns about the fairness of comparisons. To ensure fairness, we have made efforts to align the experimental settings as closely as possible. Specifically,

  • We fine-tuned foundation models on the same datasets used for our unified model to ensure a consistent basis for evaluation.
  • The evaluation metrics and benchmarks were identical across all models to provide a direct comparison of their performance.

Q3: For Table 2, I understand your experiment is based on UniTime. However, I cross-checked the UniTime paper and found that the list of datasets is not matched. Why is Exchange missing in Table 2 while described in Section 4, and why is Illness never appeared?

A3: Thank you for your detailed review. Accroding to the statement in the PatchTST[1] paper, ‘Exchange-rate is the daily exchange-rate of eight different countries,’ and they proved that by simply repeating the last value in the look-back window, the MSE loss on exchange-rate dataset can outperform or be comparable to the best results. And for the ILI dataset, it has a completely different sampling frequency (sampling frequency of 7 days) compared to ETTm (sampling frequency of 15 min), Weather (sampling frequency of 10 min), and Electricity (sampling frequency of 1h), which means that the ILI consists of nearly long-term trends without the traditional temporal fluctuation patterns. Based on the above discussion, we decided to exclude the Exchange-rate and ILI datasets from our experiments. Additionally, several recent works (iTransformer[2], TimeXer[3] and TimeMixer[4]) also includes the same benchamrks as UTSD on long-term forecasting tasks.

[1] A time series is worth 64 words: Long-term forecasting with transformers. ICLR 2023.

[2] itransformer: Inverted transformers are effective for time series forecasting. ICLR 2024.

[3] Timexer: Empowering transformers for time series forecasting with exogenous variables. NeurIPS 2024.

[4] Timemixer: Decomposable multiscale mixing for time series forecasting. ICLR 2024.

评论

Dear Reviewer eVfc:

Thank you once again for your valuable and constructive review. As the author-reviewer discussion period is coming to a close, we kindly request your feedback on our rebuttal provided above. Please let us know if our responses have sufficiently addressed your concerns. If there are any remaining issues, we would be happy to engage in further discussion, as we believe that an in-depth exchange will help strengthen our paper.

Best regards!

Authors

审稿意见
5

The model presents a Unified Time Series Diffusion (UTSD) model, which is established for the first time to model the multi-domain probability distribution, utilizing the powerful probability distribution modeling ability of Diffusion.

优点

1 The condition-denoising architecture is novel. 2 The performance improvement compared to listed baselines. 3 The paper presents cross-domain generalizability.

缺点

1 The comparison did not include some interesting baselines. 2 The cross-domain transfer is limited to similar datasets. 3 I did not see obvious advantages compared to DiffusionTS.

问题

1 The evaluation is very limited int he time series prediction. As listed in DiffusionTS, we should include other different metrics (like Discriminative Score and so on) to evaluate the performance. MSE/MAE evaluation is very biased evaluation metrics. For example, a straight average line achieve reasonable performances but is not meaningful. 2 The model should compare with DiffusionTS in Table 1/2, which I did not see in this paper. I would highly recommend the comparison since it is also diffusion based methods. 3 DiffusionTS can handle both conditional and unconditional generation. Is there a way for this method also be applicable to unconditional time series generation? 4 The paper claims this paper to be a foundation model for TS area. I would like to see the performance of a short forcasting model applied to long context prediction and vice versa. 5 For Table 3, the performance of ETHh1 to ETHh2 should be needed. 6 The performances of the other two datasets: Solar Forecasting and Mujoco Imputation should be included since it is widely benchmarked.

评论

Q6: The performances of the other two datasets: Solar Forecasting and Mujoco Imputation should be included since it is widely benchmarked.

A6: Thank you for your valuable feedback. We conducted forecasting and imputation experiments on the Solar dataset and MuJoCo dataset as you suggested, respectively, based on the publicly available code provided by DiffusionTS[1], and the detailed experimental results are shown in the following table. To ensure the fairness of the experimental results, the exact same data preprocessing operation and metrics as in DiffusionTS were utilized. Specifically, our approach achieves the best performance on the Solar forecasting task, 80% masking and 90% masking interpolation tasks. Surprisingly, compared with the state-of-the-art DiffusionTS, UTSD achieves an improvement of up to 24.5% on the prediction task, which demonstrates the superior prediction capability of the proposed method.

Solar ForecasintgUTSDDiffusion-TSSSSDCSDITLAETransMAFGP-copula
MSE2.83e2±0.51e13.75e2±3.6e15.03e2±1.06e19.0e2±6.1e16.8e2±7.5e19.30E+029.8e2±5.2e1
MuJoCo ImputationUTSDDiffusion-TSSSSDCSDINRTSINAOMILatent-ODENeuralCDEODE-RNNRNN GRU-D
70% Mask0.25(2)0.37(3)0.59(8)0.24(3)0.631.4638.359.8611.34
80% Mask0.37(2)0.43(3)1.00(5)0.61(10)1.222.322.9510.7112.0914.21
90% Mask0.48(5)0.73(12)1.90(3)4.84(2)4.064.423.613.5216.4719.68

[1] Xinyu Yuan and Yan Qiao. Diffusion-TS: Interpretable diffusion for general time series generation. In The Twelfth International Conference on Learning Representations, 2024.

Thank you very much for your constructive comments and insightful feedback. Your valuable suggestions are immensely helpful for enhancing the quality of our manuscript. We will incorporate and address all relevant points in the revision.

评论

Q4: The paper claims this paper to be a foundation model for TS area. I would like to see the performance of a short forcasting model applied to long context prediction and vice versa.

A4: Thank you for your insightful feedback. We appreciate your interest in evaluating the model's adaptability across different forecasting scenarios. We introduce further experiments about ‘short-term models applied to long-term forecasts’ and ‘long-term models applied to short-term forecasts’. Specifically, ETTm1, 336 \to 96 is defined as ‘Long-term forecasting model applying short-context time series’, and correspondingly ETTm1, 96 \to 336 is defined as ‘Short-term forecasting model applying long-context time series’. Results, including the largest Traffic dataset, demonstrate that our model demonstrates state-of-the-art performance compared to other methods. Detailed results are presented in the table below.

DatasetMetricUTSD (MSE)UTSD (MAE)TimeLLM (MSE)TimeLLM (MAE)LLM4TS (MSE)LLM4TS (MAE)GPT4TS (MSE)GPT4TS (MAE)PatchTST (MSE)PatchTST (MAE)TimesNet (MSE)TimesNet (MAE)
ETTm196 \to 3360.3260.3810.3840.4110.3780.4200.4210.4430.4420.4550.4620.461
ETTm1336 \to 960.2990.3330.2720.3340.2850.3430.2920.3460.3440.3730.3380.375
ETTm296 \to 3360.2450.3270.2950.3610.2950.3560.3200.3840.3420.3850.3620.398
ETTm2336 \to 960.1910.2840.1610.2530.1650.2540.1730.2620.1770.2600.1870.267
ETTh196 \to 3360.3990.4390.4620.4660.4530.4540.4910.4910.5590.5220.5510.537
ETTh1336 \to 960.2740.3010.3620.3920.3710.3940.3760.3970.4040.4130.3840.402
ETTh296 \to 3360.3120.3820.3990.4440.3820.4320.4260.4520.4870.4950.5100.518
ETTh2336 \to 960.2410.3010.2680.3280.2620.3320.2850.3420.3120.3580.3400.374
Electricity96 \to 3360.1530.2580.1750.2670.1790.2770.1900.2950.2300.3250.2230.338
Electricity336 \to 960.1280.2210.1310.2240.1280.2230.1390.2380.1860.2690.1680.272
Traffic96 \to 3360.3200.2230.4190.2960.4410.3020.4620.3340.4370.2980.7090.376
Traffic336 \to 960.2840.2030.3620.2480.3720.2590.3880.2820.3600.2490.5930.321
Weather96 \to 3360.2160.2660.2850.2990.2620.3000.2920.3180.3100.3340.3200.349
Weather336 \to 960.1330.1950.1470.2010.1470.1960.1620.2120.1770.2180.1720.220

[1] Yong Liu, Haoran Zhang, Chenyu Li, Xiangdong Huang, Jianmin Wang, and Mingsheng Long. Timer: Generative pre-trained transformers are large time series models. In ICML, 2024c.

[2] Gerald Woo, Chenghao Liu, Akshat Kumar, Caiming Xiong, Silvio Savarese, and Doyen Sahoo. Unified training of universal time series forecasting transformers. In ICML, 2024.

[3] Mononito Goswami, Konrad Szafer, Arjun Choudhry, Yifu Cai, Shuo Li, and Artur Dubrawski. Moment: A family of open time-series foundation models. In ICML, 2024.

Q5: For Table 3, the performance of ETHh1 to ETHh2 should be needed.

A5: As your suggestion, we report the comparative results on ETTh1\toETTh2 in the following Table. Concretely, the average MSE is reduced by 22.7%, 32.8% and 28.1% compared to the existing TimeLLM, GPT4TS and PatchTST. For convenience, the following table shows the comprehensive performance (MSE) of the UTSD and the pre-existing model.

DatasetMetricUTSD (MSE)UTSD (MAE)TimeLLM (MSE)TimeLLM (MAE)LLMTime (MSE)LLMTime (MAE)GPT4TS (MSE)GPT4TS (MAE)DLinear (MSE)DLinear (MAE)PatchTST (MSE)PatchTST (MAE)
ETTh1\toETTh2960.2590.3210.2790.3370.5100.5760.3350.3740.3470.4000.3040.350
1920.2950.3490.3510.3740.5230.5860.4120.4170.4470.4600.3860.400
3360.3210.3710.3880.4150.6400.6370.4410.4440.5150.5050.4140.428
7200.3360.4320.3910.4202.2961.0340.4380.4520.6650.5890.4190.443
Avg.0.3030.3680.3530.3870.9920.7080.4060.4220.4930.4880.3800.405
评论

Q2: The model should compare with DiffusionTS in Table 1/2, which I did not see in this paper. I would highly recommend the comparison since it is also diffusion based methods.

A2: Thank you for your constructive suggestions. We adjust the DiffusionTS to time series forecasting task in comparison with Table 1/2 as you suggested, and the results are reported in following tabel.

DatasetsMetricsUTSD across datasets (MSE)UTSD across datasets (MAE)UTSD from scratch (MSE)UTSD from scratch (MAE)Diffusion-TS (MSE)Diffusion-TS (MAE)
ETTh2960.3210.3620.2410.3010.4410.417
1920.4170.4250.2750.3750.5060.473
3360.4260.4370.3020.3720.5960.554
7200.4730.4740.3230.3860.6240.561
Avg.0.4090.4250.2850.3580.5550.506
ETTm2960.1950.2890.1910.2840.3190.371
1920.2410.3220.2210.3060.4150.421
3360.2860.3450.2350.3140.4310.439
7200.3710.4040.2830.3530.4530.477
Avg.0.2730.3400.2330.3140.4490.481
Electricity960.1830.2980.1280.2210.2460.282
1920.1920.3020.1470.2400.2610.292
3360.2030.2980.1490.2440.3220.335
7200.2300.3330.1720.2720.4300.407
Avg.0.2020.3080.1490.2440.3580.375
Traffic960.3090.2140.2840.2030.4350.299
1920.3200.2400.2930.2110.4450.309
3360.3280.2410.3080.2150.4490.320
7200.3310.2600.3190.2230.4600.339
Avg.0.3220.2390.3010.2130.4270.297

As shown in the table above, UTSD exhibits the best overall performance compared to the existing time series foundation and proprietary models, as well as the diffusion baseline model DiffusionTS. This is due to the fact that UTSD includes a number of enhanced conditional prediction components, whereas DiffusionTS is used as a baseline model for the generation task, and thus DiffusionTS does not show a clear advantage on the forecasting task.

Q3: DiffusionTS can handle both conditional and unconditional generation. Is there a way for this method also be applicable to unconditional time series generation?

A3: Thank you for your insightful comment. UTSD also supports the unconditional cross-domain generation task, the implementation of which is demonstrated in Figure 8 (Page. 18 in Revision).

The unconditional generation paradigm of UTSD can be divided into two phases:

  1. In the multi-domain pre-training phase, time series Y1:HTY_{1:H}^{T} from different domain backgrounds are successively obtained a set of noised sequence {Y1:H1,...,Y1:HT1,Y1:HT}\begin{Bmatrix}Y_{1:H}^1,...,Y_{1:H}^{T-1},Y_{1:H}^T\end{Bmatrix}. In each iterative step, the noised sequences Y1:HTY_{1:H}^T as the input of encoder block-a,b,c and a set of multi-scale condition variables is captured from it, and the rich fluctuation pattern information embedded is used to reconstruct the denoised result Y1:HT1Y_{1:H}^{T-1} in Decoder;
  2. In the downstream data domain fine-tuning phase, the pre-training weights in Encoder and Decoder are frozen. An additional set of adapter block and zero convolution is introduced to align the multi-domain uniform representation space with the specified domain representation space.

In UTSD-generation, the trainable Adapter is connected to the locked pre-trained weights. The zero convolution with weights initialised to zero are designed to ensure that they grow progressively during training. This architecture ensures that harmful noise is not added to the deep features of large diffusion models at the start of training, and protects the large-scale pre-trained backbone in the trainable replicas from such noise.

To further validate the performance of UTSD on the unconditional generation task, we also follow the design of DiffusionTS and demonstrate the leading performance of the proposed method on two datasets, ETTh1 and Energy. Specific experimental results are shown in A1.

评论

Thank you for your thoughtful review and valuable feedback. We response your concerns as following.

Q1: The evaluation is very limited int he time series prediction. As listed in Diffusion-TS, we should include other different metrics (like Discriminative Score and so on) to evaluate the performance. MSE/MAE evaluation is very biased evaluation metrics. For example, a straight average line achieve reasonable performances but is not meaningful.

A1: We appreciate the suggestion to incorporate a broader range of evaluation metrics beyond MSE and MAE. In view of this, we refer to the generative-based metrics in Diffusion-TS[1] to design additional experiments for evaluating the performance of the proposed UTSD on the time series generation task. A comparison of the performance of UTSD and other diffusion baselines on the time-series generation task is shown in the following Table. Specifically, the generation metrics Context-FID Score, Correlational Score, Discriminative Score and Predictive Score is reduced by 18.2%, 16.2%, 18.5% and 17.8% on ETTh dataset compared to the existing Diffusion-TS. Besides, compared with Diffusion-TS, the average improvement of all indicators of UTSD on ETTh and Energy datasets is 17.6% and 8.2%, respectively.

DatasetMetrics (scores)LengthUTSDDiffusion-TSTimeGANTimeVAEDiffwaveDiffTimeCot-GANImprove(%)
ETTh1Context-FID640.522±.0310.631±.0581.130±.1020.827±.1461.543±.1531.279±.0833.008±.27717.2
1280.633±.0290.787±.0621.553±.1691.062±.1342.354±.1702.554±.3182.639±.42719.5
2560.347±.0100.423±.0385.872±.2080.826±.0932.899±.2893.524±.8304.075±.89417.9
ETTh1Correlational640.070±.0020.082±.0050.483±.0190.067±.0060.186±.0080.094±.0100.271±.00714.5
1280.072±.0020.088±.0050.188±.0060.054±.0070.203±.0060.113±.0120.176±.00618.4
2560.054±.0030.064±.0070.522±.0130.046±.0070.199±.0030.135±.0060.222±.01015.8
ETTh1Discriminative640.087±.0170.106±.0480.227±.0780.171±.1420.254±.0740.150±.0030.296±.34818.1
1280.120±.0230.144±.0600.188±.0740.154±.0870.274±.0470.176±.0150.451±.08016.8
2560.048±.0110.060±.0300.442±.0560.178±.0760.304±.0680.243±.0050.461±.01020.6
ETTh1Predictive640.098±.0030.116±.0000.132±.0080.118±.0040.133±.0080.118±.0040.135±.00315.6
1280.087±.0030.110±.0030.153±.0140.113±.0050.129±.0030.120±.0080.126±.00120.8
2560.090±.0060.109±.0130.220±.0080.110±.0270.132±.0010.118±.0030.129±.00017.1
EnergyContext-FID640.136±.0140.135±.0171.230±.0702.662±.0872.697±.4180.762±.1571.824±.144-0.5
1280.084±.0150.087±.0192.535±.3723.125±.1065.552±.5281.344±.1311.822±.2713.6
2560.122±.0190.126±.0245.032±.8313.768±.9985.572±.5844.735±.7292.533±.4673.4
EnergyCorrelational640.618±.0120.672±.0353.668±.1061.653±.2086.847±.0831.281±.2183.319±.0628.1
1280.426±.0310.451±.0794.790±.1161.820±.3296.663±.1121.376±.2013.713±.0555.4
2560.341±.0390.361±.0924.487±.2141.279±.1145.690±.1021.800±.1383.739±.0895.6
EnergyDiscriminative640.066±.0050.078±.0210.498±.0010.499±.0000.497±.0040.328±.0310.499±.00115.0
1280.127±.0380.143±.0750.499±.0010.499±.0000.499±.0010.396±.0240.499±.00111.4
2560.252±.0470.290±.1230.499±.0000.499±.0000.499±.0000.437±.0950.498±.00413.2
EnergyPredictive640.225±.0010.249±.0000.291±.0030.302±.0010.252±.0010.252±.0000.262±.0029.6
1280.221±.0010.247±.0010.303±.0020.318±.0000.252±.0000.251±.0000.269±.00210.7
2560.214±.0010.245±.0010.351±.0040.353±.0030.251±.0000.251±.0000.275±.00412.8

[1] Xinyu Yuan and Yan Qiao. Diffusion-TS: Interpretable diffusion for general time series generation. In The Twelfth International Conference on Learning Representations, 2024.

评论

Dear Reviewer YBBT,

Thank you very much for devoting time on reviewing our manuscript. As the author-reviewer discussion process is ending soon, we are wondering whether our responses have well addressed your concerns. If you have any further questions regarding our manuscript, please let us know and we are glad here to provide further discussion and clarification to improve the quality of this manuscript.

Thank you for your contributions,

Authors of Paper 2173

评论

Dear Reviewer YBBT,

Thank you very much for your time and valuable comments on our manuscript. It is heartening to know that a modified version is ready with the help of valuable comments from reviewers, and we would like to know if our modified version has solved your problem. Our improvements are centered around the following suggestions:

  1. Adding additional evaluation metrics: To evaluate the quality of time series in the generation task, we refer to the metrics from DiffusionTS and the full experimental results are presented in Table 12, Page 22 of the revision. Specifically, the generation metrics Context-FID Score, Correlational Score, Discriminative Score and Predictive Score is reduced by 18.2%, 16.2%, 18.5% and 17.8% on ETTh dataset compared to DiffusionTS. More experimental analysis can be found in L1167-1170.
  2. Introducing additional baselines and benchmarks: We adjust the DiffusionTS to time series forecasting task in comparison with Table 1/2 as reviewer suggested, and the results are reported in A2,A4,A6.
  3. Applying UTSD to unconditional generation: In uncondi-UTSD, the pre-training weights in Encoder and Decoder are frozen, and an additional set of adapter block and zero convolution is introduced to align the multi-domain uniform representation space with the specified domain representation space. We show the diagram of uncondi-UTSD in the modified Figure 2, and uncondi-UTSD shows better unconditional generation performance on the two datasets ETTh and Energy, in Table 12, Page 22.
  4. Adding additional zero-shot prediction experiments: As reviewer's suggestion, we report the comparative results on ETTh1 and ETTh2 in the Table 3,9, Page 8,20. Concretely, the average MSE is reduced by 22.7%, 32.8% and 28.1% compared to the existing TimeLLM, GPT4TS and PatchTST.

Sincerely thank you for your contribution and responsibility. If you have any further concerns about our manuscript, please let us know and we are happy to provide further discussion here.

Author of paper 2173

评论

Dear Reviewer YBBT:

Thank you once again for your valuable and constructive review. As the author-reviewer discussion period is coming to a close, we kindly request your feedback on our rebuttal provided above. Please let us know if our responses have sufficiently addressed your concerns. If there are any remaining issues, we would be happy to engage in further discussion, as we believe that an in-depth exchange will help strengthen our paper.

Best regards!

Authors

审稿意见
6

This paper introduces UTSD (Unified Time Series Diffusion), a novel diffusion-based model for multi-domain time series forecasting. The key contributions include:

  • A condition-denoising architecture that captures multi-scale temporal patterns
  • Direct modeling in actual sequence space rather than latent space
  • An improved classifier-free guidance strategy for conditional generation
  • A transfer-adapter module for efficient domain adaptation The model achieves state-of-the-art results on multiple benchmarks in different forecasting scenarios including across-domain training, zero-shot learning, and probabilistic forecasting.

优点

This study introduce an innovative condition-denoising architecture along with an actual-space diffusion and denoising process, complemented by a classifier-free guidance strategy for conditional generation. This study represents the first establishment of a foundational time series model leveraging diffusion techniques. The model demonstrates exceptional cross-domain generalization capabilities, indicating its potential to redefine paradigms in time series analysis.And the Unified Time Series Diffusion (UTSD) model achieves state-of-the-art (SOTA) results across various benchmarks. The originality of this model, coupled with its high-quality outcomes, clarity in methodology, and significant implications for the field, underscores the importance of this findings.

缺点

  • There are too few types of evaluation indicators, as a model generated by the diffuison process, more indicators such as those used in the DiffusionTS paper, such as Context-FID Score, Correlational Score, Discriminative Score, Predictive Score, etc.
  • Theoretical analysis of why the model works better in actual sequence space vs latent space could be stronger
  • Memory and computational complexity analysis is missing
  • Limited discussion of model limitations and failure cases
  • Some architectural choices (like number of blocks) need better justification
  • Comparison with some recent relevant works could be expanded

问题

  1. Can you please add evaluation indicators other than MSE and MAE?
  2. As the author said in the paper “the average MSE is reduced by 17.9%, 18.6% and 22.4% compared to the existing TimeLLM, LLM4TS and GPT4TS, which indicates that the proposed method can fully utilize a small amount of data for efficient training.” Could you please explain why a decrease in MSE means that less data can be used for training?
  3. In order to train a across-domain model, a mixed dataset needs to be established. However, will the training cost increase after the dataset is enlarged?
  4. How does the computational complexity compare to existing approaches?
  5. What are the main limitations/failure cases of the model?
  6. How sensitive is the model to hyperparameter choices?
  7. Can the authors provide theoretical justification for why modeling in actual sequence space is better?
  8. How does the model perform on extremely long sequences?
评论

Q7: Can the authors provide theoretical justification for why modeling in actual sequence space is better?

A7: The recent Q-DM [1] confirms our view and proposes a develop a Timestep-aware Quantization (TaQ) method to effectively eliminate the accumulated quantization error caused by the multi-step denoising process.

To further validate our view, we designed a control experiment with the Latent-UTSD model for iterative denoising in the hidden representation space, which ‘trains the self-encoder in an end-to-end manner using a mixed multidomain synthetic dataset, and subsequently reconstructs the embeddings of the target sequences directly in the hidden space’. The experimental results in the following table show that Latent-UTSD exhibits relatively poor performance, which validates our point from an experimental perspective. For implementation details, refer to the appendix section Section C. For convenience, we show the comprehensive performance of Latent-UTSD on the four datasets in the following table.

MdoelsMetricsETTh1(96)ETTh1(192)ETTh1(336)ETTh1(720)ETTh1(Avg)ETTm1(96)ETTm1(192)ETTm1(336)ETTm1(720)ETTm1(Avg)
UTSDMSE0.2740.2900.3830.3870.3340.2990.3040.3120.3170.308
Latent-UTSDMSE0.3230.3420.4500.4540.3920.3520.3650.3650.3700.363
UTSDMAE0.3010.3390.4240.4280.3830.3330.3580.3650.3680.356
Latent-UTSDMAE0.3480.3860.4930.4920.4300.3900.4180.4350.4430.422
MdoelsMetricsTraffic(96)Traffic(192)Traffic(336)Traffic(720)Traffic(Avg)Weather(96)Weather(192)Weather(336)Weather(720)Weather(Avg)
UTSDMSE0.2840.2930.3080.3190.3010.1330.1840.2070.2640.202
Latent-UTSDMSE0.3140.3160.3410.3430.3290.1540.2110.2380.3130.229
UTSDMAE0.2030.2110.2150.2230.2130.1950.2370.2580.3130.246
Latent-UTSDMAE0.2230.2360.2420.2420.2350.2260.2710.2970.3590.288

This is due to differences in the focus of the inductive bias to capture patterns between the two stages of training, e.g., the encoder loses some of the information critical to the reconstruction process during the projection of the original sequence space into the hidden representation space. Subsequently, in the iterative denoising process the model targets ‘incomplete hidden representations’ for reconstruction and therefore the noise reduction model is unable to capture all the fluctuating pattern distributions in each round of reconstruction, which results in a natural accumulation of errors in the latent spatial inference.

[1] Li, Yanjing et al. “Q-DM: An Efficient Low-bit Quantized Diffusion Model.” Neural Information Processing Systems (2023).

Q8: How does the model perform on extremely long sequences?

A8: Thank you for your valuable feedback. We introduce more visualisation results in the revision appendix section of the revised version, including the ‘extremely long term sequences’ as you suggested. Figure 10 (Page. 24 in Revision) shows real extremely long multi-periodic series sampled from the ECL and Traffic datasets, and in addition, Figure 11 (Page. 25 in Revision) shows real short period non-periodic sequences sampled from the ETT dataset. Specifically, UTSD also shows good performance in extremely long sequence prediction, which verifies that UTSD has the ability to capture extremely long term dependencies, which is crucial for practical applications. In addition, UTSD also shows satisfactory prediction results in short-term non-periodic sequences.

评论

Q5: What are the main limitations/failure cases of the model?

A5: We appreciate you raising this important question about UTSD's limitations. Our experiments revealed several key areas where the model faces challenges:

High arithmetic overhead and time cost. As mentioned earlier, compared to traditional deep models, diffusion models have the defect of unstable and slow convergence during training and require multiple rounds of iterative denoising processes during inference. These drawbacks lead to the fact that diffusion-based time-series prediction models require high arithmetic overhead and time cost, both in the training and inference phases. However, UTSD exhibits lower computational overheads compared to existing time series foundation models, which due to three innovative designs that, patch-wise attention mechanism, iterative denoising in real sequence space and de-Markovised DDIM algorithm. Detailed discussion can be found in Appendix A4.

Challenges applied to discriminantive-based tasks. Diffusion models excel in generative-based paradigms, such as forecasting, imputation, and anomaly detection. However, compared to general-purpose deep models like TimesNet and Moment, they do not show significant advantages in discriminative tasks like classification. We anticipate that the pre-trained UTSD should not only generate high-quality prediction samples but also possess strong representational learning and discriminative capabilities. This is crucial for developing a "multi-task generalized spatiotemporal diffusion model," which will be the focus of our future work.

Q6: How sensitive is the model to hyperparameter choices?

A6: In the UTSD architecture, all components are composed four minimal blocks (encoder block, decoder block, middle block and adapter block). The basic construction of each block is illustrated in Figure 2, however, in the concrete implementation we allow consecutive L-layer block residues to be stacked together to form stacked-block and used to form the Condition Net, Denoising Net and Adapter. Furthermore, the number of input and output channels of the middle block is denoted as D, and accordingly the number of output channels in the three pairs of encoder-decoder blocks are D4\frac D4, D2\frac D2, and DD, respectively. In order to verify the sensitivity of UTSD to hyperparameter selection, the following table shows the performance of UTSD with different parameter scales on multiple benchmarks. Where the model parameter combinations include (L,D) = (2,128), (2,256), (3,128), (3,256), (4,128), (4, 256), with a fixed forecasting window of 96, and a performance metric of MSE.

By default, the hyperparameters of UTSD are fixed to L=3 and D=256, and all experimental results presented follow this setting. In the scratch forecasting task, the number of output channels in the three pairs of Encoder-Decoder blocks is 64,128,256, respectively, and number of input and output channels of the Middle Block is fixed to 256.

HyperparameterETTh1ETTh2ETTm1ETTm2WeatherECLTraffic
(L=2,D=128)0.3520.3070.3350.2440.2150.1530.311
(L=2,D=256)0.3440.3020.3190.2540.2110.1580.325
(L=3,D=128)0.3280.2740.3210.2390.2030.1540.296
(L=3,D=256)0.3340.2850.3080.2330.2020.1490.301
(L=4,D=128)0.3220.2800.3020.2350.1980.1450.298
(L=4,D=256)0.3400.3130.3190.2340.2140.1600.306
评论

Q3: In order to train a across-domain model, a mixed dataset needs to be established. However, will the training cost increase after the dataset is enlarged?

A3: Thank you for your insightful comments. Indeed, all data-driven models, whether they are traditional deep models or recent predictive models based on diffusion theory, are challenged with large computational overheads and training costs as the size of the dataset becomes larger. However, this is a necessary cost for all pretrain fine-tune paradigms, including Large Language Models (GPT, Llama) and Large Visual Models (SD1.5, SDXL).

An encouraging news is that UTSD has made great efforts to reduce the time overhead during training and inference. Firstly, the pretrain phase occurs entirely offline, and after cross-dataset pretrained, the fine-tuned model weights can quickly converge to the desired state. In addition, UTSD's hybrid architecture supports naturally efficient fine-tuning through the ‘plug-and-play’ Adapter. Specifically, pre-training to obtain a Condition-Denoising Net with a large number of weights is completely frozen, and only a small number of weights in the Adapter component need to be optimised. The effectiveness of the fine-tuning strategy can be intuitively explained by the fact that the pre-trained Condition Net is responsible for capturing generic fluctuation patterns from observed sequences as conditional information, the Denoising Net is required to reconstruct sequence samples from noise in the target domain based on specific fluctuation patterns, and the fine-tuning-enabled Adapter is used to connect the unified representation space with the proprietary representation space.

Moreover, compared to other methods (TimeLLM, LLMTime, GPT4TS), our pretrain corpus is smaller (as in the previous A1), which further illustrates the efficiency and lost cost of our UTSD.

Q4: How does the computational complexity compare to existing approaches?

A4: Thanks to the reviewer for raising an issue that is critical to a realistic scenario. A detailed discussion of the computational complexity of UTSD is as follows:

The computational complexity of UTSD is O(((L2+H2)Pd+kPd(H+L))T)O((\frac{(L^2+H^2)}{P_d}+k\cdot P_d\cdot(H+L))\cdot T), where LL and HH denote the length of the input and output sequences, PdP_d is the dimension of the patch-wise tokenizer, KK denotes the depth of the encoder-decoder, and TT denotes the number of rounds of iterative denoising operation, respectively.

To address scalability concerns, we conducted experiments on runtime and memory usage (ETTh1 dataset, lookback window=336, forecast window=96). Results, including the large foundation time series model and diffusion model, demonstrate that our model requires less time and memory compared to other approaches. Detailed results are presented in the table below.

UTSD utilizes the de-Markovised DDIM algorithm, which means that UTSD only requires a smaller number of iterations, and thus the computational complexity of UTSD in the training and inference phases is far lower than that of the traditional temporal diffusion model. In addition, UTSD does not require the high-quality autoencoder, which greatly reduces the overall training overhead of UTSD.

Metrics(ETTh1-96)Total Param. (M)Mem. (MiB)Speed (s/iter)
UniTime439.5220740.335
TimeLLM3623.7145370.184
UTSD313.7918090.270
Metrics(ETTh1-96)Total Param. (M)Mem. (MiB)Speed (s/iter)
CSDI54.299830.245
DiffusionTS258.6415390.119
UTSD313.7918090.270
评论

Thank you for your thoughtful review and valuable feedback. We appreciate recognition of our innovative work in classifier-free guidance condition-denoising architecture, actual-space denoising process, and incorporating diffusion models into time series foundation models. We response your concerns as following.

Q1: Can you please add evaluation indicators other than MSE and MAE?

A1: Thank you for your valuable suggestions. To evaluate the quality of generated time series, we refer to the metrices (i.e., Context-FID, Correlational, Discriminative, and Predictive) from DiffusionTS and perform the comparison experiments. The results are shown in the following tables. Specifically, the generation metrics Context-FID Score, Correlational Score, Discriminative Score and Predictive Score is reduced by 18.2%, 16.2%, 18.5% and 17.8% on ETTh dataset compared to DiffusionTS. Besides, the average improvement of all indicators of UTSD on ETTh and Energy datasets is 17.6% and 8.2%, respectively.

DatasetMetrics (scores)LengthUTSDDiffusion-TSTimeGANTimeVAEDiffwaveDiffTimeCot-GANImprove(%)
ETTh1Context-FID640.522±.0310.631±.0581.130±.1020.827±.1461.543±.1531.279±.0833.008±.27717.2
1280.633±.0290.787±.0621.553±.1691.062±.1342.354±.1702.554±.3182.639±.42719.5
2560.347±.0100.423±.0385.872±.2080.826±.0932.899±.2893.524±.8304.075±.89417.9
ETTh1Correlational640.070±.0020.082±.0050.483±.0190.067±.0060.186±.0080.094±.0100.271±.00714.5
1280.072±.0020.088±.0050.188±.0060.054±.0070.203±.0060.113±.0120.176±.00618.4
2560.054±.0030.064±.0070.522±.0130.046±.0070.199±.0030.135±.0060.222±.01015.8
ETTh1Discriminative640.087±.0170.106±.0480.227±.0780.171±.1420.254±.0740.150±.0030.296±.34818.1
1280.120±.0230.144±.0600.188±.0740.154±.0870.274±.0470.176±.0150.451±.08016.8
2560.048±.0110.060±.0300.442±.0560.178±.0760.304±.0680.243±.0050.461±.01020.6
ETTh1Predictive640.098±.0030.116±.0000.132±.0080.118±.0040.133±.0080.118±.0040.135±.00315.6
1280.087±.0030.110±.0030.153±.0140.113±.0050.129±.0030.120±.0080.126±.00120.8
2560.090±.0060.109±.0130.220±.0080.110±.0270.132±.0010.118±.0030.129±.00017.1
EnergyContext-FID640.136±.0140.135±.0171.230±.0702.662±.0872.697±.4180.762±.1571.824±.144-0.5
1280.084±.0150.087±.0192.535±.3723.125±.1065.552±.5281.344±.1311.822±.2713.6
2560.122±.0190.126±.0245.032±.8313.768±.9985.572±.5844.735±.7292.533±.4673.4
EnergyCorrelational640.618±.0120.672±.0353.668±.1061.653±.2086.847±.0831.281±.2183.319±.0628.1
1280.426±.0310.451±.0794.790±.1161.820±.3296.663±.1121.376±.2013.713±.0555.4
2560.341±.0390.361±.0924.487±.2141.279±.1145.690±.1021.800±.1383.739±.0895.6
EnergyDiscriminative640.066±.0050.078±.0210.498±.0010.499±.0000.497±.0040.328±.0310.499±.00115.0
1280.127±.0380.143±.0750.499±.0010.499±.0000.499±.0010.396±.0240.499±.00111.4
2560.252±.0470.290±.1230.499±.0000.499±.0000.499±.0000.437±.0950.498±.00413.2
EnergyPredictive640.225±.0010.249±.0000.291±.0030.302±.0010.252±.0010.252±.0000.262±.0029.6
1280.221±.0010.247±.0010.303±.0020.318±.0000.252±.0000.251±.0000.269±.00210.7
2560.214±.0010.245±.0010.351±.0040.353±.0030.251±.0000.251±.0000.275±.00412.8

Detailed experimental results are shown in Table 12, page 22 of our revised manuscript.

Q2: Explain the confused statement about why a decrease in MSE means that less data can be used for training.

A2: Thank you for your careful review, and we apologize for any confusion the original presentation may have caused. As we have discussed in the related work, TimeLLM, LLM4TS and GPT4TS are based on pre-trained large language models that are fine-tuned for application in the time series domain, and thusthese models use a huge corpus (more than 15,000,000 million timesteps) in pre-train. In contrast, UTSD is only pre-trained on a mixed dataset (only 27.5 million timesteps) obtained by merging 7 small datasets, while the experimental results on the right hand side of Table 2 are obtained by pre-training UTSD from scratch on a specific single small dataset (0.12~1.52 million timesteps). Compared to TimeLLM, which uses a huge corpus pre-training and then aligns to the temporal space, UTSD shows excellent data utilisation and can learn sufficiently data patterns from a small number of corpus for inference on downstream datasets, achieving better results on the prediction task. We will refine the expression in the revised manuscript to avoid situations that might confuse readers.

评论

Dear Reviewer AAQN,

Thank you very much for devoting time on reviewing our manuscript. As the author-reviewer discussion process is ending soon, we are wondering whether our responses have well addressed your concerns. If you have any further questions regarding our manuscript, please let us know and we are glad here to provide further discussion and clarification to improve the quality of this manuscript.

Thank you for your contributions,

Authors of Paper 2173

评论

Dear Reviewer AAQN,

Thank you very much for your time and valuable comments on our manuscript. It is heartening to know that a modified version is ready with the help of valuable comments from reviewers, and we would like to know if our modified version has solved your problem. Our improvements are centered around the following suggestions:

  1. Adding additional evaluation metrics: To evaluate the quality of time series in the generation task, we refer to the metrics from DiffusionTS and the full experimental results are presented in Table 12, Page 22 of the revision. Specifically, the generation metrics Context-FID Score, Correlational Score, Discriminative Score and Predictive Score is reduced by 18.2%, 16.2%, 18.5% and 17.8% on ETTh dataset compared to DiffusionTS. More experimental analysis can be found in L1167-1170.
  2. Data efficiency confusion: We rethought our description in revised L366-368 to avoid confusion for the reader.
  3. Complete complexity analysis: We provide a theoretical analysis of the computational complexity of UTSD and show the time cost and memory footprint of UTSD in A4. Compared with the advanced UniTime, TimeLLM, Diffusion, UTSD has obvious advantages. The full complexity analysis will be further included into the revised manuscript.
  4. Sensitivity analysis of model hyperparameters: We demonstrate the model performance for different groups of residual stacking layers L and model dimension D. The results can be found in revision Table8, Page 19.
  5. More ablation experiments: To validate our modeling strategy in real sequence space, the Latent-UTSD model is designed to iteratively reduce noise in the hidden representation space. Detailed ablation experiments are presented in revision Table 6, Page 16.
  6. More visualization experiments: We introduce more visualisation results in the revision appendix section (Figure 10,11 Page 24,25), including the extremely long term sequences as reviewer suggested.

Sincerely thank you for your contribution and responsibility. If you have any further concerns about our manuscript, please let us know and we are happy to provide further discussion here.

Author of paper 2173

评论

Dear Reviewer AAQN:

Thank you once again for your valuable and constructive review. As the author-reviewer discussion period is coming to a close, we kindly request your feedback on our rebuttal provided above. Please let us know if our responses have sufficiently addressed your concerns. If there are any remaining issues, we would be happy to engage in further discussion, as we believe that an in-depth exchange will help strengthen our paper.

Best regards!

Authors

审稿意见
5

The paper proposes a Unified Time Series Diffusion (UTSD) model for time series tasks (mainly focusing on forecasting tasks). The model utilizes a diffusion-based approach rather than traditional autoregressive methods. In particular, the conditional network, adaptor fine-tuning strategy are considered. Numerical experiments and sample codes are provided.

优点

  1. This paper considers an interesting research topic which trying to apply the diffusion model into time-series analysis domain.

  2. Various experimental tasks are considered and cover diverse domains and performance comparisons against both foundation models and domain-specific models.

缺点

  1. My primary concern lies in the suitability of employing a non-autoregressive diffusion model for forecasting tasks. Forecasting is typically a regression problem rather than a generative one, and autoregressive behavior is often observed in certain domains, such as climate and finance. Given this, the current draft would benefit from further justification regarding the choice of a non-autoregressive diffusion model in this context.

  2. The architecture of the proposed model requires further discussion. Section 3 describes the proposed model as a combination of convolution and attention blocks. However, existing literature shows that models based solely on either convolutional or attention networks are already effective for forecasting tasks. Could the authors clarify why a hybrid structure is necessary here? Specifically, how does this configuration address any limitations of diffusion models in forecasting?

Minor Issue: In Appendix E.2, the benchmarks also include imputation and classification tasks, but these seem to be missing from the main draft.

问题

Please see the weakness section.

At this stage, the novelty of the work is insufficient for acceptance at a top-tier machine learning conference, so I tend to recommend rejection. However, I am open to reconsidering my evaluation based on further discussion.

评论

Thank you for your thoughtful review and valuable feedback. We response your concerns as following.

Q1: My primary concern lies in the suitability of employing a non-autoregressive diffusion model for forecasting tasks. Forecasting is typically a regression problem rather than a generative one, and autoregressive behavior is often observed in certain domains, such as climate and finance. Given this, the current draft would benefit from further justification regarding the choice of a non-autoregressive diffusion model in this context.

A1: Thank you for your insightful feedback. We adapt the non-autoregressive diffusion model to handle substantial distribution shifts across different domains. Traditional autoregressive models, while effective at capturing temporal dependencies within individual domains, often face cumulative error challenges in long-term forecasting and struggle with cross-domain generalization due to their reliance on domain-specific conditional probabilities derived from historical patterns.

Notably, several well-known time series analysis models, such as Informer and PatchTST, have successfully employed non-autoregressive architectures, achieving remarkable predictive performance. These models have become widely recognized baselines in the field, further validating the effectiveness of non-autoregressive approaches for time series forecasting. Our extensive experiments demonstrate that the UTSD model outperforms existing autoregressive models on several benchmarks, showcasing its effectiveness in both zero-shot and fine-tuned settings. This indicates the suitability of our approach for forecasting tasks that involve complex, multi-domain data.

We will incorporate a more detailed justification and discussion of these points in the revised manuscript to clarify the rationale behind our model choice. Thank you again for your valuable feedback.

Q2: The architecture of the proposed model requires further discussion. Section 3 describes the proposed model as a combination of convolution and attention blocks. However, existing literature shows that models based solely on either convolutional or attention networks are already effective for forecasting tasks. Could the authors clarify why a hybrid structure is necessary here? Specifically, how does this configuration address any limitations of diffusion models in forecasting?

A2: Thank you for your insighful comments. In our UTSD, convolutional layers are highly effective for capturing local temporal patterns and reducing noise through feature extraction. By integrating convolutional blocks, our model can efficiently process local dependencies and enhance the robustness of the representation. On the other hand, attention mechanisms excel in capturing long-range dependencies and providing dynamic context awareness. By incorporating attention blocks, our model can adaptively focus on relevant parts of the time series, which is crucial for accurate cross-domain forecasting. The diffusion model's denoising process benefits from the hybrid structure by gaining a comprehensive understanding of both local and global patterns. This enhances the model's ability to generate accurate predictions across diverse domains, where patterns and dependencies can vary significantly. Traditional diffusion models may struggle with capturing complex temporal dependencies, especially in varied domains. Our hybrid approach ensures that both local fluctuations and broader trends are effectively modeled, improving the diffusion process's stability and accuracy in generating predictions.

评论

Q3: Minor Issue: In Appendix E.2, the benchmarks also include imputation and classification tasks, but these seem to be missing from the main draft.

A3: Thank you for pointing out this oversight. The UTSD model is designed to be a versatile tool for various time-series tasks, following the generative-based pretrain paradigm. By learning a unified representation that captures both local and global patterns, UTSD effectively generalizes across different domains. Its ability to incorporate multi-scale conditional contexts allows it to adapt dynamically to task-specific demands. The adapter mechanism facilitates seamless fine-tuning, enabling efficient adaptation to various downstream tasks with minimal adjustments. These features position UTSD as a comprehensive model capable of addressing diverse time-series challenges.

We conduct additional experiments on imputation tasks to assess UTSD's performance. The results, shown in the table below, indicate that the average MSE is reduced by 17.0%, 20.1%, and 38.7% compared to GPT4TS, TimesNet, and PatchTST, respectively. This improvement demonstrates that the diffusion model, with its reconstruction paradigm, is well-suited for imputation tasks. Notably, UTSD performs exceptionally well on the ECL dataset, characterized by multi-periodic patterns, aligning with the multi-scale representation mechanism of our Condition-Denoising component.

DatasetsMask RatiosUTSD (MSE)UTSD (MAE)TimeLLM (MSE)TimeLLM (MAE)GPT4TS (MSE)GPT4TS (MAE)TimesNet (MSE)TimesNet (MAE)LLMTime (MSE)LLMTime (MAE)
ETTm112.5%0.0190.0770.0170.0850.0230.1010.0410.1300.0960.229
25.0%0.0220.0920.0220.0960.0250.1040.0470.1390.1000.234
37.5%0.0280.1100.0290.1110.0290.1110.0490.1430.1330.271
50.0%0.0350.1170.0400.1280.0360.1240.0550.1510.1860.323
Avg.0.0260.0990.0280.1050.0270.1070.0470.1400.1200.253
ETTm212.5%0.0180.0790.0170.0760.0180.0800.0260.0940.1080.239
25.0%0.0190.0820.0200.0800.0200.0850.0280.0990.1640.294
37.5%0.0210.0850.0220.0870.0230.0910.0300.1040.2370.356
50.0%0.0240.0940.0250.0950.0260.0980.0340.1100.3230.421
Avg.0.0200.0850.0210.0840.0220.0880.0290.1020.2080.327
ETTh112.5%0.0400.1370.0430.1400.0570.1590.0930.2010.1260.263
25.0%0.0530.1550.0540.1560.0690.1780.1070.2170.1690.304
37.5%0.0700.1750.0720.1800.0840.1960.1200.2300.2200.347
50.0%0.0930.2020.1070.2160.1020.2150.1410.2480.2930.402
Avg.0.0640.1670.0690.1730.0780.1870.1150.2240.2020.329
ETTh212.5%0.0400.1240.0390.1250.0400.1300.0570.1520.1870.319
25.0%0.0430.1310.0440.1350.0460.1410.0610.1580.2790.390
37.5%0.0490.1430.0510.1470.0520.1510.0670.1660.4000.465
50.0%0.0530.1550.0590.1580.0600.1620.0730.1740.6020.572
Avg.0.0470.1380.0480.1410.0490.1460.0650.1630.3670.436
Electricity12.5%0.0430.1290.0800.1940.0850.2020.0550.1600.1960.321
25.0%0.0490.1420.0870.2030.0890.2060.0650.1750.2070.332
37.5%0.0560.1510.0940.2110.0940.2130.0760.1890.2190.344
50.0%0.0650.1650.1010.2200.1000.2210.0910.2080.2350.357
Avg.0.0530.1470.0900.2070.0920.2100.0720.1830.2140.339
Weather12.5%0.0240.0400.0260.0490.0250.0450.0290.0490.0570.141
25.0%0.0260.0430.0280.0520.0290.0520.0310.0530.0650.155
37.5%0.0300.0470.0330.0600.0310.0580.0810.1800.0580.121
50.0%0.0330.0520.0370.0650.0340.0620.0380.0630.1020.207
Avg.0.0280.0460.0310.0560.0300.0540.0600.1440.0760.171

We have supplemented the modified version with experiments on imputation task, the detailed results of which are presented in Table 11, page. 21, in addition to further analyses of the experimental results (L1076-1137).

评论

Dear Reviewer jtdw,

Thank you very much for devoting time on reviewing our manuscript. As the author-reviewer discussion process is ending soon, we are wondering whether our responses have well addressed your concerns. If you have any further questions regarding our manuscript, please let us know and we are glad here to provide further discussion and clarification to improve the quality of this manuscript.

Thank you for your contributions,

Authors of Paper 2173

评论

Dear Reviewer jtdw,

Thank you very much for your time and valuable comments on our manuscript. It is heartening to know that a modified version is ready with the help of valuable comments from reviewers, and we would like to know if our modified version has solved your problem. Our improvements are centered around the following suggestions:

  1. Applicability of non-autoregressive diffusion models: We have added more about the pitfalls of autoregressive mechanisms in cross-domain generalization scenarios and zero-shot tasks in L86-93 of the revised manuscript, which will help the reader understand UTSD's motivation for establishing unified diffusion models in the time-series domain.
  2. Hybrid architecture addresses the limitations of diffusion models in prediction: We have added detailed design about the innovative 1 ×\times 1 Conv1D and attention mechanism of Adapter in revision L332-337. The former is utilized to align the number of tokens in the observation space and the prediction space to achieve flexible input and output in short time, and the latter is utilized to capture the dependencies between all tokens, establishing a connection between the context learning stage and the noise reduction reconstruction stage.
  3. Supplementary imputation task: The UTSD model is intended to be a versatile tool for various time-series tasks, so we added the imputation experiments on all six benchmarks. Detailed experimental results are presented in Table 11, Page 21 and the experimental analyses were in the revision L1076-1137

Sincerely thank you for your contribution and responsibility. If you have any further concerns about our manuscript, please let us know and we are happy to provide further discussion here.

Author of paper 2173

评论

We sincerely thank all the reviewers for their insightful reviews and valuable comments, which are instructive for improving our paper further.

In this paper, a Unified Time Series Diffusion (UTSD) model is established for the first time to model the multi-domain probability distribution, utilizing the powerful probability distribution modeling ability of Diffusion. Unlike existing models that capture the conditional probabilities of the prediction horizon to the historical sequence, we use a diffusion denoising process to model the mixture distribution of the cross-domain data and generate the prediction sequence for the target domain directly utilizing conditional sampling. The proposed UTSD contains several pivotal designs: The condition network captures the multi-scale fluctuation patterns from the observation sequence, which are utilized as context representations; The adaptor-based fine-tuning strategy is utilized for downstream tasks in target domains before multi-domain joint pretraining, and supports flexible observation and prediction horizons through token-level alignment. We conduct extensive experiments on mainstream benchmarks, and UTSD outperforms existing foundation models on all data domains, exhibiting superior zero-shot generalization ability. Empirical results validate the potential of UTSD as a time series foundational model.

The reviewers generally expressed positive opinions of our paper, noting that the method proposed "an interesting research topic," coupled with "its high-quality outcomes, clarity in methodology, innovative architecture," demonstrates "significant implications for the field," and achieves "state-of-the-art performance." Additionally, they highlighted its "potential to serve as the time-series unified diffusion model."

The reviewers also raised insightful and constructive concerns. We have made every effort to address all the concerns by providing sufficient evidence and requested results. Here is the summary of the major revisions:

  1. Employing non-autoregressive diffusion model for forecasting (Reviewers jtdw, eVfc): Diffusion models exhibit good cross-domain generalisation by modelling integrated conditional probability distributions across multiple domains, rather than temporal dependencies within individual domains. Extensive experiments indicate the suitability of our approach for forecasting tasks that involve complex, multi-domain data.
  2. Add multi-tasking experiment results (Reviewers jtdw, YBBT): The UTSD model is designed to be a versatile tool for various time-series tasks, following the generative-based pretrain paradigm. We conduct additional experiments on imputation and unconditional cross-domain generation tasks to assess UTSD's performance.
  3. Add more evaluation metrics and baseline models (Reviewers AAQN, YBBT): To evaluate the quality of generated time series, we refer to the metrices from DiffusionTS and perform the comparison experiments.
  4. Compare our method and previous Diffusion4TS models (Reviewers YBBT, eVfc): We provide a comprehensive analysis of the differences of our method and other Diffusion4TS models, focusing on cross-domain generalization, efficient fine-tuning paradigm, flexible timesteps of input and output.
  5. More visualisation results for challenging time series forecasting (Reviewers AAQN, pg2F): We introduce more visualisation results in the revision appendix section of the revised version. UTSD also shows good performance in extremely long sequence prediction, which verifies that UTSD has the ability to capture extremely long term dependencies, which is crucial for practical applications. In addition, UTSD also shows satisfactory prediction results in short-term non-periodic sequences.

The valuable suggestions from reviewers are very helpful for us to improve our paper. All the changes I've made in our revision (highlighted in blue). We'd be very happy to answer any further questions.

Looking forward to the reviewer's feedback.

评论

Dear Reviewers,

Thank you for your valuable contributions to the review process so far. As we enter the discussion phase, I encourage you to actively engage with the authors and your fellow reviewers. This is a critical opportunity to clarify any open questions, address potential misunderstandings, and ensure that all perspectives are thoroughly considered.

Your thoughtful input during this stage is greatly appreciated and is essential for maintaining the rigor and fairness of the review process.

Thank you for your efforts and dedication.

AC 元评审

(a) Summary of Scientific Claims and Findings This paper introduces the Unified Time Series Diffusion (UTSD) model for multi-domain time series tasks. UTSD leverages a diffusion denoising process to model joint probability distributions across diverse domains, bypassing the autoregressive approach. The paper proposes three key contributions: A condition network to capture multi-scale patterns from observation sequences as context for denoising. An adapter-based fine-tuning strategy to enhance domain adaptability while reusing universal representations learned in pretraining. A diffusion process performed in the actual sequence space with classifier-free guidance to improve stability and performance. Extensive experiments demonstrate that UTSD outperforms baseline models in forecasting, imputation, and other tasks, with claims of superior zero-shot and cross-domain generalization.

(b) Strengths of the Paper Novelty: UTSD introduces a novel approach by integrating diffusion models into time series foundational models and emphasizing sequence-space denoising. Extensive Experiments: The authors present results across multiple benchmarks and tasks, showcasing improved performance over existing baselines. Practical Contributions: The adapter-based fine-tuning mechanism provides computational efficiency during domain-specific adaptation. Visualization and Analysis: The paper includes visualizations and ablation studies to support claims about the model’s effectiveness in long- and short-term dependencies.

(c) Weaknesses of the Paper Theoretical Insufficiency: The paper lacks rigorous theoretical justification for the advantages of sequence-space modeling over latent-space alternatives. While some empirical evidence is provided, the theoretical foundation remains weak. Limited Cross-domain Generalization: The model's cross-domain performance is restricted to datasets with similar characteristics, reducing its applicability in more diverse, real-world scenarios. Inadequate Comparisons: While comparisons to certain baselines (e.g., DiffusionTS) are included, the absence of benchmarks with other state-of-the-art models like MOIRAI or UniTS undermines the claim of being a comprehensive foundational model. Evaluation Metrics: The initial reliance on MSE/MAE as the primary evaluation metrics is problematic for generative models. Although additional metrics were added during rebuttal, this should have been included from the outset. Computational Overhead: While UTSD optimizes diffusion processes through design choices like de-Markovization, diffusion models inherently require higher computational resources, especially for long sequences and multiple iterative denoising rounds. Presentation and Clarity: Some sections, particularly on the hybrid architecture and diffusion process, lack sufficient explanation or empirical backing, leading to questions about the necessity and effectiveness of the design.

(d) Reasons for Rejection Scope and Applicability: The UTSD model is innovative but lacks generalizability to a broad range of time series tasks. Its reliance on diffusion models limits its practical impact. Theoretical Gaps: The paper does not provide sufficient theoretical insights into the proposed approach, particularly regarding the choice of sequence-space modeling and hybrid architectures. Benchmarking Limitations: Missing comparisons with key state-of-the-art models and insufficient diversity in datasets weaken the empirical claims. Reviewer Concerns: Despite an active rebuttal period, several critical concerns raised by reviewers—including cross-domain generalization, theoretical contributions, and computational efficiency—remain inadequately addressed.

审稿人讨论附加意见

  1. Theoretical Justification of Sequence-space Diffusion Concern: Reviewers questioned the theoretical basis for performing diffusion in the actual sequence space rather than latent space, as well as the claimed advantages of this approach. Author Response: The authors cited experimental comparisons showing that sequence-space modeling outperformed latent-space approaches (e.g., Latent-UTSD). However, the theoretical explanation was limited to a brief discussion of information preservation during denoising. Evaluation: While the experimental evidence partially supported the authors’ claims, the lack of deeper theoretical insights left this concern unresolved.

  2. Cross-domain Generalization Concern: Reviewers noted that the cross-domain generalization of UTSD was constrained to datasets with similar patterns, which limits the model’s applicability to diverse real-world domains. Author Response: The authors acknowledged this limitation and proposed extensions in future work but did not offer new experiments or solutions in the rebuttal. Evaluation: The acknowledgment was appreciated but insufficient to address the core issue, which undermines the paper’s claims of being a universal model.

  3. Benchmark Comparisons and Metrics Concern: The reviewers criticized the lack of comprehensive comparisons with state-of-the-art models (e.g., MOIRAI, UniTS) and the initial reliance on MSE/MAE, which are insufficient for evaluating generative tasks. Author Response: The authors added additional metrics (e.g., Context-FID, Discriminative Score) and conducted comparisons with DiffusionTS. They argued that some missing comparisons were due to unavailable code or datasets. Evaluation: While the new metrics strengthened the empirical evaluation, the lack of broader comparisons remained a significant drawback.

  4. Hybrid Architecture Justification Concern: The necessity of combining convolutional and attention-based layers was questioned, as standalone architectures have demonstrated effectiveness in prior work. Author Response: The authors explained that convolutional layers capture local dependencies while attention layers handle long-range patterns, enhancing the denoising process. However, they did not provide ablation studies to quantify the specific contributions of each component. Evaluation: The explanation provided was plausible but lacked empirical support, reducing confidence in the hybrid design’s necessity. Computational Complexity and Scalability

Despite the authors’ efforts during rebuttal, the paper requires further theoretical and empirical refinement to address these concerns adequately. For these reasons, the paper was recommended for rejection.

最终决定

Reject