PaperHub
5.3
/10
Rejected4 位审稿人
最低5最高6标准差0.4
6
5
5
5
4.3
置信度
ICLR 2024

VMFTransformer: An Angle-Preserving and Auto-Scaling Machine for Multi-horizon Probabilistic Forecasting

OpenReviewPDF
提交: 2023-09-24更新: 2024-02-11

摘要

关键词
time series forecastingprobabilistic forecasting

评审与讨论

审稿意见
6

This paper proposes a direct approach for multi-horizon probabilistic forecasting that characterizes the dependence across future horizons. The model treats the multi-horizon target as a random vector, using the von Mises-Fisher (VMF) distribution to characterize the direction and a truncated normal distribution for the magnitude. Additionally, it introduces the concept of "Angle&Scale" similarity to replace the traditional dot-product similarity in self-attention. The model's performance is evaluated on three datasets and shows superiority over some previous methods.

优点

  • The paper is well-written and easy to follow.
  • The motivation behind modeling dependencies among future horizons is clearly articulated and logically sound.
  • The idea of decoupling the target vector into angle and scale is both interesting and innovative.

缺点

  • Overall, this work proposes two orthogonal methods: and objective function based on VMF distribution and the “Angle&Scale” similarity to replace dot-product in self-attention. The motivation behind the latter is not as evident. The computation process, i.e. equations at the bottom of page 5, are confusing due to the misleading subscripts. And a figure for visualization is missing here.

  • Baselines for comparison are somewhat outdated.

问题

  • Could you provide a more detailed and clear explanation of the "Angle&Scale" similarity, specifically the shape of these matrices and computation process of variables Z,Q,K,VZ, Q, K, V?
  • As the objective and “Angle&Scale” similarity are orthogonal, the ablation stduy should compare four models: 1.VMF loss+Dot product; 2.VMF loss+Angle&Scale; 3.MSE loss+Dot product; 4.MSE loss+Angle&Scale. Please complete 3 and 4 to show the effectiveness of the proposed objective function.
  • Please compare with some recent models, such as DLinear and PatchTST.
  • Consider evaluating the speed and memory usage of the "Angle&Scale" similarity with the Dot-product method, as the constant terms in complexity analysis can significantly impact practical efficiency.
评论

Thanks for your thoughtful feedback.

W1: We appreciate the reviewer raising this important point about clearly explaining the intuition and implementation details behind the proposed "Angle&Scale" similarity measure.

The key motivation for Angle&Scale similarity is to better match the structure of our proposed VMF-Transformer objective function, which models the multi-horizon forecast target as a random vector characterized by both an angle/direction and a magnitude/scale. As explained in Section 3.1, the VMF distribution captures the dependence structure across horizons through the angle or direction of the target vector. The scale of the vector represents the magnitudes of the target at each timestep.

Accordingly, the Angle&Scale similarity is designed to evaluate relevance based on two facets - angle and scale. For two vectors in the attention module, we first compute the cosine of their angle to assess similarity in orientation. We additionally multiply this by a Gaussian kernel applied on the difference in vector lengths. This accounts for whether two vectors have a similar scale or magnitude.

Intuitively, this allows the model to match historical sequences in the attention module that have both a similar direction and scale profile to the forecast target. We empirically demonstrate in Section 4 that using Angle&Scale similarity outperforms baseline dot product attention across multiple metrics and datasets. The results suggest that incorporating both angle and scale information is beneficial for this forecasting approach.

The shape of ZZ is (t+1)×(Hd+1)(t+1)\times (H*d+1), where t+1t+1 is the length of the history (we allow the index of the time series sequence to start from 0 instead of 1.). The ii-th row of ZZ is a vector [yixi+1Txi+HT][y_i \circ \mathbf{x}^T_{i+1}\circ \cdots \circ \mathbf{x}^T_{i+H}], where yiy_i is the ground truth at time ii, and xi+1T,xi+HT\mathbf{x}^T_{i+1}, \cdots \mathbf{x}^T_{i+H} are features or known inputs corresponding to the future. Each xi+1\mathbf{x}_{i+1} is a d-dimension vector, and they are packed into one row vector by simply transpose and concatenation together with the real value yiy_i . Since the prediction length is HH, there are a total number of HH input vectors x\mathbf{x}. Hence the length of each row vector of ZZ is H×d+1H\times d+1. And therefore, ZZ is a (t+1)×(H\*d+1)(t+1)\times (H \* d+1) matrix. For QQ ,KK, and VV, they are all originated from ZZ by applying different convolutional kernels to ZZ, where the convolutional kernels are learned during the training stage. The kernels are all 1D, and hence do not change the 'length' of ZZ, but compress the 'width' of ZZ from (H\*d+1)(H \* d+1) to dq,dkd_q, d_k and dvd_v for QQ ,KK, and VV respectively.

W2, W3, and W4: we thank you for your advice. The revised draft is presented in general responses (from part 1 to part 5).

审稿意见
5

To mitigate the accumulation of forecast errors associated with the recursive strategy and the conditional assumption of the direct strategy, the authors propose a novel approach for multi-horizon probabilistic forecasting. This new strategy effectively captures the interdependence across future horizons by modeling the multi-horizon target as a random vector. The direction of this vector represents temporal dependence, while its length measures the overall scale across each horizon. The authors assume that the angle and magnitude of these vectors follow a von Mises-Fisher (VMF) and a truncated normal distribution, respectively. The authors conducted experiments on three benchmark datasets, including an ablation study, demonstrating the advantages of their proposed method over six state-of-the-art techniques.

优点

  • The paper addresses an important challenge in time series forecasting, providing new forecasting strategies for multi-horizon probabilistic forecasting.

  • The approach of modeling the direction and length of future random vectors is innovative in the context of multi-horizon forecasting strategies.

  • The proposed method is compared against multiple state-of-the-art methods.

缺点

  • The proposed method appears to be relatively complex, yet it does not seem to offer substantial improvements over simpler existing methods. Additionally, the motivations behind various design choices lack clarity in the paper.

    • The method involves intricate elements such as the Bessel function and modeling assumptions like the von Mises-Fisher and truncated normal distributions.
    • In the field of time series forecasting, simpler methods often yield satisfactory results. The paper does not convincingly justify the necessity of such complexity in deep time series forecasting.
    • The authors should thoroughly discuss their modeling assumptions and their impact relative to alternative methods. For instance, why opt for a truncated normal distribution? Are there other alternatives?
    • The rationale behind incorporating a multi-head convolutional self-attention mechanism should be explained.
    • While the authors aim to model interdependence across future horizons, they have not employed multivariate scoring rules to evaluate this aspect.
    • The paper asserts that recursive and direct strategies can yield high forecast errors, but it remains unclear how the proposed strategy compares with other forecasting approaches that use the same underlying model.
  • Experiments:

    • The choice of only three datasets for experimentation raises questions, as many machine learning papers on time series forecasting examine more extensive benchmark datasets. For example, see the Monash Time Series Forecasting Archive (https://arxiv.org/abs/2105.06643)

    • Many time series datasets exhibit a pronounced seasonal component that dominates the signal. Consequently, it may be challenging to surpass simpler methods that effectively estimate this seasonal component. The paper should address the strength of the seasonal component in the considered datasets.

    • The authors have not included simple benchmarks such as auto.arima or exponential smoothing in their comparisons.

    • Learning curves should be provided to demonstrate the stability of the training procedure.

    • The proposed method is trained using maximum likelihood estimation. The authors should also provide the negative log-likelihood (NLL) for all density-based methods.

    • The authors did not report standard errors and information regarding the number of runs performed.

    • The results for specific forecast horizons (e.g., h = 1, 2, 3, etc.) should be reported to assess if the procedure increases forecast error for initial horizons while improving the average error across horizons.

问题

  • Refer to the "Weaknesses" section for questions.

    • Typos and Improvements:
      • Equation 6 is rimarily computing the logarithm of expression (5). For clarity, it may be better to move it to the appendix.
      • On page 6, the figure number is missing.
      • The sentence beginning with "It should be noted that in Equation equation 7," is not clear.
      • "the provide the proof." Furthermore, clarify that you are citing a reference and not presenting a proof.
评论

We thank the reviewer for offering valuable feedback. We have addressed each of the concerns raised by the reviewer as outlined below.

  1. The reviewers rightly point out that many time series exhibit strong seasonal patterns that can dominate the signal. We analyzed the strength of the seasonal component in the electricity, solar, and traffic datasets using autocorrelation analysis. The electricity data shows a strong daily seasonal pattern with autocorrelation remaining above 0.6 at the 24 hour lag. The solar data shows both daily and yearly seasonal patterns due to the underlying physical processes. The traffic data has relatively weak seasonality. However, basic time series methods such as auto.arima and exponential smoothing often cannot effectively model complex real-world time series. ARIMA, exponential smoothing, and Prophet on our used datasets achieve relatively poor performance compared to more flexible machine learning approaches [1, 2]. These methods make strong assumptions about the underlying data generative processes and do not directly capture uncertainties. Our model demonstrates excellent performance across datasets with varying degrees of seasonality. By jointly modeling the sequential dependence, overall scale, and capturing uncertainty, our method provides robust performance without needing an explicit seasonal component model. We provide quantitative metrics showing our model's ability to accurately forecast across a diverse range of conditions.

[1] Lim, Bryan, et al. "Temporal fusion transformers for interpretable multi-horizon time series forecasting." International Journal of Forecasting 37.4 (2021): 1748-1764.

[2] Wang, Y., Smola, A., Maddix, D., Gasthaus, J., Foster, D., & Januschowski, T. (2019, May). Deep factors for forecasting. In International conference on machine learning (pp. 6607-6617). PMLR.

  1. Learning curves (Figure 3) of our model are provided in the updated version to demonstrate the stability of the training procedure.

  2. A comparison of metrics corresponding to different horizons where H ranges from 1 to 720 (Figure 4) of our model is provided in the updated version to demonstrate the stability of forecasting performance.

  3. This revised draft incorporates extensive additional experiments in general responses (from part 1 to part 5), now encompassing the latest baseline models (such as PatchTST and DLinear) across varying and longer horizons (24, 168, and 720). Every key result includes the mean and standard deviation to quantify uncertainty.

  4. We actually used the negative log-likelihood (NLL) in the training stage. As we explained in our paper ' The training object is to maximize Equation 6 with respect to Θ\Theta, or equivalently, minimize the inverse l(Θ)-l(\Theta)', where l(Θ)-l(\Theta) refers to the NLL. Thanks for noticing. We now realize this may be a confusing statement and we will clarify it to enhance readability and comprehension.

评论

We thank the authors for their response and clarifications

For the seasonality, I meant that by accurately accounting for seasonality, the residuals often amount to mere noise.

Also, I do not agree with the statement that "These methods ... do not directly capture uncertainties". You can capture uncertainties easily with ARIMA models.

The following statement is also very strong: "Our model demonstrates excellent performance across datasets with varying degrees of seasonality." The evidence is drawn from only three datasets. Given the empirical focus of the paper, this limited dataset selection may compromise the strength of the claims made about the new forecasting method.

Furthermore, I agree with other reviewers that your paper does not align with the standard experimental setups used in other works within the forecasting literature.

In light of these considerations, I maintain my original score. I believe there is still room for improvement in the paper,

审稿意见
5

The paper contributes to time series forecasting by proposing a new model, the VMFTransformer, which tackles the inherent problems in iterative and direct forecasting methods. Iterative methods suffer from cumulative errors, and direct methods often incorrectly presume independence between future time points. The VMFTransformer addresses these issues by conceptualizing forecasts as random vectors, utilizing the von Mises-Fisher distribution to maintain temporal directionality and a truncated normal distribution for magnitude, accurately preserving time-dependent relationships. Benchmarked against several methods, the VMFTransformer demonstrates enhanced predictive performance, showing its effectiveness, and adaptability for a range of time series forecasting applications.

优点

  1. Error Accumulation Mitigation: Effectively addresses the issue of error accumulation inherent in iterative forecasting methods.

  2. Temporal Independence: Overcomes the unrealistic assumption of temporal independence used by direct forecasting methods.

  3. Directional Dependencies: Employs the von Mises-Fisher distribution to accurately capture the directional dependencies in time series data.

  4. Magnitude Characterization: Utilizes a truncated normal distribution to model the magnitude of forecasts, enhancing predictive accuracy.

缺点

  1. Lack of Recent Benchmarks: The VMFTransformer has not been compared with the most recent benchmarks such as PatchTST [1], or with most comparisons made against older methods. for instance please consider looking at recnet benchmarks provided by https://github.com/timeseriesAI/tsai

  2. Limited comparison for different horizons: The scope of comparison is limited, which may not adequately reflect the model's performance across a broader range of forecasting scenarios.

  3. Formatting Issues: There is room for improvement in the formatting of the equations, which could enhance readability and comprehension.

  4. Presentation Clarity: The paper could better articulate the main contribution, as the current presentation may be challenging to follow, possibly obscuring the model's innovation.

  5. Insufficient Experimentation: Without more extensive experimentation, it's difficult to ascertain the paper's meaningful contribution to the community and its practical applicability.

[1] A Time Series is Worth 64 Words: Long-term Forecasting with Transformers, ICLR 2023

问题

Please consider the comments I provided above.

评论

Thank you for taking the time to review our paper and provide constructive feedback. We appreciate the reviewers recognizing the potential of our proposed VMFTransformer model for multi-horizon probabilistic time series forecasting.

W1, W2, and W5: this revised draft incorporates extensive additional experiments in general responses (from part 1 to part 5), now encompassing the latest baseline models (such as PatchTST and DLinear) across varying and longer horizons (24, 168, and 720). Every key result includes the mean and standard deviation to quantify uncertainty.

W3: We will revise the formatting of all equations in the paper to enhance readability and comprehension. Equations are centered on their own lines with consistent spacing around operators. Variable names have been standardized and vector/matrix notation clarified. We believe these changes significantly improve the formatting.

W4: The main contribution of our work is the novel probabilistic forecasting model based on the von Mises-Fisher distribution and Transformer architecture. To better articulate this, we will add an explicit contribution statement in the introduction highlighting:

  1. The VMFTransformer model itself which captures temporal dependence in a multi-horizon target.
  2. The Angle&Scale similarity measure for the self-attention module
  3. An efficient method to optimize the likelihood function with the Bessel function

Additionally, we will reorganize several sections to improve logical flow and aid comprehension of the technical details underpinning the model innovation. A graphical abstract will also be added to provide an intuitive overview. We believe these changes address the need to better communicate the main contributions of our work.

评论

Dear Authors,

Thank you for providing additional experimental results in your rebuttal. I appreciate your efforts to address the concerns raised in the initial review. However, after careful consideration, I believe there is still room for improvement, particularly in terms of the experimental setup and the consistency of results.

One key issue is that your paper does not align with the standard experimental setups used in other works within the literature, especially regarding the forecasting horizon. Notably, your results for a 720-horizon forecast appear to be inconsistent with those reported by the original authors in the following two referenced papers. To facilitate a more straightforward and meaningful comparison, I suggest that you consider adjusting your experimental setup to adhere to the standard benchmarks prevalent in the literature. This would not only enhance the clarity of your results but also their comparability with existing studies.

I have included references to two pertinent papers below for your consideration. These papers should serve as a guide for the standard numerical values and methodologies typically employed in this field of research.

Given these considerations, I have decided to maintain my original review score. I believe that addressing these concerns will significantly strengthen the impact and relevance of your work in the context of existing literature.

Looking forward to seeing the revised version of your manuscript.

Best regards, Link to github repo of the papers. https://github.com/cure-lab/LTSF-Linear

https://github.com/yuqinie98/PatchTST

审稿意见
5
  1. This paper addresses the challenges in time series forecasting, specifically in multi-horizon probabilistic forecasting, by proposing a direct approach that effectively characterizes the dependence across future horizons. Technically, the authors consider the multi-horizon target as a random vector and apply the von Mises-Fisher (VMF) distribution and the truncated normal distribution to model the angle and magnitude of the target vector.

  2. The performance of the proposed framework is evaluated on three benchmarks, demonstrating its superiority over six state-of-the-art methods in different time series forecasting tasks.

优点

  1. The paper is well-written and exhibits clarity in its presentation. The visual results help to comprehend the proposed framework.
  2. Authors propose a novel similarity measurement termed “Angle&Scale” similarity for the attention module and show that the Angle&Scale similarity outperforms the dot-product similarity in most cases in ablation studies.
  3. The proposed VMFTransformer consistently outperforms all baselines in MSE and q-risk.

缺点

  1. The current version of the paper solely presents the average value obtained from five trials without including information about the standard deviation. It is highly recommended to include error bars.
  2. Why use the VMF distribution and the truncated normal distribution to characterize the angle and magnitude of the target vector? The motivation behind this is unclear to me.
  3. Metrics used to evaluate uncertainty are not sufficiently convincing, a more commonly used metric, CRPS [1], was not used in the experiment.
  4. Some probabilistic time series baselines are not compared with the proposed method in the experiment, such as TransMAF [2], [3].

References:

[1] Tilmann Gneiting and Adrian E Raftery. Strictly proper scoring rules, prediction, and estimation. Journal of the American statistical Association, 102(477):359–378, 2007.

[2] Binh Tang and David S Matteson. Probabilistic transformer for time series analysis. Advances in Neural Information Processing Systems, 34:23592–23608, 2021.

[3] Kashif Rasul, Abdul-Saboor Sheikh, Ingmar Schuster, Urs Bergmann, and Roland Vollgraf. Multivariate probabilistic time series forecasting via conditioned normalizing flows. arXiv preprint arXiv:2002.06103, 2020.

问题

Please see my comments in Weaknesses.

I would also appreciate if the authors can respond, if they can, to the weaknesses.

评论

We deeply appreciate the thoughtful feedback provided by the reviewers. Their suggestions have helped strengthen our work substantially. In response:

W1 and W4: Per the reviewers' recommendations, we have added extensive additional experiments with the latest baseline models across longer time horizons. Every key result now includes the mean and standard deviation to rigorously quantify uncertainty. We believe these comprehensive additions showcase our method's strengths.

W2: You raises an important point regarding the motivation behind using the von Mises-Fisher (VMF) distribution and truncated normal distribution to model the multi-horizon forecasting vector. The key intuition is that the VMF distribution can effectively characterize vectors distributed around a specific direction, which fits our assumption that when dependence exists among the multi-horizon targets, they will concentrate around a certain direction. The truncated normal distribution provides a flexible prior for the scale parameter to make the VMF likelihood tractable. Compared to simply using independent normal or Laplace distributions on each horizon, the VMF and truncated normal distribution can jointly capture the dependence structure via the shared direction parameter in VMF. We will add more discussion on the motivation in the revised paper to clarify this modeling choice.

W3: We carefully re-evaluated TransMAF, but unfortunately found it performed poorly on our datasets - significantly worse than the other baseline models we assessed. As such, we decided not to include these inferior results and instead added evaluations using the popular PatchTST and DLinear models, providing readers with a more representative view of the state-of-the-art. Regarding evaluation metrics, we utilize the community standards of MAE, MSE, Q50-Loss, and Q90-Loss - prevalent metrics used across probabilistic time series works.

We thank the reviewers again for their outstanding suggestions. We believe the additions made in response (from part 1 to part 5 in general response) further demonstrate the competitiveness and rigor of our proposed approach across both model architectures.

评论

Dear authors,

Thanks for your response and further experiments. I appreciate your efforts to address my concerns. For evaluation metrics, I still think CRPS (i.e., accumulation of the quantile loss at different quantiles) is essential for uncertainty evaluation. I maintain my original score based on these considerations and other reviewers' comments.

Looking forward to the improved version of your manuscript.

评论

We deeply appreciate the reviewers taking the time to provide thoughtful feedback on our work. We are grateful that most reviewers saw the potential for this research to make a meaningful contribution. Carefully considering each comment, we have revised the paper to address the concerns raised. Our detailed responses below explain how we have enhanced the manuscript.

In particular, this revised draft incorporates extensive additional experiments, now encompassing the latest baseline models across varying and longer horizons. Every key result includes the mean and standard deviation to quantify uncertainty.

Horizon = 24MAEMSE
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR0.0167 ±\pm 0.00040.2619 ±\pm0.0030\underline{0.1667}$$\pm0.00860.0187±\pm0.00090.3243±\pm0.00910.2042±\pm0.0069
SimpleFeedForward0.0152 ±\pm 0.00070.3023 ±\pm0.00260.2751 ±\pm0.00280.0113±\pm0.00190.4213±\pm0.01010.3775±\pm0.0022
TFT0.0251 ±\pm 0.00100.2748 ±\pm0.00640.2164 ±\pm0.00690.0363±\pm0.00270.3189±\pm0.00560.2912±\pm0.0093
Transformer0.0138 ±\pm 0.00220.3031 ±\pm0.00600.1796 ±\pm0.00910.0144±\pm0.00060.4502±\pm0.01560.2050±\pm0.0031
Autoformer0.1236 ±\pm 0.08560.2106 ±\pm0.01380.2612 ±\pm0.00840.0252±\pm0.08990.1684±\pm0.0684\underline{0.1541}$$\pm0.0443
PatchTST0.0610\underline{0.0610} ±\pm 0.00240.1938 ±\pm0.01090.2746 ±\pm0.0145\underline{0.0081}$$\pm0.00080.0936±\pm0.00920.2251±\pm0.0050
Informer0.0844 ±\pm 0.00950.2607 ±\pm0.01720.4620 ±\pm0.01940.0144±\pm0.02260.1823±\pm0.00720.6446±\pm0.0074
DLinear0.1069 ±\pm0.00040.2110 ±\pm0.0050.2317 ±\pm0.00550.0309±\pm0.00030.1336±\pm0.0010.2720±\pm0.0027
VMFTransformer0.0116 ±\pm0.00150.2092\underline{0.2092} ±\pm0.00740.1567 ±\pm0.00080.0077±\pm0.0147\underline{0.1333}$$\pm0.03080.1185±\pm0.0012
Horizon = 168MAEMSE
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR0.0224 ±\pm 0.00290.3076 ±\pm0.02460.1712±\pm0.00880.0300±\pm0.00810.3806±\pm0.04000.2115±\pm0.0104
SimpleFeedForward0.0210 ±\pm 0.00040.3228 ±\pm 0.00580.2590 ±\pm0.00370.0281±\pm0.00120.4122±\pm0.01960.2901±\pm0.0043
TFT0.0286 ±\pm 0.00160.3335 ±\pm 0.00390.1670 ±\pm0.00530.0468±\pm0.00700.5125±\pm0.0380.1952±\pm0.0070
Transformer0.0198\underline{0.0198} ±\pm 0.00930.3463 ±\pm 0.01200.1918 ±\pm0.01170.0295±\pm0.01210.4903±\pm0.02430.2081±\pm0.0234
Autoformer0.1197 ±\pm 0.01830.2784 ±\pm 0.02150.2231 ±\pm0.01440.0241±\pm0.00780.1843±\pm0.02540.1746±\pm0.0622
PatchTST0.0957 ±\pm 0.00210.1836 ±\pm 0.00090.0898 ±\pm0.00220.0169±\pm0.00070.0942±\pm0.00370.0179±\pm0.0075
Informer0.1460 ±\pm 0.02300.2274 ±\pm 0.00790.5549 ±\pm0.00240.0335±\pm0.00470.1616±\pm0.01370.8068±\pm0.0021
DLinear0.1559 ±\pm 0.00160.2129\underline{0.2129} ±\pm0.00260.2633 ±\pm0.00850.0595±\pm0.0002\underline{0.1250}$$\pm0.00840.3096±\pm0.0022
VMFTransformer0.0128 ±\pm0.01570.2136 ±\pm0.00050.1567\underline{0.1567} ±\pm0.0019\underline{0.0204}$$\pm0.03730.1390±\pm0.0017\underline{0.1185}$$\pm0.0029
评论
Horizon = 720MAEMSE
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR0.0323 ±\pm 0.00670.2220 ±\pm0.00590.2958±\pm0.03590.0897±\pm0.0301\underline{0.2004}$$\pm0.00020.4112±\pm0.0399
SimpleFeedForward0.0239 ±\pm0.00060.2342 ±\pm0.01690.2890 ±\pm0.00670.0491±\pm0.00250.2226±\pm0.00990.3501±\pm0.0104
TFT0.0313 ±\pm0.00090.2027\underline{0.2027} ±\pm0.00150.2069\underline{0.2069} ±\pm0.00660.0809±\pm0.01080.2141±\pm0.00670.3467±\pm0.0080
Transformer0.0173\underline{0.0173} ±\pm0.00280.2170 ±\pm0.00430.2047 ±\pm0.0140\underline{0.0321}$$\pm0.00200.2536±\pm0.01250.3921±\pm0.0174
Autoformer0.3857 ±\pm0.0050.4082 ±\pm0.08650.3897 ±\pm0.07360.2345±\pm0.05780.2574±\pm0.0587\underline{0.3277}$$\pm0.0976
PatchTST0.0843 ±\pm0.00420.3112 ±\pm0.00190.5836 ±\pm0.01030.0623±\pm0.00080.2975±\pm0.00280.4726±\pm0.0092
Informer0.1698 ±\pm0.00100.2647 ±\pm0.0080.6165 ±\pm0.03950.0457±\pm0.00050.2599±\pm0.00590.8749±\pm0.0395
DLinear0.2529 ±\pm0.00660.2249 ±\pm0.00370.4537 ±\pm0.00310.1401±\pm0.00250.1322±\pm0.00150.5847±\pm0.0038
VMFTransformer0.0147 ±\pm0.00040.2018 ±\pm0.00900.2583±\pm0.00130.0136±\pm0.00240.2198±\pm0.00850.3029±\pm0.0009
Horizon = 24q-50-lossq-90-loss
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR\underline{0.0734} $$\pm0.00270.4626 ±\pm0.0054\underline{0.1526}$$\pm0.00790.0580±\pm0.00210.3097±\pm0.0118\underline{0.1082}$$\pm0.0014
SimpleFeedForward0.0784 ±\pm0.00410.5341±\pm0.00470.2518±\pm0.0025\underline{0.0433}$$\pm0.00080.3591±\pm0.00490.2011±\pm0.0017
TFT0.1107 ±\pm0.00550.4855 ±\pm0.01130.1981±\pm0.006330.0588±\pm0.00140.2471±\pm0.01220.1478±\pm0.0020
Transformer0.0789 ±\pm0.01260.5354 ±\pm0.01070.1644±\pm0.00830.0500±\pm0.00120.3502±\pm0.01670.1087±\pm0.0029
Autoformer0.1224 ±\pm0.01340.2233\underline{0.2233} ±\pm0.09460.2156±\pm0.08560.0653±\pm0.0023\underline{0.2112}$$\pm0.01120.1454±\pm0.0012
PatchTST0.1840 ±\pm0.00880.5611 ±\pm0.02520.4690±\pm0.02970.2291±\pm0.00910.5678±\pm0.10100.4367±\pm0.0415
Informer0.2167 ±\pm0.02340.3456 ±\pm0.06170.6055±\pm0.02840.1816±\pm0.02940.2480±\pm0.01810.6649±\pm0.0151
DLinear0.1161 ±\pm0.01240.2319±\pm0.00620.3418±\pm0.00820.1180±\pm0.01410.2129±\pm0.00510.3462±\pm0.0085
VMFTransformer0.0722 ±\pm0.00970.2125±\pm.01130.1490±\pm0.01750.0427±\pm0.00100.1371±\pm0.03460.0935±\pm0.0089
评论
Horizon = 168q-50-lossq-90-loss
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR0.1463 ±\pm0.00610.5613 ±\pm0.00390.1553±\pm0.00810.0862±\pm0.01090.2349±\pm0.01620.1012±\pm0.0018
SimpleFeedForward0.1397 ±\pm0.00220.5890 ±\pm0.01070.2348 ±\pm0.00340.0730±\pm0.00080.2557±\pm0.00380.1360±\pm0.0020
TFT0.2242 ±\pm0.04970.6085 ±\pm0.0220\underline{0.1514} $$\pm0.01060.1122±\pm0.00500.2030±\pm0.0662\underline{0.1019}$$\pm0.0122
Transformer0.1172 ±\pm0.04970.6318 ±\pm0.02200.1739 ±\pm0.0106\underline{0.0670}$$\pm0.00500.3115±\pm0.06620.1063±\pm0.0122
Autoformer0.1342 ±\pm0.0092\underline{0.2354} $$\pm0.00420.2559 ±\pm0.00210.0790±\pm0.0024\underline{0.1588}$$\pm0.00290.1948±\pm0.0086
PatchTST0.1171\underline{0.1171} ±\pm0.00070.3462 ±\pm0.00370.3041 ±\pm0.0050.1189±\pm0.00440.3976±\pm0.01870.3019±\pm0.0094
Informer0.3753 ±\pm0.00620.5434 ±\pm0.00830.7194 ±\pm0.00280.2440±\pm0.02940.2960±\pm0.00570.6800±\pm0.0052
DLinear0.1750 ±\pm0.01560.2458 ±\pm0.00300.3879 ±\pm0.00710.1874±\pm0.00330.2326±\pm0.00510.3968±\pm0.0083
VMFTransformer0.0839 ±\pm0.00420.2318±\pm0.00110.1225 ±\pm0.00990.0597±\pm0.01720.1473±\pm0.00500.1099±\pm0.0011
Horizon = 720q-50-lossq-90-loss
ModelElectricitySolarTrafficElectricitySolarTraffic
DeepAR0.1805 ±\pm0.03790.3543 ±\pm0.00950.2724±\pm0.03300.1762±\pm0.0659\underline{0.1451}$$\pm0.01350.1883±\pm0.0336
SimpleFeedForward0.1336 ±\pm0.00380.3738 ±\pm0.02710.2661 ±\pm0.00610.0854±\pm0.00310.1603±\pm0.02390.1579±\pm0.0018
TFT0.1748 ±\pm0.00550.3234 ±\pm0.00250.1906\underline{0.1906} ±\pm0.00610.0975±\pm0.00910.1084±\pm0.0061\underline{0.1266}$$\pm0.0052
Transformer\underline{0.1246}$$\pm0.00970.3463 ±\pm0.01720.1885 ±\pm0.0112\underline{0.0829}$$\pm0.00510.1600±\pm0.00910.1226±\pm0.0067
Autoformer0.3384 ±\pm0.00380.6827 ±\pm0.00670.3751 ±\pm0.06670.3138±\pm0.00120.5837±\pm0.00380.2969±\pm0.0807
PatchTST0.4222 ±\pm0.00130.5221 ±\pm0.01700.7957 ±\pm0.01710.2998±\pm0.00340.2723±\pm0.02940.4704±\pm0.0218
Informer0.3897 ±\pm0.00460.5601 ±\pm0.02450.8969 ±\pm0.01360.2496±\pm0.00490.3097±\pm0.00290.7540±\pm0.0108
DLinear0.2802 ±\pm0.00740.2598 ±\pm0.00430.6692 ±\pm0.00450.3319±\pm0.01380.2319±\pm0.00140.7092±\pm0.0054
VMFTransformer0.1028 ±\pm0.0028\underline{0.2556} $$\pm0.02420.5738±\pm0.00740.0798±\pm0.02880.5455±\pm0.04380.3917±\pm0.0032
评论

A sample of 50,000 data points was selected from the Solar dataset to test the time consumed by our model across different horizons.

Encoding lengthTime consumed (seconds)
1285.352
48122.309
72249.441
120444.464
1681629.578
2402699.469
3606767.158

A sample of 50,000 data points was selected from the Solar dataset to test the memory consumed by our model across different horizons.

Encoding lengthMemory usage (Byte)
122498048
483061248
723462656
1204314624
1685248512
2406776320
3609700864
评论

We conducted additional ablation studies to prove the effectiveness of our designed Angle & Scale similarity measurements in the attention module when used in conjunction with our proposed VMF-Loss. However, we found that these similarity measurements were not as effective when the model was trained using only the standard MSE-loss. This suggests that our Angle & Scale similarity design has a beneficial interaction specifically with the VMF-Loss function during training. To further analyze this relationship, we performed experiments altering components of both the loss function and attention module. We found that removing either the angle or scale measurement from the attention with VMF-Loss hurt performance, confirming that both aspects contribute to its effectiveness under VMF-Loss. Furthermore, incorporating only one of the two measurements along with MSE-Loss did not lead to significant improvements, verifying that the combination of both measurements works synergistically with properties unique to the VMF-Loss function.

VMF-LossMetricAngle&ScaleDot-product
Horizon=24MAE0.23390.2389
MSE0.18280.1749
50-loss0.32380.3317
90-loss0.26670.2707
Horizon=168MAE0.20140.2389
MSE0.13070.1316
50-loss0.28480.2950
90-loss0.17240.2250
MSE-LossMetricAngle&ScaleDot-product
Horizon=24MAE0.73890.7179
MSE0.47920.5061
50-loss0.50590.5010
90-loss0.81530.8650
Horizon=168MAE0.71790.7179
MSE0.50750.5061
50-loss0.50100.5010
90-loss0.86310.8685
评论

Learning curves (Figure 3) of our model are provided in the updated version to demonstrate the stability of the training procedure.

Comparison of metrics corresponding to different horizons where H ranges from 1 to 720 (Figure 4) of our model is provided in the updated version to demonstrate the stability of forecasting performance.

AC 元评审

The reviewers raised multiple concerns, notably the experimental setups. After the rebuttal, the reviewers acknowledge that there is still room for improvement, and I agree with their expertise. Nevertheless, the angle and scaling idea is interesting, and the authors are encouraged to continue working towards this direction.

为何不给更高分

The experimental setups are not good.

为何不给更低分

n/a

最终决定

Reject