Conditional Lagrangian Wasserstein Flow for Time Series Imputation
摘要
评审与讨论
The authors introduce the conditional Lagrangian Wasserstein flow method for time series imputation. The time series imputation task is treated as a conditional data generation problem. The authors use flow matching to learn an ODE sampler for generating the missing time series data. They further propose to enhance the imputation performance via Rao-Blackwellization. The method is tested on the real-world datasets.
优点
S1. The approach of using Lagrangian Wasserstein flow for time series imputation is interesting and novel.
S2. Compared to other diffusion model-based time series imputation methods, this method seems more efficient.
S3. The experimental results given in the paper are satisfactory.
缺点
W1. The imputation process is implemented using an ODE sampler. Why not use an SDE sampler?
W2. Why use the Euler method instead of other higher-order solvers, such as the Runge-Kutta methods, for solving the ODE?
W3. Can the proposed method also be applied to time series forecasting tasks?
W4. The dynamic described in Eq (15) is deterministic, which is inconsistent with Eq (1). Why the diffusion term is missing in Eq (15)?
问题
Please see my weaknesses.
And there is another question, Q. Can the VAE-based potential function be used alone for time series imputation?
局限性
CRPS is not used as the evaluation metric in the experiments.
Many thanks for the insighful comments.
Weakness 1:
Our proposed method can also be implemented via an SDE sampler by adding a time-dependent volatility term (please see Eq. (12), Sec 3.2, Page 4), but in order to achieve the optimal empirical performance in terms of RMSE and MAE on the time series imputation tasks, we opt for an ODE sampler.
Weakness 2:
In practice, we found that the Euler method has similar performance to other numerical methods while requiring less memory and computational time. Therefore, we chose to use the Euler method in our implementation.
Weakness 3:
Yes, the proposed method can be used for time series forecasting tasks as well. To this end, we can view the forecasting task as a conditional generation problem, in which one can just use the future states as the data generation target and the current state as the conditional information.
Weakness 4:
Eq. (15) is derived based on the assumption of an overdamped Langevin dynamic system, which still represents the drift term in Eq.(1). For computational convenience, we choose to omit the diffusion term to formulate an ODE sampler. Nevertheless, the theoretical analysis given in the paper still applies to this case.
Questions:
Although the VAE-based potential function can be seen as a sufficient statistic for the imputation target, it is generally not accurate enough to be used alone to represent the data generation dynamics in time series imputation tasks.
Limitations:
Since our method is implemented via an ODE sampler CRPS may not be a suitable evaluation metric.
Furthermore, [1] and [2] point out that CRPS may overlook the model's performance on each dimension, which could lead to an incorrect assessment of overall model performance.
Reference: [1] Koochali, Alireza, Peter Schichtel, Andreas Dengel, and Sheraz Ahmed. "Random noise vs. state-of-the-art probabilistic forecasting methods: A case study on CRPS-Sum discrimination ability." Applied Sciences 12, no. 10 (2022): 5104.
[2] Biloš, Marin, Kashif Rasul, Anderson Schneider, Yuriy Nevmyvaka, and Stephan Günnemann. "Modeling temporal data as continuous functions with stochastic process diffusion." In International Conference on Machine Learning, pp. 2452-2470. PMLR, 2023.
Thanks for your rebuttal. The authors have addressed all my concerns and I will maintain my score at this stage.
In this paper, the authors introduce a new time series imputation model based on the conditional Lagrangian Wasserstein flow. Different from previous diffusion-based models, the proposed model leverages the optimal transport theory and Lagrangian dynamics to improve the data generation performance.
优点
- The overall presentation of paper is clear.
- The idea of using the conditional Lagrangian Wasserstein flow for time series imputation is novel.
- The proposed method is assessed on the real-world datasets and shows competitive performance compared to the state-of-the-art methods.
缺点
- In Sec 3.2, the technique used for projecting the interpolants into the Wasserstein space is unclear. The authors should elaborate more on this.
- Compared to CSDI and other diffusion-based imputation methods, the proposed method’s data generation process seems deterministic. Hence, it cannot be used to quantify the uncertainty of the prediction.
- The proposed sampler is implemented through an ODE, which is inconsistent with the Schrodinger Bridge problem in sec 2.3.
- There are some inconsistencies in the notations, please double-check.
问题
What is the difference between the data generation processes in Euclidean space and Wasserstein space?
局限性
The ablation study on the effectiveness of Rao-Blackwellization should be conducted on more than one dataset (PM 2.5).
Many thanks for the valuable feedback .
Weakness 1:
To project the intermediate samples from the Euclidean space to the Wasserstein space, We adopt the technique descried in [1] by drawing mini-batches of the initial noise and target samples to compute the corresponding optimal transport maps.
In terms of Wasserstein metrics, we use the earth mover distance (EMD) for compute the deterministic optimal transport maps, which turns out to perform better than other metrics.
Alternatively, we can also adopt the Sinkhorn distance to compute the stochastic optimal transport maps with entropy regularization.
For the practical implementation, we use the Python Optimal Transport (POT) library [2].
Weakness 2:
This is because we adopt an ODE sampler in our method. Although ODE samplers are deterministic, they have other advantages compared to SDE samplers. For example, they are usually easier to implement than SDE samplers. Moreover, compared to CSDI and other diffusion-based methods, our method requires less simulation steps and samples to achieve good performance on time series imputation tasks.
Weakness 3:
We can also implement the proposed framework via an SDE sampler. The optimization objective becomes the KL divergence between the target SDE and the prior Wiener process, which is aligned with the frame work of the Schrodinger Bridge problem. However, in practice, we found that the ODE sampler has better empirical performance in terms of RMSE and MAE compared to the SDE sampler.
Weakness 4:
We have proofread the manuscript again and corrected the typos.
Questions:
Since the Euclidean distance is not suitable for measuring the distance between two probabilistic distributions, meaning that we cannot find the shortest path between =the initial distribution and the target distribution. As a result, performing the data generation processes in Euclidean space may lead to slow convergence and poor sample quality. Therefore, we need to perform the data generation processes in the Euclidean space instead of the Wasserstein space.
Limitations:
We conduct additional experiment using the synthetic dataset. And the experimental results demonstrates the effectiveness of the Rao-Blackwellization. Please take a look at the uploaded PDF for the details.
Test imputation results on synthetic data ( trials, values are multiplied by .
| Method | Synthetic 0.4 | Synthetic 0.6 | Synthetic 0.8 | |||
|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CLWF(no RB) | ||||||
| CLWF(with RB) |
References:
[1] Alexander Tong, Kilian FATRAS, Nikolay Malkin, Guillaume Huguet, Yanlei Zhang, Jarrid Rector-Brooks, Guy Wolf, and Yoshua Bengio. Improving and generalizing flow-based genera431 tive models with minibatch optimal transport. Transactions on Machine Learning Research,432, 2024.
[2] RFlamary, Rémi, Nicolas Courty, Alexandre Gramfort, Mokhtar Z. Alaya, Aurélie Boisbunon, Stanislas Chambon, Laetitia Chapel et al. "Pot: Python optimal transport." Journal of Machine Learning Research 22, no. 78 (2021): 1-8. Website: https://pythonot.github.io/
Since all of my questions have been adequately answered by the authors, I recommend acceptance.
Inspired by optimal transport and Lagrangian dynamics, this work proposes to use the conditional Lagrangian Wasserstein flow to impute time series data. The method requires less model evaluation steps to generate high quality samples compared to existing diffusion models. Moreover, a task-specific energy function is used to further improve the model’s performance.
优点
- The contributions of this paper are significant.
- This work is built on a solid theoretical foundation.
- The experimental results and relevant ablation study findings are provided.
缺点
- How should the values of the hyperparameters (e.g., the variance of the potential function) be chosen? How will these choice affect the model’s performance?
- The reason for choosing the VAE model to construct the potential function is unclear. Are there other functions can also be used as the task-specific potential function?
- The authors proposed to use a VAE model to enhance the data sampling procedure. However, if the VAE model performs poorly, can it still help with the data generation process?
问题
Can the proposed method be used to compute the likelihood of the generated data samples?
局限性
The deterministic sampling process may introduce bias, potentially failing to accurately reflect the target marginal distribution.
Many thanks for the insightful comments.
Weakness 1:
The value of the potential function's variance do effect the Rao-Blackwellized sampler. If the variance is too small the effect of the Rao-Blackwellization is negligible, if the variance is too large the performance will decrease. In future research, we plan to explore more advanced methods for formulating the potential function for time series imputation tasks.
Weakness 2:
Since the VAE model is trained on clean target data, we can leverage its denoising/self-correction effect to force the reconstructed intermediate samples closer to the ground truth during the sampling process. An ideal choice of the potential function should be convex.
Weakness 3:
If the VAE model's performance is too poor, it cannot be considered a sufficient statistic and, consequently, will not aid in the data generation process.
Questions:
Unfortunately, in the current implementation we cannot compute the likelihood of the samples obtained because the sampler is deterministic. We will leave this problem for future research.
Limitations:
This is true, although we tried the methods proposed in existing literature, we do not find a satisfactory method to estimate the diffusion term to achieve competitive performance on time series imputation tasks. Therefore, we plan to explore better methods for estimating the diffusion term in the flow matching model as part of our future work.
This work trains a conditional flow model from noise to time series data. A VAE is used to estimate the data density then perform interleaved flow and density gradient ascent steps to generate new time series. This is referred to as a Rao-Blackwellization procedure. It is shown empirically that the model performs well on time series imputation tasks across two datasets, and that the VAE-based guidance helps in one case.
优点
- Originality: Time series imputation is an important problem and I have not seen any flow matching works in towards this application. Presents a new improvement to the standard flow matching framework with a “Rao-Blackwellization”.
缺点
- Quality: Experiments are limited in scope and the it is not clear whether the “Rao-Blackwellization” (the main methodological novelty) step reliably improves performance.
- In order to claim “Less sampling steps” it would be great to understand how performance changes with the number of steps for both diffusion and flow-based models.
- More than 2 time series datasets would allow an understanding of when this “Rao-Blackwellized” sampler is valid.
- No error bars (even though this is claimed in the checklist)
- The empirical support is lacking for the effectiveness of the “Rao-Blackwellization” step. Currently it is shown to slightly benefit (although without error bars this is difficult to tell how much) on a single dataset. In this case, an improvement in RMSE from 18.2 to 18.1 on a single dataset seems insignificant without additional detail. While it is claimed that “the PhysioNet dataset does not have enough non-zero datapoints to train a valid VAE model”, a Rao-Blackwellized sampler should always work regardless of the performance of the estimate? It would be good to see if it helps at all on this dataset and others.
- Clarity: It is not clear to me that this even is a “Rao-Blackwellized” sampler. As far as I can tell this “Rao-Blackwellized” sampler is fundamentally biased (while the original sampler is not), and it therefore cannot possibly be a Rao-Blackwellized sampler. It would be great if the authors could prove that Algorithm 2 is really a Rao-Blackwellization (at least under some conditions).
- Significance: Without further experimental benchmarking especially with regards to the novel sampler, the applicability of this method is very limited.
问题
How is this a Rao-Blackwellization, and what does the paper
Minor comments:
Shrodinger —> Schrodinger throughout
paragraph line 146 is confusing. this equation is strange given that p(x^tar) and p(x^input) are dependent if taking batches.
局限性
I'm a bit confused about the limitations. It is claimed that
One limitation of CLWF is that the samples obtained are not diverse enough as we use ODE for inference, which results in slightly higher test (continuous ranked probability score) CRPS compared to previous works, e.g., CSDI. Therefore, for future work, we will seek suitable approaches to accurately model the diffusion term in the SDE.
Other ODE models have had no problem with diversity. I would suggest that this is actually due to performing gradient ascent on the density of the VAE.
Limitations:
''' Other ODE models have had no problem with diversity. I would suggest that this is actually due to performing gradient ascent on the density of the VAE.''' Other ODE models also have the problem with diversity to some extent. The usage of the VAE is not relevant on this issue, which, on the contrary, will slightly increase the randomnesses of the sampling procedure as the VAE is a probabilistic model. To clarify the statement, let's consider an ODE sampler described by
$
X_T = X_0 + \int_0^T \mu (X_t,t)dt,
$
and an SDE sampler described by
$
X_T = X_0 + \int_0^T \mu (X_t,t)dt + \int_0^T \sigma (X_t,t)dW_t.
$
From the above equations you can see that the randomness of the ODE sampler only arises from (the initial random noise) as the sampling process is deterministic. Thus the diversity of the samples obtained is limited as the velocity model is biased towards the training data, which may suffer the overfitting issue. Please note that the training targets are only some finite samples of the target distribution but not the target distribution itself. On the contrary, the SDE-based sampling procedure can result in different samples even the initial value is the same datapoint due to the existence of the diffusion term .
Test imputation results on synthetic data ( trials, values are multiplied by .
| Method | Synthetic 0.4 | Synthetic 0.6 | Synthetic 0.8 | |||
|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CLWF(no RB) | ||||||
| CLWF(with RB) |
Test imputation results on PM 2.5 with different simulation steps ( trials)
| Method | 5 steps | 10 steps | 15 steps | 20 steps | ||||
|---|---|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CSDI | ||||||||
| CLWF |
Test imputation results on PhysioNet 0.1 with different simulation steps ( trials).
| Method | 5 steps | 10 steps | 15 steps | 20 steps | ||||
|---|---|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CSDI | ||||||||
| CLWF |
Test imputation results on PhysioNet 0.5 with different simulation steps ( trials).
| Method | 5 steps | 10 steps | 15 steps | 20 steps | ||||
|---|---|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CSDI | ||||||||
| CLWF |
Test imputation results with error bars ( trials).
| Method | PM 2.5 | PhysioNet 0.1 | PhysioNet 0.5 | |||
|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | RMSE | MAE | |
| CLWF(no RB) | ||||||
| CLWF(with RB) |
I thank the authors for their detailed response, and additional experiments. Particularly when a small number of steps are used.
Justification of Rao-Blackwellization
After the authors' response, I am fairly convinced this not a valid Rao-Blackwellization. The provided is not a valid proof that this is a valid Rao-Blackwellization, only that Rao-Blackwell samplers in general are valid.
More specifically I believe at least two things are necessary to show that this is a valid Rao-Blackwellization:
Why is ? This does not seem possible to me. Lets take the simple example of an OT-FM model with and then the optimal which we will call . . In fact any non-zero quantity will bias our estimate of Furthermore, following the suggested “Rao-Blackwellized” estimator, will not lead to . It will be over concentrated towards the mean at every step.
Why is a sufficient statistic? It seems impossible as the VAE does not depend on for it to be a sufficient statistic for which is time dependent.
Other ODE models also have the problem with diversity to some extent.
This is simply not true. For every SDE there is an equivalent ODE with equivalent diversity through the probability flow ODE for any diversity metric measured on .
The experimental results are still extremely limited. The fact that this “Rao-Blackwellization” can only be tuned to work on one of the two datasets is not convincing. Also why are the provided results different than those in the paper?
Note that the PhysioNet dataset does not have enough non-zero data points to train a valid VAE model, therefore we only construct the Rao-Blackwellized sampler for the PM 2.5 dataset.
If this is a true Rao-Blackwellization, then the estimate should be no worse than the simple estimator. As the authors state, it may not help but it should not hurt. This is further evidence to me that this theoretical result is invalid.
My feeling is that that CLWF without (”Rao-Blackwellization”), which is equivalent to OT-CFM (Tong et al. 2023) / Minibatch-OT FM (Pooladian et al. 2023), should work quite well in this setting. However, the “Rao-Blackwellization” is not proven anywhere theoretically and the response does not justify it sufficiently. Furthermore the experimental results are only applied on a single dataset for a method that should be no worse than the baseline. I keep my current score of 2. However, am open to substantially raising it if the authors can show that this is a valid Rao-Blackwellization.
Q: Why is ?
A: The proposed sampling step is not (which the reviewer framed), but , which is fundamentally different.
Q: the optimal which we will call
A: Why the optimal velocity is in your case ( and )? You certainly cannot generate the target distribution via (not converging, more iterations always result in larger values).
Q: every SDE there is an equivalent ODE
A: An ODE is always deterministic, which certainly limits its diversity to some extent.
Q: It will be over concentrated towards the mean at every step.'''
A: That is why this is an ODE sampler.
Q: It seems impossible as the VAE does not depend on for it to be a sufficient statistic for which is time dependent.
A: Please refer to the concepts of gradient flow and Langevin dynamics.
Q: The experimental results are still extremely limited. The fact that this “Rao-Blackwellization” can only be tuned to work on one of the two datasets is not convincing. Also why are the provided results different than those in the paper?"
A: We added additional experiments, please check the results carefully.
Q: Also why are the provided results different than those in the paper?
A: Because the learning process and sampling process are random like any other stochastic learning methods.
Thank you for the quick response.
The proposed sampling step is not (which the reviewer framed), but, , which is fundamentally different
Okay great. I agree this is fundamentally different. This is closer, but still needs to be proven in my opinion. I also don't understand how this lines up with Algorithm 2 which clearly states . How is this implemented practically?
Why the optimal velocity is ...
This is the optimal velocity for a flow model which is integrated from time zero to time one.
Because
An ODE is always deterministic, which certainly limits its diversity to some extent.
Yes, conditioned on an initial point , an ODE has limited diversity, however conditioned on , it does not.
Please refer to the concepts of gradient flow and Langevin dynamics.
How does this prove that the VAE is a sufficient statistic for ? Or does this not need to be shown.
We added additional experiments, please check the results carefully.
I have. There are no additional experiments on the "Rao-Blackwellization" that I can see.
Thanks for the quick reply.
How is this implemented practically?
At each step we first update the sample via , then obtain the new sample using the previous sample with (you can take this as a correction step).
Yes, conditioned on an initial point $x_0$, an ODE has limited diversity, however conditioned on $p_0$, it does not.
It still does, because in this case the diversity of the samples obtained only relies on the initial noise (which also appears in the training samples), which certainly causes that 1) it has lost some level of diversity during the sampling process compared SDE, 2) and the overfitting issue is more serious (especially when the variance of is small).
How does this prove that the VAE is a sufficient statistic for $X_{t+1}$? Or does this not need to be shown.
It suggests that the potential function does not need to be time-dependent function (as in our case). And that the derivative of the potential function can represent the sampling dynamics, thus it is a sufficient statistic for .
There are no additional experiments on the "Rao-Blackwellization" that I can see.
Please see the new results from the first Table (in the previous comment and the uploaded PDF) with CLWF(no RB) and CLWF(with RB).
Previous comments
$\nabla_x U(X_t) = \nabla \log N(t, 1) = t - x_t$.
It should be .
any non-zero quantity will bias our estimate $X_{t+1} | X_t = \mu_t^\star(x) = 1.$
in this case the velocity is optimal, then the derivative of the potential function will be zero, which will not affect the sampling result.
Thank you very much for the constructive feedback and insightful comments
Weaknesses (Quality) 1:
We have conducted the ablation study on the impacts of sampling steps of CSDI and CLWF. The results show that CLWF has better performance when the simulation steps are small. Please also take a look at the uploaded PDF for the detailed experimental results.
Weaknesses (Quality) 2:
We have conducted additional experiment using the synthetic dataset. And the experimental results demonstrates the effectiveness of the Rao-Blackwellization, please see the uploaded PDF.
Weaknesses (Quality) 3:
We have added the error bars to the experimental results as suggested, please see the uploaded PDF.
Weaknesses (Quality) 4:
Compared to the SOTA methods, whose performances are already nearly optimal, the performance of our method has been significantly improved. In this situation, the Rao-Blackwellization can further enhance performance. Even without substantial improvements in values, it still demonstrates the effectiveness of our method.
Weaknesses (Clarity):
As far as I can tell this “Rao-Blackwellized” sampler is fundamentally biased (while the original sampler is not), and it therefore cannot possibly be a Rao-Blackwellized sampler.
The Rao-Blackwell theorem only requires the base/original sampler is unbiased and the statistic is sufficient (not necessarily unbiased). According to the law of Total Expectation, the new sampler is also unbiased. Hence, one cannot say the new sampler is not Rao-Blackwellized.
Justification for Rao-Blackwellization:
Let's first consider a dynamic system which can be described by the following SDE:
$
dX_t &= \mu_t(X_t,t)dt + \sigma_t dW_t,
$
where is the velocity and can be solved via flow matching.
Alternatively, if we assume the above system is a over-damped Langevin dynamic system, then system can be described by the following SDE:
$
dX_t &= \nabla V_t(X_t,t)dt + \sigma_t dW_t
$
where is the drift.
Essentially, we have two different mathematical interpretations for the same system.
Note that (using flow matching) and (using the potential function) are computed via different numerical methods, but ideally, they both suffice to represent the same system independently.
We also assume the Gaussian potential function constructed via the VAE model is accurate enough.
Since is the parameter we want to estimate, according to the Rao-Blackwell theorem, we construct the base sampler based on the flow matching model which is unbiased for . And based on is a sufficient statistic for , meaning that contains all the necessary information for estimating but does not dependent on . Afterwards, we define the new Rao-Blackwellized sampler by letting the base sampler condition on the sufficient statistic.
Now we prove that the new sampler is unbiased, i.e., , based on the Law of Total Expectation (or the Tower Rule).
$
\mathbb{E}[\mathcal{S}^*] = \mathbb{E}[\mathbb{E}[\mathcal{S}|\mathcal{T}]] & = \iint s p(s|t)ds p(t)dt \nonumber \ & = \iint s p(s|t) p(t)dsdt = \iint s p(s,t) dsdt = \int s p(s) dt = \mathbb{E}[\mathcal{S}].
$
Since is unbiased, is unbiased as well.
By doing so, we can leverage the denoising/self-correction effect of the VAE model (considering that it is trained on clean data), which enables the reconstructed intermediate samples closer to the ground truth during the sampling process.
If the above explanation is still not intuitive enough, one can think about a Kalman filter, whereby the first model is the system's dynamics and the second model is the noisy observation.
The noise observation can be used to improve the trajectory estimation during tacking as long as it is sufficient.
Please see new experimental results reported in the uploaded PDF.
From the experiments results, we can see that the variances are reduced, and the improvement can be verified by calculating the RMSE. For the future work, we will explore new techniques to further improve the stability of the Rao-Blackwellized sampler.
Weaknesses (Significance):
In this paper, we propose a novel framework based on Lagrangian dynamics and optimal transport for time series imputation and a novel sampler to enhance its performance. Our contributions are both theoretical and applied. We have also added new experimental results as the reviewer suggest, please refer to the uploaded PDF for more details.
Questions 1:
```
How is this a Rao-Blackwellization, and what does the paper
```
Since the comment does not fully display, we will only reply the first part of the question. Please see the above response to the issue regarding the Rao-Blackwellization.
Questions 2:
We have corrected the typos as suggested.
Questions 3:
paragraph line 146 is confusing. this equation is strange given that $p(x^tar)$ and $p(x^input)$ are dependent if taking batches.
The authors are confused by the comment as paragraph line 146 is not an equation. Doe the reviewer refers to Eq. (14)? This is the probabilistic coupling in context of optimal transport. Please refer to [1].
Reference:
[1] Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow straight and fast: Learning to generate and transfer data with rectified flow. In The Eleventh International Conference on Learning Representations, 2022.
The paper presents Conditional Lagrangian Wasserstein Flow (CLWF), a new method for time series imputation. Using (entropic) optimal transport theory and Lagrangian mechanics, CLWF generates high-quality samples. Enhanced with a Rao-Blackwellized sampler, CLWF incorporates prior information through a variational autoencoder. Experiments on real-world datasets show that CLWF performs competitively.
优点
- The paper is well-written, clear, and easy-to-follow.
- The use of a Rao-Blackwellized sampler and a variational autoencoder to integrate prior information enhances the model's performance, providing a more robust imputation process.
缺点
- The conditional generation process is trained through the simulation-free training of the Schrödinger Bridge, which limits the novelty of the CLWF.
- The paper mentions that the samples obtained using the ODE-based inference method may lack diversity, potentially leading to higher continuous ranked probability scores (CRPS) compared to some previous works.
- While the method is tested on two real-world datasets, broader evaluation across more diverse and challenging datasets could strengthen the validation of the approach.
问题
I have no other questions.
局限性
Yes
Thank you for your valuable feedback.
Weakness 1:
Indeed, CLWF is trained through the simulation-free training of the Schrödinger Bridge. However, in this work, based on the Lagrangian mechanics framework, we specifically proposed the Rao-Blackwellized sampler to further enhance the model's performance for time series imputation tasks.
Weakness 2:
The ODE sampler is adopted to optimize the performance of our method for the time series imputation tasks. However, an SDE sampler is also applicable within our proposed framework.
Since our method is implemented via an ODE sampler CRPS may not be a very suitable evaluation metric.
[1] and [2] suggest that CRPS could lead to an incorrect assessment of overall model performance due to the overlook of the model's performance on each dimension of the data.
For the future work, we plan to explore more advanced methods to improve the diversity of the samples to have better empirical performance regarding CRPS .
Weakness 3:
The datasets used in our experiments, PM 2.5 and Physio, are both highly representative in the field of time series imputation. We have also included additional experimental results using synthetic data to further demonstrate the advantages of our methods. Please refer to the uploaded PDF for more detailed experimental information.
References:
[1] Koochali, Alireza, Peter Schichtel, Andreas Dengel, and Sheraz Ahmed. "Random noise vs. state-of-the-art probabilistic forecasting methods: A case study on CRPS-Sum discrimination ability." Applied Sciences 12, no. 10 (2022): 5104.
[2] Biloš, Marin, Kashif Rasul, Anderson Schneider, Yuriy Nevmyvaka, and Stephan Günnemann. "Modeling temporal data as continuous functions with stochastic process diffusion." In International Conference on Machine Learning, pp. 2452-2470. PMLR, 2023.
The authors' reply addresses most of my issues. I appreciate the clarification made by the authors, and I have no other concerns.
We would like to thank all the reviewers for their insightful comments and try our best to address all the questions regarding this work.
To this end, we have included additional experimental results from an ablation study. Specifically, we conducted new experiments using synthetic data to validate the effectiveness of the Rao-Blackwellization. We also employed the real-world datasets to demonstrate that CLWF achieves better performance with fewer simulation steps.
Furthermore, all results are presented with trials, showing the corresponding means and standard deviations. Please refer to the uploaded PDF file for detailed experimental results.
This paper deals with the problem of time series imputation. It introduces a method that uses tools from transport theory, Lagrangian mechanics, and optimal transport to generate samples.
This paper received mixed reviews (1x accept, 2x weak accept, 1x borderline accept, 1x strong reject). After the authors' rebuttal, some reviewers stated that their concerns had been addressed and maintained their score. Reviewer 3hN3, however, points to some remaining flaws of the paper. After the rebuttal period, only Reviewer 3hN3 participated in the follow-up discussion between reviewers and AC, and thus a consensus among the reviewers was not reached.
Because of the mixed opinions, I had a careful look at the paper myself. Unfortunately, I found it does not meet the acceptance bar and therefore I encourage the authors to improve their manuscript using the received feedback and submit to another conference.
Besides the issues found by Reviewer 3hN3, I noted that the paper contains basic mistakes in the mathematical equations. As an example, consider Eq. (7) (in the "Background" section), which is: (where the inner integral is over ).
- First, note that the square bracket isn't closed - I assume that's a simple typo and that the closing bracket should be after .
- Second, note that the inner integral has but there is no in the integrand - I assume that it should be instead.
- Third, if we integrate out , then the expectation would not depend on at all, and therefore it can be safely ignored as everything inside is a constant - which should obviously not be the case.
- Fourth, after we integrate out both and , the result does not depend on either or , whereas the LHS of the equation clearly depends on both - it is written as .
As another example of wrong mathematics, Eq. (8) suffers from similar issues.
Besides these mistakes, I would like to note that the paper contains some unsupported claims. Two examples are:
- In Section 3.5 ("Rao-Blackwellized sampler"), the text says: "Note that the ODE sampler alone is good enough to generate high-quality samples for time series imputation". This leaves it unclear why switching from SDE to ODE is okay, as it's never discussed further.
- In Section 3.4 ("Potential function"), the text says: "In terms of practical implementation, we parameterize via a Variational Autoencoder (VAE)". Once again, it remains unclear why that is an okay thing to do, since the (time-independent) VAE is fitted to the observed data only, and it might suffer from some biases that would propagate further. Reviewer 3hN3 also raised this point: "More fundamentally I'm unclear why effectively taking steps towards data density of a VAE trained on a related (but fundamentally different) data distribution is even a good idea [...]".
Another point, raised by Reviewer 3hN3 (and which I agree with), is the lack of justification of the Rao-Blackwellized sampler. Section 3.5 does not clarify why this is a valid sampler, since the expectation is not correct. Unfortunately, despite this being one of the main contributions of the paper, this concern remained unaddressed after the rebuttal phase.
Even if the Rao-Blackwellized sampler were not theoretically valid, it could still be empirically useful; however, the experimental results do not support that conclusion (there is only a single result on a single dataset with marginally better results). It is also suspicious that, despite considering two datasets, the Rao-Blackwellized sampler is consctructed for one of them only, ignoring the second one.
To conclude, I think the paper would need a substantial changes before it is ready for acceptance at a conference like NeurIPS.