On Error Propagation of Diffusion Models
We present a theoretical framework to explain why error propagation happens to diffusion models and a regularization to address this problem
摘要
评审与讨论
This paper presents a theoretical framework to quantify per iteration error in variational inference within the backward pass of difussion models. The authors introduce metrics to theoretically quantify these errors and convincingly demonstrate their increase with iteration number. Consequently, the model's performance deteriorates as the number of iterations increases. To mitigate this issue, the authors propose to add the error as a regularization term upon the original loss function. Empirical experiments are conducted to validate the effectiveness of this proposed methodology.
优点
-
The paper is very well written and I really enjoys reading this paper.
-
The concept of error quantification presented by the authors is both innovative and well-conceived.
缺点
-
Modular Error Definition: The modular error is defined as the expected KL divergence from to . However, the KL divergence is not symmetric. Could the authors please explain why the modular error is defined in such a way? Why not using the reversed KL divergence as in the original loss function?
-
Assumptions in Theorem 3.1: The assumptions regarding the output distribution of the neural network (NN) following a standard Gaussian distribution and the entropy reduction with iteration number appear quite strong. Can the authors justify these two assumptions empirically?
-
Technical Issues in the Proof:
-
In the proof of Proposition A.1 in Appendix A, equation (19) holds only in the limit as . The current proof falls short of demonstrating Proposition A.1 for finite values of .
-
Several typographical errors in equation (24) in Appendix B raise concerns about the validity of the proof. Specifically, the term in both the numerator and denominator should be consistent and denoted as in order for equation (29) to hold. Additionally, the first term in the second line of (24) appears to be missing a logarithmic notation, and the third line of (24) appears inconsistent with the current notation.
-
-
Hyper-Parameter Selection: The authors should provide further clarification on the process for selecting hyper-parameters. Notably, the weight assigned to the regularization term is a critical factor that requires elucidation. How to select the hyper-parameters? One such important factor is the weight of the regularization term, could the authors please clarify?
-
Minor Issues:
-
The notation for conditional expectation should be corrected throughout the paper to .
-
In Section 3.2, second paragraph, "get ride of" should be corrected to "get rid of".
-
The proof for (14), establishing the boundedness of the KL divergence by the MMD, rests on the assumption that is well-defined, which only holds when . The authors should address and clarify this point.
-
问题
Please refer to the questions in the previous section.
Comment-3: In the proof of Proposition A.1 in Appendix A … The current proof falls short … for finite values of T.
Answer-3: Thanks for pointing out this. We address your concern by introducing a weaker precondition, which still makes our proposition hold and is also supported by current works [1,2].
In our proof to Proposition A.1, the use of that precondition is to let . Considering the fact that
,
we can achieve the goal with a much weaker assumption: . Notably, this new assumption is a standard configuration in current diffusion models (e.g., DDPM [1] and SGM [2]), which lets contain no information about .
Comment-4: Several typographical errors in equation (24) in Appendix B … both the numerator and denominator should be consistent … the first term in the second line of (24) … the third line of (24) appears inconsistent …
Answer-4: Thanks for pointing out this. The correct forms of the first two lines of our Eq. (24) are
.
We also fix the last line of Eq. (24) as
.
We will add the above corrections to the revised version.
Comment-5: … clarification on the process for selecting hyper-parameters … the weight assigned to the regularization term …
Answer-5: We applied the grid search to hyper-parameter selection. For the weights of regularization terms, we respectively construct candidate sets and for and . As mentioned in Sec. 6.1 of our paper, the experiments turned out that this combination performed the best.
Comment-6: Minor Issues: … The notation for conditional expectation … "get ride of" should be corrected to "get rid of" … rests on the assumption that log(1 - x/4) is well-defined, which only holds when x < 4 …
Answer-6: Thanks for raising these points. We will accept your suggestions about the notation and the typo in the revised version.
For your comment “x < 4”, we prove that such that . Recall that the definition of is as
,
where . According to the Triangle Inequality, we have
,
Since is a convex function, we can apply Jensen's inequality to the above equation:
,
showing that our claim holds.
Besides the above proof, the experiments in Fig. 2 of our paper also empirically confirmed that .
References
[1] Ho et al., Denoising Diffusion Probabilistic Models, NeurlPS-2020.
[2] Song et al, Score-Based Generative Modeling through Stochastic Differential Equations, ICLR-2021.
Dear Reviewer ebh3,
We thank you for your kind, constructive, and comprehensive feedback. In the following, we have answered all your concerns in a point-by-point manner.
Comment-1: Modular Error Definition … the KL divergence is not symmetric. Could the authors please explain why the modular error is defined in such a way? …
Answer-1: Thanks for raising this insightful point. We did consider the asymmetry of KL divergence. The key is that the error propagation happens to the learnable backward process (instead of the predefined forward process ). Therefore, it makes more sense to define the module error as
,
such that expectation integrals are operated on the distributions of the backward process to average the distribution gap . For your mentioned reversed case, the expectation operations will be mistakenly applied to the distributions of the forward process.
Comment-2: Assumptions in Theorem 3.1 … Can the authors justify these two assumptions empirically?
Answer-2: Thanks for raising this point. For our assumption on , we first weaken it into a new assumption from a theoretical angle and then perform experiments to empirically verify its validity.
Since neural network is designed to fit Gaussian noise in the loss function , we previously supposed that the output distribution of follows a standard Gaussian. In our proof to Theorem 3.1, that assumption is applied to indicate that ( is the vector dimension). However, considering Eq. (27) and Eq. (28) in the proof, our theorem still holds for .
To derive a new assumption, we first set a term
,
which indicates the prediction error of neural network . The term will vanish to if is fully accurate in backward denoising. According to the Triangle Inequality, we then have
.
Since the second moment of Gaussian distribution is , we further have
.
Finally, consider the fact [1] that can be reparameterized as and let the prediction error vanish, the above equation motivates us to make the following new assumption:
,
which is much weaker than the previous assumption but still makes our Theorem 3.1 hold.
We conduct experiments on 2 datasets to verify our new assumption. For every dataset, we sample trajectories from a well-trained diffusion model. In this process, we collect outputs from neural network at every step and use them to estimate . Below are the experiment results.
| Dataset | t=1000 | t=800 | t=600 | t=400 | t=200 | t=1 |
|---|---|---|---|---|---|---|
| CIFAR-10 | 1.17 | 1.19 | 1.15 | 1.11 | 1.08 | 1.05 |
| CelebA | 1.25 | 1.21 | 1.23 | 1.17 | 1.11 | 1.09 |
From this table, we can see that the estimated value of is consistently larger than at every backward iteration for both datasets. Therefore, our new assumption has firm empirical support.
Because it is computationally infeasible to estimate the probabilistic density for discrete-time diffusion models (e.g., DDPM [1]), we find it hard to empirically verify our assumption on the entropy . However, we believe that assumption is very intuitive because the uncertainty of variable (i.e., entropy) should decrease as the denoising process progresses. For example, if we are sampling from a diffusion model and find that seems like a dog, it’s less likely that the final outcome (that is denoised from ) will be a cat.
We will add the above new assumption, derivation, experiment, and discussion to the revised version.
The authors have addressed my concerns and I will keep my rating as before.
Thank you, and thanks again for your constructive review!
This paper analyses the error propagation/accumulation in DMs across iterations. The authors prove and empirically verify that the error in DMs cumulatively increases. To minimize this cumulative error as regularization, the authors prove tractable estimates which tightly bound this error, which is then used as a proxy. The authors empirically show the proposed method reduces the cumulative error, and increases generation quality across multiple datasets.
优点
- The theoretical framework and the bounds on error propagation through DMs are useful for analyzing robustness of DMs.
- The proposed method results in strong significant improvements across a range of datasets.
- The proposed method successfully decreases cumulative error in DMs
缺点
- The proposed method requires significant compute overhead, so gains need to be weighed against this increase in compute.
问题
- Contemporaneous work[1] (released after submission deadline) also analyses the sensitivity of DMs to error propagation, and aims to bound this error by scaling the long skip-connections. While the method in [1] is sufficiently different to the proposed method, could the authors comments on this? It would appear that [1] also bounds the error, without the computational overhead.
- The proposed method has significant computational overhead (Figure 4). How does the comparisons to baselines (Table 1) change at equal wall-clock time?
Dear Reviewer hknr,
We thank you for your kind and constructive feedback.
Comment-1: Contemporaneous work[1] (released after submission deadline) also analyses the sensitivity of DMs to error propagation … could the authors comments on this? …
Answer-1: Thanks for pointing out this interesting paper. The paper studied how the coefficient scales of skip connections affect the training stability of U-Net and introduced two new scaling schemes (one is predefined scales and the other is learnable ones) to better model the skip connections. For the experiment, the paper showed that their proposed methods stabilized and accelerated the training of U-Net.
Your mentioned paper differs from our work in the following 3 points:
- In terms of research topic, the paper focused on the training stability of U-Net (which is inside the denoising model ), while our work concentrates on the error accumulation of the backward process (which is outside the model);
- In terms of method, the paper proposed to better scale the skip connections of U-Net (which makes its training more stable and faster), while our work introduces a regularization loss to reduce the error propagation of the backward process (which improves the generation quality);
- In terms of theory, the paper developed theorems that estimated the feature norms and gradient magnitudes of U-Net (which is inside the model ), while our work builds a theoretical framework to analyze the error propagation of backward process (which is inside the model).
For your comment “[1] also bounds the error”, note that the error defined in that paper (i.e., the sensitivity of U-Net w.r.t. the input perturbation) is different from our defined errors (i.e., the denoising error and its accumulation along the backward process). Regarding your concern about the computational overhead, our model indeed has a higher time cost for training, but the generation quality of diffusion models is also improved accordingly.
We will cite your mentioned paper and include the above discussion in the revised version.
Comment-2: … gains need to be weighed against this increase in compute. The proposed method has significant computational overhead (Figure 4). How does the comparisons to baselines (Table 1) change at equal wall-clock time?
Answer-2: Thanks for raising this point. A reminder is that the y-axis in Fig. 4 of our paper actually represents the time cost of training steps, rather than just step. We apologize if this mistake has misled you to the impression that our method is very inefficient.
As you suggested, we have compared our model with the two most efficient baselines (i.e., DDPM and DDIM) on CIFAR-10 with the same training time. For Consistent DM and FP-Diffusion, our model performs better than them and has a lower time cost per training step. The experiment results of FID scores and time costs are as follows.
| Model | Training Steps | Total Training Time | FID Score |
|---|---|---|---|
| DDPM | 200K | 10hrs | 3.61 |
| ADM-IP | 190K | 10hrs | 3.29 |
| DDPM w/ Our Method | 98K | 10hrs | 3.16 |
| DDPM w/ Our Method | 200K | 20hrs | 2.93 |
From the above table, we can see that our proposed method still outperforms the baselines with the same amount of training time. For example, our method reduces the FID score of DDPM on CIFAR-10 by 12.47%, using no more than half of its training steps. Besides, while leading to a higher time cost per training step, our method has no impact on the inference speed of diffusion models, which is more important in practical applications.
This work analyze the error propagation of diffusion models by introducing the modular error (KL divergence between the reverse conditional distribution at each step), the cumulative error (KL divergence between the marginal distribution at each step). And then introduce an regularization loss based on MMD estimation for the cumulative loss to reduce the cumulative error and improve the sample quality of the trained diffusion model.
优点
- The proposed method is easy to understand and the writing is clean and easy to follow.
- The error propagation of diffusion models is an important question and worth to be studied. The topic is important.
缺点
Major:
- The assumption in the core theorem is absolutely wrong.
- "suppose that the output of neural network follows a standard Gaussian", which cannot be true. Because the noise-pred model corresponds to the denoising score matching loss, it is proved that the ground truth of such model is propotional to the score function of the distribution, i.e., . For a small , such score function is quite complex and cannot be a simple and single-mode Gaussian distribution, and is far different.
- Remark 3.2 is not rigorous. The proof requires goes to infty, but it is not true in practice.
- Lack of detailed settings of experiments: what is the sampling algorithm for obtaining the FID results? What is the detailed network structure (e.g., layer structure and number of hidden neurons) and amount of parameters?
问题
-
Please address and fix the proof of the main theorem.
-
Please add more detailed descriptions of the experiment settings to ensure reproducibility.
=====================
I've carefully read the authors' responses to other reviewers and AC. I deeply appreciate the authors' efforts to address the concerns, and now I don't have more questions. I think it is a quite interesting and useful technique for improving the training procedure of diffusion models, with the validation of the fine-tuning experiment results and the EDM results. So I raise the score to 6.
Comment-2: Remark 3.2 is not rigorous. The proof requires goes to infty, but it is not true in practice.
Answer-2: Thanks for pointing out this. We address your concern by introducing an alternative precondition that is weaker than the previous one but still makes our proposition hold. Importantly, our new precondition is supported by current works [1,2].
In our proof to Proposition A.1, the use of that precondition is to let . Considering the fact [1] that
,
we can achieve the goal with a much weaker assumption: . Notably, this new assumption is a standard configuration in current diffusion models (e.g., DDPM [1] and SGM [2]), which lets contain no information about .
Comment-3: Lack of detailed settings of experiments: what is the sampling algorithm for obtaining the FID results? What is the detailed network structure (e.g., layer structure and number of hidden neurons) and amount of parameters?
Answer-3: For computing the FID scores, we adopt the same sampling procedure as DDPM (see Algo. 2 of [1]) and generate samples with backward iterations. For the network structure, we typically use U-Net with layers, channels, and channel multipliers as , leading to a model size of 300M. We use Adam as the optimization algorithm with a learning rate of and set the dropout ratio as .
Sec. 6.1 of our paper has some other details of the experiment setup. We will merge this answer into that section in the revised version.
References
[1] Ho et al., Denoising Diffusion Probabilistic Models, NeurlPS-2020.
[2] Song et al, Score-Based Generative Modeling through Stochastic Differential Equations, ICLR-2021.
Dear Reviewer e2Ed,
Thanks for your review and the constructive feedback.
In the following, we have answered all your concerns in a point-by-point manner. Importantly, to address your concern about the two preconditions of our theorems, we introduce alternative assumptions that are much weaker but still make the theorems hold. Our new assumptions are also very solid because we derive them from a theoretical angle and find support from either empirical experiments or existing works [1,2].
Comment-1: … “suppose that the output of neural network follows a standard Gaussian", which cannot be true. Because the noise-pred model corresponds to the denoising score matching …
Answer-1: Thanks for raising this point. In this answer, we first explain the role of your mentioned assumption in Theorem 3.1 and then replace it with a much weaker assumption, which is derived from a theoretical perspective and still makes the theorem hold. Lastly, we perform experiments to show that our new assumption is indeed valid in practice.
Since neural network is tasked to fit Gaussian noise in the loss function , we previously assumed that the output distribution of follows a standard Gaussian. In our proof to Theorem 3.1, that assumption is applied to indicate that ( is the vector dimension). However, considering Eq. (27) and Eq. (28) in the proof, our theorem still holds for .
To derive a new assumption, we first set a term
,
which indicates the prediction error of neural network . The term will vanish to if is fully accurate in backward denoising. According to the Triangle Inequality, we then have
.
Because the second moment of Gaussian distribution is , we further have
.
Finally, consider the fact [1] that can be reparameterized as and let the prediction error vanish, the above equation motivates us to make the following new assumption:
,
which is much weaker than the previous assumption but still makes our Theorem 3.1 hold.
We perform experiments on 2 datasets to verify our new assumption. For every dataset, we sample trajectories from a well-trained diffusion model. In this process, we collect outputs from neural network at every step and use them to estimate . Below are the experiment results.
| Dataset | t=1000 | t=800 | t=600 | t=400 | t=200 | t=1 |
|---|---|---|---|---|---|---|
| CIFAR-10 | 1.17 | 1.19 | 1.15 | 1.11 | 1.08 | 1.05 |
| CelebA | 1.25 | 1.21 | 1.23 | 1.17 | 1.11 | 1.09 |
From the above table, we can see that the estimated value of is consistently larger than at every backward iteration for both datasets. Therefore, our new assumption has strong empirical support.
Dear Reviewer e2Ed,
We thank you for your time in reviewing our paper. With only 2 days left, we would like to know whether our previous response has addressed your concerns. Looking forward to your feedback!
Best regards,
The Authors
I've carefully read the authors' responses to other reviewers and AC. I deeply appreciate the authors' efforts to address the concerns, and now I don't have more questions. I think it is a quite interesting and useful technique for improving the training procedure of diffusion models, with the validation of the fine-tuning experiment results and the EDM results. So I raise the score to 6.
Thank you, and thanks for your constructive review!
Perhaps the authors could simply measure the gradients, and compare its histogram's R2 to the best fit gaussian? That may provide an alternative path to using the gaussian assumption.
Thanks for your kind suggestion. We did consider this way, but we finally weakened that assumption to make our theorem more solid.
In this paper, the authors investigate error propagation in diffusion models. They develop a framework to define error propagation for diffusion and connect error propagation to generation quality. This enables them to use the measured error as a regularization term during the diffusion model training, improving the generation results of models in small-scale experiments.
优点
-
The proposed method is a novel approach to measuring error propagation in diffusion models and offers a new perspective on diffusion model training. The authors argue that apart from making the denoiser network more accurate, which has been the main focus of the literature so far, it is also important to regularize such that the denoiser is also robust to errors in the input during inference. This could have significant impacts on the broader diffusion generative model community.
-
The presented methodology is principled and well-explained. The experiments clearly demonstrate the success of the proposed solution in mitigating error propagation in diffusion models.
缺点
- The authors briefly address the trade-off between increased training time and reduced error propagation (resulting in better FID) for the 32x32 images of CIFAR and ImageNet but do not mention their CelebA experiments on 64x64 images. It is not clear if the benefits scale with the image sizes without an increased overhead as it is possible that the error estimate requires more samples or a larger sampling length .
问题
- Would it be possible to fine-tune pre-trained diffusion models with the regularization term to mitigate this error propagation a-posteriori? If the training time of adding the regularization makes training larger models prohibitive it would be interesting to explore whether it is possible to tune the denoiser network after having trained with just the loss.
Dear Reviewer MiNJ,
Thank you for your kind and constructive comments. In particular, the idea of fine-tuning a trained diffusion model with our proposed regularization is very interesting and useful, which can potentially reduce the training time of our model. We have conducted experiments to show that your idea indeed works in practice.
Comment-1: … do not mention their CelebA experiments on 64x64 images … It is not clear if the benefits scale with the image sizes without an increased overhead …
Answer-1: Thanks for pointing out this. As a reminder, the y-axis in Fig. 4 of our paper in fact represents the time cost of training steps (not just step). We apologize if this mistake might mislead you to the impression that our method is very inefficient.
To address your concern, we have performed an additional trade-off study (showing how the FID score and the time cost change w.r.t. increasing bootstrapping steps ) on CelebA. The results are in the following:
| Bootstrapping Steps | 2 | 3 | 4 | 5 | 6 | 7 | 8 |
|---|---|---|---|---|---|---|---|
| FID Score | 1.51 | 1.39 | 1.29 | 1.22 | 1.18 | 1.16 | 1.15 |
| Decrement of FID Score | N.A. | 0.12 | 0.10 | 0.07 | 0.04 | 0.02 | 0.01 |
| Time Cost Per Training Step | 0.37s | 0.45s | 0.55s | 0.63s | 0.72s | 0.81s | 0.92s |
From the above table, we can see the FID scores tend to converge after , which is similar to the situations of CIFAR-10 and ImageNet. The results imply that the performance improvements contributed by our proposed regularization scales with the image size. Importantly, while our method incurs an extra time cost per training step, it has no impact on the sampling speed of diffusion models, which is more important in practice.
Comment-2: Would it be possible to fine-tune pre-trained diffusion models with the regularization term … it would be interesting to explore whether it is possible to tune the denoiser network …
Answer-2: Thanks for providing such an interesting idea. To verify the feasibility of your idea, we have trained a diffusion model on CIFAR-10 in a way that only the last 10% training steps involve our proposed regularization. The results are as below.
| Model | All Training Steps | Steps with Regularization | Training Time | FID Score |
|---|---|---|---|---|
| DDPM | 200K | 0 | 10hr | 3.61 |
| DDPM w/ Our Proposed Reg. | 200K | 20K | 11hr | 3.21 |
| DDPM w/ Our Proposed Reg. | 200K | 200K | 20hr | 2.93 |
From the above table, we can see that the diffusion model is still improved much (i.e., a reduction of the FID score by 11.08%) by partly applying our method, with a minor increase in training time. Therefore, your idea indeed works empirically and is very useful in improving the efficiency of our proposed regularization.
We will include the above experiment and discussion in the revised version.
I would like to thank the authors for taking the time to address my comments. I would strongly encourage you to include the fine-tuning results in the main text as it significantly helps with the presentation of the method. Given that my score is already high, I will be maintaining my rating.
Sure, we will include your suggestion and the new experiment results in the revised version. Thank you and thanks again for your constructive review.
My examination of the reviews and rebuttal indicates that assumptions were indeed made regarding the optimized denoising network output that may not hold true. The authors have tried to address this concern, as highlighted by Reviewer e2Ed, by weakening the assumptions.
Reviewer e2Ed:
Given that only two reviewers expressed clear confidence in evaluating the merits of the paper and you were the sole reviewer to identify the aforementioned issue, I strongly encourage your participation in responding to the authors and engaging in discussions with the other reviewers and AC.
Authors:
Q1: Have you tested your algorithm on a more recent diffusion codebase, such as EDM (https://github.com/NVlabs/edm), which has demonstrated superior performance compared to your lowest FID reported on CIFAR 10, using a significantly fewer number of reverse diffusion steps? I believe conducting these additional experiments is crucial to instill confidence in the reader that your method not only addresses the limitations of the DDPM codebase, but is also generalizable to diffusion models as a whole.
Q2: Utilizing MMD to align and appears computationally expensive. Have you encountered challenges in tuning the MMD kernel? Additionally, have you considered exploring alternative distribution matching methods, such as VAE and GAN? I would also suggest establishing connections with hybrid models like VAE+Diffusion and GAN+Diffusion, as they share similarities in the need to match and .
Dear AC,
We thank you for managing the review process and for your help in improving our paper. We have also answered your questions in the following.
Response to Q1
Thanks for raising this point. As you suggested, we have conducted an experiment with the codebase of EDM. Below are the results for backward iterations.
| Method | Result Source | FID score on CIFAR-10 |
|---|---|---|
| VE SDE w/ EDM | Table 2 in EDM [1] | 1.98 |
| VE SDE w/ EDM | Our Experiment | 2.16 |
| VE SDE w/ EDM, Our Method | Our Experiment | 1.86 |
The F1 score of our trained baseline (i.e., VE SDE w/ EDM) is very close to the result reported in the original paper [1] and our proposed regularization improves this score by 13.89%. The results indicate that our method not only improves DDPM but also is applicable to diffusion models in general.
Thank you for helping us further highlight the merits of our proposed method.
Response to Q2
Thanks for your questions. MMD is much more efficient after applying our proposed bootstrapping trick (see Sec. 5.2 of our paper). Importantly, as we highlighted in the response to Reviewer hknr, our method still significantly improved the diffusion models and outperformed previous baselines with the same training time. Reviewer MiNJ also provided a very interesting idea to make our method more efficient: only applying our regularization for the last 10% training steps. Our experiments showed that this idea indeed works and we are very grateful to the reviewer.
We did consider using the adversarial loss (i.e., GAN) to align with . However, we finally chose MMD instead for the following two reasons:
- GANs are known for their training instability [2] and their sensitivity to hyper-parameter selection [3]. In contrast, MMD is parameter-free and more stable;
- As proved in our paper (i.e., Proposition 4.1), MMD tightly bounds our defined cumulative loss from below and above. By using MMD, our theory and method are connected.
To our knowledge, DDGAN [4] and Latent Diffusion [5] are two popular works that respectively combine diffusion models with GAN and VAE. However, they use GAN or VAE either as the denoising model or to pre-compress the training data. Compared with these models, our proposed regularization mainly aims to minimize the cumulative error (see Def. 3.2 of our paper), reducing error propagation along the backward process.
Once again, we thank the AC and all the reviewers for all their helpful suggestions.
References
[1] Karras et al., Elucidating the Design Space of Diffusion-Based Generative Models, NeurIPS-2022.
[2] Arjovsky and Bottou, Towards Principled Methods for Training Generative Adversarial Networks, ICLR-2017.
[3] Lucic et al., Are GANs Created Equal? A Large-Scale Study, NeurIPS-2018.
[4] Xiao et al., Tackling the Generative Learning Trilemma with Denoising Diffusion GANs, ICLR-2022.
[5] Rombach et al., High-Resolution Image Synthesis with Latent Diffusion Models, CVPR-2022.
To enhance the training and sampling processes of diffusion models (DMs), the paper explores error propagation in reverse diffusion and presents a framework to quantify and address this challenge. The authors introduce a regularization term aimed at mitigating error propagation, showcasing its effectiveness in refining DMs across three different low-resoltion image datasets. The study underscores the significance of comprehending and minimizing error propagation to elevate the sample quality in trained diffusion models.
为何不给更高分
The paper's score may not be higher due to initial critical errors in the theoretical analysis, which might have impacted the overall credibility and robustness of the presented findings. Additionally, the experimental design initially relied on the DDPM sampler with 1000 sampling steps, and although the authors addressed concerns by including results on CIFAR10 with EDM and reverse steps, there still exists a level of concern regarding the generalizability of the proposed method to diffusion models as a whole. The need for more comprehensive evaluations is highlighted, suggesting that further testing and analysis could provide a clearer understanding of the method's performance and applicability.
为何不给更低分
The score is not lower because all reviewers recommended acceptance. Although the AC has concerns about the generalizability of the observed benefits, s/he is willing to give the paper the benefit of the doubt.
Accept (poster)
Thank you very much for your excellent work!
I noticed the strong Gaussian assumption in the paper (Remark 3.3) and saw that you improved the proof during the rebuttal phase.
Could you please merge these updates into the PDF file?
Thank you again!
Dear Shikun,
Thanks for your comment.
We indeed did. Please refer to Appendix D.
Best wishes,
Yangming