PaperHub
6.0
/10
Poster5 位审稿人
最低6最高6标准差0.0
6
6
6
6
6
3.4
置信度
正确性2.8
贡献度3.0
表达2.2
NeurIPS 2024

FasterDiT: Towards Faster Diffusion Transformers Training without Architecture Modification

OpenReviewPDF
提交: 2024-04-24更新: 2024-11-06

摘要

关键词
Generative ModelDiffusion TransformersFlow Matching

评审与讨论

审稿意见
6

The paper presents FasterDiT, a method intended to speed up the training of Diffusion Transformers (DiT) without making changes to the architecture. By utilizing insights from the Probability Density Function (PDF) of Signal-to-Noise Ratio (SNR), FasterDiT improves training strategies and the effectiveness of supervision. The paper also includes a thorough set of experiments and a new supervision approach called FasterDiT.

优点

The paper introduces an innovative approach to training strategies by considering the Probability Density Function (PDF) of Signal-to-Noise Ratio (SNR), aiming to improve training efficiency. It includes a thorough empirical analysis with significative experimental results, providing robust evidence to support the findings. FasterDiT achieves significant acceleration in training Diffusion Transformers with a notable increase in training speed, making it a relevant and practical solution for enhancing the efficiency of training large-scale generative models.

缺点

The paper would benefit from additional experiments to demonstrate the generalizability of the proposed approach. A more comprehensive comparison with existing methods for accelerating training of Diffusion Transformers would provide a broader context for evaluating the effectiveness of FasterDiT. Additionally, the scalability of FasterDiT to larger datasets or more complex tasks is not sufficiently discussed, potentially limiting its applicability. The absence of theoretical results and proofs in the paper limits the depth of understanding of the proposed method.

问题

No questions.

局限性

A more extensive discussion regarding the drawbacks of the proposed approach is needed.

作者回复

Question1:\color{red}{Question1:} The paper would benefit from additional experiments to demonstrate the generalizability of the proposed approach. Additionally, the scalability of FasterDiT to larger datasets or more complex tasks is not sufficiently discussed, potentially limiting its applicability.

Response1:\color{blue}{Response1:} Thanks for your suggestion. As requested, we have conducted three types of additional experiments to further explore the generalization of FasterDiT. Please refer to Response to All Reviewers for details.

Question2:\color{red}{Question2:} A more comprehensive comparison with existing methods for accelerating training of Diffusion Transformers would provide a broader context for evaluating the effectiveness of FasterDiT.

Response2:\color{blue}{Response2:} Previous well-known acceleration methods, such as MaskDiT [1], MDT [2], and MDTv2 [3], have achieved significant performance improvements by modifying architectures and redesigning the training pipelines of DiT. In contrast, our work approaches the problem from a different perspective. Our aim is to improve DiT training with minimal modifications and without any changes to the architecture. This approach is complementary to previous methods. In the future, we plan to explore how combining FasterDiT with more structured improvements can achieve even more efficient generation performance.

Question3:\color{red}{Question3:} The absence of theoretical results and proofs in the paper limits the depth of understanding of the proposed method.

Response3:\color{blue}{Response3:} Thank you for your suggestions. Our work is to analyze the relationship between SNR PDFs and generation performance through a large number of experiments. We try to derive empirical results and insights to accelerate diffusion training. We will further explore this issue from a theoretical perspective in the future.

Limitations:\color{red}{Limitations:} A more extensive discussion regarding the drawbacks of the proposed approach is needed.

Response:\color{blue}{Response:} The main limitation of this paper lies in the lack of exploration of large-scale experiments, such as 2K high-resolution images, text-to-image generation, and video generation. Among these, we particularly focus on the text-to-image generation aspect. Specifically, in the class-conditional generation described in this paper, the DiT block only needs to process image tokens. However, in some text-to-image architectures, such as SD3 [4], self-attention needs to handle sequences combining text and visual tokens. The features from different sources (such as T5 and VAE) may lead to potential instability. We plan to further investigate this issue in the future. Besides, we will update our section on 'Limitations' in the next version of our paper. Thanks for your suggestion.

Refs:

[1] Zheng H, Nie W, Vahdat A, et al. Fast training of diffusion models with masked transformers[J]. arXiv preprint arXiv:2306.09305, 2023.

[2] Gao S, Zhou P, Cheng M M, et al. Masked diffusion transformer is a strong image synthesizer[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision. 2023: 23164-23173.

[3] Gao S, Zhou P, Cheng M M, et al. MDTv2: Masked Diffusion Transformer is a Strong Image Synthesizer[J]. arXiv preprint arXiv:2303.14389, 2023.

[4] Esser P, Kulal S, Blattmann A, et al. Scaling rectified flow transformers for high-resolution image synthesis[C]//Forty-first International Conference on Machine Learning. 2024.

审稿意见
6

This paper propose FasterDiT, a diffusion model training method that considers the data distribution in the definition of signal-to-noise ratio. It formulates the SNR in a new framework, estimates the PDF of the SNR, and then employs it to improve the training efficiency of DiT. Experimental results show that FasterDiT achieves relative FID but with 1/7 training time.

优点

  1. The idea of re-formulating the SNR in the diffusion model is a very interesting and novel strategy for me. I believe this paper has a good technical contribution.
  2. The experimental results of DiT show both good acceleration and generation performance.

缺点

  1. This paper has a bad quality in writing. There are some typos and the explanation to the SNR is not clear to me. I'm confused for many questions: (1) What does the std means in Line 109. Does it mean the std of the value of pixels in the images? (2) Why can we assume std^2 approximate K(I)/std^2 as a constant C(I)? (3) Authors mention "robustness" for many times in the paper for "training robustness", "data robustness". What does it mean in detail? (4) In Figure 6, the caption inside the figure is "multi-step balance", while the caption after the figure is multiple-step balance. Please use the same description. What does multi-step balance mean? Does that mean the new training strategy with SNR? The "single-step supervision" is also not a good choice here. It should be something like "directionality of velocity" or "single-step supervision (ours)" since there has already been single-step supervision in the traditional training method.
  2. Does the proposed method generalize well to the diffusion models besides DiT, such as latent diffusion models? Please discuss on this. If so, experimental results should be provided for this since only experiments on DiT & ImageNet is not very convincing.

问题

  1. Does x_* in the paper indicate the x_0 (real images)? If so, I advise to replace it with x_0 to align with previous works.
  2. Section 4: Improving DiT Training. The "." should be removed.
  3. The space between line 140 and 141 are excessively reduced. In summary, I think this paper may have good technical contribution. But it's really difficult for me to understand the details of this paper.

局限性

n/a

作者回复

Question1:\color{red}{Question1:} What does the stdstd means in Line 109. Does it mean the stdstd of the value of pixels in the images?

Response1:\color{blue}{Response1:} The stdstd here refers to the standard deviation of the input. For traditional diffusion, it refers to the standard deviation of pixel values. For latent diffusion, it refers to the standard deviation of the latents after VAE encoding. In this paper, our experiments are primarily based on latent diffusion.

Question2:\color{red}{Question2:} Why can we assume std2std^2 approximate K(I)/std2K(I)/std^2 as a constant C(I)C(I)?

Response2:\color{blue}{Response2:} The approximation maintains our definition as an extension of the previous one with minimal changes. Specifically, in previous definitions, the standard deviation (stdstd) is the most crucial factor in quantifying the Signal-to-Noise Ratio (SNRSNR). For ease of analysis, previous works [1, 4] treat both the input and noise as normal distributions, resulting in:

SNRpre=(αt×stdinput)2(σt×stdnoise)2=αt2σt2,stdinput=stdnoise=1SNR_{pre} = \frac{(\alpha_t \times std_{input})^2}{(\sigma_t \times std_{noise})^2}=\frac{\alpha_t^2}{\sigma_t^2}, std_{input}=std_{noise}=1

In FasterDiT, our definition extends this from the following two aspects: Firstly, we take the standard deviation of input back into consideration instead of assuming it to be one. Secondly, some work [5] has shown that SNRSNR relates to more properties of inputs besides stdinputstd_{input} such as resolution. Here, we denote these additional properties as K(I)/stdinput2K(I)/std_{input}^2. Thus, we obtain:

SNR=(αt×stdinput)2σt2×K(I)stdinput2SNR = \frac{(\alpha_t \times std_{input})^2}{\sigma_t^2}\times\frac{K(I)}{std_{input}^2}

In our work, we pay more attention to the former aspect and relax the definition. We assume the stdinput2std_{input}^2 is a decoupled parameter in K(I)K(I) for ease of analysis similar to the previous definition. It helps us simplify the modeling of the SNR PDF with minimal impact on the qualitative results. We plan to conduct a more in-depth exploration of this in our future work.

Question3:\color{red}{Question3:} Authors mention "robustness" for many times in the paper for "training robustness", "data robustness". What does it mean in detail?

Response3:\color{blue}{Response3:} Here, "robustness" refers to the stability of training performance despite variations in the standard deviation of the input data. We will standardize this terminology in future versions.

Question4:\color{red}{Question4:} In Figure 6, the caption inside the figure is "multi-step balance", while the caption after the figure is multiple-step balance. Please use the same description. What does multi-step balance mean? Does that mean the new training strategy with SNR? The "single-step supervision" is also not a good choice here. It should be something like "directionality of velocity" or "single-step supervision (ours)" since there has already been single-step supervision in the traditional training method.

Response4:\color{blue}{Response4:} The "multi-step balance" refers to modifying the SNR PDF by adjusting the proportion of different timesteps during training. We will standardize the terminology and improve the naming of the direction loss to enhance clarity. Thank you for your suggestions.

Question5:\color{red}{Question5:} Does the proposed method generalize well to the diffusion models besides DiT, such as latent diffusion models? Please discuss on this. If so, experimental results should be provided for this since only experiments on DiT & ImageNet is not very convincing.

Response5:\color{blue}{Response5:} Thanks for the suggestion. Yes, it does. The method proposed in FasterDiT is decoupled from the model architecture. It could be interesting to explore its performance alongside DiT. As requested, we employed the U-Net used in LDM [2] and another well-known architecture, U-ViT [3], to investigate the influence of our method. The results demonstrate that FasterDiT continues to enhance the convergence rate of these models. Please refer to Response to All Reviewers for more details.

Question6:\color{red}{Question6:} (1) Does xx_* in the paper indicate the x0x_0 (real images)? If so, I advise to replace it with x0x_0 to align with previous works. (2) Section 4: Improving DiT Training. The "." should be removed. (3) The space between line 140 and 141 are excessively reduced. In summary, I think this paper may have good technical contribution. But it's really difficult for me to understand the details of this paper.

Response6:\color{blue}{Response6:} Thank you very much for your careful review of our submission and your detailed suggestions. In this paper, we explore DiT training with both diffusion and flow matching. We didn't choose x0x_0 to denote the input because of the different timestep definitions. Different from DDPM, in Rectified Flow, x(t=0)x (t=0) refers to the pure noise. We will consider clearer expressions in future versions of our paper. Besides, we will address all the other issues you raised and thoroughly review the entire manuscript to prevent similar issues from occurring in the future.

Refs:

[1] Salimans T, Ho J. Progressive Distillation for Fast Sampling of Diffusion Models[C]//International Conference on Learning Representations.

[2] Rombach R, Blattmann A, Lorenz D, et al. High-resolution image synthesis with latent diffusion models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 10684-10695.

[3] Bao F, Nie S, Xue K, et al. All are worth words: A vit backbone for diffusion models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023: 22669-22679.

[4] Hang T, Gu S, Li C, et al. Efficient diffusion training via min-snr weighting strategy[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision. 2023: 7441-7451.

[5] Hoogeboom E, Heek J, Salimans T. simple diffusion: End-to-end diffusion for high resolution images[C]//International Conference on Machine Learning. PMLR, 2023: 13213-13232.

评论

After reading the response from the authors, I decide to increase my rate from 5 to 6. I believe this paper has real novelty in the training of diffusion models and it still can be greatly improved by better writing. Please keep polishing it.

审稿意见
6

This work aims at solving the slow training convergence of Diffusion Transformer, from the perspective of Signal-to-Noise Ratio (SNR). Different from other works, the authors formulate the probability density function (PDF) of SNR during training, and then leverage such SNR PDF to analyze the association between training performance and robustness across some common pipelines of DiT. The finding of the trade-off between method robustness and performance makes authors to propose Fast-DiT which significantly accelerate DiT's training convergence.

优点

  1. The motivation is very clear. DiT suffers from slow convergence and this work successfully solved this issue without any changes in model architecture.

  2. The formulation of SNR PDF is novel. With this analysis tool, authors can check the robustness and performance among various DiT training pipeline.

  3. The emipirical experiments are sufficient and the final results compared with SOTA are convincing for me.

缺点

  1. Current Table 1 compares the CFG (classifier-free guidance) results. Please compare the Class-conditional results of Faster-DiT with DiT and SiT, under the same setting in Table 1 (or even more iterations of Faster DiT). Usually we need both CFG and Class-conditional results to check the training convergence. Currently Figure 1 right with small iterations is not convincing.

  2. I am still unclear how you change the std of data without changing dataset. Please specify this point.

问题

Listed in weakness.

局限性

The authors has discussed the limations of the best performance of FastDiT without sufficient GPUs.

作者回复

Question1:\color{red}{Question1:} Current Table 1 compares the CFG (classifier-free guidance) results. Please compare the Class-conditional results of Faster-DiT with DiT and SiT, under the same setting in Table 1 (or even more iterations of Faster DiT). Usually we need both CFG and Class-conditional results to check the training convergence. Currently Figure 1 right with small iterations is not convincing.

Response1:\color{blue}{Response1:} Thank you for your suggestion.

  1. As requested, we have added the following content: (1) We continue training FasterDiT up to 2000k iterations. (2) We additionally report the generation results without CFG (classifier-free guidance). The results indicate that FasterDiT still demonstrates a significant advantage. Please refer to the Response to All Reviewers for details.

  2. All experiments in Figure 1 are conducted on ImageNet and trained for 100k iterations. This setting stems from our observation of the training process that the trend of the training FID at 100k is similar to the trend of the training FID over a longer period of time. For example, the table below shows one set of our experiments on ImageNet. It shows that the trend of FID-50k at 100k is similar to the trend of FID-50k at 200k.

    TypeStdFID-50 (100k)FID-50k (200k)
    DDPM cosine schedule0.543.2222.08
    0.661.1525.70
    0.764.1825.96
    0.872.3928.94
    0.978.4530.45
    1.098.9142.41
    1.181.2638.10
    1.2111.8342.44

Question2:\color{red}{Question2:} I am still unclear how you change the std of data without changing dataset. Please specify this point.

Response2:\color{blue}{Response2:} The stdstd here refers to the standard deviation of the input. We can achieve the scaling of the standard deviation simply by multiplying a scale factor.

评论

Thanks for your experiments and answers! These answers have solved my problems. I will keep my ratings.

审稿意见
6

The paper focuses on accelerating the training process of Diffusion Transformers (DiT) without modifying their architecture. The authors identify two primary issues: inconsistent performance of certain training strategies across different datasets, and limited effectiveness of supervision at specific timesteps.

Key contributions include:

  1. Extended Definition of Signal-to-Noise Ratio (SNR)
  2. Extensive experiments and empirical findings for SNR PDF
  3. A new supervision method

优点

  • FasterDiT achieves competitive results (2.30 FID on ImageNet 256 resolution at 1000k iterations) while being seven times faster in training compared to traditional DiT (2.27 FID).
  • The paper presents a large number of experiments to empirically validate the proposed method. The paper presents various discussions and insights of SNR PDF.
  • By generalizing the definition of SNR and analyzing various training strategies through SNR PDFs, the paper shows promising outcomes of faster training.

缺点

  • The experiments were conducted on 256-resolution ImageNet only. It would be interesting to validate the proposed method on larger resolutions (such as 512, 1024, etc.). Acceleration is more critical to those scenarios.
  • The figures could be improved. The rendered text is not easy to read.

问题

see weakness

局限性

n/a

作者回复

Question1:\color{red}{Question1:} The experiments were conducted on 256-resolution ImageNet only. It would be interesting to validate the proposed method on larger resolutions (such as 512, 1024, etc.). Acceleration is more critical to those scenarios.

Response1:\color{blue}{Response1:} Thanks for your suggestion. As requested, we have conducted experiments with higher resolution (ImageNet 512x512). It shows that FasterDiT could still achieve faster convergence speed compared to the original DiT. Please refer to our Response to All Reviewers for more details. Due to the limited time available for rebuttal, we will explore higher resolution and larger training scales in the future.

Question2:\color{red}{Question2:} The figures could be improved. The rendered text is not easy to read.

Response2:\color{blue}{Response2:} Thank you very much for your kind suggestions. We will improve our figures in the future version of our manuscript.

评论

Thank you for the experiments. I hope the authors will add the high-res results to the final paper.

审稿意见
6

This paper presents observations on training strategies for Diffusion Transformers (DiT) using Signal-to-Noise Ratio (SNR) Probability Density Function (PDF) analysis. Through extensive experiments, the authors derive insights into training performance and robustness. Based on these findings, they propose a method to accelerate the training process of DiT without modifying the model architecture.

优点

  1. The problem addressed is practical, focusing on the significant computational demands of training Diffusion Transformer models.
  2. The authors propose a simple yet effective technique to improve the training efficiency of Diffusion Transformer models without modifying the architecture
  3. The paper is well-structured with clear motivation, methodology, and results.

缺点

  1. The paper lacks experiments with higher resolutions such as 512x512 or 1024x1024, which could provide insights into the method's scalability for more complex image generation tasks.
  2. The hyperparameter 'std' requires tuning to find the appropriate setting, which may introduce additional complexity in implementing the method across different datasets or tasks.

问题

Why does adding cosine similarity loss for velocity direction supervision accelerate the convergence rate of the model?

局限性

The scalability of this method is not thoroughly evaluated, including its effectiveness for higher resolutions, other datasets, or different types of generative tasks (e.g., text-to-image, video generation).

作者回复

Question1:\color{red}{Question1:} The paper lacks experiments with higher resolutions such as 512x512 or 1024x1024, which could provide insights into the method's scalability for more complex image generation tasks.

Response1:\color{blue}{Response1:} Thanks for the suggestion. As requested, we have conducted experiments on a higher resolution (ImageNet 512x512) using two different model scales (DiT-B and DiT-L). The results show that FasterDiT still achieves faster convergence at higher resolutions. For more details, please refer to our Response to All Reviewers.

Question2:\color{red}{Question2:} The hyperparameter 'std' requires tuning to find the appropriate setting, which may introduce additional complexity in implementing the method across different datasets or tasks.

Response2:\color{blue}{Response2:} Thank you for the question. We will explain it from two perspectives.

Firstly, we believe that an appropriate standard deviation (std) of the diffusion input can influence training performance in certain cases. This observation is similar to the concept of a 'magic number' or 'scale factor' in previous well-known studies [1, 2]. Using the CIFAR10 toy setting as an example, when we apply the most common image normalization, the standard deviation (std) of the data is approximately 0.4, and the training FID result is 51.40. However, when we adjust the std to 0.8, the result of the same training method improves significantly to 36.31. This demonstrates the importance of std adjustment.

ModelStdTraining SamplesFID-10k
DiT-S/20.450k x 12851.40
DiT-S/20.850k x 12836.31 (-15.09)

Secondly, the complexity brought by tuning the 'std' will decrease as research on diffusion deepens. For instance, [2] suggested that as resolution increases, training tends to favor a smaller std. In our work, the SNR PDF provides a new analytical tool and offers insights into this issue. We hope it can facilitate a deeper understanding of the diffusion process.

Question3:\color{red}{Question3:} Why does adding cosine similarity loss for velocity direction supervision accelerate the convergence rate of the model?

Response3:\color{blue}{Response3:} The direction loss of velocity serves as an auxiliary form of supervision that enables a more fully supervised training process. To provide further insights, we have conducted an ablation study. We train two DiT-B models: one with FasterDiT and the other with FasterDiT without direction loss. We then visualize the training loss at different timestep intervals, which can be seen in our submitted PDF. The results show that the model with directional loss demonstrates a significantly better training effect, particularly during the diffusion period of relative low-noise (0.4<t<0.90.4 < t < 0.9). The end-to-end performance comparison is shown in the table below.

MethodModelsTraining SamplesResolutionFID-10k (100k)FID-10k (200k)
FasterDiT (w/o direction loss)DiT-B/2100k x 128512x51283.2963.11
FasterDiTDiT-B/2100k x 128512x51277.8558.33

Limitations:\color{red}{Limitations}: The scalability of this method is not thoroughly evaluated, including its effectiveness for higher resolutions, other datasets, or different types of generative tasks (e.g., text-to-image, video generation).

Response:\color{blue}{Response:} Thanks for your suggestion. We have provided experiments with higher resolutions and different models of FasterDiT. We will explore more complex tasks you mentioned such as T2I generaiton and video generation in our following work.

Refs:

[1] Rombach R, Blattmann A, Lorenz D, et al. High-resolution image synthesis with latent diffusion models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 10684-10695.

[2] Chen T. On the importance of noise scheduling for diffusion models[J]. arXiv preprint arXiv:2301.10972, 2023.

评论

Thank you for your response. These additional experiments address most of my concerns. Based on this, I am willing to upgrade my score.

作者回复

Response to All Reviewers

We appreciate all of your valuable feedback. Your valuable suggestions have significantly improved our manuscript.

All reviewers have suggested that we conduct additional experiments to demonstrate the effectiveness and generalization ability of our method. As requested, we have added the following three types of additional experiments:

  1. Training FasterDiT with longer iterations and performance without CFG (R_dNCD).
  2. Training FasterDiT with higher resolution (R_fAgp, R_qPr3)
  3. Applying the FasterDiT strategy to different diffusion models (R_ttWn, R_FhtE).

Due to the limited time available for the rebuttal, we have conducted experiments to the maximum extent that our resources allow. The larger scale experiments mentioned by the reviewers, such as 1024 resolution, text-to-image generation ,and video generation, cannot be completed within this short period. We plan to explore these in future work. Thank you for all the suggestions.

1. Longer Training of FasterDiT:

We further improve the performance of FasterDiT with longer training iterations and report more results of FID-50k on ImageNet 256 resolution.

It shows that even without any structural modifications, FasterDiT could achieve comparable results to the original DiT at 1000k iterations. Furthermore, when the training period is extended to 2000k iterations, FasterDiT's performance further improves, achieving an FID-50k score of 2.03, demonstrating the effectiveness of our approach.

Performance without Classifier-Free Guidance (CFG)

MethodModelsTraining SamplesFID-50k
DiTDiT-XL/27000k x 2569.60
SiTDiT-XL/27000k x 2568.60
FasterDiTDiT-XL/21000k x 2568.72
1500k x 2568.22
2000k x 2567.91

Performance with Classifier-Free Guidance (CFG)

MethodModelsTraining SamplesFID-50k
DiT (cfg=1.5)DiT-XL/27000k x 2562.27
SiT (cfg=1.5)DiT-XL/27000k x 2562.06
FasterDiT (cfg=1.5)DiT-XL/21000k x 2562.30
1500k x 2562.12
2000k x 2562.03

2. Higher Resolution:

As requested, we explore FasterDiT on higher resolution generation experiments to demonstrate its generalization ability. Specifically, we apply our method to DiT-B/2 and DiT-L/2 for ImageNet 512x512 generation training. The results are shown in the table below.

FasterDiT achieves faster convergence across all experiments. When trained with 200k iterations, we improve the FID-10k performance of DiT-B/2 by 18.78 and DiT-L/2 by 17.93. This demonstrates the effectiveness of our method in high-resolution tasks.

MethodModelsTraining SamplesResolutionFID-10k
DiTDiT-B/2100k x 128512x51293.36
200k x 128512x51277.11
FasterDiTDiT-B/2100k x 128512x51277.85 (-11.51)
200k x 128512x51258.33 (-18.78)
DiTDiT-L/2100k x 64512x51287.24
200k x 64512x51267.29
FasterDiTDiT-L/2100k x 64512x51271.58 (-15.66)
200k x 64512x51249.36 (-17.93)

3. Different Architectures:

As requested, we explore our training method with other diffusion models besides DiT, such as Latent Diffusion Models (UNet architecture) [1] and U-ViT [2]. The results are shown in the table below.

It shows that with our training methods, the performance of U-ViT and UNet both improve. It demonstrates that our method might have opportunities to be generalized to more diffusion architectures.

ModelTraining SamplesFID-10k
U-ViT-L200k x 12850.22
U-ViT-L + Ours200k x 12837.12
UNet200k x 12866.73
UNet + Ours200k x 12860.07

Refs:

[1] Rombach R, Blattmann A, Lorenz D, et al. High-resolution image synthesis with latent diffusion models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 10684-10695.

[2] Bao F, Nie S, Xue K, et al. All are worth words: A vit backbone for diffusion models[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023: 22669-22679.

最终决定

The paper introduces an innovative approach to accelerate the training of Diffusion Transformers (DiT). The authors present a thorough empirical analysis, validating their method across various experiments and demonstrating competitive performance on ImageNet, while being significantly faster than the traditional DiT. The reviewers were unanimous in the decision to accept the paper and the AC agrees with the decision. The AC recommends that the authors put extra effort in polishing the paper, as well as adding the additional experiments provided in rebuttal.