PaperHub
5.5
/10
Poster3 位审稿人
最低1最高5标准差1.6
3
5
1
ICML 2025

SADA: Stability-guided Adaptive Diffusion Acceleration

OpenReviewPDF
提交: 2025-01-24更新: 2025-07-24
TL;DR

Training-free acceleration of generative models by adpatively exploit sparsity with principled approximation.

摘要

Diffusion models have achieved remarkable success in generative tasks but suffer from high computational costs due to their iterative sampling process and quadratic‐attention costs. Existing training-free acceleration strategies that reduce per-step computation cost, while effectively reducing sampling time, demonstrate low faithfulness compared to the original baseline. We hypothesize that this fidelity gap arises because (a) different prompts correspond to varying denoising trajectory, and (b) such methods do not consider the underlying ODE formulation and its numerical solution. In this paper, we propose **Stability-guided Adaptive Diffusion Acceleration (SADA)**, a novel paradigm that unifies step-wise and token-wise sparsity decisions via a single stability criterion to accelerate sampling of ODE-based generative models (Diffusion and Flow-matching). For (a), SADA adaptively allocates sparsity based on the sampling trajectory. For (b), SADA introduces principled approximation schemes that leverage the precise gradient information from the numerical ODE solver. Comprehensive evaluations on SD‐2, SDXL, and Flux using both EDM and DPM++ solvers reveal consistent $\ge 1.8\times$ speedups with minimal fidelity degradation (LPIPS $\leq 0.10$ and FID $\leq 4.5$) compared to unmodified baselines, significantly outperforming prior methods. Moreover, SADA adapts seamlessly to other pipelines and modalities: It accelerates ControlNet without any modifications and speeds up MusicLDM by \(1.8\times\) with \(\sim 0.01\) spectrogram LPIPS. Our code is available at: [https://github.com/Ting-Justin-Jiang/sada-icml](https://github.com/Ting-Justin-Jiang/sada-icml).
关键词
Diffusion ModelEfficient AlgorithmTraining-Free Acceleration

评审与讨论

审稿意见
3

This paper proposes SADA, a novel paradigm that unifies step-wise and tokenwise sparsity decisions using a shared criterion based on the denoised latent x0. By aligning with modern numerical solvers that rely heavily on x0, SADA offers more stable pruning decisions and preserves important visual details throughout the denoising trajectory. Extensive experiments on SD 2 and SDXL demonstrate that SADA significantly accelerates inference without compromising image quality.

给作者的问题

Please see the weakness

论据与证据

Yes

方法与评估标准

Yes

理论论述

Yes

实验设计与分析

Yes

补充材料

Yes

与现有文献的关系

This paper proposed a new cache-based accelerating method in the diffusion model area.

遗漏的重要参考文献

No

其他优缺点

Strengths:

  1. The paper is well-organized and includes comprehensive technical details.
  2. The authors provide sufficient theoretical analysis for the proposed method, including the proof of error bound.
  3. The images generated by the proposed method are consistent with those produced by the original diffusion models.

Weakness:

  1. This paper only presents experimental results on U-Net-based diffusion models. Since the state-of-the-art image diffusion models now primarily use DiT or MM-DiT architectures, it is essential to demonstrate the effectiveness of SADA on models like PixArt, Flux, or SD 3.
  2. The evaluation setting for DPM-Solver++ uses 50 sampling steps. However, the main advantage of DPM-Solver is its ability to achieve high-generation quality with fewer sampling steps. Therefore, it would be more reasonable to set the sampling step of DPM-Solver++ to 20.
  3. The proposed SADA performs worse than AdaptiveDiffusion when evaluated with DPM-Solver++. However, DPM-Solver++ holds greater practical value compared to the Euler solver.
  4. As shown in the results, the acceleration ratio of SADA is around 1.5×. Can its speedup ratio be extended to 2× or beyond at the cost of some performance?

其他意见或建议

N/A

作者回复

We sincerely thank the reviewer for the insightful and constructive comments.

Q1: Can its speedup be extended to 2×2 \times or beyond?

A1: Yes, it can. To apply an faster configuration, we leverage the inherent stability of the per‐step data reconstruction x0tx_0^t. When the x0tx_0^t trajectory demonstrates high stability (e.g., the second half of Fig.2), we could employ larger step sizes compensated by higher-order approximations. Building on this insight, we implement a uniform step-wise pruning strategy after the stability of the denoising process with Lagrange interpolation for correction.

For example, consider a 50‐step process. To achieve a step-wise pruning interval of 4 after stable (i.e., compute every 4th step fully and interpolate the skipped steps via Lagrange), we store x^0t\hat{x}_0^t every 4 steps before stabilization. Their indices define the fixed-size set II, which is updated dynamically to limit memory usage. For any skipped tt:

x^0tiIjIittjtitjx^0ti\hat{x}_0^{t}\gets\sum _{i \in I}\prod _{j\in I\setminus \\{ i \\}}\frac{t-t_j}{t_i-t_j}\hat{x}_0^{t_i}

Under this setting, we yield a 1.8×\geq 1.8 \times speedup regardless of models or solvers. The acceleration would be even more aggressive if further increasing the step size. To balance the degradation, we raise the Adams-Moulton approximation from second to third order, allowing x0tx_0^t to leverage information from the previous three steps (instead of two), thereby improving numerical accuracy and robustness. Our updated result in Table 1 demonstrates the effectiveness of the above improvements. Notably, we achieve a 2.02×2.02 \times speedup on the most powerful Flux.1 model with impressive 0.060.06 LPIPS and 1.951.95 FID.

Table 1: Quantitative results on MS-COCO 2017

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
SD2DPM++DeepCache17.700.2717.831.43
AdaptiveDiffusion24.300.1004.351.45
SADA26.340.0944.021.80
SD2EulerDeepCache18.900.2397.401.45
AdaptiveDiffusion21.900.1737.581.89
SADA26.250.1004.261.81
SDXLDPM++DeepCache21.300.2558.481.74
AdaptiveDiffusion26.100.1254.591.65
SADA29.360.0843.511.86
SDXLEulerDeepCache22.000.2237.362.16
AdaptiveDiffusion24.330.1686.112.01
SADA28.970.0933.761.85
FluxFlow-matchingTeaCache19.140.2164.892.00
SADA29.440.0601.952.02

Q2: Ablations on DPM-Solver++

We appreciate the reviewers for highlighting the practical importance of DPM++ and its ability to achieve high quality with fewer steps.

a. SADA performs worse than AdaptiveDiffusion?

To counterbalance the aggressive configuration, we increased the order of the Adams-Moulton approximation from second to third order. This enhancement incorporates additional information from the previous denoising trajectory, which in turn improves both accuracy and stability. As shown in Table 1, SADA now significantly outperforms AdaptiveDiffusion when used with DPM++.

b. Set the sampling step to 20?

Table 2 presents a comprehensive ablation study across various sampling steps. Our method achieves a 1.5× acceleration in the 25-step scenario with negletable difference. Furthermore, we observe that as the number of inference steps increases, the images generated by DPM++ initially change dramatically before converging when the base step is set to 25. An illustration, along with additional generation examples and comparisons, is available at the following link:

https://drive.google.com/file/d/168ovZu9fxcfY5PfE8F5AgkN4la6dvH9f/view?usp=sharing

Table 2: Ablation studyon sampling steps

ModelSchedulerStepsPSNRLPIPSFIDSpeedup Ratio
SD-2DPM++5026.340.0944.021.80
2528.150.0733.131.48
1529.840.0723.051.24
Euler5026.250.1004.261.81
2526.830.0883.871.48
1529.340.0763.701.25
SDXLDPM++5029.360.0843.511.86
2530.840.0732.801.52
1531.910.0732.541.29
Euler5028.970.0933.761.85
2529.420.0853.131.50
1531.280.0843.261.26

Q3. SADA for Flow-matching & DiT Architecture

Under the flow matching objective, the model directly predicts the transportation vector field dx/dtdx/dt between noise and data distributions. Since the denoising trajectory is ODE-based, our criterion effectively measures its stability. Table 3 on Flux (DiT) shows that our method significantly outperforms the most recent work suggested by reviewer DBkM.

Table 3: Quantitative results on MS-COCO 2017

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
FluxFlow-matchingTeaCache19.140.2164.892.00
SADA29.440.0601.952.02
审稿人评论

Thanks for the author's rebuttal. I raise my score to 3 weak accept.

作者评论

We sincerely appreciate your updated rating and positive recognition. We are genuinely pleased with your emphasis that our paper is well-organized and has solid theoretical proofs. We are particularly grateful for your support regarding our objective of training-free acceleration, which maintains consistency with the original diffusion model.

Your insightful feedback has encouraged and inspired us to further investigate SADA's potential. We have raised the Adam-Moulton method to third-order, leading to significantly improved generative performance and efficiency compared to previous baselines. Meanwhile, we have verified the superior performance of SADA on state-of-the-art flow-matching models with MM-DiT architecture

We are committed to open-sourcing the SADA plug-in package (diffuser & comfyUI), enabling training-free acceleration of existing diffusion models (and their variants) with just a single line of configuration code.

Thank you again for the time and efforts put in reviewing. Should you have additional questions or suggestions, please do not hesitate to reach out.

审稿意见
5

The paper proposes SADA (Stability-guided Adaptive Diffusion Acceleration), a method to accelerate diffusion models by jointly optimizing step-wise and token-wise sparsity using a unified criterion based on the denoised latent x0x_0. Key contributions include: (1) alignment of pruning decisions with x0x_0-based solvers for stability, (2) a second-order Adams-Moulton approximation for skipped steps, and (3) a token cache mechanism to mitigate information loss. Experiments on Stable Diffusion 2 and SDXL demonstrate up to 1.52× speedup while maintaining image quality (e.g., LPIPS of 0.118 on SDXL). The method outperforms baselines like DeepCache and AdaptiveDiffusion in metrics such as FID and LPIPS.

给作者的问题

  1. Few-Step Improvement: Why does LPIPS improve with fewer steps (Table 2)? Could you provide some insightful explanations?
  2. Baseline Comparison: It would be better to add ToMeSD or concurrent methods such as DiT-FastAttn for comparison.
  3. Some of the latest diffusion models are trained based on flow matching loss, and whether this method is also suitable for such models.
  4. The diffusion model of DiT architecture has also received a lot of attention recently, whether this method is also applicable to this architecture, and if so, increasing the experimental results of this architecture will help to improve the quality of the paper.

论据与证据

The claims are largely supported by experiments, but some aspects need clarification:

  • The assertion that x0x_0-based pruning is "more stable" than xtx_t-based methods is validated via metrics (Table 1), but direct ablation studies comparing x0x_0 vs. xtx_t criteria are missing.

方法与评估标准

  • Methods: Combining step/token pruning via x0x_0-alignment is novel and sensible. The Adams-Moulton approximation and token cache are well-motivated.
  • Evaluation: COCO-2017, SD2/SDXL, and standard metrics (LPIPS, FID) are appropriate. However, user studies or qualitative examples (beyond Fig. 5) would strengthen claims about preserved visual details.

理论论述

Overall, the theoretical proof is solid. There are two possible concerns:

  • Theorem 3.1 (global token average): Proof in Appendix A.1 applies Lindeberg-Feller CLT but assumes independent tokens, which diffusion latents may not satisfy?
  • Theorem 3.2 (error bound): The proof assumes Lipschitz continuity of ϵθ\epsilon_\theta, which is standard but not empirically verified. Or it would be better to have some literature support.

实验设计与分析

  • Table 1 shows strong results, but baselines like ToMeSD or concurrent methods (e.g., DiT-FastAttn) are omitted.
  • The ablation study (Table 2) reports improved quality with fewer steps. The authors should clarify if this stems from their method’s stability or experimental setup.

补充材料

Reviewed appendices:

  • Appendix A: Proofs for Theorems 3.1 and 3.2 are detailed but lack empirical validation or literature support of assumptions(e.g., Lipschitz continuity).
  • Appendix B: Analysis of token merging/pruning as low-pass filters is insightful but needs empirical validation or literature support.

与现有文献的关系

The work builds on diffusion acceleration via step skipping (DPM-Solver++, AdaptiveDiffusion) and token reduction (ToMeSD, DeepCache). It unifies these paradigms, addressing limitations in prior isolated approaches. The x0x_0-alignment aligns with modern ODE solvers (Karras et al., 2022), extending their utility to sparsity decisions.

遗漏的重要参考文献

N/A. However, the following works could be included in the related work section or used as baselines to enhance the quality of the paper, as they also focus on training-free acceleration of diffusion models.

  • Delta-DiT: A Training-Free Acceleration Method Tailored for Diffusion Transformers
  • Faster Diffusion: Rethinking the Role of UNet Encoder in Diffusion Models
  • Cache Me if You Can: Accelerating Diffusion Models through Block Caching

其他优缺点

  • Strengths: Solid theoretical proof, Novel unification of step/token pruning, strong empirical results, and practical speedup.
  • Weaknesses: No analysis of computational overhead from the cache mechanism.

其他意见或建议

N/A

作者回复

We sincerely appreciate the reviewer's thoughtful feedback and kind support for our work.

Based on the suggestions from Reviewer 4E6r and DBkM, we implement an aggressive version of SADA:

  1. Implementing uniform step-wise pruning when the x0tx_0^t trajectory is stable, using Lagrange interpolation.

  2. Mitigating degradation by upgrading the Adam-Moulton Approximation from second to third order.

Detailed motivation and formulation are provided in our response to Reviewer 4E6r, and updated results are shown in Table 1.

Q1: Comparison between x0x_0 and xtx_t-based criterion

A1: We compare our x0x_0 driven paradigm with AdaptiveDiffusion, which leverages the third-order difference of xtx_t as acceleration criterion. As shown in Table 1, our method consistently delivers superior generation quality—achieving higher PSNR, lower LPIPS and FID—while maintaining a stable speed-up ratio of 1.8×\geq 1.8 \times regardless of model and scheduler.

The x0x_0 representation is naturally aligned with the final output, capturing essential semantic structures and enabling a more robust criterion than xtx_t.

Table 1: Comparison between x0x_0 based and xtx_t based criterion

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
SD2DPM++AdaptiveDiffusion24.300.1004.351.45
SADA26.340.0944.021.80
SD2EulerAdaptiveDiffusion21.900.1737.581.89
SADA26.250.1004.261.81
SDXLDPM++AdaptiveDiffusion26.100.1254.591.65
SADA29.360.0843.511.86
SDXLEulerAdaptiveDiffusion24.330.1686.112.01
SADA28.970.0933.761.85

Q2 Theoretical claims

A2-1 (CLT): The independence assumption holds for xtx_t as ϵt\epsilon_t is sampled i.i.d. from Gaussian. For x^t=αˉtx^0t+1αˉtϵ^t\hat{x}_t=\sqrt{\bar{\alpha}_t}\hat{x}_0^t+\sqrt{1-\bar{\alpha}_t}\hat{\epsilon}_t, we write ϵ^t=ϵt+(ϵ^tϵt)\hat{\epsilon}_t=\epsilon_t+(\hat{\epsilon}_t-\epsilon_t). The first term is i.i.d. Gaussian with zero mean by Law of Large Number (LLN). For the second, the training objective Eϵϵ^t2E| \epsilon-\hat{\epsilon}_t|^2 implies ϵ^tE[ϵxt,t]\hat{\epsilon}_t\to E[\epsilon\mid x_t,t], so E[ϵ^tϵt]0E[\hat{\epsilon}_t-\epsilon_t]\to 0, and by LLN, the sample mean ϵ^tϵt0\overline{\hat{\epsilon}_t-\epsilon_t}\to 0.

A2-2 (Lipschiz): The Lipschiz continuity of ϵθ\epsilon_\theta is widely assumed by preliminary works such as Adaptive Diffusion and DPM-Solver.

Q3 Computational overhead

a. Memory: For step-wise pruning with third-order Adam-Moulton, after reformulate we only need to store 1 previous x0tx_0^t and 2 previous dx/dtdx/dt in the cache. For token-wise pruning, we store 1 previous representation xtl\mathbf{x}^l_t for only transformers with the highest resolution. For example, we observe only a neglectable increase in memory usage in the SD-XL model (from 14981 MB to 15127 MB).

b. Complexity: All computation in the SADA framework is addition and scaling, O(N)O(N). Note that SADA does not include any quadratic complexity computation (e.g. cosine similarity, matrix calculation) as in previous works.

Q4. Qualatative examples & Fewer step generation

We provide the following link for more generation samples and comparison with previous strategy. In Addition, the CLIP score for generation quality is provided.

https://drive.google.com/file/d/168ovZu9fxcfY5PfE8F5AgkN4la6dvH9f/view?usp=sharing

We appreciate for pointing out the better similarity when decreasing sampling steps. We believe the extent accumulation of error decrease when reducing sampling steps. This trend could be clearly found in our ablation table.

Q5. Comparison with other token-wise sparisty strategies

Table 2 shows that our method significantly outperforms ToMeSD. DiTFastAttention is limited to traditional Diffusion Transformer because its windowed attention cannot handle mixed-modality inputs (e.g., MM-DiT modules in SD-3 and Flux). In contrast, our approach easily adapts to these architectures, as demonstrated later.

Table 2: Quantitative results on MS-COCO 2017

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
SD2DPM++ToMeSD16.290.4113.701.10
SADA26.340.0944.021.80

Q6. SADA for Flow-matching & DiT Architecture

Under the flow matching objective, the model directly predicts the transportation vector field dx/dtdx/dt between noise and data distributions. Since the denoising trajectory is ODE-based, our criterion effectively measures its stability. Table 3 on Flux (DiT) shows that our method significantly outperforms the most recent work suggested by reviewer DBkM.

Table 3: Quantitative results on MS-COCO 2017

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
FluxFlow-matchingTeaCache19.140.2164.892.00
SADA29.440.0601.952.02
审稿人评论

Thank you for your response. My concerns have been mostly addressed. The adaptive mechanism in diffusion models has rarely been studied before and holds great significance; therefore, I consider this work a valuable contribution to the diffusion model community. The additional experimental results provided in the rebuttal further validate the effectiveness of the proposed method. As a result, I am inclined to raise my score.

作者评论

We sincerely thank you for your thoughtful review and support. We deeply appreciate your recognition of the novelty and great significance of adaptive mechanisms in diffusion models, the uniqueness of dynamic allocation of token-wise and step-wise sparsity, and the solid theoretical proof — which makes our proposed approach accelerate generative modeling by dynamically adjusting configurations for different prompts while best preserving faithfulness.

We are delighted that our additional experimental results and analysis have addressed your concerns. Your constructive feedback is very important for us to refine and improve our approach.

We look forward to releasing the SADA package to the diffusion community in the camera-ready phase!

审稿意见
1

The paper proposes SADA, a training-free acceleration method for diffusion models that unifies step-wise (temporal) and token-wise (spatial) sparsity using a stability criterion based on the denoised latent x0x_0. Specifically, the paper uses a unified x0x_0-guided sparsity criterion for step skipping and token pruning, leveraging x0x_0's structural alignment with modern ODE solvers. A second-order Adams-Moulton method to approximate skipped steps and a token cache to reconstruct pruned tokens. Experiments on Stable Diffusion 2 and XL show speedups of up to 1.5× while maintaining image quality.

update after rebuttal

Thank you to the authors for their response and additional experiments, which have provided me with a deeper understanding of SADA's effectiveness. However, the baselines compared in this paper are somewhat non-repetitive, and some important papers on cache-based DiT acceleration, such as Learning-to-Cache and ΔDit\Delta - Dit, were not included in the comparison. Additionally, my concerns about the novelty of this paper remain. The caching method proposed in the paper does not differ fundamentally from previous approaches. Although the work most similar to this paper, [1], was published during the review period, earlier works like [2, 3, 4] also bear significant similarity in methodology, especially TeaCache [4]. A comprehensive experimental comparison with these papers, along with a detailed explanation of the differences in approach, is necessary. I still recommend rejecting this paper.

[1] Token-aware and step-aware acceleration for stable diffusion

[2] Cached Adaptive Token Merging: Dynamic Token Reduction and Redundant Computation Elimination in Diffusion Mode

[3] Accelerating diffusion transformers with token-wise feature caching

[4] Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model

给作者的问题

see Weaknesses

论据与证据

  • Claim 1: x0x_0-based pruning improves stability and aligns with solvers.

    Evidence: Theoretical analysis (Theorem 3.1) links x0x_0 to step stability; experiments show lower LPIPS/FID than xtx_t-based methods (Table 1).

    Problems: Limited comparison to other x0x_0-aligned methods; no ablation on x0x_0 vs. xtx_t.

  • Claim 2: Unified sparsity outperforms isolated strategies.

    Evidence: SADA outperforms DeepCache/AdaptiveDiffusion in FID/LPIPS (Table 1).

    Problems: Missing comparisons to recent works[1,2,3,4]

[1] Token-aware and step-aware acceleration for stable diffusion

[2] Cached Adaptive Token Merging: Dynamic Token Reduction and Redundant Computation Elimination in Diffusion Mode

[3] Accelerating diffusion transformers with token-wise feature caching

[4] Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model

方法与评估标准

  • Methods: The x0x_0-aligned criterion and reconstruction mechanisms are well-motivated. Adams-Moulton provides a principled ODE-based approximation.
  • Evaluation: COCO-2017 benchmarks with standard metrics (LPIPS, FID) are appropriate. But no clipscore or pick-socre reported, The former evaluated the text alignment of images generated by models after acceleration, while the latter assessed aesthetic scores.

理论论述

  • Theorem 3.1 (Lindeberg condition): Correct under the assumption of token independence, but real-world spatial correlations in images may affect validity.
  • Theorem 3.2 (Error bound): Relies on Lipschitz continuity of ϵθ\epsilon_\theta, which is not empirically validated.

实验设计与分析

  • Strengths: Broad evaluation across schedulers (DPM++, Euler) and models (SD2, SDXL).
  • Weaknesses:
    • No analysis of computational overhead from the token cache or varying pruning ratios.
    • No clipscore or pick-socre reported, The former evaluated the text alignment of images generated by models after acceleration, while the latter assessed aesthetic scores.
    • Missing comparisons to recent works[1,2,3,4]
    • The acceleration effect obtained is not significant compared to previous work; it's around 1.5x, which is incremental in nature. The ideas presented in this article do not differ significantly from those based on cache methods previously [1,2,3,4], with contributions being incremental as well.

[1] Token-aware and step-aware acceleration for stable diffusion

[2] Cached Adaptive Token Merging: Dynamic Token Reduction and Redundant Computation Elimination in Diffusion Mode

[3] Accelerating diffusion transformers with token-wise feature caching

[4] Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model

补充材料

Reviewed Appendix A (proofs) and B (token merging/pruning analysis). Theorems are logically derived but lack empirical validation of assumptions (e.g., Lipschitz continuity).

与现有文献的关系

Aligns with ODE-based solvers (DPM-Solver++) and token reduction (ToMe). Missing discussion of previous similar work[1,2,3,4]

[1] Token-aware and step-aware acceleration for stable diffusion

[2] Cached Adaptive Token Merging: Dynamic Token Reduction and Redundant Computation Elimination in Diffusion Mode

[3] Accelerating diffusion transformers with token-wise feature caching

[4] Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model

遗漏的重要参考文献

The following references should be discussed because they are similar in method to this paper.

[1] Token-aware and step-aware acceleration for stable diffusion

[2] Cached Adaptive Token Merging: Dynamic Token Reduction and Redundant Computation Elimination in Diffusion Mode

[3] Accelerating diffusion transformers with token-wise feature caching

[4] Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model

其他优缺点

no

其他意见或建议

This paper is poorly written and difficult to read.

作者回复

We thank the reviewer for comprehensive comments.

Q1: Aggressive configuration of SADA

A1: To implement an aggressive version, we leverage the inherent robustness of the per‐step data reconstruction x0tx_0^t. When the x0tx_0^t trajectory demonstrates high stability (e.g., the second half of Fig.2), we could employ larger step sizes compensated by higher-order approximations. Building on this insight, we implement a uniform step-wise pruning strategy after the stability of the denoising process with Lagrange interpolation for correction.

For example, consider a 50‐step process. To achieve a step-wise pruning interval of 4 after stable (i.e., compute every 4th step fully and interpolate the skipped steps via Lagrange), we store x^0t\hat{x}_0^t every 4 steps before stabilization. Their indices define the fixed-size set II, updated dynamically to limit memory usage. For any skipped tt:

x^0tiIjIittjtitjx^0ti\hat{x}_0^{t}\gets\sum _{i \in I}\prod _{j\in I\setminus \\{ i \\}}\frac{t-t_j}{t_i-t_j}\hat{x}_0^{t_i}

Under this setting, we yield a 1.8×\geq 1.8\times speedup regardless of models or solvers. To balance the degradation, we raise the Adams-Moulton approximation from second to third order, allowing x0tx_0^t to leverage information from the previous three steps (instead of two), thereby improving numerical accuracy and robustness. Our updated result in Table 1 demonstrates the effectiveness of the above improvements.

Table 1: Quantitative results on MS-COCO 2017

ModelSchedulerMethodsPSNRLPIPSFIDSpeedup Ratio
SD2DPM++DeepCache17.700.2717.831.43
AdaptiveDiffusion24.300.1004.351.45
SADA26.340.0944.021.80
SD2EulerDeepCache18.900.2397.401.45
AdaptiveDiffusion21.900.1737.581.89
SADA26.250.1004.261.81
SDXLDPM++DeepCache21.300.2558.481.74
AdaptiveDiffusion26.100.1254.591.65
SADA29.360.0843.511.86
SDXLEulerDeepCache22.000.2237.362.16
AdaptiveDiffusion24.330.1686.112.01
SADA28.970.0933.761.85

Q2: Comparison between x0x_0 and xtx_t-based criterion

A2: To our best knowledge, we are the first work that considers x0x_0 as acceleration criterion and approximation objective. We compare our paradigm with AdaptiveDiffusion, which leverages the third-order difference of xtx_t, as demonstrated in Table 1.

Residing in an image representation space, x0x_0 demonstrates structural alignment with the final output while evolving in a more robust trajectory (as shown in Fig. 2). It captures semantics and thus yields a more consistent sparsity allocation decision.

Q3: Comparisons to recent works

A3: We thank the reviewers for listing the four recent works, and we will cite and discuss them in the camera-ready version. The four works explore different Caching mechanisms within diffusion architectures that accelerate sampling. However, we believe our work fundamentally differs from the four works listed. Note that [1] is published after submission, which is impossible be addressed at the moment.

a. Methodology: The four works mentioned above accelerate diffusion in a fixed configuration (e.g., fixed caching interval and pruning ratio), while SADA is adaptive to different prompts. In addition, to our best knowledge, SADA is the first work that unifies token- and step-wise sparsity by a single criterion from the perspective of the ODE-solver process, achieving a multi-granularity adaptive acceleration strategy. The novelty is strongly supported by Reviewer km7j.

b. Motivation: SADA formulates the acceleration of the ODE-based generative modeling (e.g., Diffusion, Flow-matching) as a stability measure of the denoising trajectory, while the four works focus only on the redundancy of the denoising architecture with relatively weak theoretical justification.

c. Experiment: We believe the objective of post-training acceleration is to preserve the similarity (faithfulness) between original generated and accelerated samples while maximizing speed. Therefore, we evaluate using LPIPS and FID computed between these samples—unlike previous work, which only compares the FID of accelerated samples against the dataset. As shown in Table 2 of our response to Reviewer km7j, our method significantly outperforms [4] on FLUX.1 in terms of faithfulness at the same speed-up ratio.

Q4: CLIP/Pick Score

A4: Our objective is to preserve the original generation quality through our sparsity framework—metrics like these do not reflect that goal. For completeness, we have provided the requested metrics, along with generation samples and comparisons, via the link below:

https://drive.google.com/file/d/168ovZu9fxcfY5PfE8F5AgkN4la6dvH9f/view?usp=sharing

Q5: Computational overhead analysis & validation of Theoretical claim

Please refer to Q2, Q3 in our response to Reviewer km7j.

最终决定

The paper introduces SADA (Stability-guided Adaptive Diffusion Acceleration), a method designed to accelerate diffusion models by jointly optimizing step-wise and token-wise sparsity through a unified criterion based on the denoised latent variable x0x_0. SADA is evaluated on Stable Diffusion 2 and SDXL, demonstrating up to a 1.5× speedup while preserving image quality.

The review scores are 1–3–5. The main concerns raised by Reviewer DBkM relate to the similarity of the proposed approach to existing work (see references [1]–[4] in the corresponding review). Although these references were unpublished or unavailable at the time of the ICML submission deadline, the authors should address and discuss them in the updated manuscript.

Reviewer 4E6r’s concerns appear to have been addressed. We encourage the authors to reflect these updates in the revised version. Additionally, the reviewers suggest further improvements, and we recommend that the authors incorporate relevant points from their rebuttal to strengthen the final version of the paper.