PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
4
4
4
2.8
置信度
创新性2.5
质量2.5
清晰度3.0
重要性2.3
NeurIPS 2025

VETA-DiT: Variance-Equalized and Temporally Adaptive Quantization for Efficient 4-bit Diffusion Transformers

OpenReviewPDF
提交: 2025-05-08更新: 2025-10-29

摘要

Diffusion Transformers (DiTs) have recently demonstrated remarkable performance in visual generation tasks, surpassing traditional U-Net-based diffusion models by significantly improving image and video generation quality and scalability. However, the large model size and iterative denoising process introduce substantial computational and memory overhead, limiting their deployment in real-world applications. Post-training quantization (PTQ) is a promising solution that compresses models and accelerates inference by converting weights and activations to low-bit representations. Despite its potential, PTQ faces significant challenges when applied to DiTs, often resulting in severe degradation of generative quality. To address these issues, we propose VETA-DiT (**V**ariance-**E**qualized and **T**emporal **A**daptation for **Di**ffusion **T**ransformers), a dedicated quantization framework for DiTs. Our method first analyzes the sources of quantization error from the perspective of inter-channel variance and introduces a Karhunen–Loève Transform enhanced alignment to equalize variance across channels, facilitating effective quantization under low bit-widths. Furthermore, to handle the temporal variation of activation distributions inherent in the iterative denoising steps of DiTs, we design an incoherence-aware adaptive method that identifies and properly calibrates timesteps with high quantization difficulty. We validate VETA-DiT on extensive image and video generation tasks, preserving acceptable visual quality under the more aggressive W4A4 configuration. Specifically, VETA-DiT reduces FID by 33.65 on the DiT-XL/2 model and by 45.76 on the PixArt-$\Sigma$ model compared to the baseline under W4A4, demonstrating its strong quantization capability and generative performance. Code is available at: https://github.com/xululi0223/VETA-DiT.
关键词
Diffusion TransformerPost-training QuantizationImage and Video GenerationEfficient Inference

评审与讨论

审稿意见
5

This paper identifies two key challenges in quantizing DiT models: (1) large inter-channel variance caused by outliers, and (2) significant shifts in activation distributions across denoising timesteps. To address these issues, the authors propose a KLT-enhanced Hadamard transform for variance alignment and an incoherence-aware adaptive calibration method to manage temporal dynamics. Experimental results demonstrate strong performance compared to existing baselines.

优缺点分析

Strengths

  • S1: The paper is clearly written and easy to follow. The presentation is organized and the explanations are generally accessible, which helps the reader understand the main ideas without much difficulty.
  • S2: The proposed method is simple and straightforward, making it easy to implement. This practical aspect increases its usability and potential for adoption in real-world applications.
  • S3: The experimental results are strong and well-supported. The model is evaluated across multiple datasets, tasks, and metrics, showing its effectiveness and generality. Notably, the clearer performance gap observed when further quantizing from W4A8 to W4A4 provides additional evidence of the proposed method’s effectiveness.

Weaknesses:

  • W1: For a quantization method like this, I would expect an analysis of hardware resource usage. Unfortunately, the paper does not include such an evaluation.
  • W2: The paper could be further strengthened by evaluating the proposed method at even lower bit settings, such as W2A8 or W4A2.

问题

Q1: Can you provide more details on the practical implementation aspects of your method, especially regarding hardware compatibility or deployment on resource-constrained devices?

Q2: Given the clear performance differences observed when quantizing from W4A8 to W4A4, do you anticipate similar trends at lower bit-widths? What challenges do you foresee when pushing the quantization limits further?

局限性

yes

最终评判理由

The authors have addressed my main concerns. I think it is reasonable to keep the current score as accept.

格式问题

There are no major formatting issues.

作者回复

Thank you sincerely for your thoughtful and positive feedback on our work. Below, we have provided a detailed explanation for your remaining concern as follows. Please do not hesitate to let us know if you have any further questions.

W1 & Q1: Hardware resource usage evaluation.

We thank the reviewer for pointing this out. We have updated the manuscript to include an evaluation of hardware resource usage. In DiTs, linear layers dominate both computation and memory consumption [1, 2]. To address this, we implement our INT4 matrix multiplications using CUTLASS on Tensor Cores (CUDA 12.4) and evaluate our custom kernels on NVIDIA A100 GPU.

We fuse most KLT matrices into model weights during offline calibration, thereby eliminating the need for runtime transformation. For layers that require runtime KLT to preserve mathematical equivalence [3], we use fast online Hadamard kernels, which are widely available and efficient.

Despite incorporating both online quantization and runtime KLT transforms, we still observe 1.95× \times speedup and 3.74×\times memory savings over FP16. Compared to the W4A8 implementation in ViDiT-Q, which adopts higher activation precision to preserve generation quality, our approach provides 1.23×\times speedup and 1.81×\times memory reduction, demonstrating the hardware efficiency of our method:

Bit-WidthMethodLatency (ms)SpeedupMemory (MB)Reduction
W16A16FP6.167 ± 0.045-38.53-
W4A8ViDiT-Q3.879 ± 0.0121.59×\times18.682.06×\times
W4A4Ours2.960 ± 0.0062.08×\times9.683.98×\times
W4A4 + Online TransformOurs3.163 ± 0.0071.95×\times10.313.74×\times

We believe this provides a comprehensive view of both computational and memory efficiency. We will add the results to the revised manuscript to provide a comprehensive view of runtime and memory efficiency.

[1] Solving Oscillation Problem in Post-Training Quantization Through a Theoretical Perspective. CVPR 2023.

[2] PTQ4DiT: Post-training Quantization for Diffusion Transformers. NeurIPS 2024.

[3] SliceGPT: Compress Large Language Models by Deleting Rows and Columns. ICLR 2024.

W2 & Q2: Evaluation under lower bit settings.

We conducted additional experiments on DiT-XL/2 under the W3A3 quantization setting, which is already highly challenging. As shown below, our method significantly outperforms Q-DiT, but performance degradation is still notable under such low bit-widths. Introducing mixed-precision, by keeping some sensitive layers in 8-bit, yields substantial quality improvements.

Bit-WidthMethodIS \uparrowFID \downarrowsFID \downarrowPrecision \uparrow
W16A16FP164.614.8617.650.80
W3A3Q-DiT1.65266.97402.470.01
W3A3Ours45.8355.4234.420.32
W3A3(Mixed Precision)Ours111.6213.7122.250.66

At even lower bit-widths, quantizing becomes extremely lossy due to limited representational range. In such cases, offline KLT-based variance equalization becomes insufficient, and online KLT computation may be required to maintain generative fidelity. However, this would introduce non-negligible runtime cost. We will explore this trade-off in future work.

评论

Thank you for the detailed rebuttal. I appreciate the authors’ clarifications, which adequately address the main concerns. I continue to find the paper acceptable and maintain my score.

评论

Dear Reviewer,​​

We sincerely appreciate your constructive comments and valuable suggestions.

We are pleased to address all your questions and deeply grateful for your recognition of our work. Thank you again for your contributions to improving the quality of our manuscript.

​​Best, Authors

审稿意见
4

This paper proposes VETA-DiT, a post-training quantization framework for Diffusion Transformers (DiTs) under W4A4 setting. The method addresses two major challenges, large inter-channel variance due to outliers and timestep-dependent activation shifts during iterative denoising. To mitigate these, the authors introduce a KLT enhanced Hadamard transform for better variance alignment and an incoherence-aware temporal adaptation strategy. VETA-DiT outperforms existing PTQ baselines across image and video generation tasks, achieving high visual quality under low-bit quantization settings without retraining.

优缺点分析

Strengths

  1. The paper is well-structured, with clear visualizations and an accessible algorithm description, aiding reproducibility and understanding.
  2. This paper integrates the Karhunen–Loève Transform (KLT) with the Hadamard transform in the context of DiT for the first time, and demonstrates outstanding generative quality in low-precision settings.
  3. The incoherence-aware temporal adaptation and synthetic data generation strategy are straightforward and well-validated through diverse empirical experiments.
  4. The proposed scheme is applied to large-scale networks for both image and video generation, demonstrating its generality and robustness.

Weaknesses

  1. The combination of KLT with the Hadamard transform has already been presented in MambaQuant (ICLR 2025), which predates this NeurIPS submission. Thus, the novelty of this paper is somewhat reduced, although its contribution to the DiT domain remains valid and meaningful.
  2. As a result, the main contribution of this work lies in the design of the incoherence-aware adaptation. While the idea is interesting, comparisons with alternative temporal calibration or sampling strategies are insufficient.
  3. Since the method combines KLT with the Hadamard transform, the computational overhead is likely higher than using Hadamard alone. The paper should provide practical validation of inference speedups on real hardware.
  4. While the paper analyzes the time-varying activation distributions for image generation, it gives limited attention to similar temporal characteristics in video synthesis tasks.

问题

  1. The combination of KLT with the Hadamard transform appears similar to the method proposed in MambaQuant (ICLR 2025). Can the authors clearly articulate how their usage is different in either formulation, application, or motivation specific to DiTs?

  2. While the paper addresses temporal variance in image generation (across denoising steps), it is unclear whether similar temporal adaptation is used or needed in video synthesis. Do video generation tasks show similar incoherence trends, and is the adaptation equally effective?

  3. How does your incoherence-aware adaptation with synthetic data generation compare to previous methods built upon KLT-Hadamard transforms, in terms of effectiveness and novelty?

  4. What is the actual runtime improvement or slowdown introduced by your method, including the overhead of the KLT-H computation and calibration?

  5. You mention disabling mixed-precision in ViDiT-Q for fair comparison. Could you elaborate on the implications of this choice, and whether similar mixed-precision designs could benefit VETA-DiT as well?

局限性

Yes

最终评判理由

During the rebuttal process, the authors provided additional supporting evidence. I think the novelty is somewhat limited, but the paper is above the decision boundary.

格式问题

No.

作者回复

We greatly appreciate the reviewer's constructive comments on our paper. We will respond to the reviewer's feedback with detailed explanations for each point.

W1 & Q1: Difference from MambaQuant in the use of KLT.

We also noticed that MambaQuant applies KLT to state space models, as described in line 147 of our paper. MambaQuant focuses on Mamba-based state space models and addresses the numerical instability caused by outlier amplification in PScan operations. In contrast, our work is dedicated to Transformer-based diffusion models (DiTs), which exhibit a naturally large variance across activation dimensions and significant temporal variation due to iterative denoising.

We introduce KLT to equalize the intrinsic variance distribution across channels and time steps in DiTs. However, applying a single, time-step-independent KLT matrix across all denoising steps, as MambaQuant, leads to a severe degradation in generation quality due to the significant temporal variation in DiT activations. On the other hand, using fully time-step-specific KLT matrices incurs excessive load/offload overhead during inference. To address this trade-off, we propose an incoherence-aware adaptation strategy that enables efficient and temporally-sensitive KLT usage in DiTs, effectively mitigating both representational and computational inefficiencies.

We will include a more explicit discussion of the differences between our approach and MambaQuant in the revised manuscript.

W2: Effectiveness compared to alternative temporal sampling strategies.

Prior methods, such as ViDiT-Q [1], adopt a shared balancing mask ss across all denoising steps, which fails to account for the dynamically changing data distributions inherent in DiTs, effectively corresponding to uniform timestep sampling without adaptation. Currently, there is a lack of viable strategies specifically designed to handle this temporal variation.

Therefore, we provide a comprehensive evaluation of our incoherence-aware adaptation in Table 3 and Appendix D.2 of our paper, assessing both its impact on final generation quality and its effectiveness in reducing quantization error during inference. The results demonstrate that our method significantly outperforms uniform and random sampling strategies.

[1] ViDiT-Q: Efficient and Accurate Quantization of Diffusion Transformers for Image and Video Generation. ICLR 2025.

W3 & Q4: Runtime and calibration overhead.

The calibration process is conducted once offline and does not affect inference latency. On NVIDIA A100 GPUs, calibrating DiT-XL/2 takes approximately 35 minutes, and calibrating Open-Sora v1.2 (1.1B parameters) takes around 2.5 hours.

During inference, we use CUTLASS INT4 kernels optimized for Tensor Cores (CUDA 12.4) and fuse most KLT matrices into the model weights. Only a few blocks (e.g., out-proj, down-proj) require lightweight online transforms, implemented via fast Hadamard kernels commonly available in existing libraries. We still achieve 1.95× \times speedup and 3.74×\times memory savings over FP16. Compared to the W4A8 implementation in ViDiT-Q, which adopts higher activation precision to preserve generation quality, our approach provides 1.23×\times speedup and 1.81×\times memory reduction, demonstrating the hardware efficiency of our method:

Bit-WidthMethodLatency (ms)SpeedupMemory (MB)Reduction
W16A16FP6.167 ± 0.045-38.53-
W4A8ViDiT-Q3.879 ± 0.0121.59× \times18.682.06×\times
W4A4Ours2.960 ± 0.0062.08×\times9.683.98×\times
W4A4 + Online TransformOurs3.163 ± 0.0071.95×\times10.313.74×\times

These results confirm that our method maintains efficiency while incorporating KLT-H adaptations. We will add the results to the revised manuscript to provide a comprehensive view of runtime and memory efficiency.

W4 & Q2: Temporal incoherence in video generation.

Temporal incoherence also arises in video synthesis. We analyzed activation statistics at different denoising steps in Open-Sora and observed similar patterns of variation as in image generation. The table below presents varying incoherence values across timesteps, confirming the presence of temporal drift in diffusion-based video generation:

TimestepMeanMaxIncoherence
t1-0.004143.20110.821
t5-0.001403.26211.109
t100.000423.28313.344
t150.000733.40811.008
t200.001083.77312.742
t250.001174.19114.078
t300.002675.39115.906

As shown in Table 2 of our paper, our incoherence-aware adaptation continues to deliver substantial improvements in video generation quality, confirming its effectiveness in modeling temporal variations in both image and video diffusion models.

Q3: Effectiveness and novelty of incoherence-aware adaptation.

MambaQuant applies a shared KLT transform across all time steps. However, these assumptions break down in DiTs, where iterative denoising introduces highly dynamic and non-stationary activation patterns.

Our incoherence-aware adaptation departs from this assumption by designing a temporally-sensitive calibration strategy specifically tailored for DiTs. Rather than relying on handcrafted or static transforms, we leverage synthetic data generation to adaptively construct calibration sets that capture evolving activation statistics across time steps—a capability absent in prior KLT-Hadamard quantization methods.

Moreover, our approach precomputes transform matrices offline and fuses them into model weights wherever possible, minimizing runtime overhead without sacrificing temporal sensitivity.

This combination of temporal modeling, data-driven calibration, and runtime efficiency distinguishes our method both in novelty and practical utility.

Q5: Mixed-precision considerations in comparison with ViDiT-Q.

In the official implementation of ViDiT-Q, certain Linear layers are retained in FP16, which substantially reduces quantization-induced degradation but also limits efficiency gains, especially since linear layers are bottlenecks in DiTs [1, 2]. For a fair comparison, we disabled mixed-precision in ViDiT-Q.

To test whether our method can benefit similarly, we evaluated W3A3 quantization on DiT-XL/2 with and without mixed precision. Although W3A3 is a highly aggressive setting, our method benefits significantly from a mixed-precision strategy:

Bit-WidthIS \uparrowFID \downarrowsFID \downarrowPrecision \uparrow
W16A16164.614.8617.650.80
W3A345.8355.4234.420.32
W3A3(Mixed Precision)118.6213.7122.250.66

This indicates that VETA-DiT is also compatible with hybrid-precision schemes. However, as in ViDiT-Q, these approaches reduce theoretical efficiency gains and often require specialized CUDA kernels or hardware to realize end-to-end improvements.

[1] Solving Oscillation Problem in Post-Training Quantization Through a Theoretical Perspective. CVPR 2023.

[2] PTQ4DiT: Post-training Quantization for Diffusion Transformers. NeurIPS 2024.

评论

Thank you for your thorough analysis. Your responses have addressed most of my concerns, and I will raise the score accordingly.

评论

Dear Reviewer,​​

We sincerely thank the reviewer for acknowledging our efforts in addressing the concerns.

We truly appreciate the decision to reconsider the score based on our responses. The constructive feedback has been instrumental in improving the quality of our manuscript.

Please feel free to let us know if you have any further questions or concerns.

​​Best, Authors

审稿意见
4

This paper introduces VETA-DiT, a post-training quantization framework tailored for Diffusion Transformers (DiTs) to enable efficient low-bit inference (W4A4 and W4A8) while maintaining high generative quality. 1) To address inter-channel variance caused by outlier activations, the authors propose a Karhunen–Loève Transform (KLT) enhanced Hadamard transformation that equalizes channel variance and improves quantization robustness. 2) To handle the temporal variation in activation distributions during the iterative denoising process, they design an incoherence-aware adaptive calibration method that constructs a synthetic calibration set by prioritizing timesteps with high quantization difficulty. Experiments on image and video generation benchmarks demonstrate that VETA-DiT outperforms prior PTQ methods and remains effective under aggressive 4-bit settings without requiring model retraining.

优缺点分析

Strengths:

  1. Proposes a novel KLT-enhanced Hadamard transform that effectively equalizes inter-channel variance and improves quantization performance under low-bit settings.

  2. The writing is clear and easy to follow.

  3. Demonstrates strong empirical performance across multiple datasets (ImageNet, COCO, VBench) and models (DiT-XL/2, PixArt-Σ, STDiT3), especially under challenging W4A4 configurations.

Weaknesses:

  1. Efficiency metrics are missing. The authors only report accuracy results, despite claiming improved runtime and efficiency as core motivations. Without latency and memory comparisons—e.g., against FP16 on TorchScript or TensorRT—the work lacks self-contained evidence of practical gains.

  2. Diffusion timesteps can be distilled to a small number. It is unclear whether the proposed method remains effective in such scenarios, such as with DMD [1,2]. How much improvement does it bring in terms of both accuracy and efficiency when applied on top of already distilled, smaller models?

  3. Although multiple models are evaluated, the core approach is mainly benchmarked on relatively small-scale models except OpenSora (should also clarify how many parameters in models). The scalability of the method to larger models remains unclear. E.g., how does it perform on models such as Wan2.1-T2V-1.3B or Wan2.1-VACE-14B?

  4. The method requires computing eigenvectors of covariance matrices, which may introduce non-negligible overhead during calibration for large models. It would be helpful to quantify this preprocessing cost.

[1] Improved Distribution Matching Distillation for Fast Image Synthesis [2] One-step Diffusion with Distribution Matching Distillation

问题

See weaknesses. The main concerns are 1) missing efficiency comparison, 2) effectiveness on distilled smaller models or larger-scale foundation models, and 3) overhead analysis.

局限性

N/A

最终评判理由

The authors have addressed most of my concerns. I'll increase the score.

格式问题

N/A

作者回复

We sincerely thank the reviewer for providing valuable feedback. We detail our response below point by point. Please kindly let us know whether you have any further concerns.

W1 & Q1: Efficiency metrics.

We have extended the manuscript to include latency and memory comparisons. Our implementation of VETA-DiT is built upon CUDA 12.4 with CUTLASS to support INT4 matrix multiplication on Tensor Cores. We evaluated the performance of the linear layers on an NVIDIA A100 GPU, which are the main bottlenecks in DiT [1, 2].

Most KLT matrices are fused into model weights to eliminate runtime overhead. For layers such as out-proj and down-proj, we apply online transformations using fast Hadamard kernels. Even with online quantization and KLT, our method still achieves 1.95× \times speedup and 3.74×\times memory savings compared to FP16. Compared to the W4A8 implementation in ViDiT-Q, which adopts higher activation precision to preserve generation quality, our approach provides 1.23×\times speedup and 1.81×\times memory reduction, highlighting its superior efficiency:

Bit-WidthMethodLatency (ms)SpeedupMemory (MB)Reduction
W16A16FP6.167 ± 0.045-38.53-
W4A8ViDiT-Q3.879 ± 0.0121.59×\times18.682.06×\times
W4A4Ours2.960 ± 0.0062.08×\times9.683.98×\times
W4A4 + Online TransformOurs3.163 ± 0.0071.95×\times10.313.74×\times

We will add the results to the revised manuscript to provide a comprehensive view of runtime and memory efficiency.

[1] Solving Oscillation Problem in Post-Training Quantization Through a Theoretical Perspective. CVPR 2023.

[2] PTQ4DiT: Post-training Quantization for Diffusion Transformers. NeurIPS 2024.

W2 & Q2: Effectiveness with distilled timesteps.

While methods such as DMD reduce the number of denoising steps, temporal variation in activation statistics still exists, which is central to our approach. We analyze activations from randomly selected Transformer blocks and observe notable differences in mean, max, and incoherence across timesteps:

TimestepMeanMaxIncoherence
t1-0.0002354.2005.680
t2-0.0004194.3755.871
t3-0.0001794.9564.316
t4-0.0005735.3637.242

Our incoherence-aware adaptation is designed to capture and address such temporal variation, regardless of the total number of steps. We quantized the Transformer blocks in the DMD2-SDXL-4step model and evaluated on COCO. Our method achieves better results than Q-DiT under lower bit-widths:

Bit-WidthMethodIS \uparrowFID \downarrowsFID \downarrowCLIP \uparrowIR \uparrow
W16A16FP42.3954.64263.930.270.89
W4A8Q-DiT41.6756.80274.340.270.90
W4A8Ours41.4054.89264.860.260.88
W4A4Q-DiT32.9374.19293.940.230.43
W4A4Ours40.1458.39277.150.250.87

Our approach can be effectively combined with timestep distillation techniques like DMD without additional runtime cost, thus improving efficiency.

W3 & Q2: Scalability to larger models.

The Open-Sora v1.2 model contains 1.1B parameters, which is comparable in scale to Wan2.1-T2V-1.3B. For even larger models like Wan2.1-VACE-14B, which adopt deeper Transformer stacks, the architecture remains consistent across layers.

It has been observed that larger models are typically more robust to PTQ degradation [3]. We conducted a block-wise MSE analysis on Wan2.1-14B, and our method shows significantly lower quantization error compared to RTN and Q-DiT:

Bit-WidthMSE(RTN)MSE(Q-DiT)MSE(Ours)
W4A80.025460.007660.00374
W4A40.074240.015300.00714

[3] Scaling Laws for Precision. ICLR 2025.

W4 & Q3: Calibration cost of KLT.

The KLT calibration is a one-time offline process and does not incurs inference-time overhead. We report the calibration time on NVIDIA A100 GPU:

  • DiT-XL/2: ~35 minutes
  • Open-Sora v1.2 (1.1B): ~2.5 hours
评论

The reviewer thanks the authors for the provided results, which address some of my concerns, such as applicability to distilled models, larger-scale models (though only partial MSE is reported), and KLT overhead. However, for the latency and memory comparisons, it is unclear why only a single layer's performance is reported. Which layer was evaluated? Since the model contains diverse layer types and dimensions with multiple dataflows, a single-layer benchmark may not reflect the overall model performance. Additionally, the evaluation does not include comparisons against optimized FP16 inference frameworks like TensorRT or TorchScript.

评论

We appreciate the reviewer's detailed feedback, which is valuable for improving our work.

Regarding the concern about evaluating only a single layer for latency and memory, we clarify that our original analysis focused on the q/k/v projections and output projection within the attention block. The MLP linear layers in the FFN also exhibit similar trends. These components account for the dominant computational cost in the DiT architecture.

To provide a more comprehensive evaluation as suggested, we have now conducted benchmarking that reflects the performance characteristics of the DiT/XL-2 Transformer backbone. The latency and memory statistics are summarized below:

ConfigurationLatency (ms)SpeedupPeak Memory (MB)Reduction
FP16 (Original)4351482
FP16 (TensorRT)3721.17× \times13771.07× \times
INT4 (Ours)2321.88× \times6842.17× \times

These results demonstrate that, despite the additional overhead introduced by the KLT transformation matrices and online transformation operations, our method still achieves a 1.88× \times speedup and 2.17× \times memory reduction. This clearly shows that our approach not only surpasses the standard FP16 execution but also outperforms optimized inference pipelines like TensorRT.

Please do not hesitate to let us know if further clarification is needed.

评论

The reviewer thanks the authors for the provided end-to-end measurement results. I'll increase the score.

评论

We are delighted to see that the major concerns raised by the reviewer have been successfully addressed. We sincerely appreciate the reviewer’s decision to raise the score. We are truly grateful for the reviewer’s thoughtful comments, constructive suggestions, and the time and effort devoted to evaluating our work.

审稿意见
4

This paper introduces VETA-DIT, a 4-bit quantization framework for Diffusion Transformers (DiTs) that tackles activation variance imbalance and temporal dynamics. It uses a KLT-enhanced Hadamard transform to equalize variance and a temporally adaptive strategy to handle shifting statistics across diffusion timesteps. This dual approach allows VETA-DIT to maintain high-quality generation in aggressive W4A4 settings, making powerful DiT models practical for resource-constrained devices.

优缺点分析

Strengths:

  1. It introduces an innovative KLT-enhanced Hadamard transform to effectively mitigate the challenging issue of channel-wise variance imbalance in activations.
  2. The proposed temporally adaptive strategy intelligently handles the dynamic statistical shifts of activations across different diffusion timesteps.
  3. VETA-DIT achieves state-of-the-art performance, preserving high generation quality even in the highly aggressive W4A4 quantization setting.

Weaknesses:

  1. Its temporally adaptive strategy relies on the quality of the calibration dataset, which could limit its effectiveness when dealing with extreme or out-of-distribution data.

问题

  1. How much computational overhead does the KLT-enhanced Hadamard transform introduce during inference?
  2. How sensitive is the final model performance to the specific method used for calculating the timestep importance scores?
  3. Since the KLT is data-dependent, how well does the learned transform generalize to datasets with different statistical properties?
  4. What are the anticipated challenges in applying this framework to even lower bit-rates, such as 3-bit or 2-bit quantization?

I will consider raising the score if the authors can address my concerns.

局限性

Yes.

最终评判理由

I appreciate the author's effort in the rebuttal and believe it has addressed my main concerns. I am willing to increase my score.

格式问题

None.

作者回复

Thanks for your time in dealing with our work. We will answer the question and discuss point by point as follows. We hope that our response satisfactorily addresses the issues you raised. Please feel free to let us know if you have any additional concerns or questions.

W1 & Q3: Generalization concern with calibration data.

KLT is indeed data-dependent. We address this concern by employing an incoherence-aware adaptation strategy to construct the calibration set for guiding the KLT process. Details regarding the calibration set composition and settings are provided in Appendix. A of our paper. Specifically:

  • For the DiT-XL/2 model, we randomly select 16 classes from ImageNet for calibration, and perform generation across all 1000 classes.
  • For PixArt-Σ \Sigma and Open-Sora models, we adopt prompts provided in their official toolkits for calibration and conduct generation on COCO and VBench datasets, which are not used during calibration.

We empirically validate that this strategy ensures the learned KLT transform generalizes across different data distributions in Section 4.2 of our paper. To further verify its generalization, we evaluate the variance-equalization effect both within and outside the calibration set:

Dataset ScopeVar(Before Transform)Var(After Transform)
In-Distribution0.2328 ± 0.00060.0885 ± 0.0001
Out-of-Distribution0.3275 ± 0.00350.1009 ± 0.0003

The demonstrate show that even on out-of-distribution data, our method significantly reduces variance, indicating its strong generalization capability.

Q1: Inference overhead from KLT-Hadamard transform.

The KLT-enhanced Hadamard matrix is defined as TKLT-H=KH \mathbf{T}_{ \text{KLT-H}} = \mathbf{KH}, where K \mathbf{K} is a data-driven KLT matrix, precomputed offline using the calibration set, and H\mathbf{H} is a random Hadamard matrix.

During inference, this transformation can be fused into model weights, incurring no additional cost for most layers. For layers like out-proj and down-proj, online transformation is necessary to preserve computational invariance [1]. We evaluated the associated overhead on DiT-XL/2 via implementing a custom CUDA kernel on NVIDIA A100 GPU, targeting matrix multiplication operations within these layers.

Bit-WidthMethodLatency (ms)SpeedupMemory (MB)Reduction
W16A16FP6.167 ± 0.045-38.53-
W4A8ViDiT-Q3.879 ± 0.0121.59× \times18.682.06×\times
W4A4Ours2.960 ± 0.0062.08×\times9.683.98×\times
W4A4 + Online TransformOurs3.163 ± 0.0071.95×\times10.313.74×\times

Due to the availability of efficient Hadamard transform implementations, the online transform introduces only 6.8% overhead, while still achieving 1.95×\times acceleration and 3.74×\times memory savings over FP16. Compared to the W4A8 implementation in ViDiT-Q, which adopts higher activation precision to preserve generation quality, our approach provides 1.23×\times speedup and 1.81×\times memory reduction, highlighting its superior efficiency. We will add the results to the revised manuscript to provide a comprehensive view of runtime and memory efficiency.

[1] SliceGPT: Compress Large Language Models by Deleting Rows and Columns. ICLR 2024.

Q2: Sensitivity to timestep importance estimation.

Section 4.3 presents our evaluation of the sensitivity to different timestep-wise importance scoring strategies. Table 3 shows that assigning uniform scores across timesteps significantly degrades performance (FID increases from 4.86 to 48.32), leading to severe degradation in image quality. In contrast, applying our incoherence-aware adaptation reduces FID to 7.90, substantially mitigating performance loss.

Additionally, Appendix D.2 demonstrates that our strategy consistently outperforms random timestep sampling in reducing quantization error.

Q4: Challenges in extending to lower bit-widths.

We conducted generation experiments on DiT-XL/2 under the challenging W3A3 setting. We summarize the results in the table below:

Bit-WidthMethodIS \uparrowFID \downarrowsFID\downarrowPrecision\uparrow
W16A16FP164.614.8617.650.80
W3A3Q-DiT1.65266.97402.470.01
W3A3Ours45.8355.4234.420.32
W3A3(Mixed Precision)Ours118.6213.7122.250.66

Although our method significantly outperforms Q-DiT, performance degradation is still observed under such extreme quantization. To mitigate this, we maintain 8-bit precision in sensitive layers, achieving better results in mixed-precision settings.

Looking forward, as bit-widths decrease further (e.g., to 2-bit), offline KLT becomes insufficient due to severely limited data representation. Online KLT may be required to dynamically equalize variance during inference, but this would incur non-trivial computational cost, which we plan to explore in future work.

评论

Dear Reviewer,

I hope this message finds you well. As the discussion period is approaching its end, we wanted to check in to ensure that we have adequately addressed your main concerns. If there are any remaining questions or additional feedback you would like us to consider, please don’t hesitate to let us know. Your insights are highly valued, and we are keen to further improve our work based on your suggestions.

Thank you again for your time and effort in reviewing our paper.

Best, Authors

最终决定

This paper introduces VETA-DiT, a novel PTQ framework to enable efficient 4-bit inference for DiTs. The paper's primary strengths, as noted by all reviewers, are its clear motivation, technical novelty, and strong empirical results on both image and video generation tasks under aggressive W4A4 settings. Initial weaknesses raised by reviewers (CJnV, eHfZ, 5kuo) were significant, focusing on the critical omission of hardware efficiency metrics, questions about novelty relative to prior work (eHfZ), and scalability (iU2r, CJnV). However, the authors provided an exemplary and thorough rebuttal, adding comprehensive end-to-end latency and memory benchmarks that demonstrated substantial speedups over FP16 and prior art. They also successfully clarified the novelty of their temporally adaptive approach and provided evidence of scalability. This discussion was highly effective, convincing reviewers CJnV, iU2r, and eHfZ to raise their scores. I recommend acceptance.