PaperHub
7.8
/10
Spotlight4 位审稿人
最低4最高5标准差0.4
5
4
5
5
3.3
置信度
创新性2.8
质量2.8
清晰度3.0
重要性2.5
NeurIPS 2025

QSVD: Efficient Low-rank Approximation for Unified Query-Key-Value Weight Compression in Low-Precision Vision-Language Models

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29

摘要

Vision-Language Models (VLMs) are integral to tasks such as image captioning and visual question answering, but their high computational cost, driven by large memory footprints and processing time, limits their scalability and real-time applicability. In this work, we propose leveraging Singular-Value Decomposition (SVD) over the joint query (Q), key (K), and value (V) weight matrices to reduce KV cache size and computational overhead. We in addition introduce an efficient rank allocation strategy that dynamically adjusts the SVD rank based on its impact on VLM accuracy, achieving a significant reduction in both memory usage and computational cost. Finally, we extend this approach by applying quantization to both VLM weights and activations, resulting in a highly efficient VLM. Our method outperforms previous approaches that rely solely on quantization or SVD by achieving more than $10$% accuracy improvement while consuming less hardware cost, making it better for real-time deployment on resource-constrained devices. We open source our code at https://github.com/SAI-Lab-NYU/QSVD.
关键词
Vision language modelsingular value decompositionquantization

评审与讨论

审稿意见
5

This paper proposes QSVD, a framework that leverages joint Singular Value Decomposition (SVD) of QKV weight matrices and quantization to reduce computational overhead and KV cache size in Vision-Language Models (VLMs). Key innovations include:

  1. Joint QKV-SVD: Concatenating Q/K/V weight matrices for unified low-rank decomposition, reducing parameters and computation

  2. Dynamic Rank Allocation: Singular value truncation based on impact on VLM accuracy (Eq. 10).

  3. Low-Rank Activation Quantization: Orthogonal transformations to suppress activation outliers (Eq. 11-12).

Experiments show QSVD outperforms baselines (ASVD, SVD-LLM) on LLaVA and SmolVLM, reducing KV cache to 9.38% in W8A4 quantization with <1% accuracy drop on ScienceQA.

优缺点分析

Strengths:

  • Efficiency: Joint SVD+quantization reduces KV cache (η<sub>qsvd</sub> = rL vs. η<sub>fp</sub> = 2LE) and computation (FLOPs ↓ when r < 0.75E).
  • Novel Ranking Strategy: Global singular value importance scoring (Eq. 10) outperforms uniform/Fisher-based allocation (Table 3).
  • Strong Validation: Tested on 5 VLMs and 3 benchmarks with W4A4 robustness (Table 2).

Weaknesses:

  • Attention-Layer Limitation: No optimization for FFN layers (Sec. 5).

问题

  1. Does the O(E²) cost of Eq. 10 affect real-time performance?

  2. How does r selection adapt to long sequences (e.g., high-res images)?

  3. Could visual-textual feature misalignment hinder joint SVD?

局限性

Yes

最终评判理由

5

格式问题

N/A

作者回复

Thank you for your insightful comments. Below, we summarize the key concerns and questions raised in your review and provide detailed responses to each point.

Q1: Attention-Layer Limitation: No optimization for FFN layers (Sec. 5).

Ans: As noted in earlier works [1-4], self attention layers, especially the key and value (KV) vectors, contribute significantly to the overall latency, energy consumption, and memory usage of large models and can severely limit inference speed. Therefore, optimizing the QKV matrices using SVD is critically important and can lead to substantial reductions in latency for large language models.

In addition, as noted in Line 315 of the paper, quantization is applied to the FFN layers and already yields significant computational savings. When applying SVD to the FFN layers, we observed that FFN are highly sensitive to SVD, so we chose to use quantization alone, consistent with the observation of prior works [5–7]. Future research can further explore methods to better optimize the FFN layers.

[1] Hooper, Coleman, et al. "Kvquant: Towards 10 million context length llm inference with kv cache quantization." Advances in Neural Information Processing Systems 37 (2024): 1270-1303.

[2] Kim, Minsu, et al. "Oaken: Fast and Efficient LLM Serving with Online-Offline Hybrid KV Cache Quantization." Proceedings of the 52nd Annual International Symposium on Computer Architecture. 2025.

[3] Kwon, Woosuk, et al. "Efficient memory management for large language model serving with pagedattention." Proceedings of the 29th symposium on operating systems principles. 2023.

[4] Zhang, Zhenyu, et al. "H2o: Heavy-hitter oracle for efficient generative inference of large language models." Advances in Neural Information Processing Systems 36 (2023): 34661-34710.

[5] Yuan, Zhihang, et al. "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

[6] Chang, Chi-Chih, et al. "xkv: Cross-layer svd for kv-cache compression." arXiv preprint arXiv:2503.18893 (2025).

[7] Chang, Chi-Chih, et al. "Palu: Compressing kv-cache with low-rank projection." arXiv preprint arXiv:2407.21118 (2024).

Q2: Does the O(E2)\mathcal{O}(E^2) cost of Eq. 10 affect real-time performance?

Ans: Thank you for the question. The O(E2)\mathcal{O}(E^2) cost in Equation 10 only applies during the offline calibration phase, where we compute importance scores to allocate ranks using a small calibration dataset. We benchmarked the complete QSVD pipeline for LLaVA-v1.5-13B on a single A100 80G GPU, using the same 256-sample calibration set as described in the paper. The total time is approximately 96 minutes. This is much shorter than the normal training time (204 GPU hours for LLaVA-v1.5-13B [1]). QSVD generates low-rank and quantized weights, which are stored and efficiently loaded during inference, in line with approaches such as AWQ [2] and GPTQ [3]. There is no need to recompute any part of the calibration during inference, making QSVD a one-time lightweight process with minimal runtime overhead.

[1] Liu, Haotian, et al. "Improved baselines with visual instruction tuning." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2024.

[2] Lin, Ji, et al. "Awq: Activation-aware weight quantization for on-device llm compression and acceleration." Proceedings of machine learning and systems 6 (2024): 87-100.

[3] Frantar, Elias, et al. "Gptq: Accurate post-training quantization for generative pre-trained transformers." arXiv preprint arXiv:2210.17323 (2022).

Q3: How does r selection adapt to long sequences (e.g., high-res images)?

Ans: Thank you for the thoughtful comment. To evaluate the adaptability of our rank selection method to long sequences, we conduct experiments on the HRBench4K dataset [1], which consists of 4K resolution images, and use VLMEvalKit [2] for evaluation. We follow the same setup as the original paper and report the “Average All” accuracy metric in HRBench4K. Both LLaVA-v1.6-13B and LLaVA-v1.5-13B are evaluated under the QSVD-noQ setting and compared with ASVD and FP16 baselines. The results are shown below, where R1 and R2 follows the same definition as Equation 15 in the paper:

LLaVA-v1.6-13B

MethodAcc.Hw costAcc.Hw costAcc.Hw costAcc.Hw cost
ASVD44.38%R1: 63.30%; R2: 22.50%44.12%R1: 60.00%; R2: 20.00%43.12%R1: 56.70%; R2: 17.50%42.62%R1: 53.30%; R2: 15.00%
QSVD-noQ44.88%R1: 60.00%; R2: 22.50%44.12%R1: 53.30%; R2: 20.00%43.88%R1: 46.70%; R2: 17.50%44.12%R1: 40.00%; R2: 15.00%
FP1645.63%

LLaVA-v1.5-13B

MethodAcc.Hw costAcc.Hw costAcc.Hw costAcc.Hw cost
ASVD39.12%R1: 63.30%; R2: 22.50%38.62%R1: 60.00%; R2: 20.00%36.62%R1: 56.70%; R2: 17.50%37.25%R1: 53.30%; R2: 15.00%
QSVD-noQ39.88%R1: 60.00%; R2: 22.50%38.75%R1: 53.30%; R2: 20.00%39.00%R1: 46.70%; R2: 17.50%39.00%R1: 40.00%; R2: 15.00%
FP1639.12%

As shown above, QSVD-noQ outperforms ASVD, and the overall trends observed on HRBench4K are consistent with those on ScienceQA-IMG and VizWiz (see Table 1 in our paper). This indicates that our rank selection method generalizes well to long sequences resulting from high-resolution images.

[1] Wang, Wenbin, et al. "Divide, conquer and combine: A training-free framework for high-resolution image perception in multimodal large language models." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 39. No. 8. 2025.

[2] Duan, Haodong, et al. "Vlmevalkit: An open-source toolkit for evaluating large multi-modality models." Proceedings of the 32nd ACM international conference on multimedia. 2024.

Q4: Could visual-textual feature misalignment hinder joint SVD?

Ans: Thank you for this insightful question. Visual-textual feature misalignment is a well-known challenge in VLMs and can significantly impact overall model accuracy. However, it does not directly hinder the effectiveness of our proposed joint SVD approach in QSVD, as QSVD is designed to maintain the original VLM's performance even under feature misalignment, while simultaneously enhancing computational efficiency. Furthermore, by compressing Q, K, and V jointly, QSVD inherently smooths parameter variations across modalities. This may even help mitigate alignment-induced instability by regularizing the QKV projection layer and encouraging a more cohesive representation space. Additionally, the evaluation in Section 4 shows consistent performance across diverse VLMs (e.g., SmolVLM, LLaVA) and datasets (e.g., ScienceQA, VizWiz), despite differences in visual-textual alignment quality. This suggests that joint SVD is robust to potential modality misalignments at the embedding level.

Finally, we simulate the effect of feature misalignment by injecting Gaussian noise into the visual features generated by the visual encoder. Specifically, we perturb the output of the projection layer with Gaussian noise at varying levels, using a standard deviation of 0.1. The corrupted visual features are then concatenated with the textual embeddings and passed through the subsequent layers for processing. Following the experimental setup in the main paper (Table 4), we use the LLaVA-v1.5-7B model and evaluate performance on the ScienceQA benchmark.

The results, shown in the table below, indicate that QSVD exhibits a certain degree of robustness to feature misalignment, performing noticeably better than ASVD under the same conditions. This observation suggests the potential resilience of joint SVD to modality noise, which warrants further investigation in future work.

SettingFP16-ASVDFP16-QSVD-noQW4A4-QSVD
W/o noise in feature66.3468.4655.16
W/ noise in feature46.2450.0248.29
评论

5->5. I wish to see all these comments and supplementary experiments can be posed on the camera-ready version.

审稿意见
4

This paper proposes an efficient method for compressing Vision-Language Models (VLMs) by applying Singular Value Decomposition (SVD) to the QKV weight matrices, reducing KV cache size and computation. A dynamic rank allocation strategy is introduced to balance compression and accuracy. The approach is further enhanced with quantization of both weights and activations. Experiments show that the method achieves better accuracy–efficiency trade-offs than prior work, making it suitable for real-time, resource-constrained settings.

优缺点分析

Strengths:

  1. The paper is well-structured and easy to follow.
  2. The experiments and ablations are thorough and clearly demonstrate the effectiveness of the proposed method.

Weaknesses:

  1. My main concern is that the method appears to lack VLM-specific design—it primarily targets the QKV weights, which is a general approach applicable to LLMs as well. This raises the question: why is the method not directly evaluated on standard LLM benchmarks and compared against existing LLM compression techniques, if it's not tailored to vision-language settings?
  2. The paper lacks comparisons with state-of-the-art post-training quantization methods in terms of inference time, memory usage, and accuracy.
  3. The figures need to be more refined. For example, Figure 2 alone is hard to interpret—the connections and meanings of the different subplots are unclear.

问题

See Weakness

局限性

Yes

最终评判理由

Thank you for the authors’ response. I feel that most of my concerns have been addressed. However, the explanation of how the proposed method specifically tackles the unique challenges of VLMs remains somewhat unclear. Therefore, I have changed my rating from borderline reject to borderline accept.

格式问题

No

作者回复

Thank you for your insightful comments. Below, we summarize the key concerns and questions raised in your review and provide detailed responses to each point.

Q1: QSVD primarily targets the QKV weights, which is a general approach applicable to LLMs as well. This raises the question: why is the method not directly evaluated on standard LLM benchmarks and compared against existing LLM compression techniques, if it's not tailored to vision-language settings?

Ans: Efficiency enhancement techniques specifically targeting VLMs have not been studied as extensively as those for language models. Given the growing computational demands of VLM deployment, we believe QSVD is both timely and essential. Our joint QKV compression approach directly addresses the unique efficiency challenges of multimodal inference, setting it apart from conventional unimodal compression methods used in language models.

In addition, whether the proposed techniques can be effectively applied to LLMs requires further investigation. This is because the quantization scheme in QSVD is specifically tailored to the input activation distributions observed in VLMs. As shown in our analysis (Section 3.3, Figure 3), VLM activations display distinct channel-wise outlier patterns. The quantization methods we introduce are designed based on this distribution. Consequently, applying the same techniques to LLMs may not yield comparable effectiveness, as their activation distributions can differ significantly from those of VLMs due to the absence of multimodal integration. We plan to explore this direction in the future.

Q2: The paper lacks comparisons with state-of-the-art post-training quantization methods in terms of inference time, memory usage, and accuracy.

Ans: In our original manuscript (Table 2), we presented accuracy comparisons between QSVD and several leading post-training quantization methods, including Duquant [1] and QVLM [2]. Additionally, we extended our evaluation by comparing QSVD with more recent PTQ approaches, including MBQ [3] (targeting VLMs) and AffineQuant (designed for LLMs) [4]. The table below shows the evaluation results in terms of accuracy:

AccuracySQAVizWiz
MBQAffineQuantOurs QSVDMBQAffineQuantOurs QSVD
llava v1.5 7BW4A439.2746.2355.1635.6142.6352.05
W8A454.6259.4565.6143.8546.2252.18
llava v1.5 13BW4A450.5958.5565.8240.0045.5656.82
W8A455.5167.8370.1248.7250.2553.20
llava v1.6 7BW4A450.6953.7259.6740.1349.9052.00
W8A463.6265.8766.1045.2552.7553.72
llava v1.6 13BW4A460.0061.2463.6145.0049.1154.27
W8A463.1268.9770.4350.0156.4158.52

We observe that QSVD consistently achieves comparable or superior accuracy, particularly under aggressive quantization settings such as W8A4 and W4A4.

To provide a fair and concrete comparison with standard PTQ methods in memory usage, we conducted an experiment on one Transformer layer of LLaVA-v1.5-13B, focusing on the decoding stage. Specifically, we compare the following configurations:

  • W8A8 PTQ: Linear layers, KV cache and attention use standard INT8 quantization.

  • QSVD-8bit: Low-rank decomposition with int8 quantization, using our fused attention kernel optimized for compressed latent cache structure.

For the W8A8 PTQ baseline, we adopt the per-tensor RTN (round-to-nearest)  quantization scheme, one of the simplest and most memory-efficient PTQ strategies, with no additional quantization parameters. This sets a conservative lower bound on memory usage. Most other SOTA PTQ methods introduce additional per-channel scales, zero-points, or outlier handling, which increase memory usage and inference time beyond RTN.

We measured the peak memory usage in MB during decoding under varying sequence lengths (1024 and 2048) and batch sizes (16, 32, 160). The results are summarized below (Peak Memory Usage in MB):

Seq Len: 1024MemoryUsage(MB)
Batch Size1632160
W8A8702.76944.182875.52
QSVD-8bit545.97633.141335.66
Seq Len: 2048MemoryUsage(MB)
Batch Size1632160
W8A8943.511424.935276.27
QSVD-8bit626.97824.122275.41

As shown, under the same 8-bit precision, QSVD achieves significantly lower peak memory usage than standard W8A8 PTQ, especially under longer sequence lengths and larger batch sizes.

Figure 4 in the paper shows the inference latency comparison between QSVD-8bit and W8A8. The same advantage is observed, QSVD-8bit achieves up to 1.4×1.4\times speedup over standard W8A8 PTQ during decoding. This improvement is due to both reduced memory traffic, as QSVD compresses the attention cache and lowers key/value memory reads, and fused kernel execution, which eliminates intermediate memory writes and enables more efficient computation by directly operating on the compressed representation.

[1] Lin, Haokun, et al. "Duquant: Distributing outliers via dual transformation makes stronger quantized llms." Advances in Neural Information Processing Systems 37 (2024): 87766-87800.

[2] Wang, Changyuan, et al. "Q-vlm: Post-training quantization for large vision-language models." Advances in Neural Information Processing Systems 37 (2024): 114553-114573.

[3] Li, Shiyao, et al. "Mbq: Modality-balanced quantization for large vision-language models." Proceedings of the Computer Vision and Pattern Recognition Conference. 2025.

[4] Ma, Yuexiao, et al. "Affinequant: Affine transformation quantization for large language models." arXiv preprint arXiv:2403.12544 (2024).

Q3: The figures need to be more refined. For example, Figure 2 alone is hard to interpret—the connections and meanings of the different subplots are unclear.

Ans: We thank the reviewer for the helpful feedback. To improve clarity and readability, we have clarified the connections and meanings as follows. We will also enhance the quality of Figure 2 in the final version of the paper. Specifically

Subplot (a) presents the original QKV architecture in VLM without applying SVD.

Subplot (b) illustrates the application of SVD separately to the weight matrices of QQ , KK and VV , where each of WqW_{q} , WkW_{k} , and WvW_{v} is decomposed into a pair of down and up projection matrices.

Subplot (c) shows our proposed QKV weight concatenation followed by the SVD process.

Subplot (d) depicts the standard KV vectors computation process. At each decoding step, the input XX is multiplied with WkW_{k} and WvW_{v}, and the resulting CkC_{k} and CvC_{v} are stored in memory for future use.

Subplot (e) illustrates the recomputation process when SVD is applied separately to WkW_{k} and WvW_{v}. During decoding, the intermediate input XX must be read from memory, and the KK and VV vectors are recomputed by multiplying XX with the down-projection matrices WkdW^{d}_k and WvdW^{d}_v, respectively.

Subplot (f) demonstrates the storage cost in QSVD. During decoding, since WqW_{q}, WkW_{k}, and WvW_{v} share a common down-projection matrix WqkvdW_{qkv}^{d} , the input XX is multiplied by WqkvdW_{qkv}^{d} to produce the intermediate data CqkvC_{qkv}, which is then stored in memory for later recomputation of the KV vectors.

评论

Thank you for the authors’ response. I feel that most of my concerns have been addressed. However, the explanation of how the proposed method specifically tackles the unique challenges of VLMs remains somewhat unclear. Therefore, I have changed my rating from borderline reject to borderline accept.

审稿意见
5

This paper introduces a QSVD method that combines quantization and SVD. It cocnats QKV weight matrix together and performs a low-rank SVD on it. Then it introduces a cross‑layer, importance‑based singular‑value ranking scheme for adaptive rank allocation. This method can reduce the GPU memory usage and computation cost, and experiments on five VLMs across ScienceQA, VizWiz, and SEED‑Bench demonstrate that QSVD outperforms SVD‑only and quantization‑only baselines.

优缺点分析

  1. The joint QKV SVD formulation is clearly derived, with precise comparisons against single matrix SVD and analysis of parameter/FLOPs/cache reductions. The results support the method proposed.
  2. The figures are intuitive. A small drawback is that the order of the sub-graphs is inconsistent with the order in the text. 2. The parameter beta is only used in Quantization, so it’s confusing in Section 3.1 without explanation.
  3. QSVD works on different models and datasets, but why do we need to concatenate the three matrices together? Is it just to reduce cache overhead? Or is there a deeper reason for this design? A deeper analysis is encouraged.
  4. While prior work apply SVD to Q, K, V separately, this paper’s joint factorization over the concatenated matrix is novel (despite the lack of a deeper motivation)
  5. The knockout approach to measuring the importance of singular values is novel.

问题

  1. Analyze why we can get better results from SVD on concated QKV. Either theory or experiment is encouraged.
  2. What is the beta obtained on the calibration dataset? Can authors study the beta values on different datasets? This ablation is meaningful for practical applications.
  3. Experiment on GPU memory usage between different methods is encouraged.

局限性

Yes

最终评判理由

Post rebuttal, my concerns about method explanations, ablation experiments and GPU consumption have been addressed. Thus I am leaning to accept the paper.

格式问题

No

作者回复

Thank you for your insightful comments. We have summarized weakness and questions raised in your review below and provided detailed responses to each.

Q1: The figures are intuitive. A small drawback is that the order of the sub-graphs is inconsistent with the order in the text.

Ans: We will clarify the figure in the final version of the paper.  And please see Reviewer HD9e Q3

Q2: The parameter beta is only used in Quantization, so it’s confusing in Section 3.1 without explanation.

Ans: The parameter β\beta, although introduced earlier in Section 3.1, does not affect the outcome of the SVD itself. Specifically, when SVD is applied to a weight matrix WW, the process can be expressed as:

W=UΣVT=(UΣβ)(Σ1βVT)=WdWuW = U \Sigma V^T = (U \Sigma^\beta)(\Sigma^{1-\beta} V^T) = W_{d}W_{u}

The choice of β\beta does not alter the reconstruction result of the SVD. We note this here to support the discussion of β\beta in Section 3.3.

However, once quantization is introduced, the value of β\beta becomes impactful on the final accuracy result, as it influences the magnitudes of WdW_d, WuW_u, and the intermediate product Cqkv=XWdC_{qkv} = XW_d, and their corresponding quantized representations. Therefore, learning an appropriate value for β\beta is crucial for optimizing quantization performance.

Q3: QSVD works on different models and datasets, but why do we need to concatenate the three matrices together? Is it just to reduce cache overhead? Or is there a deeper reason for this design? Analyze why we can get better results from SVD on concated QKV. Either theory or experiment is encouraged.

Ans: In addition to reducing the KV cache size, another potential advantage of applying SVD to the concatenated QKV weights is the improved flexibility in assigning singular values compared to applying SVD to each of the Q, K, and V weights separately. For example, given a total rank budget of 9, performing SVD on Q, K, and V individually would require a fixed allocation (such as 3, 3, and 3), which may result in truncating many important singular values and lead to suboptimal performance. In contrast, applying SVD to the concatenated QKV weights allows for a global comparison of singular values across all three matrices. This enables more adaptive and effective rank allocation under the same total budget, potentially yielding better overall performance.

To illustrate the advantages of QKV concatenation over separate QKV decomposition for SVD, we provide a quantitative comparison below, summarized from Table 1 and Table 3 of our paper. The comparison uses LLaVA-v1.5-13B under different preserved parameters across both settings, the definitions of R1R_{1} and R2R_{2} follow equation 15 in the paper.

SourceSVD TypeAcc.Hw costAcc.Hw costAcc.Hw costAcc.Hw cost
From Table 1, LLaVA-v1.5 13B, ASVDSeparately SVD64.70%R1: 63.30%; R2: 22.50%56.92%R1: 60.00%; R2: 20.00%46.50%R1: 56.70%; R2: 17.50%42.79%R1: 53.30%; R2: 15.00%
From Table 3, LLaVA-v1.5 13B, URJoint SVD71.84%R1: 60.00%; R2: 22.50%70.40%R1: 53.30%; R2: 20.00%70.40%R1: 46.70%; R2: 17.50%67.72%R1: 40.00%; R2: 15.00%

As the table shows, joint SVD improves accuracy by more than 7.14% under the same compression budget. This trend holds consistently across various preserved ratios, R1 and R2 are defined in Equation 15.

Finally, concatenating the Q, K, and V weights allows for a shared down projection, which significantly reduces the KV cache size during inference. As shown in Figure 2 and discussed in Section 3.1, our design stores a single intermediate matrix CqkvC_{{qkv}} instead of computing and caching separate key and value projections for each WkW_{{k}} and WvW_{{v}}. This results in approximately a twofold reduction in KV cache size compared to traditional approaches.

Q4: What is the beta obtained on the calibration dataset? Can authors study the beta values on different datasets? This ablation is meaningful for practical applications.

Ans: In our paper. We follow the convention of QVLM [1] to use samples from ScienceQA training set as our calibration dataset, and the resulting beta is around 0.12. In contrast, MBQ [2] uses the COCO [3] dataset for its calibration process. To align with this, we extended our beta learning experiments using COCO data and observed a similar optimal beta value of approximately 0.10. This consistency indicates that beta remains stable across different calibration datasets, which significantly enhances the practicality of QSVD, since a beta value of 0.1 can be reliably applied across various datasets.

[1] Wang, Changyuan, et al. "Q-vlm: Post-training quantization for large vision-language models." NeurIPS, 2024.

[2] Li, Shiyao, et al. "Mbq: Modality-balanced quantization for large vision-language models." Proceedings of the Computer Vision and Pattern Recognition Conference. 2025.

[3] Lin, Tsung-Yi, et al. "Microsoft coco: Common objects in context." European conference on computer vision. Cham: Springer International Publishing, 2014.

Q5: Experiment on GPU memory usage between different methods is encouraged.

Ans: To provide a fair and concrete comparison with standard PTQ methods in memory cost, we conducted an experiment on one transformer layer of LLaVA-v1.5-13B, focusing on the decoding stage. Specifically, we compare the following approaches:

  • W8A8 PTQ: Linear layers, KV cache and attention use standard INT8 quantization.

  • QSVD-8bit: Low-rank decomposition with int8 quantization, using our fused attention kernel optimized for compressed latent cache structure.

For the W8A8 PTQ baseline, we adopt the per-tensor RTN (round to nearest) quantization scheme, one of the simplest and most memory-efficient PTQ strategies, with no additional quantization parameters. This sets a conservative lower bound on memory usage. Most other recent PTQ methods introduce additional per-channel scales, zero-points, or outlier handling, which increase memory usage beyond RTN.

We measured the peak memory usage in MB during decoding under varying sequence lengths (1024 and 2048) and batch sizes (16, 32, 160). The results are summarized below (Peak Memory Usage in MB):

Seq Len: 1024MemoryUsage(MB)
Batch Size1632160
W8A8702.76944.182875.52
QSVD-8bit545.97633.141335.66
Seq Len: 2048MemoryUsage(MB)
Batch Size1632160
W8A8943.511424.935276.27
QSVD-8bit626.97824.122275.41

As shown, under the same 8-bit precision, QSVD achieves significantly lower peak memory usage than standard W8A8 PTQ, especially under longer sequence lengths and larger batch sizes.

评论

Thanks for your thoughtful response. My concerns have been addressed. I will raise my score.

审稿意见
5

This paper proposes QSVD, a unified framework for compressing Vision-Language Models by applying joint Singular Value Decomposition to the combined query, key, and value weight matrices, alongside post-training quantization of weights and activations. The goal is to reduce memory, computational cost, and KV cache size without degrading model accuracy.

优缺点分析

Strengths:

  • This paper proposes a novel compression paradigm that effectively combines SVD and PTQ.
  • The buffered mechanism and auto-regressive decoding work well together.
  • The figures well demonstrate the method design.
  • Extensive experiments on popular VLMs show superior performance over existing works.

Weaknesses:

  • The method's effectiveness is quite sensitive to β\beta. And β\beta is determined by a very small calibration set. I'm not sure if the generalization ability of such pattern is enough.
  • The method can only be used on QKV matrices and can not be used on other layers like FC layers, restricting its compression capacity.

问题

  • Can we replace post-training quantization by quantization-aware training and further improve the performance?
  • What about the training time? Does QSVD add heavy additional cost compared with normal training?

局限性

Yes

最终评判理由

See comments

格式问题

No

作者回复

Thank you for your insightful comments. Below, we summarize the weaknesses and questions raised in your review and provide detailed responses to each.

Q1: The method's effectiveness is quite sensitive to β\beta . And β\beta is determined by a very small calibration set. I'm not sure if the generalization ability of such pattern is enough.

Ans: We explored the beta learning experiments on different sample size of 64, 128, 256 and 512. Furthermore, to show the effectiveness of beta learning, we also show the QSVD performances when fixing beta to 0.2, 0.4, 0.8 and 1. The results for W4A4 and W8A4 are shown in table below:

LLaVA v1.5 7B
β\betaQSVDβ\beta=0.2β\beta=0.4β\beta=0.8β\beta=1
n=64W4A452.3148.4442.2439.3230.05
W8A464.4663.3661.3858.4049.43
n=128W4A446.8542.5435.3036.7630.05
W8A465.7462.0263.9661.3847.79
n=256 (QSVD default setting)W4A455.1654.7942.3749.2240.45
W8A465.6164.0664.3560.2758.58
n=512W4A441.6533.1226.5617.0910.55
W8A463.9162.1763.7159.4413.93

The table shows that the selection of β\beta matches in different sample size settings, and this learning β\beta method generalizes well over different calibration sizes.

Moreover, it is standard in post-training quantization and low-rank compression literature to use small calibration sets to estimate sensitive hyperparameters. For example: AWQ [1] Section 5.3 Figure 8(a) use 8-256 samples to determine the scaling factor exponent. ASVD [2] Section 4.1 use only 64 samples of calibration dataset in all their experiments and hyperparameter search.

[1] Lin, Ji, et al. "Awq: Activation-aware weight quantization for on-device llm compression and acceleration." Proceedings of machine learning and systems 6 (2024): 87-100.

[2] Yuan, Zhihang, et al. "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

Q2: The method can only be used on QKV matrices and can not be used on other layers like FC layers, restricting its compression capacity.

Ans: As noted in earlier works [1-4], self attention layers, particularly the KV vectors, contribute significantly to the overall latency, energy consumption, and memory usage of large models, and can severely limit inference speed. Therefore, optimizing the QKV matrices using singular value decomposition is critically important and can lead to substantial reductions in latency for large language models.

In addition, as mentioned in Line 315 of our paper, quantization is applied to the FC layers within the feedforward network (FFN), as well as the output FC layers within the self attention mechanism. While SVD can also be applied to these FC layers, we find them to be highly sensitive to such decomposition. Applying quantization alone is sufficient to compress these layers effectively. This design choice aligns with prior work [5, 6]. Future research may explore more advanced strategies to further optimize the FFN components.

[1] Hooper, Coleman, et al. "Kvquant: Towards 10 million context length llm inference with kv cache quantization." Advances in Neural Information Processing Systems 37 (2024): 1270-1303.

[2] Kim, Minsu, et al. "Oaken: Fast and Efficient LLM Serving with Online-Offline Hybrid KV Cache Quantization." Proceedings of the 52nd Annual International Symposium on Computer Architecture. 2025.

[3] Kwon, Woosuk, et al. "Efficient memory management for large language model serving with pagedattention." Proceedings of the 29th symposium on operating systems principles. 2023.

[4] Zhang, Zhenyu, et al. "H2o: Heavy-hitter oracle for efficient generative inference of large language models." Advances in Neural Information Processing Systems 36 (2023): 34661-34710.

[5] Yuan, Zhihang, et al. "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

[6] Chang, Chi-Chih, et al. "xkv: Cross-layer svd for kv-cache compression." arXiv preprint arXiv:2503.18893 (2025).

[7] Chang, Chi-Chih, et al. "Palu: Compressing kv-cache with low-rank projection." arXiv preprint arXiv:2407.21118 (2024).

Q3: Can we replace post-training quantization by quantization-aware training and further improve the performance?

Ans: Yes, incorporating quantization-aware training (QAT) has the potential to further improve quantization performance by allowing model weights to adapt to low-precision constraints during training. However, this approach incurs significant training cost, as backpropagation must be performed across the entire vision-language model, resulting in substantial computational overhead and increased training latency. In contrast, our primary goal in this work is to develop a highly effective post-training quantization (PTQ) method that is both efficient and practical.

SpinQuant [1] demonstrates that end-to-end QAT can yield superior results compared to QuaRot under aggressive settings (e.g., W4A4), achieving up to 28.6-percent improvements (see SpinQuant Table 5, Section 4.3.4). While impressive, this comes at the cost of additional training and online adaptation. Our method, like QuaRot, is fully PTQ and no E2E training, and we believe that with modest QAT integration (e.g., low-rank-aware finetuning or selective stage-wise updates), QSVD has the potential have an even better performance, which we leave as promising future work.

[1] Liu, Zechun, et al. "Spinquant: Llm quantization with learned rotations." arXiv preprint arXiv:2405.16406 (2024).

Q4: What about the training time? Does QSVD add heavy additional cost compared with normal training?

Ans: Thank you for the question. QSVD is designed as a post-training compression method and does not require full model retraining. The additional computation involved comes from three components: (1) low-rank SVD decomposition of the QKV matrices, (2) importance score–based rank allocation, and (3) calibration-based tuning of the quantization parameter β\beta and the quantization process. The latency overhead introduced by QSVD is negligible compared to standard training procedures. Furthermore, it introduces no additional cost during inference.

To study the detailed training cost, QSVD introduces only a lightweight calibration and tuning stage, rather than full fine-tuning. For a fair comparison, we benchmarked the complete QSVD pipeline for LLaVA-v1.5-13B on a single A100 80G GPU, using the same 256-sample calibration set as described in the paper. The total time is approximately 96 minutes. The detailed breakdown is:

StepTime
Input calibration1 min
SVD (on all layers)2 min 15 s
Gradient collection & rank allocation10 min 12 s
SVD results Fusion30 s
β\beta tuning & quantization (fake quant)82 min
Total Calibration Time96 min

In comparison, the typical training cost of a vision-language model is approximately 204 GPU hours [1] on A100, making the computational overhead of QSVD negligible relative to full model training.

[1] Liu, Haotian, et al. "Improved baselines with visual instruction tuning." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2024.

评论

Thanks for the rebuttal. My concerns are mostly solved. I think this paper is technically solid and worth being spread. I keep my initial recommendation "accept".

评论

Dear reviewers,

Thank you for your efforts. The authors have provided a detailed response. Please check if any concerns remain and engage in discussion with authors.

Best, AC

评论

good

最终决定

Driven by the high computation cost and memory footprint of vision language models, the authors propose a method to reduce the KV cash size. The authors suggest to concatenate the Key, Query and Value matrices and perform a Singular Value Decomposition with an efficient rank allocation.

Strengths

All reviewers found this interesting and novel.

The paper demonstrated strong performance.

Through experiments across various models.

Weakness

The main limitation pointed by the reviewers is the applicability to the attention layers only, however the authors showed that the work can be combined with FC quantization.