PaperHub
5.3
/10
Rejected4 位审稿人
最低3最高6标准差1.3
6
3
6
6
3.8
置信度
ICLR 2024

LRQ: Optimizing Post-Training Quantization for Large Language Models by Learning Low-Rank Weight-Scaling Matrices

OpenReviewPDF
提交: 2023-09-22更新: 2024-02-11

摘要

关键词
Efficient InferencePost-Training QuantizationLarge Language Models

评审与讨论

审稿意见
6

This work propose LRQ (Low-Rank Quantization), a simple yet effective PTQ method that leverages low-rank weight scaling matrices. This enables the method to learn with fewer parameters while still able to scale weights individually. Results show that the proposed method achieved better performance than existing works on both INT8 and INT4/8 cases.

优点

(1) A thorough, detailed, easy-to-follow description of the method in section 2.

(2) A thorough evaluation of the proposed and related methods on various models and tasks.

缺点

(1) It would be helpful to add a comparison of the computation cost of the proposed and related quantization methods.

(2) Overall this is a solid work, but some of the results are not that exciting. For example, the authors made a big claim about INT4/8 in the abstract “Remarkably, we confirm for the first time the possibility of 4-bit weight and 8-bit activation quantization for LLMs with minimal accuracy loss among LLM PTQ studies.”. However, results in table 3 and 4 show that existing methods can do INT4/8 as well, sometimes even better than the proposed method.

问题

Please see my concerns in weaknesses.

评论

Dear Reviewer m3qc,

We appreciate your constructive comments.


[Q1.Comparisonofcomputationalcost]**[Q1. Comparison of computational cost]**

For a comparative analysis of FlexRound and LRQ in terms of computational cost, we measure the execution time and peak GPU memory usage while quantizing Llama 7B to W8A8 using 512512 calibration samples and a batch size of 22. FlexRound required 5 hours and 7 minutes, whereas LRQ consumed 5 hours and 22 minutes. LRQ's extended processing time is attributed to the multiplication involving L2\mathbf{L}_2 and U2\mathbf{U}_2. Despite the slightly longer runtime, LRQ demonstrates an advantage in peak GPU memory usage, utilizing 23.523.5 GB compared to FlexRound's 25.425.4 GB. This efficiency is attributed to LRQ's fewer learnable parameters in comparison to FlexRound.


[Q2.ExperimentalresultsofW4A8quantization]**[Q2. Experimental results of W4A8 quantization]**

Although LRQ always outperforms FlexRound in the case of W4A8 quantization, as highlighted by the reviewer, the W4A8 performance of FlexRound on common sense reasoning tasks and MMLU are not notably poor. However, our primary intent was to underscore that we conducted experiments specifically addressing W4A8 quantization for the first time. In the original FlexRound paper, there were no experimental results pertaining to W4A8 quantization. To enhance clarity, we revise that portion of our manuscript to state, "we for the first time conduct experiments on W4A8 quantization among LLM PTQ studies.".


审稿意见
3

The paper proposes a new weight rounding scheme via a small modification on top of FlexRound technique: whereas in the FlexRound the full matrix of elementwise reciprocals are being learned, the proposed method changes it to a low-rank matrix. In other words, if the FlexRound's quantization (rounding) function is given as Q(W;α,S)=α×WSQ(\mathbf{W}; \alpha, \mathbf{S}) = \alpha \times \lfloor \frac{\mathbf{W}}{\mathbf S} \rceil where α\alpha is the shared re-scaling parameter, S\mathbf{S} is the matrix of learned reciprocals, and \lfloor \cdot \rceil is the rounding-to-nearest function, then the proposed modification is to change the S\mathbf{S} to be of a low rank (with r{1024,2048}r \in \{1024, 2048\} ) . The parameters of this quantization function are learned in a block-wise reconstruction manner to minimize the 2\ell_2 difference between quantized and non-quantized outputs. Gradients for non-differentiable operations are computed via straight-through-estimation. The motivation for the proposed change is rooted in the fact that learning the entire matrix S\mathbf{S} on the calibration dataset might lead to overfitting, and thus low-rank S\mathbf{S} will counteract it. The paper is supplemented with many experiments showcasing the effectiveness of the proposed method and comparisons to other techniques.

优点

The proposed change requires minimal modification for any existing quantization pipelines and codebases, and thus can be quickly absorbed into actual real-world usage. Despite simplicity, the experimental evaluation suggests that the method provides measurable, albeit marginal, improvement over existing techniques.

缺点

Despite the aforementioned strengths, I am not very convinced on why these changes are helping the quantization. The original motivation to introduce low-rank S\mathbf{S} was to combat overfitting via "decreasing the number of learnable parameters". However, in the provided ablation study between the rank of the matrix S\mathbf{S} and the achieved zere-shot accuracy (figure 5a) on a task, it seems the rank has minimal effect on the accuracy (red curve is almost flat). Then, the question would be what is contributing to the improved accuracy if not lower rank?

Improvements: whereas the presented results show indeed better accuracies, I am somewhat unsure to which degree these results showcase the best of the other methdos. For instance, it seems authors used same scale of 0.8 for SmoothQuant, whereas the scale should ideally be validated on a calibration data. Also, the same comparison to SmoothQuant seems to be unfair: smoothquant is pure no-learning post-training quantization technique, but FlexQuant and the suggested methods are actually introduces a learning objective that is optimized. In many cases the difference between the suggested method and FlexQuant is such a minuscule number, that it can be attributed to a sampling noise (see table 1).

问题

  1. Can you please address weaknesses raised above?
  2. The FlexRound method as presented in the original paper does not have exponentiation on mathbfSmathbf{S}, but the proposed method does have it. Is it a typo or the additional reparametrization on top of FlexRound?
  3. What is the effect of rr and cc (eq.2)? Some ablation studies on those would be helpful
  4. Some general recommendations on choosing rank, learning rates, and training regime would be of great help to all practitioners. Also, if you can tabularize the actual experiment details presented in appendix C; it would help to reproducibility as well
  5. There are some minor typos in the paper: in section 2.1, we concentrates => we concentrate in section 4, simeple => simple,
评论

Dear Reviewer VLPY,

We appreciate your insightful comments.


[Q1.Effectofrankonaccuracyseemsminimal]**[Q1. Effect of rank on accuracy seems minimal]**

As astutely pointed out by the reviewer, the left side of Figure 5(a) might create the impression that the accuracy of LRQ (depicted by the red curve) remains relatively unchanged despite variations in the rank rr, ranging from 6464 to 20482048. This could potentially lead readers to misconstrue the significance of the rank rr in influencing accuracy. To mitigate the risk of such misinterpretation, we conduct a comprehensive evaluation of the zero-shot and five-shot accuracies of LRQ for Llama 7B on common sense reasoning (CSR) and MMLU, considering the rank rr spanning from 6464 to 81928192.

<Table A. Zero-shot and five-shot accuracies of LRQ for Llama 7B on common sense reasoning (CSR) and MMLU according to the rank $r$>
rank rr641282565121024204840968192
Zero-shot accuracy of LRQ on CSR60.6860.7260.7560.4660.7160.5560.3860.49
Five-shot accuracy of LRQ on MMLU33.9733.9034.2534.2034.4734.0733.2131.51

As outlined in Figure 9 in Appendix G of the revised manuscript, the accuracy of LRQ either remains relatively stable (Figure 9(a)) or increases gradually (Figure 9(b)) with the rise in the rank rr from 6464 to 10241024. However, as the rank rr continuously increases from 20482048 to 81928192, the accuracy of LRQ eventually declines to match that of FlexRound (indicated by the blue curve) on both CSR and MMLU. In light of these findings, it can be concluded that the employment of lower rank contributes considerably to improving the accuracy/generalization of quantized LLMs.


[Q2.Usingthescaleof0.8forSmoothQuantisunsure]**[Q2. Using the scale of 0.8 for SmoothQuant is unsure]**

Although it is challenging to definitively assert that using α=0.8\alpha=0.8 for SmoothQuant ensures its best performance, we adopted α=0.8\alpha=0.8 on the grounds that the authors of SmoothQuant utilized α=0.8\alpha=0.8 for all Llama models. Even when employing α=0.5\alpha=0.5, a value considered by the authors of SmoothQuant as a general sweet spot, SmoothQuant (α=0.5\alpha=0.5) typically exhibits lower performance compared to SmoothQuant (α=0.8\alpha=0.8), as depicted in the tables below. We therefore believe that opting for α=0.8\alpha=0.8 is not a suboptimal choice for SmoothQuant.

<Table B. Zero-shot performance of SmoothQuant ($\alpha=0.5$) and SmoothQuant ($\alpha=0.8$) on common sense reasoning tasks>
W/A/KVLlama 7BLlama 13 BLlama 33 BLlama 65 B
SmoothQuant (α=0.5\alpha=0.5)8/8/1656.9360.10\mathbf{60.10}60.1364.85
SmoothQuant (α=0.8\alpha=0.8)8/8/1657.48\mathbf{57.48}60.0060.71\mathbf{60.71}65.05\mathbf{65.05}
SmoothQuant (α=0.5\alpha=0.5)8/8/857.1959.3060.77\mathbf{60.77}65.02
SmoothQuant (α=0.8\alpha=0.8)8/8/857.25\mathbf{57.25}59.65\mathbf{59.65}60.6960.6965.24\mathbf{65.24}
<Table C. Five-shot accuracy of SmoothQuant ($\alpha=0.5$) and SmoothQuant ($\alpha=0.8$) on MMLU>
W/A/KVLlama 7BLlama 13 BLlama 33 BLlama 65 B
SmoothQuant (α=0.5\alpha=0.5)8/8/1629.5329.4341.9752.93
SmoothQuant (α=0.8\alpha=0.8)8/8/1629.61\mathbf{29.61}29.73\mathbf{29.73}44.82\mathbf{44.82}53.79\mathbf{53.79}
SmoothQuant (α=0.5\alpha=0.5)8/8/829.1829.2742.2952.63
SmoothQuant (α=0.8\alpha=0.8)8/8/829.94\mathbf{29.94}29.87\mathbf{29.87}45.07\mathbf{45.07}53.72\mathbf{53.72}

[Q3.ComparisonofLRQwithSmoothQuantseemsunfair]**[Q3. Comparison of LRQ with SmoothQuant seems unfair]**

As noted by the reviewer, SmoothQuant is a training-free weight-activation post-training quantization (PTQ) method. However, the authors of SmoothQuant showcased its state-of-the-art performance among existing LLM weight-activation PTQ methods encompassing both learning-free and learning-based approaches. Consequently, we undertake a comparative analysis between our method, LRQ, and SmoothQuant, acknowledging the latter as one of the current state-of-the-art PTQ methods for LLM weight-activation quantization.

To introduce another learning-based LLM weight-activation quantization technique, we compare LRQ with LLM-QAT, one of the state-of-the-art LLM weight-activation quantization-aware training (QAT) methods, for Llama 7B, 13B, and 33B on MMLU. As shown in the table below, LRQ demonstrates a substantial performance advantage over LLM-QAT, which reinforces the efficacy of LRQ further.

<Table D. Five-shot accuracy of LLM-QAT and LRQ on MMLU>
W/A/KVLlama 7BLlama 13 BLlama 33B
LLM-QAT4/8/830.0041.2054.40
LRQ4/8/835.04\mathbf{35.04}45.31\mathbf{45.31}57.72\mathbf{57.72}

评论

[Q4.SmalldifferencebetweenFlexRoundandLRQinTable1]**[Q4. Small difference between FlexRound and LRQ in Table 1]**

As the reviewer mentioned, in Table 1, LRQ exhibits a slightly superior zero-shot performance on common sense reasoning (CSR) compared to FlexRound. However, we believe that this advantage is noteworthy since FlexRound already achieves the zero-shot performance on CSR comparable to FP16 baselines. The close proximity in zero-shot performance between FlexRound and FP16 baselines on CSR limits the potential for a substantial performance disparity between FlexRound and LRQ. Despite LRQ approaching the zero-shot performance of FP16 baselines more closely than FlexRound, the difference in zero-shot performance between FlexRound and LRQ cannot be anticipated to be large after all.

Nonetheless, as expounded in Section 1, it is crucial to emphasize that LRQ demonstrates competitive performance relative to FP16 baselines on both common sense reasoning (CSR) and MMLU, a feat not accomplished by FlexRound that excels solely on CSR. Given the comprehensive evaluation of large language models (LLMs) across diverse benchmarks, the proficiency of LRQ in excelling across both CSR and MMLU holds significant implications in the field of LLM quantization.

Regarding the 88-bit weight quantization presented in Tables 3 and 4, the adoption of a per-token asymmetric activation quantization scheme results in even naive rounding-to-nearest (RTN) performing closely to the levels of FP16 baselines on both CSR and MMLU. As a result, while LRQ exhibits slightly higher accuracy compared to SmoothQuant and FlexRound for most Llama 2 models, it can be concluded that SmoothQuant, FlexRound, and LRQ are nearly evenly matched.

In the context of 44-bit weight quantization as presented in Table 3, FlexRound achieves zero-shot accuracy levels comparable to FP16 baselines on CSR, resulting in a relatively small zero-shot performance gap between FlexRound and LRQ, like the scenario depicted in Table 1. However, in the case of 44-bit weight quantization in Table 4, LRQ surpasses FlexRound by a margin ranging from 0.20.2 to 0.70.7 percent. Although, as noted by the reviewer, these increments in five-shot accuracy on MMLU in Table 4 may seem modest compared to those in Table 2, we believe that the rise in five-shot accuracy by 0.20.2 to 0.70.7 percent is significant. This is particularly noteworthy as it brings the five-shot accuracy gap between LRQ and FP16 baselines to less than 1.51.5 percent, while the corresponding gap between FlexRound and FP16 baselines remains more or less at two percent for Llama 2 13B and 70B. We hope that these W4A8 quantization results will contribute to narrowing the gap between W4A8 quantized LLMs and FP16 baselines to less than one percent.


[Q5.ExponentiationonSinFlexRound]**[Q5. Exponentiation on S in FlexRound]**

As the reviewer mentioned, there is no exponential notation for S_2R_>0Cout×Cin\mathbf{S}\_2 \in \mathbb{R}\_{>0}^{C_{out} \times C_{in}} in the original FlexRound paper. However, to keep every entry of S_2\mathbf{S}\_2 positive during minimizing the block reconstruction error, the authors of FlexRound express S_2\mathbf{S}\_2 as exp(S_2)\text{exp}(\mathbf{S}\_2) where S_2RCout×Cin\mathbf{S}\_2 \in \mathbb{R}^{C_{out} \times C_{in}} in the code implementation of FlexRound at the following url: https://openreview.net/attachment?id=-tYCaP0phY_&name=supplementary_material. Accordingly, we also used a form of exponential notation in explaining FlexRound and LRQ.


评论

[Q6.Ablationstudyontheeffectof**[Q6. Ablation study on the effect of \mathbf{r**_2andand\mathbf{c}_2]}

Thank you for your helpful suggestion. We compare FlexRound, FlexRound with S_2=L_2U_2\mathbf{S}\_2 = \mathbf{L}\_2 \mathbf{U}\_2, and LRQ for Llama 7B and 13B.

<Table E. Zero-shot performance of FlexRound, FlexRound with $\mathbf{S}_2 = \mathbf{L}_2 \mathbf{U}_2$, and LRQ on common sense reasoning tasks>
W/A/KVLlama 7BLlama 13 B
FlexRound8/8/1660.5362.40
FlexRound with S_2=L_2U_2\mathbf{S}\_2 = \mathbf{L}\_2 \mathbf{U}\_28/8/1660.6962.62
LRQ8/8/1660.71\mathbf{60.71}62.92\mathbf{62.92}
FlexRound8/8/860.3562.37
FlexRound with S_2=L_2U_2\mathbf{S}\_2 = \mathbf{L}\_2 \mathbf{U}\_28/8/860.4962.62
LRQ8/8/860.63\mathbf{60.63}62.76\mathbf{62.76}
<Table F. Five-shot accuracy of FlexRound, FlexRound with $\mathbf{S}_2 = \mathbf{L}_2 \mathbf{U}_2$, LRQ on MMLU>
W/A/KVLlama 7BLlama 13 B
FlexRound8/8/1630.2043.82
FlexRound with S_2=L_2U_2\mathbf{S}\_2 = \mathbf{L}\_2 \mathbf{U}\_28/8/1633.8645.48
LRQ8/8/1634.47\mathbf{34.47}45.83\mathbf{45.83}
FlexRound8/8/829.4343.60
FlexRound with S_2=L_2U_2\mathbf{S}\_2 = \mathbf{L}\_2 \mathbf{U}\_28/8/833.9645.21
LRQ8/8/834.39\mathbf{34.39}45.83\mathbf{45.83}

As evident from the tables above, FlexRound with S2=L2U2\mathbf{S}_2 = \mathbf{L}_2 \mathbf{U}_2 surpasses the performance of FlexRound but falls short of LRQ. It is noteworthy that the five-shot accuracy on MMLU can witness an increase ranging from 1.5% to 4% by simply substituting S2\mathbf{S}_2 with L2U2\mathbf{L}_2 \mathbf{U}_2, which corroborates the significance of leveraging the parameter-efficiency inherent in low-rank weight-scaling matrices.


[Q7.Recommendationfortherank,learningrates,andtrainingregimes]**[Q7. Recommendation for the rank, learning rates, and training regimes]**

Drawing from our experience, a value of r=2048r=2048 would be appropriate for LLMs beyond 30B parameters, while r=1024r=1024 would be well-suited for smaller models. Regarding learning rates, in the context of a per-tensor static activation quantization scheme, a learning rate ranging between 11e-33 and 33e-33 is recommended. Conversely, when employing a per-token activation quantization scheme, a learning rate within the range of 11e-44 to 55e-44 is deemed suitable. The remainder of LRQ's experimental settings align with those detailed in [1].

[1] Lee et al., FlexRound: Learnable Rounding based on Element-wise Division for Post-Training Quantization, ICML 2023.


[Q8.Presentationofimplementationdetailsintabularform]**[Q8. Presentation of implementation details in tabular form]**

Thank you for your practical suggestion. We present implementation details in tabular form in Appendix C of the revision.


[Q9.Typos]**[Q9. Typos]**

Thank you for pointing out typos. We fix those typos in the revised version.


审稿意见
6

This paper introduces LRQ, a post-training quantization technique for LLM (Large Language Models), which improves upon the FlexRound methods by reducing the size of weight-scaling matrices to enhance generalization performance. Through a comprehensive set of experiments conducted on Llama and Llama2, LRQ demonstrates improvements for 4-bit and 8-bit quantization in both Common Sense Reasoning (CSR) and Massive Multitask Language Understanding (MMLU) tasks.

优点

  1. The paper is well-written, and the low-rank weight-scaling method is simple and efficient, making it easy to comprehend.
  2. This paper has conducted an extensive ablation study on the size of parameters and calibration data, providing valuable empirical insights into the optimal cost of the weight-scaling method.

缺点

  1. LRQ builds upon FlexRound by introducing a low-rank matrix into the full weight-scaling method. As this is my first time as an ICLR reviewer, I am uncertain whether the novelty meets ICLR's standards.
  2. This paper extensively explores experiments on both CSR and MMLU tasks, using various quantization methods. However, it consistently shows only marginal improvements compared to FlexRound, with one exception being MMLU when employing per-tensor asymmetric static activation quantization. Yet, the performance gain with per-token asymmetric activation quantization, which offers finer granularity, remains limited.

问题

  1. It appears that LRQ applies low-rank quantization to weights. When it comes to activation quantization, is RTN used?
评论

Dear Reviewer F2mg,

We appreciate your constructive comments.


[Q1.NoveltyofLRQ]**[Q1. Novelty of LRQ]**

The use of low-rank matrices in LRQ seems to be simple and straightforward, yet we believe that it is important to underscore the underlying motivation of LRQ as the reviewer ok57 mentioned. As SmoothQuant scales the weights collectively per channel, SmoothQuant may lead to non-negligible accuracy loss after quantization. Since FlexRound learns an individual scale for every weight with limited calibration samples, FlexRound might be prone to overfitting. However, increasing calibration samples in FlexRound is not a viable option, which leads us to employ low-rank weight-scaling matrices in order to decrease the number of learnable parameters effectively while maintaining the concept of scaling weights individually by sharing learnable parameters via low-rank structure. In this regard, the novelty of our paper lies not merely in the utilization of low-rank matrices but is firmly rooted in the problem statement concerning the number of scaling parameters and a middle-ground approach to solving this challenge.


[Q2.LRQshowsmarginalimprovementscomparedtoFlexRound,withoneexceptionofMMLUwhenusingpertensorasymmetricstaticactivationquantization.]**[Q2. LRQ shows marginal improvements compared to FlexRound, with one exception of MMLU when using per-tensor asymmetric static activation quantization.]**

In Table 1, LRQ exhibits a slightly superior zero-shot performance on common sense reasoning (CSR) compared to FlexRound. However, we believe that this advantage is noteworthy since FlexRound already achieves the zero-shot performance on CSR comparable to FP16 baselines. The close proximity in zero-shot performance between FlexRound and FP16 baselines on CSR limits the potential for a substantial performance disparity between FlexRound and LRQ. Despite LRQ approaching the zero-shot performance of FP16 baselines more closely than FlexRound, the difference in zero-shot performance between FlexRound and LRQ cannot be anticipated to be large after all.

Nonetheless, as expounded in Section 1, it is crucial to emphasize that LRQ demonstrates competitive performance relative to FP16 baselines on both common sense reasoning (CSR) and MMLU, a feat not accomplished by FlexRound that excels solely on CSR. Given the comprehensive evaluation of large language models (LLMs) across diverse benchmarks, the proficiency of LRQ in excelling across both CSR and MMLU holds significant implications in the field of LLM quantization.

Regarding the 88-bit weight quantization presented in Tables 3 and 4, the adoption of a per-token asymmetric activation quantization scheme results in even naive rounding-to-nearest (RTN) performing closely to the levels of FP16 baselines on both CSR and MMLU. As a result, while LRQ exhibits slightly higher accuracy compared to SmoothQuant and FlexRound for most Llama 2 models, it can be concluded that SmoothQuant, FlexRound, and LRQ are nearly evenly matched.

In the context of 44-bit weight quantization as presented in Table 3, FlexRound achieves zero-shot accuracy levels comparable to FP16 baselines on CSR, resulting in a relatively small zero-shot performance gap between FlexRound and LRQ, like the scenario depicted in Table 1. However, in the case of 44-bit weight quantization in Table 4, LRQ surpasses FlexRound by a margin ranging from 0.20.2 to 0.70.7 percent. Although, as noted by the reviewer, these increments in five-shot accuracy on MMLU in Table 4 may seem modest compared to those in Table 2, we believe that the rise in five-shot accuracy by 0.20.2 to 0.70.7 percent is significant. This is particularly noteworthy as it brings the five-shot accuracy gap between LRQ and FP16 baselines to less than 1.51.5 percent, while the corresponding gap between FlexRound and FP16 baselines remains more or less at two percent for Llama 2 13B and 70B. We hope that these W4A8 quantization results will contribute to narrowing the gap between W4A8 quantized LLMs and FP16 baselines to less than one percent.


[Q3.Whenitcomestoactivationquantization,isRTNused?]**[Q3. When it comes to activation quantization, is RTN used?]**

Yes, we employ rounding-to-nearest (RTN) for activation quantization.


审稿意见
6

Thanks to the authors for a well written paper.

The paper proposes a weight-activation quantization method based on low-rank weight scaling matrices. It does a nice job contrasting 2 preceding works (smoothquant & flexround) via the number of scaling parameters they each have. The authors take a middle of the ground approach which they show to be empirically beneficial, and motivate it as well using a generalization argument.

优点

  • good job explaining background work in smoothquant and flex round, and how they differ via number of scaling parameters
  • nice motivation from flex round, ie using fig2 to show that increasing calibration samples is not viable, so therefore decrease learnable parameters
  • pretty thorough set of experimental results

缺点

  • how meaningful is the improvement of LRQ over FlexRound? is it possible to share std dev for the metrics you reported? For example in Table 1 Llama2 70B, the difference between LRQ has an average accuracy of 65.93 vs FlexRound 65.86. I see that sometimes the gap between LRQ and FlexRound is bigger (table 1: Llama33B), but also sometimes FlexRound is better (table 1: llama2 7b). I see that the gaps are much more sizable in Table2 (MMLU), but then smaller again for Tables 3 and 4.
  • going back to the improved generalization motivation, if this interpretation were correct then wouldn’t we expect to see better performance on the common sense reasoning and MMLU tasks? I agree that LRQ’s performance is slightly better, but based on my previous comment I’m not sure how significant the gap is.

问题

  • How many calibration samples were used to generate Fig 3? Also how large was the held out set? I understand the generalization argument, I’d just like to know how sensitive (or not) it is to the number of samples.
  • It took me a couple minutes to realize the difference between Tables 1 and 2 was the evaluation task (common sense reasoning vs mmlu). I was confused because “per-tensor” and “static” were italicized. Just a small point, to somehow make it easier to digest.
  • could you clarify your comment that LRQ is the first to confirm the possibility or 4Wa8 quantization? Because in Tables 3,4 LRQ does not do that much better than FlexRound.

Overall, I think the motivation and presentation of the method are very nice. The proposed low-rank solution makes a lot of sense, though methodologically there is nothing novel in the low-rank formulation. My main concern is in having a more careful interpretation of the empirical results (If I missed any, please point me to the paper). To me it seems like the gap between LRQ and FlexRound may only be significant in Table 2, and somewhat small elsewhere. Also, the claim that this methods makes 4W8A possible for the first time confuses me, as FlexRound is not that much worse (Tables 3&4). Is this claim really true? Please let me know if I am misunderstanding something. My reading of the results is that LRQ non-trivially increases the performance on MMLU tasks when using the setting of Table 2, and has minor improvements in other cases.

评论

[Q5.Italicizationofpertensorandstaticisconfusing]**[Q5. Italicization of “per-tensor” and “static” is confusing]**

Our initial aim was to clarify that Tables 1 and 2 present the experimental results of a per-tensor asymmetric static activation quantization scheme, while Tables 3 and 4 depict those of a per-token asymmetric activation quantization scheme. In the revised version, we refrain from italicizing "per-tensor" and "static" to prevent any potential confusion.


[Q6.LRQisthefirsttoconfirmthepossibilityofW4A8quantization]**[Q6. LRQ is the first to confirm the possibility of W4A8 quantization]**

As highlighted by the reviewer, the W4A8 performance of FlexRound on common sense reasoning tasks and MMLU are not notably poor. However, our primary intent was to underscore that we conducted experiments specifically addressing W4A8 quantization for the first time. In the original FlexRound paper, there were no experimental results pertaining to W4A8 quantization. To enhance clarity, we revise that portion of our manuscript to state, "we for the first time conduct experiments on W4A8 quantization among LLM PTQ studies.".


[Q7.Noveltyoflowrankformulation]**[Q7. Novelty of low-rank formulation]**

The use of low-rank matrices in LRQ seems to be simple and straightforward, yet we believe that it is important to underscore the underlying motivation of LRQ as the reviewer mentioned. As SmoothQuant scales the weights collectively per channel, SmoothQuant may lead to non-negligible accuracy loss after quantization. Since FlexRound learns an individual scale for every weight with limited calibration samples, FlexRound might be prone to overfitting. However, increasing calibration samples in FlexRound is not a viable option, which leads us to employ low-rank weight-scaling matrices in order to decrease the number of learnable parameters effectively while maintaining the concept of scaling weights individually by sharing learnable parameters via low-rank structure. In this regard, the novelty of our paper lies not merely in the utilization of low-rank matrices but is firmly rooted in the problem statement concerning the number of scaling parameters and a middle-ground approach to solving this challenge.


[Q8.GapbetweenLRQandFlexRoundmayonlybesignificantinTable2andsomewhatsmallelsewhere]**[Q8. Gap between LRQ and FlexRound may only be significant in Table 2 and somewhat small elsewhere]**

In Table 1, LRQ exhibits a slightly superior zero-shot performance on common sense reasoning (CSR) compared to FlexRound, which we believe is noteworthy since FlexRound already achieves the zero-shot performance on CSR comparable to FP16 baselines. The close proximity in zero-shot performance between FlexRound and FP16 baselines on CSR limits the potential for a substantial performance disparity between FlexRound and LRQ. Despite LRQ approaching the zero-shot performance of FP16 baselines more closely than FlexRound, the difference in zero-shot performance between FlexRound and LRQ cannot be anticipated to be large after all.

Nevertheless, as expounded in Section 1, it is crucial to emphasize that LRQ demonstrates competitive performance relative to FP16 baselines on both common sense reasoning (CSR) and MMLU, a feat not accomplished by FlexRound that excels solely on CSR. Given the comprehensive evaluation of large language models (LLMs) across diverse benchmarks, the proficiency of LRQ in excelling across both CSR and MMLU holds significant implications in the field of LLM quantization.

Regarding the 88-bit weight quantization presented in Tables 3 and 4, the adoption of a per-token asymmetric activation quantization scheme results in even naive rounding-to-nearest (RTN) performing closely to the levels of FP16 baselines on both CSR and MMLU. As a result, while LRQ exhibits slightly higher accuracy compared to SmoothQuant and FlexRound for most Llama 2 models, it can be concluded that SmoothQuant, FlexRound, and LRQ are nearly evenly matched.

In the context of 44-bit weight quantization as presented in Table 3, FlexRound achieves zero-shot accuracy levels comparable to FP16 baselines on CSR, resulting in a relatively small zero-shot performance gap between FlexRound and LRQ, like the scenario depicted in Table 1. However, in the case of 44-bit weight quantization in Table 4, LRQ surpasses FlexRound by a margin ranging from 0.20.2 to 0.70.7 percent. Although, as noted by the reviewer, these increments in five-shot accuracy on MMLU in Table 4 may seem modest compared to those in Table 2, we believe that the rise in five-shot accuracy by 0.20.2 to 0.70.7 percent is significant. This is particularly noteworthy as it brings the five-shot accuracy gap between LRQ and FP16 baselines to less than 1.51.5 percent, while the corresponding gap between FlexRound and FP16 baselines remains more or less at two percent for Llama 2 13B and 70B. We hope that these W4A8 quantization results will contribute to narrowing the gap between W4A8 quantized LLMs and FP16 baselines to less than one percent.


评论

Dear Reviewer ok57,

We appreciate your insightful comments.


[Q1.StandarddeviationofFlexRoundandLRQ]**[Q1. Standard deviation of FlexRound and LRQ]**

As the reviewer suggested, we carry out three random trials for Llama 2 7B, Llama 33B, and Llama 2 70B as outlined in Table 1, presenting the average and standard deviation of them.

<Table A. Average and standard deviation of zero-shot performance of FlexRound and LRQ over three random trials on common sense reasoning tasks>
W/A/KVLlama 2 7BLlama 33 BLlama 2 70B
FlexRound8/8/859.72±0.7359.72 \pm 0.7362.83±0.3662.83 \pm 0.3665.65±0.3065.65 \pm 0.30
LRQ8/8/859.90±0.18\mathbf{59.90 \pm 0.18}63.81±0.16\mathbf{63.81 \pm 0.16}65.89±0.06\mathbf{65.89 \pm 0.06}

As seen in the table above, not only does the average of LRQ surpass that of FlexRound, but the standard deviation of LRQ is also smaller than that of FlexRound, which strengthens our assertion that FlexRound might be prone to overfitting when applied to the quantization of LLMs.


[Q2.Motivationabouttheimprovedgeneralization]**[Q2. Motivation about the improved generalization]**

As the reviewer mentioned, LRQ exhibits a slightly superior zero-shot performance on common sense reasoning (CSR) compared to FlexRound. However, we believe that this advantage is noteworthy since FlexRound already achieves the zero-shot performance on CSR comparable to FP16 baselines. The close proximity in zero-shot performance between FlexRound and FP16 baselines on CSR limits the potential for a substantial performance disparity between FlexRound and LRQ. Despite LRQ approaching the zero-shot performance of FP16 baselines more closely than FlexRound, the difference in zero-shot performance between FlexRound and LRQ cannot be anticipated to be large after all.

Nonetheless, we posit two notable points. First of all, as expounded in Section 1, LRQ demonstrates competitive performance relative to FP16 baselines on both common sense reasoning (CSR) and MMLU, a feat not accomplished by FlexRound that excels solely on CSR. Given the comprehensive evaluation of large language models (LLMs) across diverse benchmarks, the proficiency of LRQ in excelling across both CSR and MMLU holds significant implications in the field of LLM quantization. Moreover, notwithstanding LRQ's marginally superior zero-shot accuracy compared to FlexRound, the smaller variance of LRQ, as evidenced in Table A in the previous question, suggests that leveraging the parameter-efficiency of low-rank weight-scaling matrices can help variance reduction. Consequently, it can be inferred that LRQ holds the potential for superior generalization compared to FlexRound.


[Q3.NumberofcalibrationsamplesusedtogenerateFigure3]**[Q3. Number of calibration samples used to generate Figure 3]**

We employed 512512 calibration samples from the C4 dataset to perform the quantization of Llama 7B to W8A8, consistent with the experimental setup detailed in Section 3. However, in Figure 3, we depicted (a) a single sample chosen from the pool of 512512 calibration samples, and (b) an individual unseen sample from both common sense reasoning and MMLU. For visual representation across various samples, three figures were incorporated in Appendix E of the revision, each depicting the accumulated RMSE for three distinct samples.


[Q4.SensitivityofFigure3tothenumberofcalibrationsamples]**[Q4. Sensitivity of Figure 3 to the number of calibration samples]**

Thank you for your helpful suggestion. We incorporate a figure illustrating the sensitivity of accumulated RMSE to the number of calibration samples in Appendix F of the revised version. Figure 8 in the revision displays the comparison of accumulated RMSE for FlexRound and LRQ at the last Transformer block of Llama 77B with the number of calibration samples varying from 6464 to 512512. As depicted in Figure 8(a), the accumulated RMSE of the last Transformer block on a calibration sample diminishes with a reduction in the number of calibration samples. This phenomenon is because FlexRound and LRQ are more likely to be fitted to calibration samples as the number of calibration samples becomes smaller. Conversely, Figure 8(b) reveals that the accumulated RMSE of the last Transformer block on each unseen sample from common sense reasoning and MMLU decreases with a larger number of calibration samples.

Notably, the pattern elucidated in Section 2.4 persists consistently across varying calibration sample sizes from 6464 to 512512. In other words, for every calibration sample size spanning from 6464 to 512512, LRQ consistently attains nearly identical accumulated RMSE to FlexRound for a calibration sample from the C4 dataset. Concurrently, the accumulated RMSE of LRQ remains markedly smaller than that of FlexRound for an unseen sample from common sense reasoning and MMLU. This observation provides additional support for the insight presented in Figure 3, as discussed in Section 2.4.


评论

Dear Reviewers,

We sincerely appreciate your time and efforts in reviewing our paper. We kindly remind you of our responses to your insightful and constructive comments and faithfully reflected them in the revised manuscript. As the discussion period will end in about 9 hours, we would appreciate if you check our responses and the revision. If you have any remaining issues or concerns, please do not hesitate to bring them to our attention.

Warm regards,

Authors

AC 元评审

Three reviewers gave a score of 6 but with fairly low confidence and little enthusiasm. One reviewer gave a score of 3 with high confidence and a more detailed review. All reviewers noted a limited methodological novelty, marginal improvements over previous work, and overall the results are not that exciting. Unfortunately, the area of neural net compression has been very crowded for years and it is hard to make an impression. Unfortunately, this paper doesn't have enough support for acceptance.

为何不给更高分

See metareview

为何不给更低分

N/A

最终决定

Reject