PaperHub
3.8
/10
Rejected4 位审稿人
最低3最高6标准差1.3
6
3
3
3
4.5
置信度
正确性2.3
贡献度2.0
表达2.8
ICLR 2025

TensorGPT: Efficient Compression of Large Language Models based on Tensor-Train Decomposition

OpenReviewPDF
提交: 2024-09-26更新: 2025-02-05
TL;DR

Language model compression based on low-rank factorization for low-end devices.

摘要

The Small Language Models (SLMs, or on-device LMs) is a concept corresponding to the Large Language Model (LLM), which has significantly fewer parameters and is typically deployed on low-end devices, like mobile phones and single-board computers (e.g. Raspberry Pi). Unlike LLMs, which utilize the increasing model size for better generalization, SLMs are expected to adjust the exact deployment environment changes. Furthermore, most edge applications have battery life concerns, which have never been considered in the GPU servers for data centres. Targeting these two issues, this paper focuses on the token embedding compression for adaptivity and low energy requirements in edge applications. We propose a training-free model compression approach based on the Tensor-Train Decomposition (TTD), whereby each pre-trained token embedding vector is converted into a lower-dimensional Matrix Product State (MPS). We then comprehensively investigate the low-rank structures extracted by this approach, regarding the compression ratio, language task performance, latency and energy consumption on a typical low-end device (i.e. Raspberry Pi). Taking the sub-billion parameter versions of GPT-2/Cerebres-GPT and OPT as examples, the model compressed with our approach can achieve a comparable language task performance to the original model with around $2.0\times$ embedding layer compression, while the energy consumption of single query drops by half.
关键词
model compressionlow-rank factorizationtensor decomposition

评审与讨论

审稿意见
6

The paper presents TensorGPT, a novel approach for compressing LLMs through tensor-train decomposition of embedding layers. The key innovation is applying tensor-train decomposition to individual token embeddings without requiring additional training data or computation. The authors evaluate their method on GPT family models (GPT-2 and CerebrasGPT), demonstrating meaningful parameter reduction while maintaining or sometimes improving model performance. The work provides comprehensive evaluations on GPT-family models and demonstrates practical applicability on edge devices.

优点

  1. Novel training-free compression method specifically targeting embedding layers. Strong practical value for edge device deployment
  2. Comprehensive experiments across multiple tasks and model sizes. Solid theoretical foundation with clear mathematical derivations. Thorough analysis of compression vs. performance trade-offs

缺点

  1. Limited comparison with existing compression methods and baselines. The comparison would be more comprehensive by comparing with more baselines which are train-free and trained.
  2. Evaluation focused mainly on GPT-family models and mainly focus on small models.
  3. It would be a great to combine the proposed embedding compression method with other model compression methods to check compatibility.

问题

  1. Could this approach be extended to show performance of other model architectures beyond the GPT family?
  2. Could this approach compatible with other model compression methods?
  3. Have you investigated the effect on model robustness like in the multilingual setting using more diverse tokens?
评论

Thank you so much for your detailed comments and suggestions, our response is as follows.

   

Weakness 1. Limited comparison with existing compression methods and baselines.

The comparison would be more comprehensive by comparing with more baselines which are train-free and trained.

Thanks for the suggestion. However, we think training is too heavy for low-end devices, since memory usage can be three to four times as much as inference [21].

We have added SVD-based compression as our baseline in General Response G3.

   

Weakness 2. Evaluation is limited on small GPT models.

Evaluation focused mainly on GPT-family models and mainly focus on small models.

We have extended our experiments to OPT series models, please refer to General Response G3.

Regarding the small model size, please refer to General Response G1.

   

Weakness 3. Compability with other model compression approaches.

It would be a great to combine the proposed embedding compression method with other model compression methods to check compatibility.

Thanks for the suggestion; we are considering quantization and plan to take it as future work.

   

Question 1. Language models outside GPT family.

We have extended to OPT, please refer to General Response G3.

   

Question 2. Could this approach compatible with other model compression methods?

Yes, quantization is the easiest one, and it is orthogonal to our approach.

The combinations with weight tying mentioned by Reviewer JBLc, and pruning like that in SliceGPT[6] are a bit difficult, but we believe it can be solved by changing product/multiplication sequences.

   

Question 3. Have you investigated the effect on model robustness like in the multilingual setting using more diverse tokens?.

Thanks for the suggestion. At the moment, we suspect that the multilingual tasks are too complex for the small language models, as we discussed in Tab.xaFK.3, Weakness 3 with Reviewer xaFK. MGSM [19] in Tab.xaFK.3 is also a multilingual dataset, but all the investigated sub-billion models cannot perform well. We feel that it is better to evaluate robustness only when the language models perform well on the task.

However, we acknowledge that multilingual tasks are rather valuable for the robustness of the edge application, and plan to take this as our future work. Yes, quantization is the easiest one and is orthogonal to our approach.

   

Reference

[19] "Language models are multilingual chain-of-thought reasoners." arXiv preprint arXiv:2210.03057 (2022).

[21] Zhao, Jiawei, et al. "Galore: Memory-efficient llm training by gradient low-rank projection." arXiv preprint arXiv:2403.03507 (2024).

审稿意见
3

The paper presents "TensorGPT," a model compression method using Tensor-Train Decomposition (TTD) to reduce the parameter size of Large Language Models (LLMs), particularly in embedding layers, without additional training. This training-free approach preserves model performance while significantly reducing memory requirements, demonstrated on low-end devices like Raspberry Pi.

优点

  1. The approach does not require extra training, making it applicable for scenarios with limited resources or when extra data is unavailable.
  2. TensorGPT achieves substantial compression with low memory overhead, retaining performance across language modeling and sentiment analysis tasks.
  3. The paper includes experiments on low-end hardware (Raspberry Pi) and larger models, evaluating compression ratios, latency, and task-specific performance.

缺点

  1. The novelty is limited. Tensor-Train Decomposition is explored in several language model compression works. What are the differences between the proposed method and existing works?
  2. Lack important baselines. Compressing token embedding layers is studied in some papers and the paper does not compare with them. For example, [1] and baselines used in [1] should be set as baselines.
  3. The empirical evaluation is primarily on language modeling and sentiment classification, with potential limitations in representing other complex NLP tasks, such as math reasoning and multi-hop question answering.

[1] Wang, Haoyu, et al. "LightToken: A task and model-agnostic lightweight token embedding framework for pre-trained language models." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023.

问题

refer to weaknesses

评论

Thank you so much for your comments, our response is as follows.

   

Weakness 1. The novelty is limited.

Tensor-Train Decomposition is explored in several language model compression works. What are the differences between the proposed method and existing works?

As far as we know, [11,16,17] are the only works that use Tensor-Train Decomposition for the language model compression. All of them require extra training, which is unrealistic for low-end devices. Our training-free approach is suitable for meeting the requirements of adaptivity and low energy in edge applications, as we discussed in General Response G0.

We would highly appreciate it if the Reviewer were aware of such work and possibly gave more references for it, so we could discuss them in our paper.

   

Weakness 2. Lack important baselines.

Compressing token embedding layers is studied in some papers and the paper does not compare with them. For example, [1] and baselines used in [1] should be set as baselines.

Thanks for the reference. We have discussed the works about token embedding layer compression ([7,10-12, 14, 17]) in the General Response G2. Given that we only focus on small language models deployed on low-end devices, only [10] solves the same problem as ours and should be compared. However, [10] still requires training a meta-model, which should be fine-tuned when the deployment environment changes. So, we only take the SVD-based approach, the same as that in LightToken [7] (reference [1] in the Reviewer's original reviews), as our baseline. The experimental comparison of this part is in the General Response G3.

   

Weakness 3. Limited evaluation.

The empirical evaluation is primarily on language modelling and sentiment classification, with potential limitations in representing other complex NLP tasks, such as math reasoning and multi-hop question answering.

Thanks for pointing out. We report the results of zero-shot reasoning in General Response G3.

We suspect that math reasoning and multi-hop question answering, which are still hard problems for LLMs, are too complex for sub-billion language models on low-end devices.

We have evaluated some sub-billion language models on MGSM [19] for math reasoning and DROP [20] for multi-hop question answering, and it came out that all the models could not perform well. Our view is that there is no need to compress models for these two tasks on low-end devices.

   

Tab. xaFK.3 Small language model scores on math reasoning (MGSM) and multi-hop question answering (DROP)

Task (metric or filter)OPT-125MOPT-350MQwen2.5-0.5BCerebrasGPT-111MCerebrasGPT-256MCerebrasGPT-590Mdistilgpt2gpt2gpt2-medium
MGSM (flexible extract)0.630.935.770.230.70.570.30.60.73
MGSM (remove whitespace)000.23000000
DROP (EM)0.090.350.0400.070.0500.070.05
DROP (F1)2.262.830.9815.31.881.881.392.723.69

   

Reference

[7] "LightToken: A task and model-agnostic lightweight token embedding framework for pre-trained language models." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023.

[10] "Direction is what you need: improving word embedding compression in large language models." arXiv preprint arXiv:2106.08181 (2021).

[11] "Tensorized embedding layers for efficient model compression." arXiv preprint arXiv:1901.10787 (2019).

[17] "Efficient gpt model pre-training using tensor train matrix representation." arXiv preprint arXiv:2306.02697 (2023).

[19] "Language models are multilingual chain-of-thought reasoners." arXiv preprint arXiv:2210.03057 (2022).

[20] "DROP: A reading comprehension benchmark requiring discrete reasoning over paragraphs." arXiv preprint arXiv:1903.00161 (2019).

审稿意见
3

The authors applied tensor train to an edge device application, and found we can use that to compress some NLP models. It reads to me that the author is simply applying a pre-existing, well-known tensor decomposition method on a problem. And I didn't see any new knowledge developed or presented (I might be wrong, so hopefully the authors could point out clearly what's the real technical contribution other than yet another application).

优点

  • Working on a problem that's important.

缺点

  • The authors claim "- As far as we know, we are the first to compress LLMs with low-rank factorization, specifically", which is not true to me. The architecture is the same for most NLP models and thus pardon me I really don't know what's added. I think all the tensor-train stuff has been developed and applied before. It's not immediately obvious to me what's new conclusions or findings drawn from this paper.

问题

  • Pardon me that I really don't know what's the biggest contribution. Other than your claim of try things on LLM, what's the real technical contribution in your paper not discussed by previous works?
评论

Thanks for your comments, our response is as follows.

   

Weakness 1. Overclaim

The authors claim "- As far as we know, we are the first to compress LLMs with low-rank factorization, specifically", which is not true to me.

Thanks for pointing out. We acknowledge that it is confusing and misleading here. We wanted to convey "As far as we know, we are the first to compress Small Language Models (SLMs) [9] for low-end devices use cases, with low-rank factorization'', which has been updated in line 78-80 of the updated submission.

The architecture is the same for most NLP models and thus pardon me I really don't know what's added.

We guess the Reviewer wanted to express that our approach has no obvious contribution in algorithms, since model compression is to work on the models with existing architectures rather than creating a new architecture.

For a short answer, our approach is specifically designed for small language models on low-end devices, to meet the requirements of adaptivity and low energy in edge applications. We systematically analysed energy, latency, etc, which common LLM compression rarely considers.

For a detailed answer, please refer to the General Response G0, G2 and our answers to Weakness 4 raised by Reviewer JBLc.

   

Weakness 2. About Tensor-Train

I think all the tensor-train stuff has been developed and applied before.

General Response G1 includes all the relevant tensor-train stuff [11,16,17] we know. All [11,16,17] contain extra training, which is unsuitable for low-end devices. In the updated submission, this part has been emphasized in line 189-193.

We would highly appreciate it if the Reviewer were aware of such work and possibly gave more references for such work, so that we could discuss them in our paper.

   

Weakness 4. Conclusion

It's not immediately obvious to me what's new conclusions or findings drawn from this paper.

Conclusion: our approach is suitable for low-end devices in edge applications.

   

Question

Pardon me that I really don't know what's the biggest contribution. Other than your claim of try things on LLM, what's the real technical contribution in your paper not discussed by previous works?

As we mentioned in the response to Weakness 2, our contribution is the first work for low-end devices to design a low-rank model compression strategy to satisfy the adaptivity and low energy requirements in edge applications.

For the comparison between relevant works and ours, please refer to General Response G2.

   

Reference

[8] "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases", ICML 2024

[9] "Small Language Models: Survey, Measurements, and Insights." arXiv preprint arXiv:2409.15790 (2024).

[11] "Tensorized embedding layers for efficient model compression." arXiv preprint arXiv:1901.10787 (2019).

[17] "Efficient gpt model pre-training using tensor train matrix representation." arXiv preprint arXiv:2306.02697 (2023).

审稿意见
3

This paper proposes using tensor-train decomposition to compress the token embedding matrix, aiming to reduce model size and accelerate inference, particularly in edge device scenarios. The technique leverages a higher-order tensor approximation method, built upon singular value decomposition (SVD), to efficiently represent embeddings while maintaining performance. This tailored approach is well-suited for resource-constrained environments, promising benefits in storage and computational efficiency.

优点

The paper is well-structured and easy to follow, with a clear presentation of the methodology and its applications.

缺点

Limited Novelty: The method largely builds upon the existing TT_SVD approach, with Algorithm 1 in this paper replicating methods already established in prior literature [1]. The methodological advancements or generalizations beyond TT_SVD appear minimal based on the methodology section.

Unconvincing Experiments: The experiments are conducted on older and relatively small models, such as GPT-2 and models up to only 1.3B parameters. This limits the relevance of the results, as they don’t reflect performance on contemporary large language models (LLMs). Additionally, the evaluation setup lacks modern benchmarking practices; for instance, LLM harness [2] would provide a more standardized evaluation framework.

Unintuitive Rationale: The foundational rationale for embedding a matrix using a tensor is unclear. The method requires reshaping the matrix into a higher-order tensor before applying tensor decomposition, but the paper does not provide an intuitive explanation for why this approach is effective or reasonable. Also, the embedding matrix only occupies a small part of the model even for a mid size of model, and a common way to reduce parameter size is through weight tie, which directly reduce the parameter sizes of embedding into half. How does you method performs in this case?

Overstatement of Contribution: Some claims appear overstated. For instance, the authors state, "As far as we know, we are the first to compress LLMs with low-rank factorization, specifically for low-end devices." However, the cited reference [3] already demonstrates compression of LLMs using SVD for similar purposes. Furthermore, there are numerous existing works that apply low-rank decomposition for LLM compression, such as [4-6].

Limited Impact on Overall Model Size: The embedding matrix occupies only a small portion of the model's parameters, even in mid-sized models. A commonly used technique, weight tying, can directly halve the embedding parameter size, offering a straightforward compression approach. The paper does not address how the proposed method compares to weight tying or performs when weight tying is already applied, raising questions about its practical impact on overall model size reduction.

[1] V. Oseledets. Tensor-train decomposition. SIAM Journal on Scientific Computing, 33(5):2295–2317, 2011. doi: 10.1137/090752286.

[2] Gao, Leo, et al. "A framework for few-shot language model evaluation." Version v0. 0.1. Sept 10 (2021): 8-9.

[3] Yen-Chang Hsu, Ting Hua, Sungen Chang, Qian Lou, Yilin Shen, and Hongxia Jin. Language model compression with weighted low-rank factorization. In International Conference on Learning Representations, 2022.

[4] Yuan, Zhihang, et al. "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

[5] Lin, Chi-Heng, et al. "MoDeGPT: Modular Decomposition for Large Language Model Compression." arXiv preprint arXiv:2408.09632 (2024).

[6] Ashkboos, Saleh, et al. "Slicegpt: Compress large language models by deleting rows and columns." arXiv preprint arXiv:2401.15024 (2024).

问题

How does your method perform with weight tying? What is the percentage of overall compression for model sizes larger than 13B?

伦理问题详情

NA

评论

(1/2)

Thank you so much for the very detailed comments and references! Our response is as follows.

   

Weakness 1. Limited Novelty

... The methodological advancements or generalizations beyond TT_SVD appear minimal based on the methodology section.

Though we did not change the exact implementation of TT-SVD, we did adjust its working unit (vectors rather than the whole matrix) and workflow to cope with the unique issues (adaptivity and low energy, as stated in General Response G0) for low-end devices. We also have a comprehensive systematic analysis regarding latency, energy and flops, and the impacts (latency and accuracy) of the tensor orders in the experimental section.

Also, we do not believe a simple methodology necessarily means no novelty. "Attention is All You Need" and "LoRA: Low-rank Adaptation of Large Language Models" are two examples. Their methodologies are rather simple (as stated in their abstract or introduction), but both revolutionised the community.

   

Weakness 2. Unconvincing Experiments:

The experiments are conducted on older and relatively small models, such as GPT-2 and models up to only 1.3B parameters...

Please refer to General Response G1, we only focus on SLMs on low-end device.

... the evaluation setup lacks modern benchmarking practices...LLM harness [2] ...

Thank you so much for the helpful evaluation resources, our new experiment results are shown in General Response G3.

   

Weaknesses 3. Unintuitive Rationale:

The foundational rationale for embedding a matrix using a tensor is unclear. The method requires reshaping the matrix into a higher-order tensor before applying tensor decomposition, but the paper does not provide an intuitive explanation for why this approach is effective or reasonable.

Tensors can model implicit high-dimensional representations (as well as the interactions among orders) of the model weights. In this sense, tensors have better expressivity than matrices, which is also a good solution for small-size models to express complex functionality with limited parameter space.

The results in G3 have empirically proved this point, as tensor-based approaches have higher chances of retaining the language task performance than the matrix-based approach.

... the embedding matrix only occupies a small part of the model even for a mid size of model, and a common way to reduce parameter size is through weight tie, which directly reduce the parameter sizes of embedding into half. How does you method performs in this case?

Firstly, apart from reasoning and language modelling, classification is also an important task for edge applications, which we have investigated in the paper. Weight tying cannot be compiled with the classification layer.

Secondly, we acknowledge that weight tying is a common approach to reducing memory, but we do not reckon it is a necessary part of LMs. Especially for the adaptivity requirements in edge applications (Section 2.1 in the updated submission), there should be a different approach to compress the fully connected layers (i.e. further amplifying the signal from the transformer) rather than directly reusing the weights of the embedding layer.

   

Weakness 4. Overstatement of Contribution:

Some claims appear overstated. For instance, the authors state, "As far as we know, we are the first to compress LLMs with low-rank factorization, specifically for low-end devices."

Thanks for pointing out. We acknowledge it is confusing and misleading. We wanted to convey "As far as we know, we are the first to compress Small Language Models (SLMs) for low-end devices use cases, with low-rank factorization'', which has been updated in line 82-83 of the updated submission.

However, the cited reference [3] already demonstrates the compression of LLMs using SVD for similar purposes. Furthermore, there are numerous existing works that apply low-rank decomposition for LLM compression, such as [4-6].

Thanks for the references. [3-6] are discussed in General Response G2.

Furthermore, we want to clarify the inconsistency of the term "LLMs". The use of the term "LLM'' in the original submission is to follow the term used in MobileLLM [8]. To avoid confusion, we emphasized that we only focus on sub-billion parameter models in the abstract and the introduction of our original submission. We have updated the paper title in our new submission accordingly, TensorSLM: Sub-billion Parameter Language Model Compression for Low-end Devices based on Tensor-Train Decomposition. This clarifies our focus is not "LLMs" but Small Language Models (SLMs) [9].

评论

(2/2)

   

Weakness 5. Limited Impact on Overall Model Size:

The embedding matrix occupies only a small portion of the model's parameters, even in mid-sized models.

Please refer to the General Response G1.

A commonly used technique, weight tying, can directly halve the embedding parameter size, offering a straightforward compression approach. The paper does not address how the proposed method compares to weight tying or performs when weight tying is already applied, raising questions about its practical impact on overall model size reduction.

Please refer to the last response to Weakness 3.

   

Question 1. How does your method perform with weight tying?

Please refer to the last response to Weakness 3.

   

Question 2. What is the percentage of overall compression for model sizes larger than 13B?

Please refer to the General Response G1.

   

Reference

[1] "Tensor-train decomposition." SIAM Journal on Scientific Computing, 33(5):2295–2317, 2011. doi: 10.1137/090752286.

[2] "A framework for few-shot language model evaluation." Version v0. 0.1. Sept 10 (2021): 8-9.

[3] "Language model compression with weighted low-rank factorization." ICLR 2022.

[4] "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

[5] "MoDeGPT: Modular Decomposition for Large Language Model Compression." arXiv preprint arXiv:2408.09632 (2024).

[6] "Slicegpt: Compress large language models by deleting rows and columns." arXiv preprint arXiv:2401.15024 (2024).

[7] "LightToken: A task and model-agnostic lightweight token embedding framework for pre-trained language models." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023.

[8] "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases", ICML 2024

[9] "Small Language Models: Survey, Measurements, and Insights." arXiv preprint arXiv:2409.15790 (2024).

评论

Thank you for your response. However, the response did not fully address my concerns regarding two fundamental aspects:

Compression of small language models (SLMs) is less general than for large language models (LLMs): LLMs are more widely used, and their compression techniques can often be applied to smaller models, whereas the reverse is not always true. For instance, methods like LLM-Pruner [1] and layer pruning strategies [2] demonstrate fast and effective compression, even for large models.

Lack of comparisons with state-of-the-art (SOTA) methods: The additional comparisons with SVD are insufficient to justify the proposed method. SVD performs significantly worse compared to other SOTA approaches, making it an inadequate baseline.

In summary, while the authors have clarified the paper's applicability, I believe the method lacks sufficient generality and empirical support to justify its contribution for publication in this venue. Therefore, I will retain my score.

[1] Ma, Xinyin, Gongfan Fang, and Xinchao Wang. "LLM-Pruner: On the Structural Pruning of Large Language Models." Version v1. May 19 (2023). arXiv:2305.11627.

[2] Men, Xin, et al. "ShortGPT: Layers in Large Language Models are More Redundant Than You Expect." Version v1. March 6 (2024). arXiv:2403.03853.

评论

(3/3)

   

Tab.JBLc.3. Zero-shot performance of OPT-1.3B after compression.

OPT-1.3BParams %ARC-cARC-eBoolQHellaS.PIQAWinoG.Avg.
Original10022.3550.8859.3332.5566.4951.9347.26
SparseGPT 2:4-19.8844.8257.3432.9763.4955.4945.67
SVD96.1620.5644.8757.3428.0463.2851.3044.23
98.1422.0150.1361.8030.6966.7656.5147.98
99.7323.5553.9660.2836.3569.5957.7750.25
SliceGPT96.2924.1553.6646.9137.1867.4655.4147.46
97.8124.1555.3947.9539.0868.7256.7548.67
99.9123.7255.2248.1338.3468.4455.7248.26
Ours96.0421.0826.3554.0425.9153.8148.9338.35
97.7125.2652.8657.9838.7869.4858.4850.47
99.5923.3855.2251.6840.437159.4350.19

       

We also give the results of SparseGPT. Since SparseGPT only freezes the weights rather than "deletes" them, we do not list its parameter ratio. From Tab Tab.JBLc.2 and Tab.JBLc.3, we can observe that the different compression approaches have different superiorities for different zero-shot reasoning tasks and compression ratios. Even SVD-based sometimes outperforms the others.

These results further indicate that compression approaches of LLMs may not be easily migrated to SLMs, as we discussed in (1/3).

References

[6] "Slicegpt: Compress large language models by deleting rows and columns." arXiv preprint arXiv:2401.15024 (2024).

[22] Ma, Xinyin, Gongfan Fang, and Xinchao Wang. "LLM-Pruner: On the Structural Pruning of Large Language Models." Version v1. May 19 (2023). arXiv:2305.11627.

[23] Men, Xin, et al. "ShortGPT: Layers in Large Language Models are More Redundant Than You Expect." Version v1. March 6 (2024). arXiv:2403.03853.

[24] Frantar, Elias, and Dan Alistarh. "Sparsegpt: Massive language models can be accurately pruned in one-shot." International Conference on Machine Learning. PMLR, 2023.

评论

(1/3)

Thank you so much for your reply and new references for LLM-Pruner [22] and ShortGPT [23]. Our responses to your reply are as follows:

       

1. Compression of small language models (SLMs) is less general than for large language models (LLMs).

LLMs are more widely used, ...

We acknowledge that "LLMs are more widely used" at the moment, but we do think SLM compression is also important. SLMs are for on-device applications (e.g. mobile phones and Raspberry Pi), and are suitable for applications without stable networks, sufficient GPU resources or continuous power charging. The compression of SLMs directly impacts battery life and, therefore, the user experience.

... and their compression techniques can often be applied to smaller models, whereas the reverse is not always true.

There has yet to be a consensus that the compression approaches of LLMs can be easily migrated to SLMs. On the contrary, there exists empirical evidence that LLM compression approaches cannot maintain the accuracy of SLMs as they can for LLMs:

  1. In Tab.2 of ShortGPT [23], for the same model series (Llama2 and Baichuan2), 13B models maintain more model accuracy than 7B models with ShortGPT, as summarized as follows

Tab.JBLc.1. Average score degradation after compression with ShortGPT.

Δ\Delta Avg.7B13B
Llama2-6.54-4.61
Baichuan2-8.59-7.88
  1. As stated in the second last paragraph of the introduction of SparseGPT [24], "larger models are more compressible". In Fig.2 of [24], with the same compression settings, the models with fewer parameters have a more severe accuracy drop (larger slope in Fig.2).

  2. In line 418-420 and line 445-448 of our current submission, we addressed that our compression approach performs better on the larger-sized models, which can be observed from Fig. 3 (a-b,d-j,m) in our current submission.

For instance, methods like LLM-Pruner [1] and layer pruning strategies ShortGPT [2] demonstrate fast and effective compression, ...

Thanks for this information. In the referred LLM-Pruner (7B as their smallest tested models) and ShortGPT (2.8B as their smallest tested models) in detail, we did not find their results on sub-billion parameter models. As we stated in line 58 of our current submission, running an uncompressed Gemma-2B on Raspberry Pi leads to a system crash. Thus we only consider models around or less than 1B.

... , even for large models.

Our paper only focuses on SLMs running on low-end devices, which normally cannot hold LLMs. Thus we do not consider the "large models".

评论

(2/3)

   

2. Lack of comparisons with state-of-the-art (SOTA) methods.

The additional comparisons with SVD are insufficient to justify the proposed method. SVD performs significantly worse compared to other SOTA approaches, making it an inadequate baseline.

Thanks for this emphasis. To find an appropriate baseline, we investigated the references in our General Response, the newly referred LLM-Pruner[22], ShortGPT[23] and a commonly used baseline SparseGPT[24]. Among these references, only SliceGPT[6] is training-free and compresses the embedding layers. The comparisons are in Tab.JBLc.2 and Tab.JBLc.3. Given that SliceGPT also compresses other layers, we only listed the results of similar overall parameter ratios after the compression. The bold numbers are the best performance for each parameter ratio setting.

Tab.JBLc.2. Zero-shot performance of OPT-125M after compression.

OPT-125MParams %ARC-cARC-eBoolQHellaS.PIQAWinoG.Avg.
Original10023.3857.1157.7441.5371.7159.3550.91
SparseGPT 2:4-19.0337.1258.5927.7758.3251.742.09
SVD77.3620.0531.1437.8326.5956.3150.5937.09
85.5118.1734.2637.8326.9657.5652.8837.94
97.7418.7743.1447.3728.5163.0051.3042.02
SliceGPT77.1519.2035.1437.8627.3855.3351.9337.81
86.2019.1138.5537.9228.0458.0050.2038.64
99.1620.3941.4640.0028.8461.5950.2840.43
Ours78.1620.2228.6639.1426.1753.5449.8836.27
84.8321.2529.5540.1526.1954.9050.1237.03
99.7620.0538.6845.4128.8661.5349.8840.74
评论

General Response (1/6)

We sincerely appreciate the time, effort, and detailed comments from the Reviewers.

We first respond to the four common issues in the reviews and then respond separately to each Reviewer.
       

G0. What is the novelty and contribution of this work?

This paper focuses on compressing the Small Language Models (SLMs) [8,9] deployed on low-end devices (i.e. Raspberry Pi) in edge applications. The edge applications pose two requirements to our compressing approach, which are not common in LLM applications:

  • Adaptivity: the approach should dynamically adjust the model to the environmental changes (e.g. tokens registered or deregistered);
  • Low energy: the computation and memory operations should consider the energy consumption (i.e. for longer battery life).

Centred on these two issues (detailed discussion is in Section 2 of the updated submission), our approach based on Tensor-Train Decomposition (TTD) is specifically designed for compressing SLMs:

  1. Adaptivity: TTD works on embedding vector level, which allows the application to update the vocabulary without operating on the whole compressed embedding matrix;
  2. Low energy: As computation operations are "cheaper" than memory operations regarding energy consumption, we chose to "exchange" the memory with computation to save energy during the forwarding passes, with negligible extra latency.

As far as we know, none of the current LLM (at least for those designed for GPUs) compression work has these concerns, though these concerns are critical for low-end devices and edge applications. TT format has an expressive and flexible form, which makes it easier for us to analyze and satisfy the two requirements.

Apart from common metrics like compression ratio and language task performance, we also give the estimated energy costs (with a similar approach in [18]) of our approach and SVD-based approach. At a quick glance, the following is the estimated inference energy costs with their best language task performance are

   

Tab.G.0 Inference energy costs of an input text of 100 tokens. (% percentage of uncompressed model energy costs, the lower, the better)

ModelOPT-125MOPT-350MDistilGPT2GPT-2GPT2-MGPT2-LCerebrasGPT-111MCerebrasGPT-256MCerebrasGPT-590M
SVD84.44%43.35%70.28%70.28%54.01%44.24%84.44%81.80%75.34%
Ours59.21%52.73%66.84%59.21%52.74%43.23%61.51%51.71%50.45%

The details of this part can be found in line 137-186, line 281-300, line 515-522 of the updated submission.

       

G1. The investigated LLMs are too small.

We feel there might be some misunderstanding here, especially since our paper focuses on the sub-billion language models (as stated in MobileLLM [8]) running on low-end devices, and we had stated this in the abstract and introduction of our original submission.

We have therefore decided to change the title of our paper to avoid this ambiguity. The new title is "TensorSLM: Sub-billion Parameter Language Model Compression for Low-end Devices based on Tensor-Train Decomposition". We have also rephrased our paper to further emphasise the focus (mainly Section 1,2,3 in the updated submission).

评论

General Response (6/6)

   

Reference

[1] "Tensor-train decomposition". SIAM Journal on Scientific Computing, 33(5):2295–2317, 2011. doi: 10.1137/090752286.

[2] "A framework for few-shot language model evaluation." Version v0. 0.1. Sept 10 (2021): 8-9.

[3] "Language model compression with weighted low-rank factorization." ICLR 2022.

[4] "Asvd: Activation-aware singular value decomposition for compressing large language models." arXiv preprint arXiv:2312.05821 (2023).

[5] "MoDeGPT: Modular Decomposition for Large Language Model Compression." arXiv preprint arXiv:2408.09632 (2024).

[6] "Slicegpt: Compress large language models by deleting rows and columns." arXiv preprint arXiv:2401.15024 (2024).

[7] "LightToken: A task and model-agnostic lightweight token embedding framework for pre-trained language models." Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 2023.

[8] "MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases", ICML 2024

[9] "Small Language Models: Survey, Measurements, and Insights." arXiv preprint arXiv:2409.15790 (2024).

[10] "Direction is what you need: improving word embedding compression in large language models." arXiv preprint arXiv:2106.08181 (2021).

[11] "Tensorized embedding layers for efficient model compression." arXiv preprint arXiv:1901.10787 (2019).

[12] "Groupreduce: Block-wise low-rank approximation for neural language model shrinking." Advances in Neural Information Processing Systems 31 (2018).

[13] "Improved Residual Vector Quantization for High-dimensional Approximate Nearest Neighbor Search." arXiv preprint arXiv:1509.05195 (2015).

[14] "Learning k-way d-dimensional discrete codes for compact embedding representations." International Conference on Machine Learning. PMLR, 2018.

[15] "Monarch: Expressive structured matrices for efficient and accurate training." International Conference on Machine Learning. PMLR, 2022.

[16] "Compute Better Spent: Replacing Dense Layers with Structured Matrices." arXiv preprint arXiv:2406.06248 (2024).

[17] "Efficient gpt model pre-training using tensor train matrix representation." arXiv preprint arXiv:2306.02697 (2023).

[18] "Addition is all you need for energy-efficient language models." arXiv preprint arXiv:2410.00907 (2024).

评论

General Response (4/6)

Tab. G.3.3 OPT-1.3B

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0023.3857.1157.7441.5371.7134.4959.3549.33
SVD (matrices)0.0521.9326.4338.7525.6653.7533.3249.835.66
25.4620.3134.8540.8926.4657.0734.2950.9137.83
50.8720.5644.8757.3428.0463.2835.5251.342.99
76.2822.0150.1361.830.6966.7634.6556.5146.08
96.6123.5553.9660.2836.3569.5934.757.7748.03
Ours (vectors)1.0721.3325.3842.6925.3953.3233.9850.2836.05
24.2221.1625.9360.4325.8854.934.5450.3639.03
49.4121.0826.3554.0425.9153.8134.5448.9337.81
70.7025.2652.8657.9838.7869.4835.3158.4848.31
94.7323.3855.2251.6840.437135.3159.4348.06




Tab. G.3.4 CerebrasGPT-111M

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0016.6437.8862.1426.7659.4133.8849.0140.82
SVD (matrices)0.1320.926.5237.8625.4652.4533.5748.9335.10
26.5717.4931.4437.7726.5356.233.5750.9936.28
53.0117.0635.2744.1626.5756.7534.0850.2837.74
79.4516.5537.0859.8826.7658.8133.8849.7240.38
92.6715.4437.9261.7726.8459.1933.6249.1740.56
Ours (vectors)2.4719.9728.2461.9326.0954.1334.5450.3639.32
29.1720.4829.8459.8526.2655.6634.750.0439.55
50.7819.831.4849.1126.7857.5133.4249.0938.17
71.8817.9234.5158.3226.7458.0534.5450.2840.05
87.1120.9924.0761.0425.6652.6733.9849.4938.27




Tab. G.3.5 CerebrasGPT-256M

256Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0016.8940.9561.527.4461.3734.2451.340.82
SVD (matrices)0.0921.1626.7337.8325.7452.2932.851.2235.40
28.2617.753338.226.4858.0533.9850.4336.84
47.0517.1535.6139.9726.8359.0334.1951.7837.79
75.2218.0939.4461.0127.460.8334.0351.6241.77
94.0018.1740.7459.9427.461.0433.6750.9141.70
Ours (vectors)2.6720.2224.6237.7425.554.5734.1950.7535.37
38.6020.927.5737.8325.853.4833.4749.9635.57
50.3718.9431.5252.3526.7656.8633.9350.6738.72
61.4020.0535.0259.9127.357.9433.8850.5940.67
98.9019.2840.7461.527.3961.4333.9852.0142.33
评论

General Response (5/6)

Tab. G.3.6 CerebrasGPT-590M

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0019.0346.4259.1729.1262.7335.3149.843.08
SVD (matrices)0.0721.3327.137.9225.7952.4534.0347.9935.23
26.9118.0935.9837.8326.825834.1450.5937.35
47.0317.2439.3137.8327.5659.7434.7550.8338.18
73.8718.8644.1149.4528.4261.6434.850.7541.15
94.0019.846.1752.7228.9762.1935.6249.8842.19
Ours (vectors)1.3723.3824.8756.6125.5952.6733.7352.1738.43
19.2119.9726.8149.8225.6652.7234.2449.1736.91
46.8819.5435.940.8927.0457.8334.5449.7237.92
66.4120.3138.0958.8728.2160.2834.2949.8841.42
94.3422.144.756.4229.0261.6435.5249.4942.70




Tab. G.3.7 CerebrasGPT-1.3B

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0022.3550.8859.3332.5566.4934.4451.9345.42
SVD (matrices)0.0521.3326.7340.0625.6552.3933.3248.8635.48
25.461838.5937.8327.3859.334.9551.8538.27
50.8719.7145.0850.4929.1362.3534.2952.2541.90
76.2819.9749.0355.630.764.8534.4949.9643.51
96.6120.9950.5153.8232.0665.6134.5450.4343.99
Ours (vectors)1.0722.5327.9540.0325.8654.3534.2450.7536.53
24.2221.4227.3639.5425.7853.3733.7850.5935.98
57.0322.6139.5653.5230.4463.0633.6247.7541.51
70.7022.6143.661.3531.465.0734.9550.0444.15
94.7322.3546.9756.3632.2165.6133.9851.7844.18
评论

General Response (3/6)

   

G3. Evaluation of more complex language tasks and LMs outside the GPT family is required.

We have extended our experiments with OPT-{125M, 350M, 1.3B} on zero-shot reasoning tasks with APIs in [1], since OPT performs well on the zero-shot reasoning tasks (as shown in Fig.1 of the updated submission). The experimental results are as follows. The bold numbers indicate the top-3 best performance cases. We also evaluated CerebrasGPT on these tasks, which is available in our updated submission.

We can observe from Tab. G.3.1 - G.3.7 that our approaches have a higher chance of maintaining the language task performance (especially in the average scores).

Tab. G.3.1 OPT-125M

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0023.3857.1157.7441.5371.7134.4959.3549.33
SVD (matrices)0.1321.525.8437.8325.8152.532.9149.0135.06
26.5720.0531.1437.8326.5956.3134.0350.5936.65
53.0118.1734.2637.8326.9657.5633.7852.8837.35
79.4518.7739.945.6327.3859.934.3450.8339.54
92.6718.7743.1447.3728.516334.1451.340.89
Ours (vectors)2.4721.3326.3937.8325.6352.9433.9850.5935.53
29.1720.2228.6639.1426.1753.5433.8349.8835.92
50.7821.2529.5540.1526.1954.933.3750.1236.50
71.8819.3735.3147.0927.7859.5833.3750.5939.01
87.1119.0339.659.5128.4161.1533.7851.1441.80




Tab. G.3.2 OPT-350M

Param (%)ARC-cARC-eBoolQHellaS.PIQASIQAWinoG.Avg.
Original100.0020.8244.1957.6832.0364.6432.9652.0943.49
SVD (matrices)0.2021.525.2537.8325.6751.3632.3249.1734.87
19.9320.8225.9338.5325.9453.9232.1151.0735.62
39.6620.2225.838.6226.2253.2632.1650.5935.41
59.3919.225.5538.526.5453.9733.0351.4635.61
79.1219.227.5337.8327.1655.9332.6249.3335.80
98.8520.7341.3737.8930.4762.9532.7349.2539.48
Ours (vectors)3.5221.0824.9245.4725.6653.133.751.1436.58
18.7520.3926.0162.1725.953.1632.1648.738.50
28.1320.0524.8762.1726.0853.5433.1449.3338.60
42.1920.0525.2548.4426.1853.732.3249.0936.58
70.3120.4825.4262.172653.1632.2751.6238.87
94.5321.4236.1545.929.5961.9233.1452.3340.21
评论

General Response (2/6)

   

G2. Lack of the comparison of relevant works.

We did address our differences with the relevant works in line 42 - 51 in our original submission. Along with the referred works addressed by the Reviewers, the comparison is as following:

Tab.G.2 Study on LM compression or relevant low-rank factorization

high-end devicelow-end deviceTraining required?MatrixTensorEmbedding layerLinear layerLLMsSLMs
GroupReduce [12]\surd\surd\surd\surd\surd
[11]\surd\surd\surd\surd\surd
[17]\surd\surd\surd\surd\surd
LightToken [7]\surd\surd\surd\surd\surd
DSVD [10]\surd\surd\surd\surd\surd
iRVQ[13]--\surd\surd----
DCQ [14]\surd\surd\surd\surd\surd
ASVD[4]\surd\surd\surd\surd
[3]\surd\surd\surd\surd\surd
ModeGPT[5]\surd\surd\surd\surd
Monarch[15]\surd\surd\surd\surd
[16]\surd\surd\surd\surd\surd
MobileLLM[8]\surd\surd----\surd
Ours\surd\surd\surd\surd

It should be noticed that though SliceGPT [6] also works with matrices, it performs a kind of pruning that exploits the sparsity rather than low-rank. Thus, it is outside the scope of low-rank factorization and the relevant works.

From Tab.G.2 we can observe that none of the relevant works has the same focus as ours. Though [10-12, 15-16] do not require fine-tuning the compressed model, they need to train a meta-model to get the compressed weights. If the input token distribution changes, the meta-models still require fine-tuning. Thus, for edge applications in this paper, the meta-learning based approaches [10-12, 15-16] are outside our scope of "training-free".

AC 元评审

This paper proposes TensorGPT, a method for compressing small language models (SLMs) using tensor-train decomposition (TTD) of the token embedding layer. The authors claim this approach is training-free and suitable for deploying SLMs on low-end devices. The method is evaluated on GPT-2, OPT, and CerebrasGPT models with up to 1.3B parameters.

The paper addresses the important issue of SLM compression for edge devices and provides experiments on low-end hardware like Raspberry Pi. The proposed training-free compression method and consideration of energy efficiency are relevant for resource-constrained scenarios.

However, the paper's novelty is limited, as tensor-train decomposition has been used before in language model compression. The lack of comparisons to state-of-the-art baselines and the focus on relatively small models limit the generalizability and impact of the work. Additionally, the initial claims about being first to compress LLMs with low-rank factorization were overstated.

The primary reasons for rejection are the paper's failure to demonstrate significant technical novelty or empirical contributions beyond applying existing techniques to a specific use case. While the work addresses an important problem, it falls short of the level of innovation and impact expected for publication at ICLR.

审稿人讨论附加意见

During the discussion, reviewers raised concerns about limited novelty, lack of comparisons to state-of-the-art baselines, focus on small models, and overstatement of claims. The authors responded by clarifying their focus on SLMs for low-end devices, adding comparisons to SVD and SliceGPT baselines, expanding experiments to include OPT models and more reasoning tasks, and emphasizing adaptivity and low-energy requirements.

Although these responses addressed some concerns, they did not fully alleviate the core issues. The expanded experiments and clarifications, while noted, did not significantly change the overall contribution. The focus on sub-1B parameter models still limits the broader applicability of the method, and the added baselines did not include some of the most recent and competitive approaches in the field.

In the final decision, the limited technical novelty and lack of substantial advances over existing compression techniques for SLMs were the primary factors for rejection. Despite addressing an important problem, the work does not meet the innovation and impact standards expected for ICLR publication.

最终决定

Reject