PaperHub
4.8
/10
withdrawn4 位审稿人
最低3最高6标准差1.1
5
5
6
3
3.8
置信度
正确性2.5
贡献度2.5
表达2.3
ICLR 2025

SlimLLaVA: Automatic Pruning for Large Vision-language Models

OpenReviewPDF
提交: 2024-09-27更新: 2024-11-13
TL;DR

The first work on pruning large vision language models maintaining generalization ability

摘要

Multimodal large language models achieve satisfying performance in complex reasoning tasks, while still suffers from high model complexity in deployment especially for resource-limited devices. In this paper, we propose an automatic pruning method of large vision-language models for efficient multimodal reasoning. Conventional methods leverage the training data of the original model to select the proper pruning ratio for different network components, while they are infeasible for large vision-language models due to the unbearable search cost caused by web-scale training corpus. On the contrary, we only use a few samples to search the desired pruning policy by maximizing its generalization ability on the unknown training data despite of the model accuracy, so that the optimal accuracy-efficiency trade-off can be obtained for large vision-language models. Specifically, we formulate the generalization gap for the pruning policy based on the structural risk minimization principle. With the task performance and the generalization ability, we iteratively search for the optimal pruning policy in the given search space and optimize the vision projector to evolve the search space with higher upper bound of performance. We conduct extensive experiments on ScienceQA, Vizwiz, MM-vet and LLaVA-Bench datasets for the task of visual question answering. With only 64 samples for pruning policy search, our method achieves 83.05% accuracy on ScienceQA and $\times$1.47 speedup compared to the dense LLaVA-v1.5-7B model.
关键词
PruneLarge vision lanugage modelGeneralization

评审与讨论

审稿意见
5

The paper presents SlimLLaVA, a method for pruning large vision-language models for efficiency. SlimLLava proposes to only use a few samples to search the desired pruning policy by maximizing its generalization ability on the unknown training data. Vision projector is optimized by iteratively searching of the optimal pruning policy. Experiments on multimodal tasks like visual question answering show it maintains high accuracy while reducing model complexity, offering a 1.47x speedup over denser models on ScienceQA.

优点

  • SlimLLaVA provides an effective way to prune large vision-language models, using a few samples to assign the optimal pruning ratio for weight matrix. The pruned model achieves a good trade-off between efficiency and accuracy.
  • The proposed method can outperform existing methods on various datasets under different pruning ratio, which shows the effectiveness of SlimLLaVA.
  • The ablation studies and analysis are thorough, which provides insights in different aspects of the proposed method.

缺点

  • The authors mentioned many times in the paper that this pruning strategy is for efficient multimodal reasoning. However, from both the design space and evaluation datasets, it's not specific to reasoning. I.e., the design is not "optimized" for reasoning, and only ScienceQA in evaluation datasets are somewhat reasoning, and the other datasets are general recognition/perception benchmarks. This claim should be toned down.
  • Only four datasets were evaluated. More extensive evaluation on different types of datasets are needed.
  • The authors can also compare the pruning methods with some quantization methods. E.g., compare the accuracy under similar inference speed or vice versa.

问题

  • Do authors find any drawbacks/failure cases of SlimLLaVA compared to previous method e.g., SparseGPT?
  • Will the code be open-sourced?
审稿意见
5

This paper presents an automatic pruning method for large vision-language models (LVLMs), aimed at reducing model complexity to facilitate deployment on resource-constrained devices. Specifically, it innovatively explores the objective of LVLM pruning from the perspective of model generalization capability and introduces the structural risk minimization principle to the searching process for optimal pruning policy. Meanwhile, the search space is iteratively updated and a weighted fitness function is designed to enhance model accuracy. The proposed method surpasses a range of existing LLM pruning approaches, demonstrating superior speed-accuracy trade-offs.

优点

+The paper is clearly written, logically rigorous, and meets the general standards of academic writing. +The issues of generalization and efficiency in pruning are indeed critical topics for the effective deployment of large models. +The proposed method is straightforward and easy to implement. The simple introduction of a term similar to regularization to enhance the generalization capability of the pruning process is enlightening for me. +Extensive experimental results demonstrate the effectiveness and generalization capability of the proposed method.

缺点

-The experiments are somewhat insufficient. a) Ablation studies are only performed on generalization considerations and the evolution of the search space, lacking a detailed analysis of specific designs, such as the application of Eq. (7). Or if it is not your original contribution, a citation for the source is needed. b) Lack of discussion on certain hyperparameter settings, including η\eta in Eq.(4) and τ\tau, kk, and nn in Algorithm 1. A more detailed analysis will enhance the credibility of the method. -The analysis of the experimental results appears somewhat disorganized. The speed up rate is mentioned prior to the discussion of the related experimental setup, yet no corresponding figures or tables are referenced. Additionally, the reported speed up effect is confusing. SparseGPT achieves a 2.16x speedup at 60% sparsity, while the proposed method reaches such result at only 50% sparsity. An explanation for this discrepancy would be appreciated. -The writing could be improved, as some critical information is not sufficiently presented. a) There is a lack of theoretical analysis of the relationship between Rademacher complexity and matrix norms. In fact, simply citing previous work adds to the burden of reading and comprehension. b) Noting that the proposed method involves evolutionary algorithms to generate candidate pruning policies, the specific schemes employed are not adequately described.

问题

(1) Why is the evolutionary objective to maximize the fitness function, which comprises the loss function and matrix norm term, rather than minimizing it? (2) For the implementation details, the meanings of "choices of pruning ratio" and "candidate pruning ratio" are unclear. What is the relationship between the them? (3) As far as I know, the PTB dataset contains only texts. How is such data used for LVLM pruning? How is the loss function calculated? (4) I want to discuss the consistency between the insight and the implementation. According to the stated motivation, the introduction of the matrix norm term in Eq.(4) aims to maximize the generalization ability of the proxy data over the training corpus. Given that the parameters of the LLM are frozen, it follows that as the pruning rate increases, this term gradually decreases. Does this imply that the proxy data's generalization ability increases along with the pruning rate? If so, why does SlimLLAVA exhibit a higher degradation rate than SparseGPT at high pruning rates, as shown in Tab.3?

审稿意见
6

This paper proposed an evolution search based pruning method for vision language models. The method works by identifying that the projector in typically vision language models like LLaVA plays an important role in pruning, and uses this projector to define a fitness function. The experiments on LLaVA-SQA demonstrate the effectiveness of the proposed pruning methods.

优点

  1. The task of pruning for vision language models can have real-world applications.
  2. The proposed method considers the architecture of vision language models, and makes it more suitable for pruning LVLMs.

缺点

  1. The experiments are only conducted on a limited set of datasets and models. It seems that the only architecture used here is LLaVA?

问题

  1. How does the proposed method generalizes beyond the LLaVA-v1.5 architecture? for example, what about models like MiniGPT4 or LLaVA-Next?
  2. It seems that the similar techniques can also be used for models for audio-language models or even more modalities, including these could greatly enhance the paper.
审稿意见
3

The paper aims to search for the best pruning ratios for each layer by minimizing the model's loss and a norm regularization term, which the latter term is related to reducing the generalization errors. To further increase the capacity of the method, the paper's method updates the projector layers as well to minimize the model's loss. The method and baselines are evaluated on LLaVA and show the proposed method is better than SparseGPT and LLM-Pruner.

优点

  1. Connecting weights norm to the model's generalizability is interesting.

  2. With the dynamic sparsity, the method outperforms SparseGPT.

缺点

  1. The comparison to LLM-Pruner is not very fair. The proposed method is apparently for pruning the model in an unstructured manner (correct me if I am wrong), but LLM-Pruner is a structured pruning approach so the performance will not be comparable.

  2. The method part is not straightforward to follow. Please see "Questions" for more details.

  3. The background and related work for pruning on LVLM are lacking, and I think they should be discussed or cited. For example,

[1] ECoFLaP: Efficient Coarse-to-Fine Layer-Wise Pruning for Vision-Language Models

[2] Upop: Unified and progressive pruning for compressing vision-language transformers

[3] MoPE-CLIP: Structured Pruning for Efficient Vision-Language Models with Module-wise Pruning Error Metric

The ECoFLaP's idea is very related to this paper, so it would be great to include it in the experiments in the paper.

  1. In lines 249-250, "However, calculating the upper bound for all pruning policies is computationally infeasible" is repeated with the previous sentence.

问题

  1. In equation (6), how to know if the loss is upper bound or not?

  2. In lines 254-255 and line 4 in Algorithm 1, what are the "candidates" mean? How to accurately find out what candidates are among the top k?

  3. In Algorithm 1, does the pruning's performance or pruning policy keep improving through steps?

  4. In line 357, does the "Auto w/o Gen" only remove the regularization term on weight's norm and keep the rest of the components the same?

  5. How to optimize equation (6) to get the updated WpW_p? Is it done by gradient descent or evolution algorithm?

撤稿通知

I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.