PaperHub
6.8
/10
Poster5 位审稿人
最低3最高4标准差0.5
3
4
4
3
4
ICML 2025

SlimLLM: Accurate Structured Pruning for Large Language Models

OpenReviewPDF
提交: 2025-01-24更新: 2025-07-24

摘要

关键词
structure pruninglarge language modelsmachine learning

评审与讨论

审稿意见
3

This work involves compressing LLMs by width pruning of sublayers in transformer blocks, where MHA and FFN are treated differently. For MHA, head pruning is performed based on the similarity between outputs with and without specific heads. For FFN, channel pruning is carried out using a Wanda-based metric from LoRAP, which additionally considers feature direction along with feature magnitude. Performance recovery is achieved through a linear regression method, and a non-uniform pruning ratio is applied across different blocks. The experiments are exclusively conducted on the LLaMA family.

给作者的问题

Please refer to the sections on <Claims And Evidence> and <Experimental Designs Or Analyses>.

I like several aspects of this work and appreciate the authors’ efforts in compressing LLMs. However, several justifications and ablation studies appear to be missing, and the experimental validation seems weak compared to relevant papers.

论据与证据

This work is largely supported by empirical evidence, and the approach to pruning MHA and FFN differently seems interesting. Although additional hyperparameters are introduced, the type of performance recovery employed is novel to me.

However, some aspects of FFN pruning remain unclear, and heuristics are used to determine non-uniform pruning ratios. The paper introduces one method for MHA pruning (A), another for FFN pruning (B), and a third for performance recovery (C). While (A) and (C) appear promising and interesting, the rationale and effectiveness of (A), (B), and (C) independently are not well justified. Additionally, several necessary ablation studies are missing.

方法与评估标准

The proposed framework aligns well with previous studies on LLM pruning. Given the emphasis on LLM compression in this work, the use of evaluation metrics seems adequate.

理论论述

While the paper does not make theoretical claims, it does provide clear descriptions of the equations used in the method.

实验设计与分析

  • Measuring the impact of per-head removal as shown in Equation (4) is intriguing, yet its effectiveness is still in question. Regarding MHA pruning, is it more effective than gradient-based importance measures like those used in LLM-Pruner, AWSVD in LoRAP, or the fluctuation metric in FLAP, when all other conditions are identical? Although Tables 1 and 2 offer comparisons with several methods, the inclusion of other factors, such as FFN pruning and non-uniform ratios, makes it challenging to clearly evaluate the effectiveness of the proposed metric for MHA.

  • Furthermore, the computations in Algorithm 1 appear to be intensive due to its greedy search nature. How long do these computations typically take, and is it possible to provide comments on this aspect?

  • I think the pruning importance metric for the FFN is quite similar to LoRAP, except it employs a sigmoid function for the down projection. Why is the down projection designed in this particular way while the gate and up projection do not follow the same design? Is it impossible to apply Eqn (7) to the gate and up projection matrices, or is there another reason?

  • Linear regression for performance recovery seems effective; however, are A and B in Eqn (10) additional, newly introduced parameters? It appears that A and B are necessary for the inference process. If this is the case, it may not integrate well with other inference engines, such as vLLM and TensorRT-LLM, because it requires modifications to the forward code and involves saving additional parameters. This could limit the practicality of the work.

  • How did the authors set r_0​ and alpha in Eqn (11)? The statement, 'Consequently, we employ a stratified pruning strategy, wherein the pruning ratios are systematically lower in the shallower layers and incrementally higher in the deeper layers,' (line 241 of page 5) seems entirely heuristic and lacks a systematic foundation. Furthermore, the generalizability to other LLMs beyond the LLaMA family is not guaranteed.

  • For the results for Llama-2 in Table 2, I am not convinced that SlimLLM outperforms LoRAP. The numerical results, such as WikiText2 PPL and downstream accuracy, appear very similar, and in some cases, LoRAP seems to perform better.

  • The number and type of calibration samples have not been investigated.

  • The experimental validation in this study seems restricted since only models from the LLaMA family are used. To better ascertain the widespread applicability and superiority of this method, I suggest conducting experimental comparisons with additional models such as OPT, Qwen, Phi, and MoE-based architectures.

  • The tests were limited to models with only 7B parameters. Including models with at least 13B parameters, as done in previous studies, is essential to provide a comprehensive assessment.

  • The methodology used to calculate latency gains in Table 3 needs further elaboration, especially since width pruning often faces challenges in achieving actual speedups in setups like ZipLM and Shortened LLaMA. Could you clarify which framework was used (e.g., Vanilla HuggingFace, TensorRT-LLM, vLLM)? It's also important to know if the method can accelerate both the prefill and decoding stages, as the current results seem to focus solely on prefill with single-token generation. Is a batch size of 1 appropriate?

补充材料

I haven't checked the supplementary material.

与现有文献的关系

Reducing computational requirements for LLMs is widely discussed, and this work tackles it in a clear and effective manner.

遗漏的重要参考文献

其他优缺点

N/A

其他意见或建议

  • I am uncertain about the relevance of the quantization and low-rank approximation methods discussed in the 'Compression technologies' part of Sect 2. Related Work to this study, as the authors do not test the compatibility of these methods with the proposed pruning technique. Additionally, there appears to be no citation for the studies mentioned in this subsection.
  • Figure 3, which illustrates the impact of each layer on LLMs, presents a concept that is already familiar in this field, as evidenced by similar graphs in ShortGPT, SLEB, and Shortened LLaMA. This raises questions about the value this graph adds to the paper. Furthermore, relevant citations are missing in line 274 on page 5, where the paper states, 'Similar to many current methods, we...'.
作者回复

Q1: Proposed metric for MHA

A1: We design similarity-based approach to enhance linear fitting by maximizing output similarity. To verify the effectiveness of our similarity-based approach, we employed the fluctuation metric from FLAP to evaluate the importance of heads, while aligning other configurations. From the results, It can be observed that our similarity-based strategy achieves a lower PPL.

methodratiowikitext2\downarrowPTB\downarrowAvg.\uparrow
fluctuation metric50%46.3487.2544.44
similarity-basd metric50%37.8967.6850.10

Q2: Computations in Algorithm 1

A2: The computational cost of greedy search is acceptable. The time spent on greedy search is approximately 2 seconds for each layer on Llama-7B.

Q3: Importance metric for the FFN

A3: Yes, there is another reason. When pruning the input channels of the down projection, each channel's input activation only affects the magnitude of its output vector, while the direction is determined by the weights. For up and gate projection, their outputs correspond to the input activations of the down projection, and thus, we have not considered their impact on the direction of the final output matrix.

Q4: Linear regression

A4: A and B can be integrated into the output projection. Here, the weight matrix of the output projection is AWoA \cdot W_{o}, and the bias term is B+BoB + B_{o}.

Q5: r_0​ and alpha

A5: Under the pruning ratio rr r_0 is calculated according to the following formula: r0=rr_{0} = r * pruned_layers / total_layers

For alpha, we provide results for various values. Other models can use these results for adjustment.

alpharatiowikitext2PTBAvg.\uparrow
150%54.1796.5744.42
450%43.3683.9144.95
750%37.8967.6850.10
1050%44.8275.8046.86

Q6: Performance

A6: The experimental results show that the advantages of our model generally decrease after fine-tuning. This is mainly because the fine-tuning strategies of the two methods cannot be fully aligned. During compression, LORAP uses low-rank decomposition for MHA's linear layers, while LORA fine-tuning does not. LORAP reconstructs the decomposed matrices into a single matrix to keep MHA's parameter count unchanged. This increases the parameter count during LORA training.

Q7: Calibration samples

A7: We have added experiments to investigate the impact of the calibration set size on model performance.

alpharatiowikitext2PTBAvg.\uparrow
150%58.8129.9546.50
450%45.8999.6449.14
850%45.9899.2547.89
1650%44.380.0645.06
3250%37.8967.6850.10
6450%39.5672.0549.31

Q8: Experiments

A8: To demonstrate the generalizability of our strategy, we have included experimental results on the Vicuna-7B.

MethodsratioBoolQPIQAHellaSwagWinoGrandeARC-eARC-cOBQAAverage
Vicuna-7B0%75.6977.9171.0467.8068.9840.742.2063.47
SLimLLM w/o tune20%74.9276.1267.9865.8267.0939.3342.6061.98
SLimLLM w/ tune20%76.1576.3969.3264.7268.5639.2541.8062.31

Q9: 13B model

A9: Thank you for your suggestion. We have added experiments on LLaMA-13B.

MethodsratioBoolQPIQAHellaSwagWinoGrandeARC-eARC-cOBQAAverage
Llama-13B0%68.4778.8976.2470.0974.5844.5442.0064.97
LoRAP w/o tune20%73.9477.3174.9369.6970.7940.4441.4064.07
SlimLLM w/o tune20%74.1377.5374.7369.3070.4542.3241.0064.21

Q10: Latency

A10: The framwork is Vanilla HuggingFace. To test decoding latency, we set the batch size to 8 and measured the latency for generating 256 tokens. The results are as follows:

ratiolatency(s)
0%13.12
20%11.48
50%9.38

Others: Thank you for comments on writing, we will revise it carefully.

审稿人评论

Thank you for your detailed rebuttal and the additional experiments. I believe my initial score may have been somewhat harsh, as I was unsure about the competitiveness of your work. However, your thorough responses and the comments from other reviewers have alleviated my concerns.

I would like to increase my score from 1 to 3 for the following reasons:

  • During the re-read, similarity-based MHA head pruning and regression-based performance recovery feel interesting again. Also, the authors describe the greedy search cost for head pruning, which does not seem to take long (A2).
  • The comparison with FLAP appears to be impressive (A1), and the additional matrices for performance recovery (A4) appear to be integrable, ensuring compatibility with existing serving frameworks.
  • Furthermore, several ablations strengthen the experimental validation.

However, I am still concerned about the settings for r_0 and alpha (A5), i.e., whether these values are applicable to other models, the effectiveness over LoRAP for Llama-2-7B and Llama-1-13B, and a concern that the models examined seem a bit outdated. Releasing the code would be beneficial for verifying reproducibility and facilitating future work.

审稿意见
4

This paper proposes SlimLLM, a structured pruning method for LLMs that evaluates redundancy via Pearson similarity-driven head pruning and PCA-guided FFN channel pruning. A lightweight linear recalibration reduces post-pruning accuracy loss, while dynamic layer sparsity optimizes resource allocation. Experiments show strong performance on LLaMA models, surpassing prior methods.

给作者的问题

N/A

论据与证据

The claims made in the paper are well-supported.

方法与评估标准

The evaluation criteria and datasets are adequate for this task.

理论论述

N/A

实验设计与分析

The experimental designs and analyses appear to be solid. The proposed method is evaluated on various benchmarks and extensive ablation studies are provided.

补充材料

No additional supplementary materials.

与现有文献的关系

SlimLLM provides a novel and accurate method for pruning LLMs.

遗漏的重要参考文献

N/A

其他优缺点

Strengths

  1. Eigenvector-guided PCA preserves critical feature directions in FFN layers, surpassing magnitude-only criteria.
  2. The proposed method demonstrates stronger performance than prior methods like LoRAP and LLM-Pruner.
  3. The linear regression strategy shows great impact for performance recovery, as demonstrated in Table 5.

Weaknesses

1.This paper should be carefully proofread. For example, what is ohio_{-hi} in Figure 1? SpS_{-p} is not easy to understand in Algorithm 1 and should be replaced by better equation format.

2.What is exactly the latency reported in Table 3? Prefill latency or Decoding latency? More details should be included.

3.The proposed method is evaluated for LLMs. Is it possible that this method could also prune vision-language models?

其他意见或建议

For LLaMA2-7B, the proposed method obtains slightly poorer performance than prior methods on pruning ratio of 50% with finetuning. More analysis and discussions are encouraged.

作者回复

Q1: This paper should be carefully proofread. For example, what is OhiO_{-h_{i}} in Figure 1? SpS_{-p} is not easy to understand in Algorithm 1 and should be replaced by better equation format.

A1: Thank you for your suggestion. OhiO_{-h_{i}} denotes the output excluding the ithi-th head. SpS_{-p} represents the set of unpruned attention head. SpS_{-p}={headihead_{i}, ii \in unpruned head index}. The detailed descriptions will be added in the paper.

Q2: What is exactly the latency reported in Table 3? Prefill latency or Decoding latency? More details should be included.

A2: Thank you for your suggestion. We show the prefill latency in Table 3 and we will add details in paper. Additionally, we conducted experiments to measure the decoding latency. The table below shows the latency results for generating 256 tokens with a batch size of 8.

ratio#Paramlatency(s)
0%6.7B13.12
20%5.4B11.48
50%3.4B9.38

Q3: The proposed method is evaluated for LLMs. Is it possible that this method could also prune vision-language models?

A3: Yes, it is. For VLM (Visual-Language Model) models, the pruning strategy needs to be adapted, such as the proportion of visual tokens to text tokens. We are currently conducting preliminary experiments on Qwen-VL-7B, and the experimental results are shown in the table. We will present more details in our future work.

Datasetsprune ratioMMEMMBench_dev_en
Qwen-VL-7B0%2292.8877.9
Qwen-VL-7B w/o tune20%2047.69174.9
审稿人评论

The authors totally address my concern, I will raise my score to 4.

审稿意见
4

This paper proposes SlimLLM for pruning large language models (LLMs). The method uniquely combines Pearson correlation analysis for attention head redundancy detection with PCA-based directional importance for FFN channel pruning. A lightweight linear calibration technique minimizes post-pruning performance degradation, while layer-specific sparsity allocation leverages input-output alignment. Evaluations on LLaMA-7B and LLaMA2-7B demonstrate significant efficiency gains while maintaining competitive accuracy on commonsense reasoning tasks.

给作者的问题

How does the computational cost of greedy search scale with model size?

论据与证据

The authors claim SlimLLM achieves state-of-the-art structured pruning for LLMs. This is substantiated by Table 1, where SlimLLM outperforms LoRAP by 2.85% average accuracy at 50% pruning without fine-tuning, and Table 3, which highlights a 3.4× speedup on LLaMA-7B.

方法与评估标准

The framework is evaluated on various tasks (BoolQ, PIQA etc.) using LLaMA-family models. Latency measurements on NVIDIA V100 GPUs confirm practical applicability.

理论论述

None explicitly stated.

实验设计与分析

This paper provides extensive experimental results and ablation studies.

补充材料

NA.

与现有文献的关系

LLM pruning is a popular topic and this paper proposes a novel method that achieves strong performance.

遗漏的重要参考文献

NA.

其他优缺点

Strengths

  • PCA-based feature importance for FFN channels addresses a critical limitation of magnitude-only pruning, preserving directional information.
  • The proposed method achieves great improvement for running speed, making it viable for real-time applications.
  • The linear regression strategy is an interesting idea and recovers performance with negligible computational overhead. Weaknesses
  1. Computational complexity of iterative head pruning (Algorithm 1) is unaddressed, raising concerns for larger models.
  2. Linear regression is applied only to output matrices. Is it possible for extending it to intermediate layers for further reducing accuracy loss.

其他意见或建议

See weakness part.

作者回复

Q1: Computational complexity of iterative head pruning (Algorithm 1) is unaddressed, raising concerns for larger models.

A1: The computational complexity of Algorithm 1 is acceptable. First, the number of heads is generally small, which limits the number of iterations in the algorithm. Second, the outputs of each head can be precomputed and each iteration only involves the calculation of similarity, which reduces the number of redundant calculations during the iterations. For each layer, the time spent on greedy search is approximately 2 seconds when pruning Llama-7B.

Q2: Linear regression is applied only to output matrices. Is it possible for extending it to intermediate layers for further reducing accuracy loss.

A2: We compensate for the pruning error by performing linear fitting between the outputs before and after pruning. This requires that the outputs of the two stages maintain the same dimensionality. Currently, the output matrices of MHA and FFN meet this requirement. However, for other linear layers, since the output channels have been pruned, this method cannot be directly applied at present.

审稿意见
3

This paper proposes a structured pruning method for large language models (LLMs) that compresses both the feed-forward (FFN) and attention layers to accelerate inference. The pruning algorithm incorporates two key techniques: (1) removing redundant attention heads based on Pearson similarity and (2) pruning FFN layers using principal component analysis (PCA) on feature representations. After pruning, the method employs linear regression to efficiently recover the pruned weights by minimizing reconstruction error. The paper evaluates the approach on the LLaMA-7B model, assessing performance across benchmarks including WikiText, PTB, and BoolQ, among others. Experimental results demonstrate that SlimLLM outperforms baselines such as LLM-Pruner, LoRAPrune, and LoRAP.

update after rebuttal

Thanks for the response. I will keep my initial positive score.

给作者的问题

Please see the weaknesses.

论据与证据

Yes

方法与评估标准

Yes

理论论述

No theoretical results in this paper.

实验设计与分析

Yes.

补充材料

Supplementary materials not available.

与现有文献的关系

N/A

遗漏的重要参考文献

N/A

其他优缺点

Strengths:

  1. SlimLLM is a simple yet highly practical structured pruning method. It requires no additional training and can be easily applied to various large language models (LLMs), making it a scalable solution for real-world deployment.
  2. The proposed fast recovery method achieves performance levels comparable to low-cost fine-tuning. For example, at a 50% pruning rate, the LoRA-based LLM-Pruner attains a perplexity (PPL) of 38.12 on WikiText, while SlimLLM achieves a slightly lower PPL of 37.89, demonstrating its effectiveness in maintaining model accuracy despite significant compression.
  3. Unlike uniform pruning approaches, SlimLLM incorporates a non-uniform pruning ratio, dynamically adjusting layerwise pruning based on cosine similarity between input and output representations.

Weaknesses:

  1. Although the proposed sub-methods have been proven effective, the relation between FFN pruning and attention layers pruning is no so clear. Are they independent methods?
  2. The paper mainly focuses on LLaMA models. It is encouraged to conduct more experiments on other models.
  3. The performance gain appears to be not so significant at a 20% pruning ratio, as shown in Table 5. This suggests that SlimLLM might be more beneficial for higher compression levels but less impactful for moderate pruning. A deeper investigation into why performance varies across different pruning ratios would be valuable.
  4. The layer-wise pruning ratio relies on empirical hyperparameter α, lacking theoretical justification and more ablation studies.

其他意见或建议

N/A

作者回复

Q1: Although the proposed sub-methods have been proven effective, the relation between FFN pruning and attention layers pruning is no so clear. Are they independent methods?

A1: In this method, both the head pruning and channel pruning strategies are designed to increase the linear correlation between the outputs before and after pruning, which benefits our linear regression strategy. Meanwhile, considering the significant difference in the number of heads and channels, we employ the PCA method in the FFN to make the pruned output as close as possible to the principal direction. In the MHA, we directly maximize the Pearson similarity to bring the similarity closer.

Q2: The paper mainly focuses on LLaMA models. It is encouraged to conduct more experiments on other models.

A2: Thank you for your suggestion. We have added the experimental results on Vicuna-7B.

MethodsratioBoolQPIQAHellaSwagWinoGrandeARC-eARC-cOBQAAverage
Vicuna-7B0%75.6977.9171.0467.8068.9840.742.2063.47
LLMPruner w/o tune20%62.8775.4164.0058.4160.9837.1239.0056.83
LLMPruner w/ tune20%60.4075.6365.4563.2263.0537.7139.0057.78
LoRAP w/o tune20%76.4276.3868.3164.9665.8237.2938.6061.11
LoRAP w/ tune20%75.8176.7768.3965.0470.0839.3339.2062.09
SLimLLM w/o tune20%74.9276.1267.9865.8267.0939.3342.6061.98
SLimLLM w/ tune20%76.1576.3969.3264.7268.5639.2541.8062.31

Q3: The performance gain appears to be not so significant at a 20% pruning ratio, as shown in Table 5. This suggests that SlimLLM might be more beneficial for higher compression levels but less impactful for moderate pruning. A deeper investigation into why performance varies across different pruning ratios would be valuable.

A3: We posit that the impact on model performance is comparatively minimal at a pruning ratio of 20%, whereas the pruning error becomes significantly more pronounced at a pruning ratio of 50%. Our proposed strategy is capable of effectively mitigating this error, thereby yielding more substantial performance improvements.

Q4: The layer-wise pruning ratio relies on empirical hyperparameter α, lacking theoretical justification and more ablation studies.

A4: Thank you for your suggestion. We have added ablation experiments on Llama-7B, and the results are as follows. When the value of alpha increases to 7, the model achieves optimal performance. Further increment of alpha leads to a decline in model performance.

alpharatiowikitext2\downarrowPTB\downarrowAvg.\uparrow
150%54.1796.5744.42
450%43.3683.9144.95
750%37.8967.6850.10
1050%44.8275.8046.86
审稿意见
4

The paper presents SlimLLM, a structured pruning approach for large language models (LLMs) that tackles channel and attention head pruning through a unified importance evaluation framework. The paper introduces several novel techniques, including using Pearson similarity to identify redundant attention heads with a greedy search strategy, evaluating FFN channel importance via PCA to prioritize feature directions, and a layerwise pruning strategy that adjusts sparsity based on input-output cosine similarity. Evaluated on LLaMA-7B/2-7B, SlimLLM achieves better accuracy and lower perplexity than baselines like LoRAP.

给作者的问题

NA

论据与证据

The claims made in the paper are well-supported by extensive comparisons and ablation studies.

方法与评估标准

The proposed method is evaluated on various benchmarks for LLaMA-7B and LLaMA2-7B. Evaluation focuses on zero-shot accuracy across commonsense reasoning datasets (e.g., BoolQ, PIQA), perplexity (WikiText2, PTB), and inference latency.

理论论述

The paper does not provide too much theoretical analysis.

实验设计与分析

The experimental setup is reasonable, with all comparative experiments conducted under the same conditions. Additionally, extensive ablation studies are provided to demonstrate the effectiveness of each proposed module.

补充材料

NA.

与现有文献的关系

The paper focuses on LLM pruning, a critical problem in the field that has garnered significant attention from both the academic community and industry.

遗漏的重要参考文献

NA.

其他优缺点

Strengths

  1. SlimLLM introduces novel criteria for structured pruning, such as Pearson similarity-based head pruning and PCA-driven feature space importance for FFN channels. These methods move beyond element-wise aggregation, capturing interdependencies within sub-modules and providing a more comprehensive assessment of redundancy. The greedy search for head combinations further enhances this approach, addressing limitations of prior works like LoRAP that rely on weight magnitude or activation norms.

  2. The proposed linear regression strategy for output matrix fine-tuning is both simple and effective. It achieves significant performance recovery with negligible computational cost, avoiding complex retraining.

  3. The paper provides comprehensive comparative experiments and ablation studies to justify the effectiveness of the proposed method.

Weaknesses:

  1. The paper focuses on pruning-based acceleration of LLMs but lacks a comparison with quantization-based acceleration methods. Including such a comparison would better demonstrate the proposed approach's advantages of LLM compressing techniques.

  2. Table 2 only compares the proposed method with LoRAP. Expanding the evaluation to include more state-of-the-art pruning or compression baselines would strengthen the demonstration of the method's effectiveness and generality.

其他意见或建议

NA

作者回复

Q1: The paper focuses on pruning-based acceleration of LLMs but lacks a comparison with quantization-based acceleration methods. Including such a comparison would better demonstrate the proposed approach's advantages of LLM compressing techniques.

A1: Thank you for your suggestion. Quantization is also a highly effective method for LLM compression. Since the directions of pruning and quantization as compression strategies are different and they can yield cumulative benefits, we have not compared our method with quantization at present.

Q2: Table 2 only compares the proposed method with LoRAP. Expanding the evaluation to include more state-of-the-art pruning or compression baselines would strengthen the demonstration of the method's effectiveness and generality.

A2: Thank you for your suggestion. In Table 2, we compared our method with the LoRAP method, which has excellent compression performance. However, due to memory limitations, we were unable to conduct experiments on LLMPruner with Llama2-7B. Below are the relevant results reproduced from other works, which show that our pruning strategy can better preserve model performance.

MethodsratioBoolQPIQAHellaSwagWinoGrandeARC-eARC-cOBQAAverage
LLaMA2-7B0%71.0478.4072.9667.1769.3240.5340.0862.89
LoRAP20%69.2476.3969.1565.1161.9935.5838.6059.44
LLMPruner20%67.9577.5871.4364.0163.5138.0539.8060.33
SLimLLM(ours)20%69.7976.2868.8863.5465.7439.0839.8060.44
最终决定

This paper puts forward SlimLLM, a structured pruning approach for large language models (LLMs). It assesses redundancy through Pearson similarity-based head pruning and PCA-directed feed-forward network (FFN) channel pruning. A lightweight linear recalibration is used to decrease the accuracy loss after pruning, and dynamic layer sparsity helps optimize resource distribution. Experiments carried out on LLaMA models demonstrate that SlimLLM performs well and outperforms previous pruning methods.

All reviewers recommend acceptance of this work. All the questions are solved during the discussion phase.

Thus, the AC recommends acceptance of this work as poster.