PaperHub
3.0
/10
withdrawn3 位审稿人
最低3最高3标准差0.0
3
3
3
4.3
置信度
正确性2.3
贡献度2.0
表达1.7
ICLR 2025

Dynamic Self-Distillation via Previous Mini-batches for Fine-tuning Small Language Models

OpenReviewPDF
提交: 2024-09-27更新: 2024-11-20
TL;DR

A model-agnostic and task-agnostic self-distillation method via the previous mini-batch's information.

摘要

关键词
Knowledge DistillationSelf-DistillationPre-trained Language ModelsLarge Language ModelsSmall Language ModelsNLPFine-tuning

评审与讨论

审稿意见
3

The work explores techniques of self-distillation in which logits from the latest mini-batch are exploited to supervise (regularize) the logits of the current batch. Empirical studies are conducted with both encoder-only and decoder-only LMs on NLU and NLG tasks.

Here are my concerns:

  1. Limited technical contributions. The proposed method bears significant similarity with the existing work [Shen et al., Self-Distillation from the Last Mini-Batch for Consistency Regularization. In CVPR'22]. The only difference lies in dynamic adaption of the hyper-parameters. While the authors have emphasized the importance of dynamic hyper-parameter and empirical results seem verify the claim, I still think the innovations are limited given the existing work. Moreover, given the existing CVPR paper, it is not appropriate to claim that "To the best of our knowledge, we are the first to propose a SelfD method called DynSDPB....". Perhaps, you are the first to apply the technique in the field of NLP.
  2. Poor organization. Related Work is an important part of an academic paper. Normally, we cannot leave related work to appendix. I also suggest the authors conduct proof-reading, as I have found many typos like the repetition of inspiration in Line 62.
  3. More analysis are needed. It is better to show how the model gradually get improved in the process of self-distillation. It is unconvincing to just show the final performance and claim that the proposed method is effective.

优点

  1. Good performance on a wide range of tasks.

缺点

  1. Limited technical innovations.
  2. Poor presentation.

问题

N/A

审稿意见
3

Knowledge distillation is a widely used method to reduce the computational cost of large language models (LLMs). However, the standard KD usually require an additional complex teacher model. Therefore, this paper focus on how to improve self-distillation in LLMs. Specifically, considering that the original self-distillation method is static, this paper introduces a dynamic SelfD method (DynSDPB) to address prediction inaccuracies during the early iterations. It dynamically adjusts the distillation influence and temperature value to enhance the adaptability of fine-tuning. Moreover, a vocabulary map matching is proposed to address the output inconsistency of the proposed method. Experimental results demonstrate the effectiveness of the proposed method.

优点

Strengths

  1. This paper introduces an adaptive method to dynamically adjust the influence based on the current states.
  2. The proposed method can be a plug-in technique to integrated into existing self-training method.
  3. Experimental results demonstrate the effectiveness of the proposed method.

缺点

Weaknesses

  1. The main idea of this paper is just an extension of the original SelfD, which makes it contributions as incremental.
  2. It seems authors only consider classification cases in NLU tasks. As authors claim their method as a task-agnostic method, so can you discuss how to adapt this method into regression tasks in NLU?
  3. Authors claim that at the early stage, model prediction could be incorrect and affect self-distillation training. So why not conduct self-distillation when the training / model prediction is stable? Can you compare your method with a baseline that applies self-distillation after some initial training and discuss the trade-off between these approaches?
  4. Why we need Output Mismatch Alignment? For decoder-only tasks, if the inputs are same, we only need to consider its logits and why we need to consider output mismatch problem during the fine-tuning? Could you provide some examples or explanations of when output mismatch occurs in practice during fine-tuning of decoder-only models.
  5. The results from some baselines are mismatched from the original paper (e.g., RoBERTa and DeBERTa). Please see my comments in questions. Could you explain these discrepancies or the experimental setups that cause these differences.

Minor issues

  1. Line 62: ''taking inspiration inspiration from DLB'' \rightarrow ''taking inspiration from DLB''.

问题

  1. How to apply the proposed method to regression tasks in NLU tasks (e.g., STS-B).
  2. Some baselines seem too weak. For example, the original baselines in RoBERTa [a] and DeBERTa [b]:
MNLIQNLIQQPRTESSTMRPCCoLA
RoBERTabase_{base}87.692.891.978.794.890.263.6
DeBERTalarge_{large}91.195.392.388.396.891.970.5
DeBERTaV3large_{large}91.896.093.092.796.992.295.3

It seems the baseline in this paper like (RTE, CoLA, SST, MRPC) are weaker than the original baseline. Please check these results.

[a] RoBERTa: A Robustly Optimized BERT Pretraining Approach.

[b] DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing.

审稿意见
3

This paper propose a model-agnostic method soft-label based self-distillation method in NLP. Since in NLP, there seems not much recent work about self-distillation, this work reference DLB (in CV) which utilize previous mini-batch for self-distillation and further adjust two hyper-parameter dynamically (temparature and CE/KL scale). Experiments on many NLU and NLG tasks shows improved performance over standard finetune and knowledge distillation-related methods.

优点

  1. Self-distillation in NLP remains under-explored. Recent soft-label-based knowledge distillation do require a large teacher model, in which either online KD or storing the probability need additional cost. This paper show Self-distillation would be a promising way to imporve the performance while remain low cost.

缺点

  1. Novelty is limited as an incremental method paper. Self-distillation using information from previous batch is an idea from DLB[1]. This paper only add a dynamic strategy about changing temparature and scale during distillation.
  2. This paper argue that DLB using fixed τ and α, which is not good. However, there lacks analysis on how dynamically change such hyper-parameters brings benefit (i.e. how this two parameters changed, and why this improve the performance intuitively)
  3. For decoder-only models, there is a lack comparison to recent soft-label-based distillation methods.
  4. For fine-tuning, it would be better to do full finetuning instead of Lora tuning. Lora tuning do not have comparable results to fine-tuning, and lora requires fewer compuation cost than your method I think, so your method would have an advantage over lora tuning which is expected. If the training resources does not support on large model, I woudl suggest to do full fine-tuning on small models such as 1B/3B.

[1]Self-distillation from the last mini-batch for consistency regularization. Yiqing Shen, Liwu Xu, Yuzhe Yang, Yaqian Li, and Yandong Guo.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11943–11952, 2022.

问题

  1. line 62: duplicated word "inspiration"
  2. It would be better to add an overall benchmark score for each method so that the overall improvement can be easily checked.
撤稿通知

I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.