PaperHub
5.0
/10
withdrawn4 位审稿人
最低3最高6标准差1.2
6
3
5
6
3.8
置信度
正确性3.0
贡献度2.3
表达2.5
ICLR 2025

Mitigating Catastrophic Forgetting in Large Language Models with Forgetting-aware Pruning

OpenReviewPDF
提交: 2024-09-13更新: 2025-01-24
TL;DR

We propose Forgetting-Aware Pruning Metric, a novel pruning-based approach to balance Catastrophic Forgetting and downstream task performance for LLMs.

摘要

关键词
Large Language ModelsCatastrophic ForgettingNeural Network Pruning

评审与讨论

审稿意见
6

This paper focuses on addressing the issue of catastrophic forgetting (CF) in large language models (LLMs) during fine-tuning, i.e., striking a balance between speciality and generality. It proposes Forgetting-Aware Pruning Metric (FAPM): A novel method that integrates the ratio of task vector magnitude to pre-trained model parameters into the pruning criteria, allowing for the mitigation of CF without modifying the model architecture or training process. FAPM successfully mitigates forgetting while retaining 99% of downstream task accuracy across various natural language tasks. The experiments show that FAPM performs competitively against state-of-the-art methods.

优点

1.This paper introduces two key concepts through preliminary experiments—“absolute change magnitude” and “relative change magnitude”—which are designed to balance the trade-off between task specialization and generality.

2.Building on these concepts, the paper proposes the Forgetting-Aware Pruning Metric (FAPM), aimed at deriving an optimal task vector to obtain the final model weights.

3.Extensive experiments demonstrate the effectiveness of FAPM, showing significant improvements over baseline methods.

缺点

1.This paper overlooks an important baseline—Wise-FT [1], a standard model parameter merging method.

2.The experimental setup for the preliminary experiments, particularly those in Sections 2 and 3, is not clearly detailed, e.g., the experimental model, raising concerns about whether similar observations would hold for other models.

3.While instruction-following is a critical aspect of generality, this paper focuses solely on world knowledge and generic reasoning, omitting this aspect.

4.The rationale and motivation for striking a balance between speciality and generality is not clearly articulated in the paper.

[1] Mitchell Wortsman, Gabriel Ilharco, Mike Li, Jong Wook Kim, Hannaneh Hajishirzi, Ali Farhadi, Hongseok Namkoong, and Ludwig Schmidt. 2021. Robust fine-tuning of zero-shot models. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7949–7961.

问题

1.The limitation mentioned in lines 81-83, stating that "methods that alter the training process or model architecture not only make the training process more difficult to control but also degrade the accuracy of downstream tasks," lacks references to prior work supporting this claim.

2.The figures in this paper appear not to be in vector graphic format, which may affect the clarity and quality of the visualizations.

3.The quotation marks in line 212 are not properly formatted according to standard usage.

评论

Thank you for Reviewer YFJm's review.

Q1: This paper overlooks an important baseline—Wise-FT [1], a standard model parameter merging method.

We appreciate the reviewer for providing an important baseline. We have added a comparison between Wise-FT and our method. The original Wise-FT paper mentions: "WiSE-FT using a fixed mixing coefficient α=0.5 with the fixed optimal mixing coefficient." Therefore, we adhere to the hyperparameter settings as described in the original paper. The experimental results can be seen in the following two tables.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
RTEPre-trained model0.43860.79220.65940.59140.62040.819
RTEFull SFT0.23110.0750.25540.00.14030.890
RTEWiSE-FT0.30460.44200.42550.56090.43320.889
RTEFAPM(Ours)0.46230.79150.64540.59750.62420.897
WikiQAPre-trained model0.43860.79220.65940.59140.62040.913
WikiQAFull SFT0.25470.00.24220.00.12420.966
WikiQAWiSE-FT0.25810.00.24580.00.12590.958
WikiQAFAPM(Ours)0.47490.79750.65630.58530.62850.964
WinograndePre-trained model0.43860.79220.65940.59140.62040.519
WinograndeFull SFT0.27920.06060.34380.00.17090.820
WinograndeWiSE-FT0.47410.44120.59670.50600.50450.830
WinograndeFAPM(Ours)0.48290.76800.64720.57310.61780.824
SQuADPre-trained model0.43860.79220.65940.59140.62040.371
SQuADFull SFT0.28060.02120.32060.00.15560.646
SQuADWiSE-FT0.43090.40090.54840.51020.47250.639
SQuADFAPM(Ours)0.47380.73100.64550.57480.60630.637

Table 1: Comparison of FAPM and WiSE-FT on the Llama3-8B model. The 'Avg.' column represents the performance on general tasks, while the 'Performance' column indicates the performance on downstream tasks.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
RTEPre-trained model0.74780.81800.68840.76820.75560.574
RTEFull SFT0.26020.0750.24230.00.14430.890
RTEWiSE-FT0.74880.76640.46380.76820.68680.888
RTEFAPM(Ours)0.75680.81040.68570.75000.75070.903
WikiQAPre-trained model0.74780.81800.68840.76820.75560.896
WikiQAFull SFT0.25100.0760.24340.00.14260.965
WikiQAWiSE-FT0.24760.00.24140.00.12220.963
WikiQAFAPM(Ours)0.75550.80360.69020.76210.75290.962
WinograndePre-trained model0.74780.81800.68840.76820.75560.558
WinograndeFull SFT0.40900.03030.29960.06090.19990.790
WinograndeWiSE-FT0.67470.53430.58210.59140.59560.780
WinograndeFAPM(Ours)0.76180.80680.68450.73950.74820.785
SQuADPre-trained model0.74780.81800.68840.76820.75560.451
SQuADFull SFT0.35310.021220.31830.00.17310.624
SQuADWiSE-FT0.67840.57430.58680.55240.59790.622
SQuADFAPM(Ours)0.74100.80060.67520.75000.74170.615

Table 2: Comparison of FAPM and WiSE-FT on the Qwen2-7B.

From the two tables above, we can conclude that WiSE-FT achieves comparable downstream task accuracy to FAPM. However, WiSE-FT still suffers from significant catastrophic forgetting.

评论

Q4: The rationale and motivation for striking a balance between speciality and generality is not clearly articulated in the paper.

The issue addressed by catastrophic forgetting research is how to enhance a model’s performance on downstream specialized tasks (speciality) without compromising its effectiveness on general tasks (generality). Existing methods [1][2][3] that tackle catastrophic forgetting maintain the model's general performance but may, to varying degrees, impair its performance on specialized tasks. Therefore, these methods usually aim to find a balance between speciality and generality. Our approach addresses the same issue as these methods, and thus our goal is also striking a balance between speciality and generality. We believe this goal is equivalent to “mitigating catastrophic forgetting,” and therefore, we use these two expressions interchangeably in our paper.

Q5: The limitation mentioned in lines 81-83, stating that "methods that alter the training process or model architecture not only make the training process more difficult to control but also degrade the accuracy of downstream tasks," lacks references to prior work supporting this claim.

We have provided references [1][4] to support the claims in our paper. See lines 81 - 83. The experimental results from these two papers demonstrate that modifying the training process (e.g., V-SoftMask [4], HAT-Adapter [5], etc.) or altering the model architecture (e.g., LoRA) can degrade the accuracy of downstream tasks. Refer to Table 2 in [1] and Table 2 in [4] for more details.

[1] Hengyuan Zhang, Yanru Wu, Dawei Li, Zacc Yang, Rui Zhao, Yong Jiang, and Fei Tan. Balancing speciality and versatility: a coarse to fine framework for supervised fine-tuning large language model. arXiv preprint arXiv:2404.10306, 2024.

[2] Jianheng Huang, Leyang Cui, Ante Wang, Chengyi Yang, Xinting Liao, Linfeng Song, Junfeng Yao, and Jinsong Su. Mitigating catastrophic forgetting in large language models with self-synthesized rehearsal. arXiv preprint arXiv:2403.01244, 2024.

[3] Yajing Kong, Liu Liu, Huanhuan Chen, Janusz Kacprzyk, and Dacheng Tao. Overcoming catas- trophic forgetting in continual learning by exploring eigenvalues of hessian matrix. IEEE Trans- actions on Neural Networks and Learning Systems, 2023.

[4] Ke, Z., Shao, Y., Lin, H., Konishi, T., Kim, G., & Liu, B. Continual Pre-training of Language Models. In The Eleventh International Conference on Learning Representations.

[5] Zixuan Ke, Hu Xu, and Bing Liu. Adapting bert for continual learning of a sequence of aspect sentiment classification tasks. In NAACL, pp. 4746–4755, 2021c.

Q6: The figures in this paper appear not to be in vector graphic format, which may affect the clarity and quality of the visualizations.

Thank you for your thorough review. To enable readers to view the images in our paper more clearly, we have replaced the images in the original paper from PNG format to vector graphic format. The revised paper has been re-uploaded.

Q7: The quotation marks in line 212 are not properly formatted according to standard usage.

We appreciate the diligent reviewers for pointing out the improperly formatted quotation marks in our paper. We have made the necessary corrections.

评论

The author have well addressed my concerns, I am happy to modify the Soundness score.

评论

Q2: The experimental setup for the preliminary experiments, particularly those in Sections 2 and 3, is not clearly detailed, e.g., the experimental model, raising concerns about whether similar observations would hold for other models.

Following the reviewer's feedback on Sections 2 and 3 of the paper, we have provided explanations for the experimental settings in Figures 2 and 3. The revised paper has been re-uploaded. In Figures 2 and 3, the model we used is Llama3-8B. To further illustrate our observations, we conducted experiments on Qwen2-7B and included the related figures (Figures 7 and 8) in the appendix, which can be found in lines 843 - 900. From Figures 7 and 8, we observed similar patterns on Qwen2-7B as those on Llama3-8B.

Q3: While instruction-following is a critical aspect of generality, this paper focuses solely on world knowledge and generic reasoning, omitting this aspect

We introduced two domain-specific instruction-following scenarios, namely the medical domain and the mathematical domain. For medical data, we chose the MedMCQA dataset, and for mathematical data, we selected the MetaMathQA dataset. We validated the effectiveness of our method on these two scenarios using Llama3-8B and Qwen2-7B models. The experimental results can be seen in the following two tables. It is worth mentioning that our method remains effective even when catastrophic forgetting occurs during LoRA fine-tuning. Specific results can be found in our response to Reviewer YFJm.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
MedMCQAPre-trained model0.43860.79220.65940.59140.62040.5169
MedMCQAFull SFT0.33150.48970.53650.44340.45020.5862
MedMCQAFAPM(Ours)0.45860.77330.66380.57310.61720.5935
MetaMathQAPre-trained model0.4386-0.65940.59140.56310.7922
MetaMathQAFull SFT0.3101-0.44110.18900.31330.8178
MetaMathQAFAPM(Ours)0.4697_0.64470.57290.56240.8121

Table 3: The results of FAPM on various datasets using Llama3-8B. The 'Avg.' column represents the performance on general tasks, while the 'Performance' column indicates the performance on downstream tasks. Since the downstream task is a mathematical task, we used gsm8k for testing. Therefore, when calculating the performance of general tasks, we did not include gsm8k, which is indicated by a '-'.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
MedMCQAPre-trained model0.74780.81800.68840.76820.75560.5208
MedMCQAFull SFT0.56620.37310.50010.64430.52090.5811
MedMCQAFAPM(Ours)0.76130.79680.67320.74450.74390.5713
MetaMathQAPre-trained model0.7478-0.68840.76820.73470.8180
MetaMathQAFull SFT0.5582-0.43190.32420.43810.8499
MetaMathQAFAPM(Ours)0.7468-0.68570.75110.72780.8484

Table 4: The results of FAPM on various datasets using Qwen2-7B.

From the table above, we can see that our method still effectively alleviates forgetting to a mere 1% while maintaining an impressive 99% accuracy on downstream tasks in both the mathematical and medical task scenarios.

审稿意见
3

The authors identified pruning as a way to mitigate catastrophic forgetting. Specifically, the authors propose to prune weight updates due to fine-tuning, using a combination of pruning metrics including absolute magnitude of change in weights and relative magnitude of change in weights relative to the magnitude of pre-trained weights.

优点

  • Using pruning to mitigate catastrophic forgetting during fine-tuning is novel.
  • This paper contains some original insights about the generality/specificity trade-off when fine-tuning LLMs.

缺点

  • This paper proposes a complicated pruning metrics that only shows marginal improvement over prior art.
  • The quality of write-up is generally poor. For example, section 3.1 is entirely unnecessary and can be in the appendix without affecting the flow of the paper.
  • No ablation study. Why not use the relative portion of the pruning metric alone? It’s a complicated scoring metric, please ablate it.

问题

  • See weakness.
评论

Thank you for Reviewer 53VB's review.

Q1: This paper proposes a complicated pruning metrics that only shows marginal improvement over prior art.

  1. We cannot agree with your comment of "a complicated pruning metrics." This is because our method is sufficiently simple. FAPM only requires pruning the incremental model after fine-tuning, without necessitating any changes to the model training process, and our pruning procedure is also simple enough. To visually demonstrate the simplicity of FAPM, we have provided the PyTorch implementation code for the FAPM pruning metrics below. The formula (1) in the paper can be implemented in just a few lines of code.
def FAPM(W_pre, W_tv, key):
    """
    Computes the FAPM for a specific parameter in a model.

    Parameters:
    -----------
    W_pre : OrderedDict
        The model state_dict of the pre-trained model.
    
    W_tv : OrderedDict
        The model state_dict of the task vectors,
    
    key : str
        The identifier to specify which parameter matrix to use from both `W_pre` and `W_tv`.

    Returns:
    --------
    score : torch.Tensor
        The importance score (FAPM) for the specified parameter. 
    """

    # Compute the absolute value of the task vector for the specified parameter
    abs_W_tv = W_tv[key].abs()

    # Compute the mean of the absolute values of the pre-trained weights for the specified parameter
    mean_abs_W_pre = W_pre[key].abs().mean()

    # Compute the normalized absolute task vector
    normalized_abs_W_tv = abs_W_tv / W_pre[key].abs()

    # Compute the importance score (FAPM)
    score = abs_W_tv - mean_abs_W_pre * normalized_abs_W_tv

    return score
  1. The improvements brought by our method are significant. FAPM effectively alleviates forgetting to a mere 1% while maintaining an impressive 99% accuracy on downstream tasks. Most importantly, compared to previous state-of-the-art methods, our greatest advantage lies in the simplicity of our approach. It does not require any changes to the model training process, does not necessitate adding extra model structures, and does not introduce additional data. Moreover, our method achieves results that are better than or comparable to the previous state-of-the-art methods (refer to Table 1 and Table 2 in the paper).

  2. The simplicity and effectiveness of our method have been fully recognized by other reviewers, e.g., "The simplicity of the technique introduced by the authors is really great" commented by Reviewer YFJm and "FAPM doesn’t complicate the model architecture or training process, making it easier to implement" commented by Reviewer HmBx. Therefore, we believe that Reviewer 53VB's statement regarding our method as "a complicated pruning metric that only shows marginal improvement over prior art" is not appropriate. We cannot agree with this assessment.

Q2: The quality of write-up is generally poor. For example, section 3.1 is entirely unnecessary and can be in the appendix without affecting the flow of the paper.

We do not believe that Section 3.1 is unnecessary because this section presents our analysis of the causes of the CF problem, elucidates our perspective on the issue, and introduces the starting point of our method. We believe these analyses are necessary and insightful. As mentioned by Reviewer YFJm, "I loved reading section 3.1 wherein they provide analysis and a clear thinking process of how they arrived at the method", this section helps readers to better understand our motivation.

评论

Thank you for Reviewer 53VB's review.

Q3: Why not use the relative portion of the pruning metric alone? It’s a complicated scoring metric, please ablate it.

In the paper, we presented comparative experiments with the pruning method that uses only “absolute change magnitude” (refer to Tables 3 and 4 in the paper). Here, we supplement the results of ablation experiments that use only “relative change magnitude” (only RCM) for pruning. We conducted the experiments on Llama3-8B, with the pruning ratio consistent with that in the paper, which is 90%.

TasksMethodscevalgsm8kmmluhumanevalAvg.Performance
RTEPre-trained model0.43860.79220.65940.59140.62040.819
RTEFull SFT0.23110.0750.25540.00.14030.890
RTEonly RCM0.43910.79150.66200.60970.62550.823
RTEFAPM0.46230.79150.64540.59750.62420.897
WikiQAPre-trained model0.43860.79220.65940.59140.62040.913
WikiQAFull SFT0.25470.00.24220.00.12420.966
WikiQAonly RCM0.43850.78920.65890.60970.62400.947
WikiQAFAPM0.47490.79750.65630.58530.62850.964
WinograndePre-trained model0.43860.79220.65940.59140.62040.519
WinograndeFull SFT0.27920.06060.34380.00.17090.820
Winograndeonly RCM0.43690.78990.65970.59750.62090.551
WinograndeFAPM0.48290.76800.64720.57310.61780.824
SQuADPre-trained model0.43860.79220.65940.59140.62040.371
SQuADFull SFT0.28060.02120.32060.00.15560.646
SQuADonly RCM0.44490.79150.65820.59750.62300.374
SQuADFAPM0.47380.73100.64550.57480.60630.637

Table 1: The results of using only RCM on the Llama3-8B model. The 'Avg.' column represents the average performance on general tasks, while the 'Performance' column indicates the performance on downstream tasks.

From the table, it can be seen that using only the relative change magnitude cannot ensure the performance on downstream tasks; the model performance is comparable to that of the pre-trained model. The reason is that pruning using the “relative change magnitude” is designed solely to address catastrophic forgetting. Using only the relative change magnitude results in retaining parameter values that are relatively small in the vector during pruning, which cannot guarantee the performance on downstream tasks.

评论

Dear Reviewer 53VB,

Thank you once again for your valuable feedback. We have submitted our responses to your comments and greatly appreciate your insights. As the discussion phase between the reviewer and the authors approaches its deadline, we kindly request that you share any further thoughts or questions at your earliest convenience.

Best regards, Authors

审稿意见
5

In this paper, the authors introduce Forgetting-Aware Pruning Metric (FAPM), a new pruning-based method aimed at striking a balance between reducing forgetting and maintaining downstream task performance. They find that the overlap between task vectors – which represent the difference between pre-trained weights and those fine-tuned on new tasks – and pre-trained model parameters is a major factor in forgetting. FAPM uses this overlap ratio as a metric for assessing forgetting and incorporates it into pruning decisions. A key advantage is that FAPM doesn't require changes to the training process or the model architecture, nor does it need any additional data. The authors conducted experiments across six datasets covering various tasks, and the results show that FAPM keeps forgetting to just 1% while achieving 99% accuracy on downstream tasks, making it very competitive against other state-of-the-art methods.

优点

  1. FAPM effectively limits forgetting to a mere 1%, which is a significant improvement.
  2. The method maintains an impressive 99% accuracy in downstream tasks, showing that it preserves performance.
  3. Unlike some existing methods, FAPM doesn’t complicate the model architecture or training process, making it easier to implement.
  4. This approach doesn’t rely on replaying original training data or using additional datasets, making it more practical in various scenarios.

缺点

  1. The reliance on the overlap between task vectors and pre-trained weights might not generalize across all types of tasks or models.
  2. While pruning is effective, it may not address all aspects of learning or forgetting, so there could be other methods worth exploring along with FAPM.
  3. While the experiments are comprehensive, testing on even more diverse datasets could strengthen the generalizability of the findings.
  4. The paper doesn’t discuss how FAPM performs in highly specialized or unusual task scenarios, which might reveal limitations.

问题

  1. This appears to be a unique model merging method; however, why is there no comparison made with other methods of a similar nature?
  2. While you have elucidated how to control FAPM to achieve a better balance between accuracy and forgetting levels, is there a more effective approach to directly estimate an optimal balance point rather than adjusting parameters through multiple rounds of performance feedback?
评论

Q3. While you have elucidated how to control FAPM to achieve a better balance between accuracy and forgetting levels, is there a more effective approach to directly estimate an optimal balance point rather than adjusting parameters through multiple rounds of performance feedback?

FAPM does not require multiple rounds of performance feedback to achieve a better balance between accuracy and forgetting levels. Instead, it directly controls the balance by calculating the mean of the absolute values of the pre-trained parameter matrix, as shown in Equation 1 by Wprei| W_{pre}^{i} |. For more details, refer to the textual description of Equation 1 (lines 257 - 272). If the "multiple rounds of performance feedback" you mentioned refer to the for loop in our pseudocode or the ii in Equation 1, this actually represents the application of our method to each module in the parameter matrix separately (or can be understood as layer-wise operations), rather than adjusting parameters through multiple rounds of performance feedback.

评论

Thank you for Reviewer HmBx's review.

Q2. This appears to be a unique model merging method; however, why is there no comparison made with other methods of a similar nature?

Reviewer aVtG also mentioned this issue, stating, "This paper overlooks an important baseline—Wise-FT [1], a standard model parameter merging method," and provided a baseline for model merging for comparison. The original Wise-FT paper notes: "WiSE-FT using a fixed mixing coefficient α=0.5 with the fixed optimal mixing coefficient." Therefore, we followed the hyperparameter settings from the original paper. The experimental results are presented in the following two tables.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
RTEPre-trained model0.43860.79220.65940.59140.62040.819
RTEFull SFT0.23110.0750.25540.00.14030.890
RTEWiSE-FT0.30460.44200.42550.56090.43320.889
RTEFAPM(Ours)0.46230.79150.64540.59750.62420.897
WikiQAPre-trained model0.43860.79220.65940.59140.62040.913
WikiQAFull SFT0.25470.00.24220.00.12420.966
WikiQAWiSE-FT0.25810.00.24580.00.12590.958
WikiQAFAPM(Ours)0.47490.79750.65630.58530.62850.964
WinograndePre-trained model0.43860.79220.65940.59140.62040.519
WinograndeFull SFT0.27920.06060.34380.00.17090.820
WinograndeWiSE-FT0.47410.44120.59670.50600.50450.830
WinograndeFAPM(Ours)0.48290.76800.64720.57310.61780.824
SQuADPre-trained model0.43860.79220.65940.59140.62040.371
SQuADFull SFT0.28060.02120.32060.00.15560.646
SQuADWiSE-FT0.43090.40090.54840.51020.47250.639
SQuADFAPM(Ours)0.47380.73100.64550.57480.60630.637

Table 3: Comparison of FAPM and WiSE-FT on the Llama3-8B model. The 'Avg.' column represents the average performance on general tasks, while the 'Performance' column indicates the performance on downstream tasks.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
RTEPre-trained model0.74780.81800.68840.76820.75560.574
RTEFull SFT0.26020.0750.24230.00.14430.890
RTEWiSE-FT0.74880.76640.46380.76820.68680.888
RTEFAPM(Ours)0.75680.81040.68570.75000.75070.903
WikiQAPre-trained model0.74780.81800.68840.76820.75560.896
WikiQAFull SFT0.25100.0760.24340.00.14260.965
WikiQAWiSE-FT0.24760.00.24140.00.12220.963
WikiQAFAPM(Ours)0.75550.80360.69020.76210.75290.962
WinograndePre-trained model0.74780.81800.68840.76820.75560.558
WinograndeFull SFT0.40900.03030.29960.06090.19990.790
WinograndeWiSE-FT0.67470.53430.58210.59140.59560.780
WinograndeFAPM(Ours)0.76180.80680.68450.73950.74820.785
SQuADPre-trained model0.74780.81800.68840.76820.75560.451
SQuADFull SFT0.35310.021220.31830.00.17310.624
SQuADWiSE-FT0.67840.57430.58680.55240.59790.622
SQuADFAPM(Ours)0.74100.80060.67520.75000.74170.615

Table 4: Comparison of FAPM and WiSE-FT on the Qwen2-7B.

From the two tables above, we can conclude that WiSE-FT achieves comparable downstream task accuracy to FAPM. However, WiSE-FT still suffers from significant catastrophic forgetting.

评论

Thank you for Reviewer HmBx's review.

Response to Weaknesses

The four issues raised by reviewer HmBx in the weakness section primarily concern the generalizability of our proposed method or its potential limitations when applied to more specialized scenarios. To better demonstrate the generalizability of our method, we introduced two more specialized scenarios: a medical scenario and a mathematical scenario. For medical data, we selected the MedMCQA dataset, and for mathematical data, we selected the MetaMathQA dataset. We validated the effectiveness of our method in these two scenarios using Llama3-8B and Qwen2-7B. The experimental results are presented in the following two tables. Notably, even when catastrophic forgetting occurs during LoRA fine-tuning, our method remains effective. The detailed results can be found in our response to reviewer YFJm.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
MedMCQAPre-trained model0.43860.79220.65940.59140.62040.5169
MedMCQAFull SFT0.33150.48970.53650.44340.45020.5862
MedMCQAFAPM(Ours)0.45860.77330.66380.57310.61720.5935
MetaMathQAPre-trained model0.4386-0.65940.59140.56310.7922
MetaMathQAFull SFT0.3101-0.44110.18900.31330.8178
MetaMathQAFAPM(Ours)0.4697_0.64470.57290.56240.8121

Table 1: The results of FAPM on various datasets using Llama3-8B. The 'Avg.' column represents the average performance on general tasks, while the 'Performance' column indicates the performance on downstream tasks. Since the downstream task is a mathematical task, we used gsm8k for testing. Therefore, when calculating the performance of general tasks, we did not include gsm8k, which is indicated by a '-'.

TasksMethodsC-EvalGSM8KMMLUHumanEvalAvg.Performance
MedMCQAPre-trained model0.74780.81800.68840.76820.75560.5208
MedMCQAFull SFT0.56620.37310.50010.64430.52090.5811
MedMCQAFAPM(Ours)0.76130.79680.67320.74450.74390.5713
MetaMathQAPre-trained model0.7478-0.68840.76820.73470.8180
MetaMathQAFull SFT0.5582-0.43190.32420.43810.8499
MetaMathQAFAPM(Ours)0.7468-0.68570.75110.72780.8484

Table 2: The results of FAPM on various datasets using Qwen2-7B.

From the table above, we can see that our method still effectively alleviates forgetting to a mere 1% while maintaining an impressive 99% accuracy on downstream tasks in both the mathematical and medical task scenarios.

评论

Thank you to the author for the patient response. I have carefully reviewed the replies from the other reviewers. I maintain my score unchanged.

评论

We sincerely appreciate your response.

We have carefully read your review, which raises the same concerns as Reviewer aVtG. The main issues are our method's generalization ability and the baseline without model merging. In their subsequent response, Reviewer aVtG mentioned, "The author has well addressed my concerns; I am happy to modify the Soundness score." We are unsure if our response and experiments have resolved your initial concerns about the paper. This would help us further improve the quality of the paper.

评论

Dear Reviewer HmBx,

Thank you once again for your valuable feedback. We have submitted our responses to your comments and greatly appreciate your insights. As the discussion phase between the reviewer and the authors approaches its deadline, we kindly request that you share any further thoughts or questions at your earliest convenience.

Best regards, Authors

审稿意见
6

This paper tries to tackle the problem of whether catastrophic forgetting (CF) due to finetuning be avoided without changing training process, without any additional data, and without altering model structure. To this end the authors come up with Forgetting-Aware Pruning Metric (FAPM) wherein the task vectors are not only pruned based on magnitude but also uses the ratio of task vector to pretrained parameters to avoid CF. The authors provide extensive experiments across various tasks and latest LLMs like Qwen and Llama3 models.

优点

  1. The simplicity of the technique introduced by the authors is really great. Also I loved reading section 3.1 wherein they provide analysis and a clear thinking process of how they arrived at the method.

  2. There are extensive experiments in the paper across various tasks and two models. The models chose are the latest ones which makes this paper a bit more relevant. Section 5 in general is pretty enjoyable to read.

  3. Although the method is not pareto optimal but the numbers are consistently high on all the tasks across both Llama3 and Qwen2.

缺点

  1. A small nitpick. It would be really great if the captions of the images and tables could be a bit longer and more informative.

问题

  1. Nowadays most people prefer to use LoRA not because it reduces CF, but instead its pretty cheap. In my experience even LoRA suffers from CF if the finetuning domain is pretty niche. So I was wondering how informative are task vectors from LoRA merged models and how will this method perform if we use LoRA to finetune.
评论

Thank you for Reviewer YFJm's review.

Q1. A small nitpick. It would be really great if the captions of the images and tables could be a bit longer and more informative.

We have revised the paper according to the reviewers' comments, with the following changes:

  1. We have provided an explanation of the experimental setup for Figures 2 and 3.
  2. We have improved the caption for Figure 4 by adding an example explanation.

The revised paper has been re-uploaded.

Q2. Nowadays most people prefer to use LoRA not because it reduces CF, but instead its pretty cheap. In my experience even LoRA suffers from CF if the finetuning domain is pretty niche. So I was wondering how informative are task vectors from LoRA merged models and how will this method perform if we use LoRA to fine-tune.

In our paper, we define task vectors as the subtraction of pre-trained weights from the weights fine-tuned on downstream tasks (WftWpreW_{ft} - W_{pre}). In LoRA fine-tuning, the initialization of LoRA weights is not based on the pre-trained model weights. Therefore, the concept of task vectors as we defined cannot be applied when using LoRA for fine-tuning.

However, through our consideration, we believe that with a simple modification, LoRA fine-tuning can still utilize our method. After training, we can merge the LoRA parameters with the pre-trained model parameters. The merging formula is Wnew=Wpre+WloraBWloraAW_{new} = W_{pre} + W_{loraB}W_{loraA}, where WloraBW_{loraB} and WloraAW_{loraA} are the LoRA matrices corresponding to a particular parameter matrix of the pre-trained model. Through the above formula, we observe that when merging the LoRA parameters with the pre-trained model parameters, it involves the dot product of the WloraBW_{loraB} and WloraAW_{loraA} matrices and then addition to the pre-trained model weights. Therefore, we can treat WloraBWloraAW_{loraB}W_{loraA} as the task vectors in our algorithm for pruning. That is, ΔW\Delta W in Equation 1 of the paper is WloraBWloraAW_{loraB}W_{loraA}.

To validate the effectiveness of our approach, we conducted experiments on the SQuAD and MedQA datasets. MedQA is a newly added medical dataset. The model used is Llama3-8B. The experimental results are shown in the table below.

datasetmodelcevalgsm8kmmluhumanevalAvg.Performance
SQuADPre-trained model0.43860.79220.65940.59140.62040.371
SQuADLoRA0.47950.72550.59140.58530.59540.648
SQuADFAPM (LoRA/0.9)0.44350.80130.65940.58530.62230.456
SQuADFAPM (LoRA/0.1)0.45020.78620.65520.59140.62070.616
SQuADFAPM (LoRA/0.05)0.45790.77930.65440.58860.62000.636
SQuADFAPM (LoRA/0.03)0.45910.78010.65110.59140.62040.644
MedQAPre-trained model0.43860.79220.65940.59140.62040.552
MedQALoRA0.42160.70890.58890.54090.56500.640
MedQAFAPM (LoRA/0.9)0.45440.78990.65900.59140.62360.588
MedQAFAPM (LoRA/0.1)0.45400.78070.64390.58750.61650.623
MedQAFAPM (LoRA/0.05)0.45310.78450.64750.59140.61910.629
MedQAFAPM (LoRA/0.03)0.45920.78070.64840.57310.61540.632

Table 1: Results of applying our method FAPM to LoRA fine-tuning on the SQuAD and MedQA dataset. The numbers after '/' indicate the pruning ratio. The 'Avg.' column represents the performance on general tasks, while 'Performance' indicates the performance on downstream tasks.

From the table, we can see that LoRA fine-tuning exhibits some forgetting phenomena on both datasets, with the forgetting being more severe on the MedQA dataset. FAPM effectively alleviates forgetting to a mere 1% while maintaining an impressive 99% accuracy on downstream tasks in both datasets. Further experiments revealed that the pruning ratio in LoRA fine-tuning needs to be relatively small to avoid significant performance degradation in downstream tasks. We speculate that this is mainly because the LoRA matrices WloraBWloraAW_{loraB}W_{loraA} are low-rank, and thus contain relatively few redundant parameters.

评论

Thank you for the results. I would like to keep my rating as 6.

评论

We thank the reviewers for their insightful comments and helpful feedback, which allowed us to significantly improve our submission. We are pleased that they highlighted the simplicity, effectiveness, and ease of implementation of our proposed method (YFJm, HmBx, aVtG) and the novelty of using the concept of pruning to address catastrophic forgetting in large language models (YFJm, HmBx, 53VB, aVtG). They mentioned that "Section 5 in general is pretty enjoyable to read and I loved reading section 3.1" (YFJm).

Here, we summarize our responses to the concerns raised by the reviewers, and we address the individual questions below each review.

How our method can be applied to LoRA fine-tuning and its effectiveness: To validate the effectiveness of our method in LoRA fine-tuning, we conducted additional experiments. The results show that our method remains effective even when catastrophic forgetting occurs during LoRA fine-tuning. The experimental results are presented in Table 1 in our response to reviewer YFJm.

Generalizability and instruction-following scenarios: We further validated the generalizability of our method in two highly specialized domains, namely medical and mathematical scenarios. The experimental results show that our method can maintain forgetting to just 1% while achieving 99% accuracy on downstream tasks in both scenarios. The detailed experimental results can be found in Table 1 and Table 2 of our response to reviewer HmBx, or in Table 3 and Table 4 of our response to reviewer aVtG.

Ablation study: We conducted an additional ablation experiment using only the relative magnitude changes. The results demonstrate that using only relative magnitude changes leads to the inability to maintain progress on downstream tasks. The results can be found in Table 1 in our response to reviewer 53VB.

Adding an Important Baseline: We have included an important baseline, "Wise-FT." Our experiments demonstrate that our method significantly outperforms Wise-FT in addressing the issue of catastrophic forgetting. The results can be found in Table 3 and Table 4 of our response to reviewer HmBx, or in Table 1 and Table 2 of our response to reviewer aVtG.

撤稿通知

I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors.