PaperHub
8.3
/10
Poster4 位审稿人
最低4最高5标准差0.4
4
4
5
4
ICML 2025

Towards Lifelong Model Editing via Simulating Ideal Editor

OpenReviewPDF
提交: 2025-01-18更新: 2025-07-24
TL;DR

We propose a general framework that restores the strong performance of standard model editing methods in a lifelong context, bridging the gap between these two paradigms for the first time.

摘要

关键词
Lifelong Model EditingLinear SystemLarge Language Model

评审与讨论

审稿意见
4

This paper introduces Simulating Ideal Editor (SimIE), a framework that extends standard parameter-modifying methods to lifelong scenarios. SimIE computes the ideal parameter shift as the minimum-norm solution of a linear system using the Moore-Penrose inverse, and allows recursive updates by truncating its limiting expression under mild assumptions. The theoretical analysis shows that even when assumptions fail, SimIE remains near-optimal or stable, balancing optimality and robustness. Extensive experiments confirm that SimIE achieves performance comparable to specialized lifelong editing methods.

给作者的问题

See weaknesses.

论据与证据

The claims is supported by experiments.

方法与评估标准

The proposed method make sense for the problem.

理论论述

I only partially checked the formulas and principles, and some parts are not very easy to understand.

实验设计与分析

The experiments is sound. I suggest that the authors include additional case studies and evaluate the method's performance by calculating metrics without relying on teacher forcing.

补充材料

There is no code.

与现有文献的关系

SimIE formulates the ideal parameter shift as the minimum-norm solution to a linear system constructed using the Moore-Penrose inverse, and it subsequently enables recursive updates by truncating the inverse’s limiting expression under two mild assumptions.

遗漏的重要参考文献

There are many papers on lifelong model editing; it is recommended to carefully discuss the differences from previous work.

其他优缺点

Strengths:

SimIE effectively connects standard model editing with lifelong model editing, leveraging advancements from both fields.

The framework is underpinned by a clear theoretical formulation using the Moore-Penrose inverse, providing insights into the optimality and stability of the solution.

By reformulating the ideal parameter shift as a minimum-norm solution, SimIE generalizes parameter-modifying methods to lifelong scenarios, potentially broadening their applicability.

Weaknesses:

The recursive update mechanism depends on two mild assumptions. If these assumptions are not met, the method may face a trade-off between optimality and stability.

The paper could benefit from a deeper comparative discussion with existing lifelong model editing approaches to highlight distinctive advantages and potential limitations.

The theoretical guarantees are tied to the specific conditions assumed; deviations in real-world scenarios might challenge the robustness of the proposed method.

其他意见或建议

No.

作者回复

We are deeply grateful to Reviewer 1PGC for the careful and insightful comments on our manuscript. Our detailed responses to your questions are outlined below.

Q1: evaluating the method's performance without relying on teacher forcing

Following your suggestion, we conduct additional experiments on the recently proposed QAEdit benchmark [1] (released on February 16, 2025), which adopted the real-world evaluation metrics that do not rely on teacher forcing. Specifically, we take FT, ROME, AlphaEdit^-, WISE, and AlphaEdit as baselines and perform T=1000T=1000 sequential edits on Llama-2. The results are summarized in the table:

MethodFTROMEROME+SimIEAlphaEdit^-AlphaEdit^-+SimIEWISEAlphaEdit
Real-world Rel0.000.000.200.000.320.140.27
Real-world Gen0.000.000.130.000.170.080.18
Real-world Loc0.000.000.340.000.350.220.03
Real-world Avg0.000.000.220.000.280.150.16

We observe that all methods, particularly lifelong editing approaches, suffered significant performance degradation in the more realistic evaluation setting. In contrast, standard algorithms enhanced with SimIE consistently outperformed these specialized lifelong editing methods, demonstrating greater robustness. Although SimIE's performance is still limited by the effectiveness of the fundamental editor, its core strength lies in its generality: SimIE enables lifelong editing to benefit directly from any future improvements in standard editors without the need for specialized redesign.

Q2: if these assumptions are not met

We provide a more in-depth empirical analysis of the two assumptions, measuring their violation in practice. Please refer to Reviewer hGQy's Q2 for the details.

Q3: a deeper discussion and code

We will expand our Related Work and Limitations sections to offer a deeper discussion, highlighting both the advantages of SimIE (e.g., leveraging advances in standard editors without specialized redesign) and its limitations (e.g., may be limited by the fundamental editor’s performance). As for the code, we have provided an anonymous GitHub repository in Section 4, which will be moved at the end of the abstract to ensure better visibility.

Thank you once again for your insightful feedback, which has significantly enhanced and refined our work.

[1] Yang, Wanli, et al. The mirage of model editing: Revisiting evaluation in the wild. arXiv preprint arXiv:2502.11177 (2025).

审稿意见
4

This paper introduces "Simulating Ideal Editor" (SimIE), a general framework that enables standard model editing methods to perform effectively in lifelong editing scenarios. The authors formulate the ideal parameter shift as the minimum-norm solution to a linear system constructed using the Moore-Penrose inverse, and develop a recursive update mechanism that approximates this solution through sequential edits. Their theoretical analysis demonstrates that even when key assumptions (over-parameterization and key-value invariance) are violated, SimIE remains either near-optimal or stable against perturbations. Extensive experiments on GPT2-XL, LLaMA-2, and Mistral models show that standard algorithms enhanced with SimIE achieve comparable performance to specialized lifelong editing methods, with minimal implementation. The framework effectively bridges the gap between standard and lifelong model editing paradigms, allowing lifelong editing to benefit from ongoing advancements in standard editing techniques.

给作者的问题

  1. How extensively are the over-parameterization and key-value invariance assumptions violated in your experiments?
  2. After comparing Table 3 with Table 1, I notice an interesting pattern: SimIE demonstrates superior performance over lifelong editing baselines on the smaller GPT2-XL (1.5B) model, but this advantage becomes less significant with the larger 7B models like LLaMA-2 and Mistral. This raises an important question: Is SimIE's effectiveness inversely related to model size?

论据与证据

Yes, I did not find problems for this part.

方法与评估标准

This part is ok. The base models, baseline methods, benchmarks, and metrics in general look good to me in Sec 5. There are some additional benchmark datasets, but not included in this paper.

理论论述

There are many theorems in this paper, and most of the proof are in Appendix. The natural language introduction of the idea behind the theorems looks reasonable to me, but I did not check the exact correctness of the theorems in Sec 4.

实验设计与分析

According to the results reported by the authors, I feel that there is a certain correlation between model size and performance. Therefore, seeing how larger models perform with the proposed SimIE method would provide a deeper understanding of its generalizability.

补充材料

Yes. "D. More experimental details and results."

与现有文献的关系

This paper proposes SimIE, a general framework that bridges the gap between standard model editing and lifelong model editing, enabling standard methods to retain strong performance in lifelong scenarios.

遗漏的重要参考文献

No.

其他优缺点

Strength: A key strength of the SimIE framework lies in its strong theoretical foundation that supports each part of the method. By offering clear mathematical proofs for each step of the algorithm, the authors build a solid basis for their approach and achieve good performance on both three backbone models.

Weakness:

  1. While SimIE is applied across multiple layers independently, the analysis doesn't address how assumption violations cascade through the network, e.g., edits to layer 5 alter the hidden states flowing to layer 6, thereby affecting key representations in layer 6. This oversight limits our understanding of why the approach succeeds in practice despite lacking theoretical guarantees for such complex inter-layer dependencies.
  2. Despite the theoretical contributions, Table 1 shows that SimIE's performance doesn't significantly outperform specialized lifelong editing methods like WISE and AlphaEdit. This raises questions about the practical necessity of the proposed approach. Without demonstrating clear advantages over existing methods, the authors' justification for "bridging the gap" between standard and lifelong editing remains primarily theoretical rather than performance-driven.

其他意见或建议

N/A

作者回复

We sincerely thank Reviewer hGQy for your insightful and constructive comments. We have carefully considered your feedback and responded to each of your points below.

Q1: larger models perform with the proposed SimIE

We conduct additional evaluations on larger LLMs, specifically Llama-3 (8B) and Qwen2.5 (7B). For detailed experimental results, please refer to Reviewer nKJ9's Q1.

Q2: How extensively are the over-parameterization and key-value invariance assumptions violated in your experiments?

We conduct two analyses to assess the extent to which the assumptions are violated in practice. Specifically, we utilize AlphaEdit+SimlE to edit Llama-2 using the ZsRE dataset. At each time step tt, we track the updated key-value pairs (kt,vt)(k_t, v_t), alongside their original counterparts (kt,vt)(k^{\prime}_t, v^{\prime}_t) from the initial unedited model.

For the over-parameterization assumption, we quantify how effectively each layer can fit the desired update by computing the optimal residual RminR_{\mathrm{min}} of the least-squares problem: min1TtTWktvt2\min \frac{1}{T}\sum_{t}^{T}\\|Wk^{\prime}_t-v^{\prime}_t \\|^2. The results are summarized as follows:

Layer45678
RminR_{\mathrm{min}}2.46e-023.36e-024.43e-025.19e-028.41e-02

Empirically, we observe that RminR_{\mathrm{min}} tends to increase in deeper editing layers, indicating a greater deviation from the over-parameterization assumption.

To assess the key-value invariance assumption, we measure the deviations ktkt\\|k_t-k^{\prime}_t\\| and vtvt\\|v_t-v^{\prime}_t\\|, which capture the extent to which edits (both from previous time steps and preceding layers) perturb the original key-value representations. For visualization, we divide the T=1000T=1000 edits uniformly into 100100 intervals and plot their average values within each. The detailed visualizations are available in the Assumption_analysis.PDF at SimIE (our code link), and a partial summary of key deviations for Llama-2 is provided below:

Layer1~10191~200391~400591~600791~800991~1000
40.0000.0000.0000.0000.0000.000
50.0880.3060.3350.3640.3760.373
60.1650.6120.6870.7600.7300.730
70.2480.8851.0251.1361.0521.137
80.3231.2271.4761.6081.4721.626

The findings demonstrate that deviations gradually increase with both editing layer depth and the time step, aligning with your speculation about the perturbations cascade phenomenon.

Overall, both assumptions are indeed violated to some degree in practical scenarios, which emphasizes the trade-off between optimality and sensitivity established by Theorem 4.3. However, the observed violations appear manageable under the current experimental setup, as they do not lead to the empirical failure of SimIE. In scenarios involving longer edit sequences or more editing layers, the violations of assumptions may become severe. These new analyses provide insight into both why SimlE currently succeeds and where it might face challenges going forward, which will be integrated into the manuscript.

Q3: practical necessity of the proposed approach

We would like to clarify that SimIE's primary practical value lies in enabling lifelong editing to benefit from ongoing advances in standard editing research. Indeed, on more challenging datasets and more realistic evaluations recently proposed, specialized lifelong algorithms can exhibit significant performance deterioration, whereas SimlE-enhanced standard methods provide more robust performance (please refer to Reviewer 1PGC's Q1).

Q4: is SimIE's effectiveness inversely related to model size?

Our results generally indicate that the performance of SimIE depends more on the fundamental editor, which may be affected by various aspects like model architecture, dataset, hyperparameters, etc. Although we do observe some performance variations at different scales, there is no definitive evidence of an inverse correlation between model size and SimIE’s effectiveness. For instance, on Llama-3 (8B), ROME+SimIE achieves an average metric of 0.750.75, outperforming the 0.630.63 average observed on the smaller Llama-2 (7B). Additionally, in our new experiments on the other 7B model, Qwen2.5, SimlE improves performance by an average of 4.84.8\\% over existing SOTA lifelong methods. These results suggest that SimIE may leverage that strength when the underlying editor performs robustly, even for larger LLMs.

Thank you once again for your thoughtful and detailed feedback, which offers valuable guidance for the improvement of our paper.

审稿意见
5

Standard model editing techniques suffer significant performance degradation in sequential editing setting due to model drift and catastrophic forgetting. To tackle the issue, this paper proposes a general framework, i.e., SimIE, to restore the performance of any standard model editing techniques in lifelong editing. The key insight is that the desired parameter shift St=BtKtT(KtKtT+λI)1S_t=B_t K_t^T(K_tK_t^T+\lambda I)^{-1} can be written to the recurrence form St=St1+ΔWtktktTPt1S_t=S_{t-1}+\Delta W_t k_tk_t^TP_t^{-1}, where ΔWt\Delta W_t is the parameter shift produced by the standard model editing method. SimIE is empirically evaluated by being applied to four standard model editing methods, i.e., MEND, ROME, MEMIT, and AlphaEdit, on zsRE and CounterFact datasets, where it significantly improves the performance of the standard model editing methods from nearly complete failure to a level comparable to the state-of-the-art lifelong editing methods.

给作者的问题

NA

论据与证据

Following Figure 2, the performance of four standard model editing methods are significantly improved in the sequential editing setting, which strongly supports the claim of the paper.

方法与评估标准

The proposed method is reasonable and the evaluation setting is standard, though I would expect to use more (recent) datasets or metrics.

理论论述

The proposed method is built upon solid theoretical analyses. I have carefully checked the core part and found on problem.

实验设计与分析

See Methods and Evaluation Criteria

补充材料

NA

与现有文献的关系

NA

遗漏的重要参考文献

NA

其他优缺点

Strengths

  • This paper is very well written and organized. The notations are clearly defined. The theoretical analyses are insightful and reasonable. I enjoy reading the paper.
  • The idea is quite novel and interesting. Instead of proposing a sequential editing method, this paper introduces a framework to convert any standard model editing methods to lifelong editing methods, making it general and potentially impactful.
  • The method is simple, which only requires a single line of code to implement, while effective.

Weakness

The main text of this paper leans more to theoretical analyses rather than empirical evaluation. I noticed that some experiments are placed in the Appendix. I recommend to incorporate less theoretical analyses and more empirical evaluation in the main text to increase the visibility (as not all readers will go through the Appendix) and to favor a broader audience (as I assume that all readers are interested in the empirical evaluation but only limited care about full theoretical justifications).

Simpler theoretical analyses may make the main text clearer. The author may directly consider the least square problem in [1], leading to the approximation after Lemma 3.5 (so that the two assumptions and Lemma 3.5 can be moved to Appendix). Then, by writing the optimal solution to the least square problem in a recurrence form, the core method of the paper, i.e., Formula 3.5, is obtained. The regularization term in the least square formulation also provides a straight-forward explanation for the phenomena observed in the ablation study.

[1] Massive Editing for Large Language Models via Meta Learning, ICLR 2024

其他意见或建议

  • I recommend to place the code link at the end of abstract for a better visibility.
  • Please unify the usage of ^\top and ^T in Lemma 3.5.
作者回复

We extend our heartfelt thanks to Reviewer 7yH3 for your thorough and thoughtful review of our manuscript. Following are our responses to each individual comment.

Q1: use more recent datasets or metrics

We conduct experiments on a new benchmark, QAEdit, which is tailored for real-world QA tasks and adopts real-world evaluation metrics (released on February 16, 2025). For further details, please refer to our response to Reviewer 1PGC’s Q1.

Q2: simpler theoretical analyses may make the main text clearer

We agree that the least-squares problem in [1] offers a more succinct path toward deriving the update rule (Equation 3.5). That said, we find that keeping two assumptions in the main text helps bridge our theoretical arguments, and retaining Lemma 3.5 can aid readers in understanding the motivation (i.e., simulating the ideal editor). Nonetheless, we will streamline the theory to retain the minimum element that is essential to guide readers through Equation 3.5.

Q3: code link and transpose symbol

We will move the code link to the end of the abstract and correct the notation ATA^T to AA^\top throughout the manuscript.

Thank you once again for your valuable time and effort in reviewing our work. These insightful feedbacks will significantly enhance the quality and clarity of our paper.

[1] Tan, Chenmien, Ge Zhang, and Jie Fu. Massive Editing for Large Language Models via Meta Learning. ICLR (2024).

审稿意见
4

The paper studies the problem of lifelong model editing, wherein a model has to sequentially incorporate new knowledge without retraining and without altering its behavior on unrelated tasks. Specifically, authors tackle the known issue where consecutive edits cause the model to forget previous edits or compromise its general accuracy - as opposed to batch edits where the model can better account for all examples. The paper proposes a framework (SimIE) for this problem that formulates the task of sequential edits as imitating batch edits, i.e. trying to perform sequential changes that add up to the same effect as though the edited examples were processed jointly. Authors design SimIE to be applied on top of multuple/any parameter-modifying editing algorithms to improve their ability to deal with lifelong(sequential) edits. They evaluate their approach on a range of popular past generation LLMs: Llama 2, Mistral, etc, where it augments multiple different editing methods (ROME, MEND, MEMIT, etc) and improves their performance in a lifelong setting - from abysmal to non-abysmal.

给作者的问题

No questions that would change my evaluations of the paper (as per the reviewing guidelines).

论据与证据

To the best of my understanding, the main claims in the paper are posed around the proposed SimIE framework

  • that it can improve the ability to perform sequential (lifelong) edits in terms of standard statistics (locality, generality, reliability)
  • that it is method-agnostic in the sense that it applies to any parameter-editing algorithms
  • the theoretical analysis of SimIE optimality under relaxed assumptions

The first two claims are, from my perspective, properly supported, though it can be enhanced by editing more recent models. The theoretical claims appear outwardly sound, but could be better supported by testing if the assumptions hold in practice (e.g. the the degree of overparameterization for latest LLMs). Thought, I believe that the paper is worth accepting even without these components.

方法与评估标准

Authors evaluate the traditional metrics (locality, generality, reliability) on popular LLM knowledge editing tasks for several models. There are always more datasets and models to further improve the analysis, but the current scope appears sufficient to verify the claims.

I am slightly concerned by the fact that the paper focuses on the arithmetic mean between locality, generality and reliability in their tables. Since the best results are, in many cases, sub-67% accuracy, this allows a method to win by abandoning one of three criteria - e.g. locality - which would make it a terrible editor. Perhaps it would be better to report (a) geometric mean? (b) conditioned metrics, e.g. best accuracy when locality > 0.8? (or whichever other means authors devise to better represent the method's utility for practitioners)

理论论述

Authors formulate a number of theorems regarding the stability and optimality of SimIE under a number of assumptions. Unfortunately, I only managed to follow the rough outline and did not verify every detail of the proof. Though, I believe that the paper is, in its current form, worth accepting even for its empirical contributions.

实验设计与分析

Their evaluation criteria appear sound, though they could be improved by using more recent models.

补充材料

I have reviewed additional experimental results in the supplementary materials, but not the proofs. I have read the supplementary code linked on L328 (right) and was able to reproduce the results for MEMIT on Llama 2 7B.

I commend the authors for a well documented supplementary code and the reproducibility techniques (e.g. specific dependency versions in requirements). While this wasn't the most important factor in my recommendation to accept the paper, it was certainly one of the factors.

与现有文献的关系

To the best of my knowledge, the contributions presented in the paper account for main backbone editing methods from broader literature / prior work. That said, there are some related works that may deserve additional discussion (e.g. https://arxiv.org/pdf/2405.03279 performs lifelong editing through prompt 'learning' - as opposed to parameter editing), but the relation is debatable.

遗漏的重要参考文献

The idea of model editing was concurrently introduced in [1] and [2]. You cite [2], but it is probably best to also include [1], since it was published about a year earlier. [1] https://openreview.net/forum?id=HJedXaEtvS [2] https://aclanthology.org/2021.emnlp-main.522

其他优缺点

The proposed framework is general, making it possible for future editing algorithms to evaluate in a lifelong without modifications. Though, there is still a gap between editor-agnostic SimIE and specialized lifelong editors.

As I stated earlier, the paper would also become more convincing if authors evaluate on newer and, importantly, more accurate LLMs, e.g. Llama-3.x, Qwen 2.5, deepseek R1 (if enough hardware), as of the time of reviewing. This is because more accurate LLMs are known to be easier to break with model perturbation (e.g. quantization https://arxiv.org/abs/2404.14047 ). Hence, it may be easier to notice a loss of generality / locality / reliability there.

其他意见或建议

Minor: you often capitalize Llama-2 as 'LLaMA-2'. This capitalization was dropped in the second version of the model ( https://arxiv.org/abs/2307.09288 ) and, to the best of my knowledge, did not reappear ever since.

作者回复

We are deeply grateful to Reviewer nKJ9 for the detailed and constructive feedback on our manuscript. Below, we address your questions point-by-point.

Q1: evaluate on newer and, importantly, more accurate LLMs

We conduct additional evaluations on Llama-3 (8B) and Qwen2.5 (7B). We select FT, ROME, AlphaEdit^-, WISE, and AlphaEdit as baselines and perform T=1000T=1000 sequential edits using the ZsRE dataset. The experimental results are presented in the following table:

Llama-3 (8B)Qwen2.5 (7B)
MethodRelGenLocAvgRelGenLocAvg
FT0.170.140.010.100.090.080.020.07
ROME0.090.080.010.060.220.220.090.18
ROME+SimIE0.750.720.800.750.920.870.720.84
AlphaEdit^-0.060.060.030.050.670.630.760.68
AlphaEdit^-+SimIE0.740.670.750.720.910.830.880.87
WISE0.510.501.000.670.540.530.990.69
AlphaEdit0.860.780.620.750.980.800.720.83

We observe that SimIE achieves competitive performance across these recent models, especially surpassing the SOTA method (AlphaEdit) by an average of 4.84.8\\% on Qwen2.5. These results are consistent with those in the main paper, further confirming the effectiveness of our proposed SimIE.

Q2: testing if the assumptions hold in practice

To provide deeper insights into our theoretical assumptions, we consider two additional analyses assessing the extent to which the assumptions are violated in practice. Please refer to hGQy’s Q2 for details.

Q3: focuses on the arithmetic mean

Although the arithmetic mean is widely adopted, it does have the potential pitfalls you mention. Inspired by your suggestion, we introduce a new composite metric defined as eα(Loc1)(RelGen).e^{\alpha(\mathrm{Loc}-1)}(\mathrm{Rel}*\mathrm{Gen}). This metric takes locality as a penalty term, serving as a smoothed condition factor. Meanwhile, by using the (squared) geometric mean of reliability and generality, it prevents methods from abandoning one of them. We will incorporate this new metric into our paper, thereby offering a clearer view of practical utility.

Q4: still a gap between SimIE and specialized lifelong editors

We recognize that SimlE may underperform some specialized lifelong approaches under certain scenarios. Nevertheless, as increasingly challenging datasets and more realistic evaluations emerge, SimlE provides greater robustness compared to elaborated lifelong algorithms (refer also to our response to Reviewer 1PGC’s Q1). These findings reinforce our central claim: SimlE enables lifelong editing to benefit directly from the ongoing advances in standard editing research, thereby bridging the gap between these two paradigms.

Q5: related works and minor mistakes

All the relevant literature [1,2] mentioned will be thoroughly integrated into our discussion, and the capitalization of Llama-2 will be consistent.

Thank you once again for your careful and insightful comments, which provide valuable insights for further refinement of our work.

[1] Chen, Qizhou, et al. Lifelong knowledge editing for llms with retrieval-augmented continuous prompt learning. arXiv preprint arXiv:2405.03279 (2024).

[2] Sinitsin, Anton, et al. Editable neural networks. ICLR (2020).

最终决定

This paper introduces a new method for lifelong model editing. By estimating and caching ideal parameter updates throughout sequential edits, their approach significantly improves the performance of parameter-updating model editing methods on two standard editing datasets.

Reviewers overall found the strengths to outweigh the weaknesses. They appreciated the novelty of the idea, the theoretical analysis, and that the method outperforms the state-of-the-art (though their approach's performance is very similar to the lifelong model editors).