PaperHub
7.0
/10
Spotlight4 位审稿人
最低6最高9标准差1.2
6
9
6
7
4.0
置信度
正确性3.3
贡献度3.3
表达3.0
NeurIPS 2024

Towards Unified Multimodal Editing with Enhanced Knowledge Collaboration

OpenReviewPDF
提交: 2024-05-07更新: 2024-11-06

摘要

关键词
vision and languagemultimodal understanding

评审与讨论

审稿意见
6

The paper introduces UniKE, a novel multimodal editing method that addresses challenges in knowledge editing for Multimodal Large Language Models. UniKE unifies intrinsic knowledge editing and external knowledge resorting by vectorized key-value memories. By disentangling knowledge representations into semantic and truthfulness spaces, UniKE promotes collaboration between intrinsic and external knowledge editing, enhancing the post-edit MLLM's reliability, generality, and locality. Extensive experiments validate UniKE's effectiveness.

优点

  1. UniKE establishes a unified framework for intrinsic knowledge editing and external knowledge resorting.
  2. Extensive experiments demonstrate that UniKE consistently maintains excellent reliability, generality, and locality

缺点

  1. The presentation is confusing. For example, in intrinsic knowledge editing, how do you actual edit the intrinsic knowledge in FFN. Do you follow the same pipeline as T-Patcher or adding an additional neural network? Although intrinsic knowledge is considered as key-value pairs but they are different in the end.
  2. The unified framework and disentanglement of knowledge representations into semantic and truthfulness spaces introduces additional complexity to the editing process. If I understand correctly from the implementation details, it requires more than 15k additional triplets to train the encoders.
  3. In the experiments, I think /beta is another hyperparameter, but I did not find an experiment discussing it.

问题

Why would you add extra 10 key-value pairs in the FFN of the last four transformer layers in intrinsic knowledge editing? It seems that \zeta is quite sensitive. Could you explain that?

局限性

N/A

评论

To facilitate your reading, we also paste our additional experimental results here, which are consistent with the tables / figures in the rebuttal PDF.

 

Table 7: The computational speed, resource utilization and performance of each method. We use the average results of five metrics (Reliability, T-Generality, M-Generality, T-Locality, and M-Locality) as the performance measure.

MethodGPU memoryEditing time for each sampleAvg performance
FT22G6.1s60.6
KE24G5.8s74.7
T-Patcher18G4.7s80.4
MEND36G5.2s90.3
IKE20G1.6s65.5
SERAC49G3.6s76.4
UniKE18G5.0s95.2

 

Table 10: Editing time cost and performance with/without encoders for UniKE. The time refers to the average editing or inference time for one sample. Gen is the average result of T-Generality and M-Generality; while Loc is the average result of T-Locality and M-Locality.

MethodGPU MemoryEditing timeInference timeRel.Gen.Loc.
w/o encoders17.7GB4.92s0.212s96.291.290.3
UniKE17.8GB5.04s0.217s97.494.693.5

 

Table corresponding to Figure 1 in Rebuttal PDF: Editing Performance on Different Value of β\beta. Generality is the average result of T-Generality and M-Generality; while Locality is the average result of T-Locality and M-Locality.

β\beta0.00.40.81.01.21.6
Rel.97.097.797.898.097.597.9
Gen.91.793.395.095.194.894.7
Loc.91.693.093.693.893.293.5

 

We hope we have effectively addressed your concerns. Discussions are always open. Thank you once again for your time and insightful comments!

作者回复

We sincerely appreciate your constructive comments. We will explain your concerns point by point.

Q1: The presentation is confusing. For example, in intrinsic knowledge editing, how do you actual edit the intrinsic knowledge in FFN?

A1: We apologize if our manuscript causes any confusion. The intrinsic knowledge editing is implemented by adding 10 knowledge neurons in the FFN of the last four transformer layers with a similar pipeline as T-patcher. The addition of neurons can be understood as an expansion of the dimension of the weight matrix in the FFN, as shown in Eq.3. This process retains the original d×dd\times d' weights in the FFN while appending a learnable parameter of dimensions d×ned\times ne as the new neurons. And Eq.2 demonstrates why intrinsic knowledge can be viewed as key-value pairs in this context.

Although this pipeline may seem similar to TPatcher, the key innovation of our paper lies in consistently transforming in-context editing to the latent level and integrating it into the self-attention, which allows both external and intrinsic knowledge editing to follow a unified paradigm within a cohesive framework. Furthermore, our intrinsic knowledge editing is actually an improvement of T-Patcher. By synergistically unifying the paradigms of both editing methods and incorporating knowledge disentangling, we identify a truthful editing direction, ζ\zeta, which further guides intrinsic knowledge into a more generalizable direction. To understand how ζ\zeta enhances intrinsic knowledge editing, please refer to Eq.6, which is inspired by the success of [Chen et al 2022].

[Chen et al 2022]AdaptFormer: Adapting Vision Transformers for Scalable Visual Recognition.Neurips2022.

 

Q2: The unified framework and knowledge disentanglement introduce additional complexity to the editing.

A2: Thank you for valuable feedback. We would like to address your concerns from two points.

(2.1) The unified framework and knowledge disentanglement almost incur no additional time or GPU memory overhead during the editing process. Both the truthfulness and semantic encoders consist of several MLP layers with only 4M parameters (MLLM contains 7B parameters), which remain frozen during editing. As shown in Table 10 in the Rebuttal PDF, these encoders enhance editing performance with NEGLIGIBLE additional time and memory overhead. You can also refer to Table 7 in the Rebuttal PDF to compare the editing speed and performance of UniKE with baselines. It shows that UniKE achieves significant performance advantages without incurring additional resource costs or computation time. Therefore, we think that the additional complexity they bring to the editing process is acceptable.

(2.2) Leveraging 15k additional triplets to train the encoders is a one-time pre-training process, taking less than an hour. Once pre-trained, the encoders are frozen for downstream knowledge editing tasks, adding no training-time cost to knowledge editing. As shown in Table 10 in the Rebuttal PDF, after removal of encoders without pre-training, the performance significantly decreases. So we think that the time cost of this one-time pre-training is acceptable. Furthermore, this pre-training is common in knowledge editing. Methods like MEND and SERAC also require pre-training, with pre-training time often exceeding one day, which is 24 times longer than UniKE pre-training.

[Please Refer to Table 10 in the Rebuttal PDF for the Results]

 

Q3: Did not find an experiment discussing /beta

A3: We apologize for neglecting to discuss the value of β\beta. In our experiments, we set β\beta=1.0. And we conduct further experiments, ranging β\beta from 0.0 to 1.6 to evaluate the performance. As shown in Figure 1 in the Rebuttal PDF, within the range of 0.8 to 1.6, UniKE's performance shows no significant variation with changes in β\beta. This indicates that ζ\zeta is robust to the change of β\beta.

[Please Refer to Figure 1 in the Rebuttal PDF for the Results]

 

Q4: Why would you add extra 10 kv pairs in the FFN of last 4 transformer layers in intrinsic knowledge editing? It seems that \zeta is quite sensitive.

A4: Thank you for the question. We would like to address your concerns from two points.

(4.1) The implementation detail for intrinsic knowledge editing is to add an extra 10 kv pairs to the FFN of the last four transformer layers. "extra" means that we preserve the original parameters of the MLLM while achieving intrinsic knowledge editing through the additionally added kv pairs. Moreover, We choose these hyperparameters because we find them to be the most suitable for multimodal editing. Of course, we maintain a comparable number of neurons and transformer layers in the T-Patcher baseline for fair comparison.

(4.2) Figure 4(c) in our paper highlights the importance of selecting an appropriate method for constructing the editing direction ζ\zeta. Specifically, using a random vector for ζ\zeta degrades performance due to the introduction of noise during editing, which hinders the MLLM from learning effectively. In contrast, our proposed method for constructing ζ\zeta shows a noticeable performance improvement (especially in generality), as we learn a generalizable editing direction from a large number of samples, addressing the inherent limitations of generality in intrinsic knowledge editing.

Moreover, once completing the construction of ζ\zeta, as seen from the above A3, adjusting its weight β\beta has little impact on editing performance. It indicates that the constructed ζ\zeta itself is robust to weight variations, although its construction method is sensitive.

 

Once again, we express our heartfelt gratitude for your valuable suggestions! In the revised version of the paper, we will improve the descriptions to make them clearer.

评论

The rebuttal have addressed most of my concerns. I will raise my score to 6.

评论

Thank you for raising the score. Your valuable suggestions greatly contribute to the quality of our paper. And we are deeply grateful for your insightful suggestions on our work!

审稿意见
9

This paper proposes UniKE, a novel multimodal editing method that establishes a unified perspective for intrinsic knowledge editing and external knowledge resorting. On this basis, the authors combine both types of knowledge editing methods, executing them in the latent space with a unified paradigm. Furthermore, this paper proposes to disentangle the knowledge representations into the semantic and truthfulness spaces, effectively enhancing the collaboration between intrinsic knowledge and external knowledge resorting. Extensive experimental results show that UniKE achieves promising results under various settings, ensuring that the post-edit MLLM maintains excellent reliability, generality, and locality.

优点

(1) I think knowledge editing for MLLMs is a relatively new topic. Previously, a few studies merely adapted existing knowledge editing methods from the NLP field into multimodal domain. To the best of my knowledge, this paper is the first to conduct a detailed and systematic analysis of the strengths and weaknesses of existing methods when applied to editing multimodal LLMs.

(2) The proposed method is very novel and effective. Previous efforts in knowledge editing show significant differences between intrinsic knowledge editing methods and external knowledge resorting methods. In this work, the authors ingeniously convert in-context editing into the format of feature shifting, achieving a unification of the editing paradigms that can operate simultaneously within the same transformer layer with synergistic correlation. I find this to be a very inspiring design. Moreover, the design of knowledge collaboration is closely integrated with this unified paradigm.

(3) The experiments are very solid and thorough, clearly demonstrating that UniKE effectively addresses multimodal knowledge editing tasks under various setups. Meanwhile, the authors have also provided the implementation code for the experiments.

(4) I commend the authors for conducting an extensive set of ablations and analyses, which are very helpful in understanding the impact of each component within UniKE.

(5) Additionally, I believe that the method proposed by the authors is not only applicable to knowledge editing tasks. By converting in-context learning into the representation space and avoiding the need to increase the context window space, it better synergizes with parameter update learning. I consider this to have significant implications for further studies on how to construct more powerful MLLMs.

缺点

(1) In the NLP community, some studies will discuss the resilience to overediting [1] of knowledge editing methods by adopting the contrastive knowledge assessment [2] (CKA). Unlike the locality property that measures whether LLMs forget previous knowledge, overediting can be understood as excessive generalized to seemingly similar but unrelated samples. Although there may be no current work on multimodal editing that discusses the phenomenon of overediting, I encourage authors to add relevant experiments for a straightforward comparison of the resilience to overediting among each method (UniKE, MEND, T-Patcher, and IKE). (2) A more challenging task of knowledge editing is counterfactual editing, where the edited answer yy to the question xx can sometimes be counterfactual to the real world. A typical counterfactual editing dataset in the NLP community is called COUNTERFACT [3], which more accurately reflects the true effectiveness of knowledge editing methods by avoiding the effects of LLMs knowing this knowledge before editing. I encourage authors to construct multimodal counterfactual editing datasets and conduct more experiments to verify whether UniKE performs better in counterfactual editing scenarios compared to MEND, T-Patcher and IKE. [1] Zheng, Ce, et al. "Can we edit factual knowledge by in-context learning?." arXiv preprint arXiv:2305.12740 (2023). [2] Dong, Qingxiu, et al. "Calibrating factual knowledge in pretrained language models." arXiv preprint arXiv:2210.03329 (2022). [3] Meng, Kevin, et al. "Locating and editing factual associations in GPT." Advances in Neural Information Processing Systems 35 (2022): 17359-17372.

问题

See the weaknesses.

局限性

The authors have adequately discussed the limitations and potential negative societal impact.

作者回复

We sincerely thank you for your valuable comments and high appreciation of our work! We are encouraged that our research is recognized as having significant implications for further studies on constructing more powerful MLLMs. We will address your concerns point by point.

Q1: In the NLP community, some studies will discuss the resilience to overediting of knowledge editing methods by adopting the contrastive knowledge assessment (CKA). Unlike the locality property that measures whether LLMs forget previous knowledge, overediting can be understood as excessively generalized to seemingly similar but unrelated samples. Although there may be no current work on multimodal editing that discusses the phenomenon of overediting, I encourage authors to add relevant experiments for a straightforward comparison of the resilience to overediting among each method (UniKE, MEND, T-Patcher, and IKE).

A1: Thank you for raising such a professional question! To construct the dataset for overediting evaluation, we randomly select a portion of the test data from the E-VQA task. We then make minor modifications to the text questions in each example so that they appear similar to the original questions but differ in semantics, thus constructing the test data for overediting. To improve experimental efficiency, we conduct experiments on MiniGPT-4 with the setup of one-step editing. The results of the CKA evaluation for each method are listed in the following table. We can see that UniKE outperforms all other methods. The results indicate that **UniKE has less influence on over-editing, which further demonstrates its robust capabilities.

Table 1: The results of CKA evaluation on MiniGPT-4 with the setup of one-step editing.

MethodT-PatcherMENDIKEUniKE
CKA1.381.331.471.58

We also present the results in Table 8 of the Rebuttal PDF.

 

Q2: A more challenging task of knowledge editing is counterfactual editing, where the edited answer to the question can sometimes be counterfactual to the real world. A typical counterfactual editing dataset in the NLP community is called COUNTERFACT, which more accurately reflects the true effectiveness of knowledge editing methods by avoiding the effects of LLMs knowing this knowledge before editing. I encourage authors to construct multimodal counterfactual editing datasets and conduct more experiments to verify whether UniKE performs better in counterfactual editing scenarios compared to MEND, T-Patcher, and IKE.

A2: Thank you for the insightful question. We first construct a counterfactual editing dataset based on the MMEdit dataset, ensuring that the MLLM did not have prior knowledge of the editing target. We also conduct experiments on MiniGPT-4 with one-step editing. The results of the counterfactual editing are shown in the following table. **It can be seen that UniKE significantly outperforms existing methods in this more challenging editing scenario, fully demonstrating the effectiveness of our method.

Table 2: Performance of counterfactual editing on MiniGPT-4.

MethodRelT-GenM-GenT-LocM-LocAvg
T-Patcher80.065.957.784.288.375.2
MEND90.683.274.193.582.184.7
IKE90.383.781.544.15.060.9
UniKE90.884.780.794.994.589.1

We also present the results in Table 9 of the Rebuttal PDF.

 

We will add these experiments to the main body or the appendix of our paper. Thank you once again for your insightful and professional feedback!

评论

Thank you for your rebuttal response. I think that over-editing evaluation and counterfactual editing are critical tasks that can reflect whether a model truly possesses the capabilities of knowledge cognitive learning in knowledge editing scenarios. It's impressive to see that UniKE performs well on these two challenging tasks. I consider UniKE to be a strong contribution, and I will raise my score to 9.

评论

Thank you for raising the score. We deeply appreciate your recognition of our work and the constructive advice you've offered!

审稿意见
6

UniKE is a unified framework for multi-modal knowledge editing that includes three main aspects: 1.Knowledge Separation: UniKE divides knowledge into factuality and semantic spaces to manage and coordinate different types of knowledge more effectively. 2.Knowledge Collaboration: In the factuality space, UniKE standardizes new knowledge based on a learned factuality distribution, enhancing reliability and generality. In the semantic space, it adjusts the integration of external knowledge based on relevance to the input samples, maintaining locality. 3.Multi-step Editing: UniKE supports single-step, multi-step sequence, and cross-task editing while maintaining high reliability, generality, and locality. Through these innovations, UniKE significantly improves performance in multi-modal knowledge editing tasks.

优点

1.The paper introduces UniKE, a novel framework that seamlessly integrates intrinsic and external knowledge editing, enhancing the model’s ability to handle complex multimodal information effectively. The motivation behind the work is clear, and the article is well structured, guiding the reader through the innovative approach and its benefits.

2.By disentangling knowledge into semantic and truthfulness spaces, the proposed method ensures robust collaboration between different types of knowledge, significantly improving the model's reliability, generality, and locality. The method's effectiveness is demonstrated through comprehensive experiments across various settings, consistently outperforming existing state-of-the-art methods.

3.UniKE's design allows for application across different multimodal models and editing scenarios, making it a versatile and robust solution for enhancing multimodal language models.

缺点

1.The paper does compare UniKE with other intrinsic knowledge editing and external knowledge resorting methods and highlights its efficiency in several aspects. However, it lacks a detailed discussion on computational speed and resource utilization.

2.The reliance on fine-tuning for knowledge updates could lead to overfitting, especially if the model is frequently updated. This could impact the model’s generalization abilities, making it less effective in unforeseen or less frequent scenarios.

3.The paper lacks detailed explanations for the evaluation metrics in Line 240. In Table 2, across multiple experiments, it is unclear why there is little difference compared to the SERAC method in the T-Loc metric, but a significant difference in the M-Loc metric.

4.The paper lacks detailed information about the parameter "n" used in the contrastive learning formula, making it difficult to understand its impact on model performance, and does not discuss how different values of "n" might influence the results and effectiveness of the method.

问题

Please address the comments in the weaknesses section.

局限性

The limitations of the article have been discussed by the authors

作者回复

Thank you for your kind feedback and valuable comments. We will explain your concern as follows.

Q1: Lacking a detailed discussion on computational speed and resource utilization.

A1: Thank you for the suggestion. In Table 7 of the Rebuttal PDF, we list the computational speed and resource utilization of each method. As can be seen, compared to most baselines, UniKE achieves significant performance advantages without incurring additional resource costs or computation time. Moreover, although external knowledge editing methods (IKE, SERAC) are faster in terms of editing speed, their performance is unacceptable, lagging 20-30 points behind UniKE. These results demonstrate the superiority of UniKE.

[Please Refer to Table 7 in the Rebuttal PDF for the Results]

 

Q2: The reliance on fine-tuning for knowledge updates could lead to overfitting.

A2: First, The objective and evaluation metrics of knowledge editing include not only successfully completing edits but also ensuring the post-edit MLLM can generalize over equivalent input neighbors (Generality) and maintain consistent output for irrelevant inputs (Locality). In fact, Generality and Locality can be considered as two aspects of evaluating overfitting: the new knowledge should generalize well without disrupting the generalization of unrelated old knowledge.

In our method, to maintain Generality and Locality and prevent overfitting, we retain the original model parameters while combining both intrinsic and external knowledge editing within a unified framework. Intrinsic knowledge editing introduces a minimal amount of tunable parameters in a small subset of FFNs to preserve locality. Meanwhile, external knowledge editing incorporates retrieved in-context representations (with no tunable parameters) into the latent space of the transformer to enhance generality. We also propose knowledge disentangling to promote knowledge collaboration. In the semantic space, intrinsic knowledge helps select appropriate external knowledge, preventing locality disruption. In the truthfulness space, external knowledge identifies a generalizable editing direction, regulating intrinsic knowledge and alleviating its restriction on generality.

The experimental results (Tables 2, 3, and 4 in our paper) demonstrate that UniKE achieves better locality and generality in various multimodal editing scenarios, and more effectively addresses the issue of overfitting compared to baselines.

 

Q3: The paper lacks detailed explanations for the evaluation metrics. In Table 2, across multiple experiments, it is unclear why there is little difference compared to the SERAC method in the T-Loc metric, but a significant difference in the M-Loc metric.

A3: We apologize for any confusion regarding the evaluation metrics. In Appendix B.1, we detail the definitions of the five evaluation metrics (see Eq.11-15). In brief, these metrics assess accuracy across different input samples: Reliability on edit targets, T-Locality on unrelated QA tasks, M-Locality on unrelated VQA tasks, T-Generality on samples that rephrase the text inputs based on the edit targets, and M-Generality on samples that redraw image inputs based on the edit targets.

Furthermore, To analyze the reasons for the little differences in the T-Loc metric and the significant differences in M-Loc, we examined some intermediate outputs of SERAC during editing. We discover that the reason is related to the pipeline of SERAC. SERAC adopts a counterfactual model while keeping the original model unchanged. It employs a scope classifier to determine if new input falls within the range of stored edit examples. If the input matches any cached edit, output the counterfactual model’s prediction based on the input and the most probable edit. Otherwise, the original model’s prediction is given.

For T-Loc, the input samples contain only text information, which differs significantly from the input format of edited samples. This makes the counterfactual model less likely to activate, resulting in outputting the original model's prediction and better locality. However, for M-Loc, the input samples, like the edited samples, contain multimodal information. As the scope classifier is trained to activate the counterfactual model for multimodal inputs even if they are unrelated to the editing target, the incorrect choices made by the scope classifier result in significantly worse M-Loc performance of SERAC, causing the observed phenomenon.

 

Q4: The paper lacks detailed information about the parameter "n" used in the contrastive learning formula, and does not discuss how different values of "n" might influence the results and effectiveness of the method.

A4: Thank you for raising an important concern! "n" is a crucial hyperparameter worth discussing. First, n represents the number of training samples used for contrastive learning during encoder pre-training. These training data come from the knowledge representations of the in-context knowledge we constructed, which we detailed in Appendix C on how to prepare and extract these in-context knowledge representations. In our experiments, we set n to approximately 16000.

Building on this, we perform an ablation study on the value of n to test how different numbers of pre-training samples affect the final editing performance. As shown in Figure 2 in the Rebuttal PDF, the performance keeps increasing when the data number increases from 0 to 8000. Beyond this, escalating the data count from 8000 to 16000 yields only marginal enhancement. It demonstrates that our encoder training is data-efficient and only requires relatively small amounts of data to achieve effective knowledge representation disentangling.

[Please Refer to Figure 2 in the Rebuttal PDF for the Results]

 

Thank you once again for your time and valuable feedback!

评论

Dear Reviewer:  Thank you very much for your kind feedback and valuable comments. I have carefully read your comments. However, it appears that the content from the "Strengths" section may have been inadvertently pasted into the "Weaknesses" section. As a result, the weaknesses you mentioned still seem to describe the strengths of our work. This makes it challenging to identify and address the specific concerns you may have intended to highlight.   Could you please review your comments once more and provide additional clarification on the weaknesses you observed? Your insights are crucial for improving our paper, and I am keen to address all your concerns effectively.   Thank you again for your time and valuable feedback!

评论

I apologize for the mistake in my previous feedback. I have reviewed my comments and provided the correct feedback on the weaknesses of your paper. I hope this clarifies my observations and helps you improve your paper.

评论

To facilitate your reading, we also paste our additional experimental results here, which are consistent with the tables / figures in the rebuttal PDF.

 

Table 7: The computational speed, resource utilization and performance of each method. We use the average results of five metrics (Reliability, T-Generality, M-Generality, T-Locality, and M-Locality) as the performance measure.

MethodGPU memoryEditing time for each sampleAvg performance
FT22G6.1s60.6
KE24G5.8s74.7
T-Patcher18G4.7s80.4
MEND36G5.2s90.3
IKE20G1.6s65.5
SERAC49G3.6s76.4
UniKE18G5.0s95.2

 

Table corresponding to Figure 2 in Rebuttal PDF: Editing performance on different value of nn. Generality is the average result of T-Generality and M-Generality; while Locality is the average result of T-Locality and M-Locality.

nn0400080001200016000
Rel.96.297.097.898.198.0
Gen.91.293.394.594.995.1
Loc.90.392.293.393.793.8

 

We hope we have addressed all of your concerns. Discussions are always open. Thank you once again for your constructive suggestions!

审稿意见
7

This paper proposes UniKE, a novel multimodal editing method that establishes a unified perspective and paradigm for intrinsic knowledge editing and external knowledge resorting. Within such a unified framework, the authors further promote knowledge collaboration by disentangling the knowledge representations into the semantic and truthfulness spaces. Extensive experiments validate the effectiveness of UniKE, which ensures that the post-edit MLLM simultaneously maintains excellent reliability, generality, and locality.

优点

  1. The paper is well-written and easy to follow and the motivation is clear and reasonable.
  2. The paper gives a unified perspective on the intrinsic refinement of knowledge and the strategic reorganization of external knowledge, which can enhance subsequent research endeavors.
  3. The experimental settings (one-step editing, sequential editing, cross-task editing) are fair.

缺点

  1. The authors only edit Qformer style MLLMs (MiniGPT4, BLIP2), more foundation models should be compared. e.g., LLava. Besides, the improvement on BLIP2 (Tab.1) seems incremental.
  2. More recent model editing baselines should be compared to prove the effectiveness of the proposed methods.
  3. The paper focuses on MLLM editing, what are the main differences between MLLM editing and LLM editing? What are the specific designs for multimodal models? I notice that some compared methods are proposed for LLM editing. Can the proposed methods be used on LLM editing? The authors should give some explanations and experimental results if possible.

问题

Please refer to the weakness below.

局限性

The paper briefly mentioned limitations.

评论

To facilitate your reading, we also paste our additional experimental results here, which are consistent with the tables in the rebuttal PDF.

 

Table 1: Performance of one-step editing on LLaVA1.5 (We average the results on E-IC and E-VQA).

MethodRel.T-Gen.M-Gen.T-Loc.M-Loc.Avg
FT67.457.954.867.163.262.1
KE72.765.455.382.267.368.4
T-Patcher89.076.569.081.281.379.6
MEND95.492.678.383.580.386.0
IKE93.485.177.927.73.257.5
SERAC96.392.485.583.37.773.0
UniKE95.792.888.486.086.489.9

 

Table 2: Performance of cross-task editing on LLaVA1.5.

MethodRel.T-Gen.M-Gen.T-Loc.M-Loc.Avg
FT66.957.251.062.354.456.4
KE69.860.352.579.362.164.8
T-Patcher81.260.057.477.476.570.5
MEND90.484.373.878.676.080.6
SERAC92.188.382.582.21.269.3
UniKE92.289.283.882.784.786.5

 

Table 3: Performance of cross-task editing on BLIP-2 OPT.

MethodRel.T-Gen.M-Gen.T-Loc.M-Loc.Avg
FT57.249.943.252.249.750.4
KE64.260.157.283.559.264.8
T-Patcher83.169.765.984.577.976.2
MEND84.282.474.991.480.282.6
SERAC90.889.284.190.01.771.2
UniKE91.190.688.291.785.689.4

 

Table 4: Comparison with recent baselines for one-step editing on MiniGPT-4 (We average the results on E-IC and E-VQA).

MethodRel.T-Gen.M-Gen.T-Loc.M-Loc.Avg
MENMET97.096.282.498.085.291.8
WISE97.292.288.798.488.293.0
UniKE97.496.692.698.888.194.7

 

Table 5: Comparison with recent baselines for cross-task editing on MiniGPT-4.

MethodRel.T-Gen.M-Gen.T-Loc.M-Loc.Avg
MENMET88.487.278.086.182.584.4
WISE89.285.483.487.883.685.9
UniKE90.788.286.890.483.888.0

 

Table 6: Performance of each method on LLM editing task (ZsRE) for one-step editing and 200-step editing.

ONE-STEP EDITING200-STEP EDITING
MethodRel.Gen.Loc.Avg.Rel.Gen.Loc.Avg.
FT77.476.735.563.219.517.25.414.0
KE20.620.181.340.77.66.865.826.7
T-Patcher97.195.096.296.181.470.691.381.1
MEND98.297.797.497.80.00.00.00.0
In-Context Editing99.497.259.285.3----
SERAC88.687.999.992.124.023.296.447.9
MENMET99.186.897.494.482.973.690.282.2
WISE98.896.399.998.382.874.795.584.3
UniKE99.597.999.699.085.176.795.685.8

 

These experiments will be integrated into the main body or the appendix of our paper. We hope we have addressed all of your concerns. Discussions are always open. Thank you again for your time and valuable suggestions!

作者回复

We sincerely thank you for the valuable comments! We are encouraged to see that our work can enhance subsequent research endeavors. We will address your concerns point by point.

Q1: More foundation models should be compared (LLava). The improvement on BLIP2 seems incremental.

A1: Thank you for raising the important concern. We first leverage LLaVA1.5 to perform multimodal editing, and then we will show that the improvement on BLIP-2 OPT is not incremental.

(1.1)Multimodal Editing on LLaVA

We conduct both one-step editing and 10-step cross-task editing on LLaVA1.5. The results are shown in Tables 1 and 2 of the rebuttal PDF. With LLaVA as the backbone, UniKE still balances all three target properties well, outperforming all baseline methods in both one-step editing and cross-task editing. This demonstrates that UniKE is model-agnostic and effective across various types of MLLMs.

[Please Refer to Tables 1 and 2 in the Rebuttal PDF for the Results]

 

(1.2)Analysis on BLIP-2 OPT

Though improvement of one-step editing on Blip-2 OPT(in Table1 of our paper) may seem incremental, we aim to demonstrate the significant performance advantage of UniKE on BLIP-2 OPT from 3 aspects:

(I) Across all ten metrics of BLIP-2 OPT one-step editing, UniKE achieves best results in five metrics, with the remaining suboptimal results only marginally lower than the best. We also calculate the average performance of each method across all ten metrics, as shown in the following Table. The average result of UniKE in BLIP-2 OPT one-step editing significantly surpasses other methods, exceeding the second-best method (MEND) by 5.1 points.

Table: Avg performance of all metrics on Blip-2 OPT one-step editing

MethodFTKETPatcherMENDIKESERACUniKE
AVG Perf.56.270.382.390.565.377.695.6

(II) In one-step editing on BLIP-2 OPT, UniKE is the only method that consistently performs well across all metrics, whereas other methods have at least one notably poor metric. For instance, MEND's M-Gen(E-VQA), M-Gen(E-IC), and M-Loc(E-IC) results are 15.7, 19.0, and 14.9 points lower than UniKE's, respectively. IKE and SERAC perform especially poorly on M-Loc, with accuracy below 10%.

(III) Additionally, the relatively low difficulty of one-step editing task may allow some baselines to maintain relatively high evaluation scores. In more challenging scenarios such as multi-step cross-task editing, UniKE's advantage becomes more apparent. Specifically, in BLIP-2 OPT cross-task editing with higher difficulty (shown in Table 3 of the Rebuttal PDF), UniKE consistently outperforms all other baselines in all metrics, demonstrating a more significant performance advantage.

[Please Refer to Table 3 in the Rebuttal PDF for the Results]

 

Q2: More recent baselines should be compared.

A2: Thank you for the suggestion. We further compare UniKE with two recent baselines proposed in 2024: MENMET, WISE (a concurrent work). We leverage MiniGPT4 to conduct both one-step editing and cross-task editing. The results are shown in Tables 4 and 5 of the Rebuttal PDF. The performance of UniKE surpasses these recent baselines, further demonstrating the effectiveness of UniKE.

[Tan et al 2024]MENMET: Massive Editing for Large Language Models via Meta Learning.ICLR2024.

[Wang et al 2024]WISE: Rethinking the Knowledge Memory for Lifelong Model Editing of Large Language Models.23 May 2024.

[Please Refer to Tables 4 and 5 in the Rebuttal PDF for the Results]

 

Q3: What are the main differences between MLLM editing and LLM editing?What are the specific designs for multimodal models?Can UniKE be used on LLM editing?

A3: Thank you for your questions. We will address them from three points:

(3.1) The main differences between MLLM editing and LLM editing lie in task difficulty. LLM stores single-modality NLP knowledge within its parameters, while MLLM stores multi-modality knowledge, making MLLM editing more challenging due to the need to edit knowledge from multiple modalities to fix errors. Although mainstream editing methods proposed for both are similar, current methods can effectively edit LLMs but not MLLMs [Chen et al 2023].

(3.2) [Chen et al 2023] attempted specific designs for editing MLLMs such as editing the unique Qformer, but found these methods less effective than directly applying LLM editing methods, as shown in the following table. This is because the bottleneck for MLLMs lies in enabling LLMs to reason with multimodal input, rather than in extracting visual information. Therefore, mainstream MLLM editing methods still focus on editing LLM and directly adopt LLM editing techniques..

However, directly applying LLM editing methods for multimodal editing still fails to balance both generality and locality. In this paper, we discuss the shortcomings of existing LLM editing methods when applied to multimodal scenarios and develop a unified framework along with knowledge disentangling, which leverages the strength and mitigates the weakness of each method type, leading to more effective multimodal editing.

Methods editing QformerMethods editing LLMUniKE
AVG Multimodal Editing Performance71.889.395.2

(3.3) Since LLM editing is merely a simpler case of MLLM editing, UniKE which successfully addresses MLLM editing, can also be effectively used for LLM editing. Following previous LLM-editing work, we use GPT-J-6B to conduct both one-step and multi-step editing on the ZsRE benchmark. As shown in Table 6 of the Rebuttal PDF, UniKE still outperforms all baselines in LLM editing.

[Please Refer to Table 6 in the Rebuttal PDF for the Results]

 

We will integrate these experiments into our paper. Thank you again for the valuable feedback!

[Chen et al 2023]Can We Edit Multimodal Large Language Models?EMNLP2023.

评论

Thanks for your response. The Rebuttal addresses most of my concerns. I will raise my score to 7.

评论

Thank you for raising the score. Your valuable suggestions greatly contribute to the quality of our manuscript. Thank you again for your precious time and valuable suggestions!

作者回复

We sincerely thank all the reviewers for their insightful and valuable comments! Overall, we are encouraged that they find that:

(1) The motivation is clear and reasonable, supported by a well-structured article. (Reviewer mp35, Reviewer w8aR, Reviewer V1xr)

(2) UniKE establishes a unified framework for intrinsic knowledge editing and external knowledge resorting, which is novel and effective. (All Reviewers)

(3) The experiments are solid and thorough, clearly demonstrating that UniKE consistently maintains excellent reliability, generality, and locality across various settings. (All Reviewers)

(4) UniKE has significant implications for further studies on constructing more powerful MLLMs, can enhance subsequent research endeavors, and is a versatile and robust solution for enhancing multimodal language models. (Reviewer mp35, Reviewer w8aR, Reviewer V1xr)

 

To address the concerns raised by the reviewers, we have conducted several additional experiments to further demonstrate the superiority of UniKE from various perspectives. And we include these experimental results in the rebuttal PDF, which contains 10 tables and 2 figures.

(1) In Tables 1-3, we perform one-step editing and cross-task editing with LLaVA 1.5 as the backbone, also conducting cross-task editing on Blip-2 OPT. The results demonstrate that UniKE is effective across various types of MLLMs.

(2) In Tables 4-5, we compare UniKE with more recent baselines, further demonstrating the proficiency of our proposed UniKE.

(3) In Table 6, we apply UniKE to LLM editing tasks, showing that UniKE can effectively address LLM editing tasks in pure NLP scenarios.

(4) In Tables 8-9, we leverage UniKE on two interesting and challenging tasks with newly constructed data (overediting evaluation and counterfactual editing), finding that UniKE minimizes the influence of overediting and effectively addresses counterfactual editing tasks.

(5) In Table 7, Table 10, and Figures 1-2, we further analyze the editing time efficiency of UniKE and provide additional ablation studies, proving the robustness of UniKE's design.

 

These experiments will be integrated into the main body or the appendix of our paper. Next, we will address each reviewer's detailed concerns point by point. We hope to address all of your concerns. Discussions are always welcome. Thank you!

最终决定

All reviewers acknowledged the novelty of this paper, as well as the solid and thorough experiments conducted. In the initial reviews, there were suggestions to further demonstrate the superiority of the proposed method from various perspectives. After rebuttal phase, the paper received positive feedback from all reviewers. AC has also carefully reviewed the paper, the reviews, and the rebuttal, and concurs with the reviewers' opinions. Therefore, AC recommends acceptance. It would be greatly beneficial to include the additional results from the rebuttal phase in the final version.