PaperHub
6.3
/10
Poster3 位审稿人
最低5最高7标准差0.9
7
7
5
3.3
置信度
COLM 2024

CoLLEGe: Concept Embedding Generation for Large Language Models

OpenReviewPDF
提交: 2024-03-23更新: 2024-08-26
TL;DR

We introduce a new method that generates concept embeddings for LLMs.

摘要

关键词
meta-learningfew-shot learningconcept learning

评审与讨论

审稿意见
7

This paper presents a framework for a concept embedding generation system. The proposed method is a rather elegant approach to learning new concept embeddings, where new concepts are learned by masking them with a frozen MLM (roBERTa), encoding example sentences mentioning them, pooling their representation, and then train an autoregressive LLM (LLama2-7b) with an augmented vocabulary of +1 new concept. Different components play an important role, such as sampling negative examples or ensuring the masked representation is as close as possible to the true (unmasked) one.

The method I found to be really neat, and the presentation and general writeup very clear. The experiments are also reasonable (verbal reasoning, definition generation and correct definition identificatino from Twitter), and show that such models could be very useful for domain-specific or slang understanding.

My main criticism is that the evaluation, while sensible, seems to be quite limited and not be well positioned with existing lexical semantics benchmarks. It seems this would be important so that we can understand, beyond comparing with GPT, etc., how this fine-tuned Llama-2 model fares with older but well tested models.

接收理由

Very nice method, interesting insight leading to the negative examples loss and knowledge distillation. This leads to flexible representations that could be used in several downstream tasks.

Very well written paper, with nice introduction and plenty of additional details in the appendices.

拒绝理由

The evaluation is a bit slim and ad-hoc. For example:

  • GRE experiment seems fine, however I was't able to find any discussion about the role of definitions, and given that increasing the number of definitions for the considered terms boosts performance significantly, it would be good to understand a bit better why. It is not clear if they all come from the same source, mixed, or if more than 4 were tested.

  • Definition generation. There is a body of work on definition generation (aka definition modeling), with standard datasets and metrics. I would have expected these works and baselines to be included (see below).

  • Twitter slang experiment seems fine, although again, there are other works and datasets that could have been referenced. Beyond this, this looks very similar to the WiC-TSV task, so I'd encourage the authors to add an experiment here and compare with appropriate baselines.

Definition generation [1] - Bevilacqua, M., Maru, M., & Navigli, R. (2020). Generationary or “how we went beyond word sense inventories and learned to gloss”. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) (pp. 7207-7221). [2] - Giulianelli, M., Luden, I., Fernández, R., & Kutuzov, A. (2023). Interpretable word sense representations via definition generation: The case of semantic change analysis. arXiv preprint arXiv:2305.11993. [3] - Mickus, T., Van Deemter, K., Constant, M., & Paperno, D. (2022). Semeval-2022 Task 1: CODWOE--Comparing Dictionaries and Word Embeddings. arXiv preprint arXiv:2205.13858.

WiC-TSV [3] - Breit, A., Revenko, A., Rezaee, K., Pilehvar, M. T., & Camacho-Collados, J. (2020). WiC-TSV: An evaluation benchmark for target sense verification of words in context. arXiv preprint arXiv:2004.15016. Mapping words to definitions and mentions

Other helpful works mapping words, definitions and examples, maybe good for swapping with the original encoder [4] - Yu, W., Zhu, C., Fang, Y., Yu, D., Wang, S., Xu, Y., ... & Jiang, M. (2021). Dict-bert: Enhancing language model pre-training with dictionary. arXiv preprint arXiv:2110.06490. [5] - Chen, P., & Zhao, Z. (2022). A unified model for reverse dictionary and definition modelling. arXiv preprint arXiv:2205.04602.

给作者的问题

Just a typo: Sect 5.1 "On an version"

作者回复

We thank the review for their helpful comments on our work and for pointing to prior work relevant to our evaluations.

The evaluation is a bit slim and ad-hoc.

All of our evaluations are done zero-shot, without any additional finetuning. We built our tasks to test a wide array of settings, including complex verbal reasoning, definition modeling, and a more natural slang setting. CoLLEGe is trained on general purpose language modeling data to enable broad transfer.

GRE experiment seems fine, however I was't able to find any discussion about the role of definitions, and given that increasing the number of definitions for the considered terms boosts performance significantly [...].

To clarify, we only present the model with one definition alongside 1–4 support sequences. For GRE, we treat each answer choice as a new concept represented with a new token.

Definition generation [...]

Thank you for the pointers to this research. With the exception of Dict-BERT, we were unaware of these, so this is very helpful. We reached out to the authors of Generationary, as suggested, but have yet to hear a response. We were able to compare our work with that of Giulianelli et. al. (2023) by generating definitions 1-shot from their Oxford dataset and replacing the target word with a new token. The results can be found here. CoLLEGe and Prompting + CoLLEGe score comparably to FLAN models trained for definition generation when the target word is replaced with a new token.

We also evaluated the FLAN models on our task, with results here. CoLLEGe outperforms in terms of ROUGE-L and performs on-par or slightly better than FLAN when evaluated with BERT-F1. Importantly, the FLAN models are trained specifically to generate definitions given an example sentence. CoLLEGe, on the other hand, does not undergo any additional training or finetuning on definition modeling and is able to perform about the same or better.

Twitter slang experiment [...]

Thank you for the suggestion. Our Twitter task is related to WiC but distinct. WiC is binary classification, but in Twitter we provide a choice between one correct and three incorrect (slang, definition) pairs. We are in the process of comparing these methods to ours on both WiC and on our Twitter task, and we plan to add these additional results during discussion.

评论

Thank you for the answers and the additional experiments, the comparison with the Flan definition generation models is insightful.

评论

Thank you for reading our comments and increasing your score.

审稿意见
7

This paper introduces a framework called CoLLEGe for learning how to embed concepts from LLMs. The framework uses a meta-learning approach and makes use of text data from pretrained LM datasets. The proposed method efficiently samples support and query sequences by using saved sentences with new tokens. The knowledge distillation process then involves matching the generated embeddings with the ground truth embeddings and logits.

接收理由

  • the paper is well-written and easy to follow
  • the proposed method is novel, and the methodology is clear
  • experiments on GRE verbal reasoning, definition generation, and slang identification show the efficiency of the proposed framework.

拒绝理由

  • Several works have been proposed recently on learning static vector representations of concepts such as MirrorBERT, Numberbach, etc. Perhaps authors should discuss these concept embeddings and use them instead of simply averaging mentions.

  • more detail about the sampling and knowledge distillation process in section 3

给作者的问题

  • Could the authors explain how the concept embedding is taken into account in the LLM to produce the definition?
  • Could the authors add more clarification about the sampling process?
作者回复

We thank the reviewer for their helpful suggestions and comments on our work.

Several works have been proposed recently on learning static vector representations of concepts such as MirrorBERT, Numberbach, etc. Perhaps authors should discuss these concept embeddings and use them instead of simply averaging mentions.

This is an interesting idea, which we may explore in future work. MirrorBERT seems like a promising embedding method and worth trying. Numberbatch embeddings may be less directly applicable, but they may be useful alongside contextual embeddings.

more detail about the sampling and knowledge distillation process in section 3

We first select words to be used as new concepts during training based on word frequency in the original training corpus, since we want words which do not appear too often, because they would be too easy to learn, or words that are too rare, because their representation would be too noisy. For each such word, we identify pairs of query and support sequences.

For distillation, we retain the query sequence with the original word form (e.g. for “My ◾is complaining I never have any time to relax”, the original would be “My beige flag is complaining I never have any time to relax”). The intuition is that CoLLEGe should represent “is complaining I never have any time to relax” in the query sequence similarly to how LLaMA would represent it in the original sequence. This encourages CoLLEGe to generate embeddings such that the output embeddings and the logits for that subsequence are close (in cosine distance and mean-squared error, respectively) to those of the LLaMA model for that same subsequence originally

Could the authors explain how the concept embedding is taken into account in the LLM to produce the definition?

We sample from the LLM conditioned on the prompt "The word '◾` is defined as'" in order to generate a definition for the new concept. CoLLEGe learns to generate an embedding for the new concept token ◾, and that embedding is projected to the input and output embedding space of the LLaMA model. The LLaMA model then makes use of those embeddings when applying the embedding the prompt (and subsequent sampled tokens) and when it computes the output logits.

Could the authors add more clarification about the sampling process?

We believe you are referring to the sampling process for the definition generation task. All of our definitions are generated using greedy decoding at temperature = 1.0.

评论

Thank you for the answers.

审稿意见
5

The paper proposed a few-shot method to enable a language model to learn new concepts on the fly.

接收理由

The problem is very interesting and it is one of the critical problems that should be solved for language models.

拒绝理由

I found the problem studied in the paper artificial. I have a few concerns:

New concepts come constantly and naturally and should be identified and learned naturally by the language model itself in the particular contexts of the concepts. Manually identifying such concepts is not practical. How do you know that the language model does not have a concept already in the system? After identifying such concepts, positive and negative examples have to be prepared in sophisticated ways in order to learn the concepts. Training even involves meta-learning, which makes the task even more difficult.

Experiments are conducted but again artificial.

The paper mainly described what the authors have done with gory details, but provided little principle.

There are two closely related areas of research that the paper has missed. The first area is model editing. A large body of literature on the topic is trying to do what you are doing. Another area is continual learning, which tries to ensure little negative interference to the language model when learning new knowledge or concepts. The following papers can give you some ideas about the existing work in these two areas.

Mitchell et al., Fast Model Editing at Scale, ICLR-2022

Wang et al. Knowledge Editing for Large Language Models: A Survey. arXiv:2310.16218 [cs.CL]

Ke et al. Continual Pre-training of Language Models. ICLR-2023.

给作者的问题

No other questions

作者回复

We thank the reviewer for their helpful comments about our work.

New concepts come constantly and naturally and should be identified and learned naturally by the language model itself in the particular contexts of the concepts.

This is an interesting direction for future research and we believe CoLLEGe is an important first step in addressing it. In order to eventually identify and learn new concepts constantly, we first develop a method which can learn identifiable new concepts on the fly.

After identifying such concepts, positive and negative examples have to be prepared in sophisticated ways in order to learn the concepts.

Our goal is to present a method which learns a new concept given only a few examples. To train our model, select words to be used as new concepts and train our model to predict query sequences using them with a few support sequences. The new concept is represented by a new token, guaranteeing that the language model has not encountered it before. Negative examples along and distillation are used to improve the generated embeddings.

Training even involves meta-learning, which makes the task even more difficult.

We choose a meta-learning objective that blends into the pretraining objective, which is not much more difficult than standard autoregressive training and allows CoLLEGe to generate task-general embeddings with rich semantic knowledge.

How do you know that the language model does not have a concept already in the system? Experiments are conducted but again artificial.

In our Twitter evaluation, we hand-select very recent slang terms. Some incorrect answers to questions about slang terms generated by ChatGPT-3.5 and LLaMA-2 can be seen here. Additionally, our Twitter and GRE evaluations contain multi-word phrases as single concepts, which LLaMA-2 model has not seen. Finally, if LLaMA could simply make use of its pretraining knowledge, then the baselines would not perform so poorly.

The first area is model editing. Another area is continual learning, [...]

Thank you for these pointers. We intend to cite them as relevant background, but do not believe they are directly relevant. Applying model editing to the new concept learning setting would mean manually editing the LM for each new concept, which seems more artificial and sidesteps our research problem. Also, our problem is not continually adapting to new domains or tasks, like in Ke et. al.

评论

I don't agree with your reply to my final point. I think they are very relevant. Model editing is not done manually. Some model editing methods also use positive and negative examples. Your method has to deal with forgetting as in continual learning or model editing.

Thanks for responding to my comments. Since you have addressed some of my concerns, I increased my score to 5.

评论

Thank you for your response and for increasing your score, as well as for suggesting we compare with Knowledge Editing methods. To address your concerns, we evaluated some of these methods on our GRE task. Here, we use the definition of the new concept for editing and also present support sequences in-context. IKE [1] uses in-context learning for knowledge editing, whereas ROME [2] and MEND [3] update the model weights. To train MEND, we use the zsRE dataset and we use the implementations provided in EasyEdit [4]. In the first table below, we provide only support sequences in-context and in the second we also provide definitions.

1 shot2 shots3 shots4 shots
CoLLEGe35.0 ± 5.640.5 ± 4.342.7 ± 3.944.5 ± 3.9
Prompting13.9 ± 4.317.7 ± 3.019.8 ± 4.221.8 ± 3.0
Prompting + MEND13.4 ± 5.010.1 ± 2.914.7 ± 5.419.4 ± 1.9
Prompting + IKE18.1 ± 2.722.3 ± 3.123.2 ± 3.227.9 ± 5.9
Prompting + ROME14.9 ± 5.216.2 ± 3.216.7 ± 5.322.3 ± 1.8
Def. + 1 shotDef. + 2 shotsDef. + 3 shotsDef. + 4 shots
CoLLEGe42.5 ± 4.346.8 ± 3.745.2 ± 3.249.3 ± 1.5
Prompting21.6 ± 5.720.5 ± 6.819.3 ± 4.824.0 ± 4.7
Prompting + MEND23.3 ± 5.010.8 ± 4.016.3 ± 1.920.2 ± 2.2
Prompting + IKE20.0 ± 6.217.2 ± 5.217.6 ± 2.323.7± 5.8
Prompting + ROME21.9 ± 6.718.6 ± 5.114.4 ± 4.723.2 ± 3.2

We find that knowledge editing methods score better than our Prompting baseline in some cases, when the definition is not provided in-context, but around the same or slightly worse when it is provided. In either case, CoLLEGe achieves much higher accuracy than all KE methods.

We also tried using the support sequences as well as the definition during editing. Unfortunately, the time it took to evaluate ROME in this setting was much longer, and we were only able to complete 2 evaluation runs (as opposed to 5 above). Nonetheless, we found that performance was sometimes a bit worse than if we just use the definition for editing. As above, in the first table we provide only support sequences in-context and in the second we also provide definitions. In both cases, definitions and support sequences are used for editing.

Method1 shot2 shots3 shots4 shots
CoLLEGe35.0 ± 5.640.5 ± 4.342.7 ± 3.944.5 ± 3.9
Prompting13.9 ± 4.317.7 ± 3.019.8 ± 4.221.8 ± 3.0
Prompting + ROME10.4 ± 1.114.0 ± 2.312.8 ± 1.220.9 ± 0.0
MethodDef. + 1 shotDef. + 2 shotsDef. + 3 shotsDef. + 4 shots
CoLLEGe42.5 ± 4.346.8 ± 3.745.2 ± 3.249.3 ± 1.5
Prompting21.6 ± 5.720.5 ± 6.819.3 ± 4.824.0 ± 4.7
Prompting + ROME26.7 ± 1.216.3 ± 4.711.6 ± 2.322.1 ± 3.5

We find that ROME performs worse than Prompting in most cases, and each achieves lower accuracy than CoLLEGe.

[1] Zheng, Ce, et al. "Can We Edit Factual Knowledge by In-Context Learning?." Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.

[2] Meng, Kevin, et al. "Locating and editing factual associations in GPT." Advances in Neural Information Processing Systems 35 (2022): 17359-17372.

[3] Mitchell, Eric, et al. "Fast model editing at scale." arXiv preprint arXiv:2110.11309 (2021).

[4] Wang, Peng, et al. "Easyedit: An easy-to-use knowledge editing framework for large language models." arXiv preprint arXiv:2308.07269 (2023).

评论

During the evaluation, we noticed that KE methods took much longer to evaluate than CoLLEGe. To quantify this, we computed the average amount of time each method took on each GRE evaluation example, which we present below. CoLLEGe and Prompting + CoLLEGe are 7x and 3x, respectively, faster than the fastest KE method, MEND. CoLLEGe is also 50x faster than the slowest method, IKE. One caveat is that we are using the implementation of these methods from EasyEdit and did not separately optimize their efficiency.

MethodAvg. Time Elapsed Per Example (s) ↓
CoLLEGe0.51
Prompting + CoLLEGe0.96
Prompting + MEND3.59
Prompting + ROMEO14.0
Prompting + IKE25.8
最终决定

This paper proposes a new method for large language models to learn new concepts on the fly to overcome current methods such as prompt-based in-context learning and few-shot fine-tuning. For achieving this, the authors design a meta-learning approach for flexible embeddings for new coming concepts.

Among three reviewers, two reviewers voted to acceptance score (7) and one reviewer gave a rejection (4) initially. All reviewers agree that interesting setup, the importance of the task that this paper addresses, novel methods, and good writing. The main concerns were less natural setup and lack of description on related research area.

During the rebuttal period, the authors tried to address the concerns and it seems that some of them were alleviated.

AC carefully read the paper, the review comments, and the authors' feedback.

Considering the discussion comments as well, AC agree the contribution of this paper, and thus recommends accepting this paper.

AC asks the authors reflect valuable comments of the reviewers in the final camera-ready version.