PaperHub
5.0
/10
Poster4 位审稿人
最低3最高8标准差2.1
3
3
8
6
3.8
置信度
正确性2.8
贡献度2.5
表达3.3
ICLR 2025

Vision and Language Synergy for Rehearsal Free Continual Learning

OpenReviewPDF
提交: 2024-09-26更新: 2025-03-02

摘要

关键词
continual learningprompt dilemmalanguage descriptorsprompt generatorcatasthropic forgetting

评审与讨论

审稿意见
3

This paper introduces a new prompt-based approach for continual learning that take advantage of language guidance. It uses task-wise generators, soft task ID prediction, and generated prompts as auxiliary data. The method outperforms state-of-the-art models on CIFAR100, ImageNetR, and CUB datasets with significant improvements in accuracy and forgetting measures.

优点

  1. The paper provides a detailed explanation of current prompt-based continual learning methods and analyzes the weakness associated with these methods, effectively leading to the introduction of the authors' proposed method.
  2. The paper presents a novel language-guided continual learning method and thoroughly demonstrates its effectiveness through extensive experiments.

缺点

  1. The comparison mentioned in lines 183-184 between "the photo of car" and "the photo of cat," stating they have 15/16 similarity but 1/16 dissimilarity, is confusing. The authors compared the similarity between these two prompts on a letter-by-letter basis and concluded that they are similar. However, the similarity between two prompts should not be evaluated in such a manner. After being processed by the tokenizer, the embeddings of the two prompts are not similar, as their meanings differ significantly.
  2. The proposed method relies on using ChatGPT to generate descriptive terms for each class and requires an additional Sentence Transformer to obtain embeddings. The introduction of these extra resources creates an unfair comparison with other methods and limits the practical applicability of the approach in real-world scenarios.
  3. Although the proposed method achieves state-of-the-art performance on multiple datasets, the improvement is minor on some, such as the comparison with HiDe-Prompt on the CUB200 dataset. This is particularly relevant given the substantial additional resources, like ChatGPT, required by the method, which other approaches do not utilize. Moreover, this suggests that the performance gains on fine-grained datasets may not be significant, as the generated descriptive terms for each class are quite similar.
  4. The authors have used excessive line spacing adjustments, which negatively impact the visual presentation of the paper, such as in lines 437-438 and 446-447. They should revise the layout of the entire paper to provide a better reading experience for the reader.

问题

Please the weakness.

评论

Q1/W1. The comparison mentioned in lines 183-184 between "the photo of car" and "the photo of cat," stating they have 15/16 similarity but 1/16 dissimilarity, is confusing. The authors compared the similarity between these two prompts on a letter-by-letter basis and concluded that they are similar. However, the similarity between two prompts should not be evaluated in such a manner. After being processed by the tokenizer, the embeddings of the two prompts are not similar, as their meanings differ significantly.

Our Response: Thank you for your concern.

  • About the writing error in lines 183-184. We apologize for our writing error. We have revised the sentences in our revised manuscript. The sentences are revised into: "LGCL produces class prototype LcnL_c^n by encoding string ”the photo of class name”. However, the prototypes could be misleading due to high similarity between different classes, e.g. the prototype of class ”Great White Shark” has 0.9 cosine similarity to the prototypes of class ”Tree Frog” and ”Iguana”, please see Appendix D.2. "

  • Our statement is supported by our empirical study on the high similarity of language embedding encoded by CLIP as presented in Appendix D.2.

Q2/W2. The proposed method relies on using ChatGPT to generate descriptive terms for each class and requires an additional Sentence Transformer to obtain embeddings. The introduction of these extra resources creates an unfair comparison with other methods and limits the practical applicability of the approach in real-world scenarios.

评论

Q3/W3. Although the proposed method achieves state-of-the-art performance on multiple datasets, the improvement is minor on some, such as the comparison with HiDe-Prompt on the CUB200 dataset. This is particularly relevant given the substantial additional resources, like ChatGPT, required by the method, which other approaches do not utilize. Moreover, this suggests that the performance gains on fine-grained datasets may not be significant, as the generated descriptive terms for each class are quite similar.

Our Response: Thank you for your concern.

  • We added a new experiment and analysis on the "class name" as descriptors where we have no class descriptors generated by GPT or other LLMs. Thus we utilize equal resources as language-guided prompting such as LGCL and even less resource than ConvPrompt and GMM. Our experiment results as presented in Appendix D4 shows that our method outperforms the existing SOTAs by a significant margin in 3 datasets despite not using the descriptors by LLMs. It proves that our method can substitute class descriptors with the class name to generate prompts.

  • Regarding comparison on CUB dataset, It may look small difference between our method and HiDe-prompt in CUB dataset if we see from final task measurements i.e. 1.24% FAA and 0.6% FFM. However, in terms of cumulative measurements, our method outperforms HiDe-Prompt with a significant margin i.e. 3.3% CAA and 1.36% CFM. The table A19 in Appendix G shows the complete numerical result of the CUB dataset. The table shows that except in the last 2 tasks, LEAPGen outperforms HiDe-Prompt by a significant margin i.e. 2.8-5.5%. In addition, HiDE-Prompt suffers from a significant drop after the first task i.e. 6% accuracy drop, while LEAPGen only experiences a 3% drop. Thus, in earlier tasks, LEAPGen outperforms HiDePrompt with more than a 5% margin. In terms of forgetting measures, LEAPGen consistently experiences lower average forgetting than HiDe-Prompt in every task with a 1-2.5% margin. In the earlier tasks, It is even close to zero. Those numerical result shows that LEAPGen outperforms HiDE-Prompt in CUB dataset.

  • Please see the running time analysis in Table A9 in Appendix D9 that shows HiDe-Prompt requires 12x higher running time than our proposed method (3.94h vs 0.32h) for the CUB dataset. This drawback is important to be considered along with the comparison of the performance.

  • As for the fine-grained dataset, we confirm that ImageNet-R also has fine-grained classes i.e.:

a. Subtypes of Dog: ['west highland white terrier', 'golden retriever', 'labrador retriever', 'cocker spaniels', 'collie', 'border collie', 'rottweiler', 'german shepherd dog', 'boxer', 'french bulldog' 'saint bernard', 'husky', 'dalmatian', 'pug', 'pomeranian', 'Chow Chow', 'pembroke welsh corgi', 'toy poodle', 'standard poodle']

b. Subtypes of fish: ['goldfish', 'great white shark', 'hammerhead','stingray', 'clown fish', 'puffer fish', 'grey whale', 'killer whale']

c. Subtypes of bird: ['hen', 'ostrich', 'goldfinch', 'junco', 'bald eagle', 'vulture,'peacock', 'lorikeet', 'hummingbird', 'toucan', 'duck', 'goose', 'black swan','flamingo', 'american egret', 'pelican']

The result in ImageNet-R as shown in Table 2 shows that our method works well in spite of the presence of fine-grained classes. Our method outperforms existing SOTA with a significant margin i.e. 12-30% FAA and 8-24% CAA in 20-task settings, 7-22% FAA and 4-18% CAA in 10-task settings, and 3-18% FAA and 2-17% CAA in 5-task settings. These numerical results given the fact that Imagenet-R contains fine-grained classes prove that fine-grained classes don't decrease the significance of our method.

Q4/W4. The authors have used excessive line spacing adjustments, which negatively impact the visual presentation of the paper, such as in lines 437-438 and 446-447. They should revise the layout of the entire paper to provide a better reading experience for the reader.

Our Response: Thank you for your suggestion. We have ensured the the clear spacing for the mentioned parts and other components to improve the visualization of our paper. Please kindly see our latest manuscript.

评论

Thanks for your reply. Would you please showcase latency metrics such as inference speed, fps, storage and latency? I am curious about it given that additional ChatGPT are utilized and it plays a crucial role in the pipeline. I do not see them explicitely in the reply and updated manuscripts

评论

Dear Reviewer bACr,

We would like to follow up on our responses. Could you please kindly review our updated manuscript or previous comments regarding your concerns?

We believe we have addressed all your concerns in our updated paper and previous responses (comments). Thus, we kindly request Reviewer bACr to reevaluate the score for our paper.

Thank you.

Best Regards,

Author of Submission 6684

评论

Thanks for your response. Answer to the first question seems to be the primary concern and I still find it not convincing.

As for the utilization of GPT, I understand what the author means, but the actual storage and consumption should not be dismissed even if it is pre-computed. Since GPT is more like an external knowledge base, I'm not sure if it is fair to compare it with existing methods. More substantial improvements are supposed to observe since the introduction of these large-scale models.

评论

We would like to follow up on our response and revised manuscript. We would appreciate it if Reviewer bACr could look at our revised manuscript, and offer additional comments.

评论

Thank you for your suggestion.

  1. We have added the advised metrics i.e. inference time only, along with total time, including additional time for descriptors generation as utilized by LEAPGen and ConvPrompt. Our method has the lowest total running time, and moderate inference time i.e. lower than ConvPrompt and higher than HiDe-Prompt. We have added the required storage, please see Appendix D.3.

  2. We would like to emphasize the following points:

  • Following ConvPrompt, the LLM such as GPT is one of the alternatives to generate descriptors before the training phase. Thus the GPT is not part of the methods. In the case of no classes' descriptors, Our method still works excellently i.e. utilizing class names as descriptors (Appendix D.4).

  • GPT is utilized via online query, thus we don't need extra storage to save it.

  • Descriptors generation by GPT indeed consumes a fair amount of time. Even though spending additional time for descriptors generations, our method still has a lower total running time than HiDe-Prompt and ConvPrompt.

评论

Dear Reviewer bACr,

Thank you for your feedback on our paper. Could you please kindly review our latest manuscript regarding your concern? Here are the pointers to your points on the use of ChatGPT:

  • Section 5.2.d Tables 2,3,4, and Figure 3: Analysis of LEAPGen-lite (lightweight generators and without descriptors) performance that proves the significant performance of our ideas is not bounded by LLM-generated descriptors or the complexity (parameters and size) of the generators.

  • Section 5.2.g: Analysis of parameters, running time, and storage

  • Appendix D.3: Detailed Analysis of Performance and Cost TradeOffs.

Thank you.

Best Regards,

Author of Submission 6684

评论

W2. Increase of Parameters:

  • Section 5.2.d Tables 2,3,4, and Figure 3: Analysis of LEAPGen-lite (lightweight generators and without descriptors) performance.

5.2.d) LEAPGen-lite’s Performance: As shown in Table 2-4 and Figure 3 Despite utilizing far smaller (2.67% params) generators and without class descriptors generated by LLM, LEAPGen-lite still outperforms the existing methods significantly i.e. 4.7-25% FAA and 3-11% CAA in CIFAR100, and 3-30% FAA and 2-26% CAA in ImageNet-R dataset. LEAPGen-lite also achieves a low for- getting rate in these 2 datasets for all task-settings. In the CUB dataset, LEAPGen-lite archives a comparable performance to HiDe-Prompt and outperforms other SOTA with a significant margin i.e. 6-20% FAA and 3.3-14% CAA. This evidence proves our ideas i.e. language embedding as input for prompt generation, task-wise generators, soft task-id predictor, and learning with auxiliary data are not bounded by the generated descriptors and the size of generators.

Here is the snapshot of the tables, please see our revised paper for the full tables.

MethodImageNet-R
FAACAAFFMCFM
5 Tasks @40c
L2P64.62 ± 0.3268.01 ± 0.423.94 ± 0.163.55 ± 0.20
DualPrompt69.71 ± 0.1172.78 ± 0.143.32 ± 0.162.78 ± 0.25
CODA-P74.89 ± 0.3679.71 ± 1.278.89 ± 0.657.65 ± 0.98
LGCL69.93 ± 0.2172.91 ± 0.193.04 ± 0.362.50 ± 0.38
HiDe-Prompt75.40 ± 0.2778.88 ± 0.043.15 ± 0.462.64 ± 0.16
PGP69.71 ± 0.1572.77 ± 0.073.36 ± 0.232.85 ± 0.25
EvoPrompt77.27 ± 0.4081.67 ± 0.181.79 ± 0.311.41 ± 0.32
CPrompt78.65 ± 0.0082.44 ± 0.006.00 ± 0.005.49 ± 0.00
ConvPrompt79.36 ± 0.0882.93 ± 0.243.42 ± 0.052.36 ± 0.16
LEAPGen-lite82.44 ± 0.6384.37 ± 0.900.43 ± 0.080.17 ± 0.06
LEAPGen82.79 ± 0.3285.06 ± 0.290.51 ± 0.040.18 ± 0.07
ImageNet-R
L2P62.50 ± 0.5167.05 ± 0.475.01 ± 0.404.41 ± 0.43
DualPrompt68.59 ± 0.2472.18 ± 0.204.61 ± 0.073.70 ± 0.18
CODA-P73.77 ± 0.5079.38 ± 1.487.94 ± 0.086.72 ± 0.79
LGCL68.65 ± 0.2572.57 ± 0.194.75 ± 0.333.38 ± 0.58
HiDe-Prompt75.75 ± 0.4079.27 ± 0.172.29 ± 0.272.33 ± 0.17
PGP68.62 ± 0.1472.19 ± 0.204.53 ± 0.403.63 ± 0.35
EvoPrompt76.00 ± 0.2680.97 ± 0.304.22 ± 0.423.59 ± 0.52
CPrompt76.32 ± 0.5381.50 ± 0.306.10 ± 0.755.60 ± 1.35
ConvPrompt77.08 ± 0.2681.47 ± 0.104.17 ± 0.043.11 ± 0.17
LEAPGen-lite82.38 ± 1.0485.14 ± 0.523.01 ± 1.192.13 ± 0.60
LEAPGen84.09 ± 0.9385.54 ± 0.651.46 ± 1.252.11 ± 1.21
ImageNet-R
L2P57.40 ± 0.3163.33 ± 0.2110.76 ± 0.457.88 ± 0.17
DualPrompt65.19 ± 0.1770.31 ± 0.297.30 ± 0.185.16 ± 0.34
CODA-P70.55 ± 0.7177.08 ± 1.028.23 ± 0.866.95 ± 0.70
LGCL64.96 ± 0.6770.18 ± 0.377.35 ± 0.655.05 ± 0.32
HiDe-Prompt-81.60 ± 0.48-2.23 ± 0.38
PGP65.24 ± 0.2570.36 ± 0.267.17 ± 0.215.09 ± 0.25
EvoPrompt74.93 ± 0.6479.92 ± 0.136.72 ± 0.905.67 ± 0.26
CPrompt74.23 ± 0.1779.82 ± 0.515.98 ± 0.245.54 ± 0.48
ConvPrompt73.93 ± 0.3678.92 ± 0.374.87 ± 0.573.57 ± 0.25
LEAPGen-lite83.67 ± 0.3985.65 ± 0.331.06 ± 0.240.47± 0.14
LEAPGen87.03 ± 0.1287.81 ± 0.482.17 ± 0.172.54 ± 0.77

Please Note: The analysis of LEAGen-lite empirically proves that our method doesn't rely on GPT/LLM or complex generator networks (high number of parameters)

Thank you.

Best Regards,

Author of Submis.6684

评论

Dear Reviewer bACr,

Thank you for your feedback on our paper. Could you please kindly review our latest manuscript regarding your concern?

Here are the pointers to your key points:

We also copied the updates (or snapshots) here, thus you can review them directly.

  • Section 5.2.g: Analysis of parameters, running time, and storage
MethodDesc#Params(M)Running Time (h)Storage (MB)
T.DescInfTr+InfTotal
HiDe-Prompt-0.15-0.0195.405.40334
ConvPromptv1.281.070.0331.042.11346
LEAPGen-lite-0.16-0.0280.530.53332
LEAPGenv6.351.070.0250.721.79567

T.Desc, Tr, and Inf denote time for generating descriptors, training, and inference respectively, detailed in Appendix D3.

5.2.g) Parameters, Running Time, and Storage: The table above compares the number of parameters, running time (ImageNet-R), and storage of our methods and existing SOTAs. Despite having a higher number of parameters and storage, LEAPGen consumes less running time than existing SOTAs both training+inference and total running time. LEAPGen-lite consumes the least costs in total running time and storage and requires relatively low parameters and inference time. LEAPGen and ConvPrompt require additional time to generate descriptors that increase their total simulation time. Despite having the least parameters, Hide-Prompt requires the longest training and total times since it needs extra operations to generate uninstructed class representations.

  • Appendix D.3: Detailed Analysis of Performance and Cost TradeOffs.

(a) Running Time for All Datasets

MethodTime (h)
CIFAR100 10TImageNet-R 10TCUB 10T
T.Desc.InfTr+InfTotalT.Desc.InfTr+InfTotalT.Desc.InfTr+InfTotal
HiDe-Prompt-0.0374.634.63-0.0195.405.40-0.0173.943.94
ConvPrompt0.530.0232.012.541.070.0331.042.111.090.0358.089.17
LEAPGen-lite-0.0161.21.2-0.0280.530.53-0.0280.380.38
LEAPGen0.530.0170.981.511.070.0250.721.791.090.0250.321.41

(a) Performance vs Storage.

MethodPerformanceStorage(MB)
FAACFA
HiDe-Prompt75.7579.27334
HiDe-Prompt-Large74.3078.56797
ConvPrompt77.0881.47346
ConvPrompt-Large74.56*80.79632
LEPGen-lite82.3885.14332
LEAPGen84.0985.54567

HiDe-Prompt-Large and ConvPrompt-Large is the large version of HiDe-Prompt and ConvPrompt that have 12.2M and 13.3M parameters respectively.

Again, we want to emphasize these points:

  • Following ConvPrompt, the LLM such as GPT is one of the alternatives to generate descriptors before the training phase (Please see figure 2 and section 4.2-4.3). Thus the GPT is not part of the methods. In the case of no classes' descriptors, Our method works excellently i.e. utilizing class names as descriptors as demonstrated by LEAPGen-lite and LEAPGen-CN(Appendix D.4).

  • GPT is utilized via online query, thus we don't need extra storage to save it.

评论

Thank you for your response.

Please kindly note that:

  • LEAPGen utilizes the same setting and resources (text encoder, GPT-generated descriptors, etc) as ConvPrompt.

  • LEAPGen-lite utilizes even less resources (without GPT-generated descriptors, fewer parameters, less running time, and less storage) that are comparable to the other SOTAs.

Considering such comparable resources (LEAPGen vs ConvPrompt, and LEAPGen-lite vs other SOTAs), we believe we have satisfied the fairness aspect.

Also please kindly note that:

  • Descriptors are optional in our method, our method works excellently and significantly outperforms existing SOTAs by utilizing class names as language modality. Thus, our method doesn't rely on GPT or LLMs.
  • GPT is utilized via online query so we don't save the GPT model. It adds fair additional running time, but not storage. Again, LEAPGen and ConvPrompt are evaluated with the same setting. Online querying GPT utilizing API and Python script is a common practice. We don't need to download and save the GPT model. This is applicable in real-application as in our simulation.

Please go thoughtfully on our revised manuscripts (mainly sec. 5.2.g. and Appendix D.3) and our previous responses (comments).

In summary, once again we emphasize that the impact of our method is not bounded by GPT, parameters, running time, and storage.

Please let us know which part you are not sure about, Thank you.

Best Regards,

Author of Submission 6684

评论

Dear Reviewer bACr,

We would like to follow up on our responses. Could you please kindly review our latest response that detailed in our updated manuscript? Could you please point out which part you are not sure regarding the fairness aspect and GPT-generated descriptors?

We believe we have addressed all your concerns in our updated paper and previous responses (comments). Thus, we kindly request Reviewer bACr to reevaluate the score for our paper.

Thank you.

Best Regards,

Author of Submission 6684

评论

Dear Reviewer bACr,

First, we thank reviewer bACr for your effort and time to review our work, rebuttal, and revised paper. We believe we have addressed all your concerns in our revised paper and previous responses (comments) Second, since the discussion phase has ended, We would like to summarize our discussion and emphasize the following points:

  1. Fairness Setting : LEAPGen utilizes the same resources as ConvPrompt and LEAPGen-lite(no descriptors with lightweight generators) utilizes less resources than ConvPrompt and comparable resources to existing SOTAs (sec. 5.1, sec. 5.2.g, Appendix E). All methods are evaluated in the same setting and run with respective best (recommended by official paper/code) hyperparameters. Thus we believe we have satisfied the fairness aspect in our study.

  2. The presence of GPT :

  • GPT is not part of our method (sec 4.2.a, and fig. 2), It is an optional part to generate descriptors following ConvPrompt. Even without GPT-generated descriptors. our method works excellently and significantly outperforms the existing SOTAs, as demonstrated by LEAPGen-lite (sec. 5.2.d) and LEAPGen-CN (Appendix D.4).

  • GPT is utilized in an online way (through the Internet), by querying it using API and Python script. It is a common practice nowadays as in our simulation (please see https://platform.openai.com/docs/guides/). Thus we never download and save the GPT model in our storage. LEAPGen indeed saves the descriptors embedding (numerical vector), but not the GPT model. The embedding adds to the whole model's storage size but is still in a reasonable amount, and we have included it in our storage analysis.

  • Descriptor generation by GPT (before training) indeed takes a fair amount of time. However, despite spending extra time on it, our method still has a lower total running time than ConvPrompt and HidePrompt.

  • LEAPGen-lite proves the significant impacts of our ideas i.e. (1) language as input for prompt generation, (2) task-wise generators, (3) soft task-id predictor, and (4)learning with auxiliary despite without GPT-generated descriptors, and utilize least parameters, running time and storage.

We believe, these highlights clarify your uncertainties about the fairness setting and (again) prove the significant impacts of our ideas, without being bound to the presence of GPT.

Best Regards,

Author of Submission 6684

审稿意见
3

This paper presents a novel approach to address the catastrophic forgetting in continual learning through a prompt-based structure. This method incorporates language inputs for prompt generation and utilizes task-wise generators and soft task-ID prediction. The authors highlight the advantages over existing methods across various datasets, showcasing substantial improvements accuracy while minimizing forgetting metrics.

优点

  • The authors systematically review the limitations of existing prompt-based methods, and Figure 1 commendably illustrates the comparison of these methods.

  • The proposed method demonstrates performance that far exceeds current methods, which is pleasantly surprising.

  • The paper introduces the novel and meaningful use of large language models to generate more refined descriptions.

缺点

  • The writing of this paper needs significant improvement. Additionally, the structure is overly compact, affecting readability. For instance, the theorem presented in the paper has weak relevance to the proposed method, and I believe it is unnecessary. I suggest moving it to the supplementary materials. Given the complexity and sophistication of the method, I encourage the authors to provide an overview initially.

  • The provided ablation study is not sufficiently convincing. I recommend that the authors validate the effectiveness of the proposed modules step-by-step starting from a standard baseline.

  • The inclusion of the ChatGPT model raises the question of how to ensure the fairness of experimental comparisons under such settings, and to what extent the performance gains are due to more accurate descriptions.

  • Considering the authors (are about to) open-source their code, I suggest making the descriptions obtained via ChatGPT publicly available for deeper analysis. I believe this will enhance the impact of the paper.

问题

Please see the comments of weaknesses.

评论

W1/Q1. The writing of this paper needs significant improvement. Additionally, the structure is overly compact, affecting readability. For instance, the theorem presented in the paper has weak relevance to the proposed method, and I believe it is unnecessary. I suggest moving it to the supplementary materials. Given the complexity and sophistication of the method, I encourage the authors to provide an overview initially.

Our Response: Thank you for your suggestion. We have improved the readability of our paper by adding an overview of our method as presented in section 4.1, moving theorems 2 and 3 into the Appendix section, and improving the general layout of our paper

W2/Q2. The provided ablation study is not sufficiently convincing. I recommend that the authors validate the effectiveness of the proposed modules step-by-step starting from a standard baseline.

Our Response: Thank you for your suggestion. We have revised our ablation study with a step-by-step scenario from fine-tuning (FT) baseline, emphasizing the clarity of each component contribution, Please see section 5.2.d.

W3/Q3. The inclusion of the ChatGPT model raises the question of how to ensure the fairness of experimental comparisons under such settings, and to what extent the performance gains are due to more accurate descriptions.

Our Responses:

  • We added a new experiment and analysis on the "class name" as descriptors where we have no class descriptors generated by GPT or other LLMs. Our experiment results as presented in Appendix D4. shows that our method outperforms the existing SOTAs by a significant margin in 3 datasets despite substituting class descriptors with the class name.

  • In our main experimental setting, We follow ConvPrompt which utilizes GPT to generate class descriptors before the training process. Thus we emphasize that our method and ConvPrompt utilize the same external resources that are considered as fair. Please note that in our method, the GPT (LLM) is solely utilized to generate descriptors, It is different from GMM which utilizes an LLM decoder in its learning process.

  • For the other methods, we can't enforce ChatGPT in them as they were designed to learn without language descriptors. The most feasible setting is evaluating the methods following their official (optimal) setting and hyperparameters that we have done in our study. Thus we believe that we maintain fairness in our evaluation.

  • As for the relation between descriptors and performance, we have extended our analysis about the performance of our method w.r.t types of descriptors and LLM (GPT, Gemini, and Llama) in comparison to ConvPrompt. The table below shows the performance of our method compared to ConvPrompt in those settings. Please kindly see Appendix D.5 for a detailed sample of descriptors, results, and analysis. In summary, we conclude that long descriptors are preferable over short and narrative descriptors as generally they carry richer visual descriptions of an object without being contaminated by unrelated words. However, our methods can utilize three of the descriptors types, and gain better performance than ConvPrompt.

W4/Q4. Considering the authors (are about to) open-source their code, I suggest making the descriptions obtained via ChatGPT publicly available for deeper analysis. I believe this will enhance the impact of the paper.

Our Response: Thank you for your suggestion, We have published all the descriptors including short, long, and narrative descriptors generated by ChatGPT, LLama, and Gemini. As for now, the descriptors can be accessed via https://anonymous.4open.science/r/xt124j05/descriptors/.

评论

We would like to follow up on our response and revised manuscript. We would appreciate it if Reviewer UGi2 could look at our revised manuscript, and we offer additional comments.

评论

Dear Reviewer UGi2,

Could you kindly review the rebuttal thoroughly and let us know whether the authors have adequately addressed the issues raised or if you have any further questions.

Best,

AC of Submission6684

评论

Dear Reviewer UGi2,

We would like to follow up on our response. Could you please kindly review our updated manuscript regarding your concerns?

Here are the pointers to your key points:

W1. Proposed Method Overview: Section 4.1

W1. Presentation Improvement: Please check our latest manuscript.

W2. Step-by-step Ablation: Section 5.2.e.

W3. The use of GPT:

  • LEAPGen-lite (lightweight generators and without descriptors) performance analysis in Section 5.2.d, Tables 2,3,4, and Figure 3 that proves the significant performance of our ideas is not bounded by LLM-generated descriptors or the complexity (parameters and size) of the generators.

  • Running time analysis in 5.2.g.

  • Detailed setting in Appendix E, LEAPGen and ConvPrompt use the same setting

W4. Open the descriptors: Please check https://anonymous.4open.science/r/xt124j05/descriptors/.

Thank you.

Best Regards,

Author of Submission 6684

评论

W2. Step-by-step Ablation: Section 5.2.e.

ComponentLossFAACAAFFMCFM
FTL\mathcal{L}_intra61.5365.315.875.62
FT+EL\mathcal{L}_intra72.9378.545.785.30
FT+E+KL\mathcal{L}_intra +Lt\mathcal{L}_t74.8879.522.843.17
FT+G(E)+KL\mathcal{L}_intra +Lt\mathcal{L}_t76.6480.481.852.44
FT+G(E)+K+LL\mathcal{L}_intra +Lt\mathcal{L}_t+Lc\mathcal{L}_c77.2381.050.501.27
FT+G(E)+K+L+AuxL\mathcal{L}_intra +Lt\mathcal{L}_t+Lc\mathcal{L}_c78.3881.980.491.15
FT+G(E)+K+L+AuxL\mathcal{L}_intra +L\mathcal{L}_inter +Lt\mathcal{L}_t+Lc\mathcal{L}_c83.7085.982.461.93
FT+G(E)+K+L+Aux+ Soft Task-ID PredictorL\mathcal{L}_intra +L\mathcal{L}_inter+Lt\mathcal{L}_t+Lc\mathcal{L}_c84.7386.140.911.42

e) Ablation Study: yable above shows the impacts of the proposed method's components on its performance. ** (1) Descriptor Embedding EE** as one of LEAPGen's main components, elevates fine tuning (FT) baseline significantly i.e. >11%>11\% FAA and >13%>13\% CAA. It shows the promising impact of our idea . (2) Task Key KtK^t and Loss Lt\mathcal{L}_t transform the method into task-wise prompting based on input to KtK^t and Lt\mathcal{L}_t similarity. They produce a better model with 1-2% FAA and CAA improvement and reduce FFM and CFM by >2%>2\%. It shows the effectiveness of task-wise decomposition as applied in the task-wise prompt approach. (3) Task-wise generator Gt()G^t() enhances the model recognition capability by 2%2\% and 1%1\% for FAA and CAA respectively. It also decreases its forgetting rate by 1%. It means the trainable generator produces a more discriminative prompt than the descriptor embedding only. (4) LctL^t_c Class-wise Key LctL^t_c and Loss Lc\mathcal{L}_c for top-k embedding selection i.e. EiE_i for i[1...k]i \in [1...k] improves the model performance with 1% margin. It also reduces the model forgetting rate with up to 1% margin. Thus, measuring the input similarity to LctL^t_c offers a better embedding selection than toEctE^t_c directly. (5) Auxiliary} embedding improves the LEAPGen's performance with a more than 1%1\% margin. It shows the secondary contribution of descriptors embedding as an auxiliary modality along with its primary role in prompt generation. 6) Inter-tasks loss L\mathcal{L}_inter improves LEAPGen accuracy significantly i.e. 5.4%5.4\% and 4.0%4.0\% for FAA and CAA respectively even though it increases FFM and CFM by 0.52%0.5-2\%. It contributes to balancing the knowledge of previously learned classes (stability) and currently learned classes (plasticity), and emphasizes the risk of forgetting previously learned classes. Soft task-id predictor} substitution to conventional task-id (as in DualPrompt) improves the accuracy with up to 1%1\% and reduces FFM and CFM by 1.5%1.5\% and 0.5%0.5\% respectively. It implies that our designed task-id predictor outperforms the existing task-id predictor both in accuracy and forgetting.

Thank you.

Best Regards,

Author of Submission 6684

评论
  • Running time analysis in 5.2.g.
MethodDesc#Params(M)Running Time (h)Storage (MB)
T.DescInfTr+InfTotal
HiDe-Prompt-0.15-0.0195.405.40334
ConvPromptv1.281.070.0331.042.11346
LEAPGen-lite-0.16-0.0280.530.53332
LEAPGenv6.351.070.0250.721.79567

T.Desc, Tr, and Inf denote time for generating descriptors, training, and inference respectively, detailed in Appendix D3.

5.2.g) Parameters, Running Time, and Storage: The table above compares the number of parameters, running time (ImageNet-R), and storage of our methods and existing SOTAs. Despite having a higher number of parameters and storage, LEAPGen consumes less running time than existing SOTAs both training+inference and total running time. LEAPGen-lite consumes the least costs in total running time and storage and requires relatively low parameters and inference time. LEAPGen and ConvPrompt require additional time to generate descriptors that increase their total simulation time. Despite having the least parameters, Hide-Prompt requires the longest training and total times since it needs extra operations to generate uninstructed class representations.

  • Detailed setting in Appendix E, LEAPGen and ConvPrompt use the same setting

E DETAILED E XPERIMENTAL S ETTING

Existing Methods: The existing SOTAs are run by executing the official implementation (code) of the respective methods. The hyper-parameters setting is chosen based on the official setting. HiDe-Prompt utilizes S-Prompt(Wang et al., 2022a) i.e. similar to DualPrompt but without the global (task-shared) prompt. LGCL and PGP don’t propose a specific prompt structure but rather utilize L2P and DualPrompt. The reported results for PGP and LGCL are obtained with DualPrompt structure which is the best from their result. The other methods i.e. L2P, DualPrompt, CODA-P, EvoPrompt, CPrompt, and ConvPrompt utilize their proposed structure. All the evaluated methods utilize ViT B/16 pre-trained on ImageNet 21K as the backbone model. LGCL utilizes pre-trained CLIP text encoder, while ConvPrompt utilizes SentenceTransformer pre-trained on BERT as its text encoder.

LEAPGen: Our proposed method is implemented on top of the ViT backbone pre-trained on ImageNet-21K. The prompt structure is as defined in section 4. The prompt length is set to 30, and the prefix tuning layers is set to 7 i.e. [0,1,2,3,4,5,6] for all main experiments. We utilize Adam optimizer with a cosine learning rate scheduler. For CIFAR100 dataset, We set 0.01 initial learning rate and 3, 10, and 10 epochs for 5-task, 10-task, and 20-task settings respectively. For ImageNet- R dataset, We choose 5, 10, and 20 epochs for 5-task, 10-task, and 20-task settings respectively. The initial learning rate is chosen from the best of 0.04,0.05,0.06. For CUB dataset, We choose 20 epochs and 0.005 initial learning rate. Similar to ConvPrompt, we utilize SentenceTransformer as a text encoder. All the pre-trained models i.e. ViT, SentenceTransformer, and CLIP(LGCL) are kept frozen (not fine-tuned).

W4. Open the descriptors: Please check https://anonymous.4open.science/r/xt124j05/descriptors/.

评论
  • LEAPGen-lite (lightweight generators and without descriptors) performance analysis in Section 5.2.d, Tables 2,3,4, and Figure 3.
MethodCUB 10-Tasks @20c
FAACAAFFMCFM
L2P66.95 ± 0.1374.03 ± 0.325.18 ± 0.117.39 ± 0.23
DualPrompt73.95 ± 0.7380.29 ± 0.157.87 ± 0.768.60 ± 0.37
CODA-P72.99 ± 0.3082.64 ± 0.6411.71 ± 1.499.91 ± 1.04
LGCL79.93 ± 0.3083.07 ± 0.235.45 ± 0.335.58 ± 0.39
HiDe-Prompt87.21 ± 0.1887.66 ± 0.011.90 ± 0.452.66 ± 0.13
PGP78.35 ± 0.6882.21 ± 0.565.76 ± 0.106.10 ± 0.26
EvoPrompt76.23 ± 0.5181.00 ± 0.403.96 ± 0.433.57 ± 0.83
CPrompt77.14 ± 1.1685.67 ± 0.5611.65 ± 0.478.69 ± 0.31
ConvPrompt80.12 ± 1.3784.70 ± 0.646.04 ± 0.974.61 ± 0.45
LEAPGen-lite86.00 ± 0.3288.03 ± 0.472.46 ± 1.612.29 ± 1.15
LEAPGen88.45 ± 0.5890.90 ± 0.811.32 ± 0.521.29 ± 0.82

Compared to other approaches

MethodImageNet-R 5TImageNet-R 5TImageNet-R 10TImageNet-R 10TImageNet-R 20TImageNet-R 20TCIFAR100 10TCIFAR100 10T
FAACAAFAACAAFAACAAFAACAA
C-LoRA75.85 ± 0.3178.85 ± 0.3471.89 ± 0.4575.33 ± 0.2865.71 ± 0.6070.63 ± 0.8582.97 ± 0.4787.06 ± 0.25
LAE73.84 ± 0.1477.29 ± 0.4571.70 ± 0.3976.71 ± 0.1066.98 ± 0.3573.72 ± 0.0588.81 ± 0.3491.59 ± 0.13
InfLoRA77.52 ± 0.3782.01 ± 0.1275.65 ± 0.1480.82 ± 0.2471.01 ± 0.4577.28 ± 0.4589.84 ± 0.0391.70 ± 0.32
SLCA--77.00 ± 0.3381.17 ± 0.64--91.53 ± 0.2894.09 ± 0.87
GMM--80.72---87.59-
LEAPGen-lite82.44 ± 0.6384.37 ± 0.9082.38 ± 1.0485.14 ± 0.5283.67 ± 0.3985.65 ± 0.3398.58 ± 0.0398.69 ± 0.10
LEAPGen82.79 ± 0.3285.06 ± 0.2984.09 ± 0.9385.54 ± 0.6587.03 ± 0.1287.81 ± 0.4898.38 ± 0.1598.15 ± 0.39

5.2.d) LEAPGen-lite’s Performance: As shown in Table 2-4 and Figure 3 Despite utilizing far smaller (2.67% params) generators and without class descriptors generated by LLM, LEAPGen-lite still outperforms the existing methods significantly i.e. 4.7-25% FAA and 3-11% CAA in CIFAR100, and 3-30% FAA and 2-26% CAA in ImageNet-R dataset. LEAPGen-lite also achieves a low for- getting rate in these 2 datasets for all task-settings. In the CUB dataset, LEAPGen-lite archives a comparable performance to HiDe-Prompt and outperforms other SOTA with a significant margin i.e. 6-20% FAA and 3.3-14% CAA. This evidence proves our ideas i.e. language embedding as input for prompt generation, task-wise generators, soft task-id predictor, and learning with auxiliary data are not bounded by the generated descriptors and the size of generators.

Please Note: The analysis of LEAGen-lite empirically proves that our method doesn't rely on GPT/LLM or complex generator networks (high number of parameters)

评论
  • LEAPGen-lite (lightweight generators and without descriptors) performance analysis in Section 5.2.d, Tables 2,3,4, and Figure 3.
MethodCIFAR100ImageNet-R
FAACAAFFMCFMFAACAAFFMCFM
10 Tasks @10c10 Tasks @20c
L2P83.84 ± 0.3288.67 ± 0.166.55 ± 0.345.16 ± 0.1462.50 ± 0.5167.05 ± 0.475.01 ± 0.404.41 ± 0.43
DualPrompt85.36 ± 0.2089.77 ± 0.205.41 ± 0.334.33 ± 0.1568.59 ± 0.2472.18 ± 0.204.61 ± 0.073.70 ± 0.18
CODA-P86.44 ± 0.1691.27 ± 0.566.38 ± 1.465.09 ± 1.1973.77 ± 0.5079.38 ± 1.487.94 ± 0.086.72 ± 0.79
LGCL85.68 ± 0.4390.16 ± 0.295.46 ± 0.224.25 ± 0.3268.65 ± 0.2572.57 ± 0.194.75 ± 0.333.38 ± 0.58
HiDe-Prompt92.89 ± 0.1195.01 ± 0.081.98 ± 0.051.56 ± 0.1575.75 ± 0.4079.27 ± 0.172.29 ± 0.272.33 ± 0.17
PGP86.36 ± 0.1990.83 ± 0.175.49 ± 0.354.28 ± 0.2768.62 ± 0.1472.19 ± 0.204.53 ± 0.403.63 ± 0.35
EvoPrompt88.17 ± 0.5192.18 ± 0.495.39 ± 0.453.97 ± 0.7376.00 ± 0.2680.97 ± 0.304.22 ± 0.423.59 ± 0.52
CPrompt86.92 ± 1.0491.73 ± 0.665.43 ± 0.744.01 ± 0.8176.32 ± 0.5381.50 ± 0.306.10 ± 0.755.60 ± 1.35
ConvPrompt88.77 ± 0.2492.71 ± 0.044.12 ± 0.442.67 ± 0.1177.08 ± 0.2681.47 ± 0.104.17 ± 0.043.11 ± 0.17
LEAPGen-lite98.58 ± 0.0398.69 ± 0.100.11 ± 0.030.06 ± 0.0382.38 ± 1.0485.14 ± 0.523.01 ± 1.192.13 ± 0.60
LEAPGen98.38 ± 0.1598.15 ± 0.390.10 ± 0.030.05 ± 0.0084.09 ± 0.9385.54 ± 0.651.46 ± 1.252.11 ± 1.21
20 Tasks @5c20 Tasks @10c
L2P81.89 ± 0.3887.16 ± 0.338.81 ± 0.106.79 ± 0.3357.40 ± 0.3163.33 ± 0.2110.76 ± 0.457.88 ± 0.17
DualPrompt82.32 ± 0.2287.47 ± 0.246.88 ± 0.355.63 ± 0.2365.19 ± 0.1770.31 ± 0.297.30 ± 0.185.16 ± 0.34
CODA-P81.29 ± 0.1687.72 ± 0.446.82 ± 1.604.98 ± 0.9570.55 ± 0.7177.08 ± 1.028.23 ± 0.866.95 ± 0.70
LGCL83.18 ± 0.4088.63 ± 0.187.22 ± 0.474.91 ± 0.5664.96 ± 0.6770.18 ± 0.377.35 ± 0.655.05 ± 0.32
HiDe-Prompt-97.62 ± 0.14-0.74 ± 0.03-81.60 ± 0.48-2.23 ± 0.38
PGP83.41 ± 0.3589.23 ± 0.137.95 ± 0.235.66 ± 0.2265.24 ± 0.2570.36 ± 0.267.17 ± 0.215.09 ± 0.25
EvoPrompt84.63 ± 0.2189.47 ± 0.219.19 ± 0.417.39 ± 0.6574.93 ± 0.6479.92 ± 0.136.72 ± 0.905.67 ± 0.26
CPrompt83.60 ± 0.0090.10 ± 0.006.47 ± 0.004.78 ± 0.0074.23 ± 0.1779.82 ± 0.515.98 ± 0.245.54 ± 0.48
ConvPrompt87.21 ± 0.2091.60 ± 0.365.47 ± 0.333.92 ± 0.3373.93 ± 0.3678.92 ± 0.374.87 ± 0.573.57 ± 0.25
LEAPGen-lite95.28 ± 3.3798.38 ± 1.131.08 ± 1.540.70 ± 0.9983.67 ± 0.3985.65 ± 0.331.06 ± 0.240.47± 0.14
LEAPGen96.51 ± 2.1698.73 ± 0.260.66 ± 0.750.52 ± 0.3987.03 ± 0.1287.81 ± 0.482.17 ± 0.172.54 ± 0.77
评论

Dear Reviewer UGi2,

We would like to follow up on our response. Could you please kindly review our updated manuscript regarding your concerns? Thank you.

Here are the pointers to your key points: We also copied the updates here, thus you can review them directly.

W1. Proposed Method Overview: Section 4.1'

In this study, we propose a novel LanguagE As Prompt Generator (LEAPGen) accommodating our main principles that are emphasized in the introduction section. The structure and flow of LEAPGen are visualized in figure 2. LEAPGen generators produce prompts from top-k selected embedding as input. LEAPGen also produces auxiliary (aux) data from the top-k embedding. The prompts are prepended into ViT MSAs while the aux is appended into input patches, thus producing feature and final prediction by ViT layers and MLP head respectively. The top-k embedding is selected based on the cosine similarities between an input and the class-wise keys. LEAPGen limits the search space into task 1 to predicted task t, by performing soft task-id prediction. In each task of the training phase, LEAPGen updates task-associated learnable parameters i.e. generator, task-wise key, and class-wise keys. In the inference phase, LEAPGen selects the generators based on the predicted task t. LEAPGen utilizes cross-entropy loss and cosine similarity loss to optimize its parameters. Task- wise generators, language embedding for prompt generation, and soft task-id prediction are unique to recent SOTAs of evolving generator methods e.g. (Roy et al., 2024) and (Kurniawan et al., 2024), task-wise fixed prompt methods such as (Wang et al., 2024a) and (Gao et al., 2024), and pool-based method (Wang et al., 2022c) in terms of prompt generation/selection, task-prediction mechanism, and modality for prompt generation. The detailed architecture, flow, and learning mechanism are presented in sub-section 4.2 and 4.3.

W3. The use of GPT:

  • LEAPGen-lite (lightweight generators and without descriptors) performance analysis in Section 5.2.d, Tables 2,3,4, and Figure 3.
MethodSplit-CIFAR100Split-ImageNet-R
FAACAAFFMCFMFAACAAFFMCFM
5 Tasks @20c5 Tasks @40c
L2P84.77 ± 0.4888.67 ± 0.306.18 ± 0.575.99 ± 0.2964.62 ± 0.3268.01 ± 0.423.94 ± 0.163.55 ± 0.20
DualPrompt86.41 ± 0.2189.95 ± 0.105.37 ± 0.214.77 ± 0.4669.71 ± 0.1172.78 ± 0.143.32 ± 0.162.78 ± 0.25
CODA-P88.22 ± 1.0692.25 ± 1.287.05 ± 2.186.06 ± 2.6674.89 ± 0.3679.71 ± 1.278.89 ± 0.657.65 ± 0.98
LGCL86.90 ± 0.4090.45 ± 0.185.01 ± 0.354.36 ± 0.1369.93 ± 0.2172.91 ± 0.193.04 ± 0.362.50 ± 0.38
HiDe-Prompt91.99 ± 0.0393.95 ± 0.092.52 ± 0.182.33 ± 0.1575.40 ± 0.2778.88 ± 0.043.15 ± 0.462.64 ± 0.16
PGP87.69 ± 0.0691.26 ± 0.135.32 ± 0.184.60 ± 0.1569.71 ± 0.1572.77 ± 0.073.36 ± 0.232.85 ± 0.25
EvoPrompt89.07 ± 0.3892.32 ± 0.265.25 ± 0.655.39 ± 0.2477.27 ± 0.4081.67 ± 0.181.79 ± 0.311.41 ± 0.32
CPrompt89.22 ± 0.0593.09 ± 0.065.02 ± 0.174.31 ± 0.3578.65 ± 0.0082.44 ± 0.006.00 ± 0.005.49 ± 0.00
ConvPrompt90.26 ± 0.4493.49 ± 0.193.64 ± 0.283.25 ± 0.1679.36 ± 0.0882.93 ± 0.243.42 ± 0.052.36 ± 0.16
LEAPGen-lite97.07 ± 0.0897.28 ± 0.100.05 ± 0.010.02 ± 0.0182.44 ± 0.6384.37 ± 0.900.43 ± 0.080.17 ± 0.06
LEAPGen96.84 ± 0.1296.85 ± 0.260.08 ± 0.070.06 ± 0.0482.79 ± 0.3285.06 ± 0.290.51 ± 0.040.18 ± 0.07
评论

Dear Reviewer UGi2,

We would like to follow up on our responses. Could you please kindly review our updated manuscript or previous comments regarding your concerns?

We believe we have addressed all your concerns in our updated paper and previous responses (comments). Thus, we kindly request Reviewer UGi2 to reevaluate the score for our paper.

Thank you.

Best Regards,

Author of Submission 6684

评论

Dear Reviewer UGi2,

First, we thank reviewer UGi2 for your effort and time to review our work. We believe we have addressed all your concerns in our revised paper and previous responses (comments) Second, since the discussion phase has ended, We would like to summarize our revision and responses, and emphasize the following points:

  1. Revision of Ablation: We have revised our ablation study in a step-by-step as you advised. It explains the contribution of each component in a more detailed way (sec. 5.2.e).

  2. Method Overview and Writing Improvement: We have added an overview explaining the general ideas of our proposed method in sec 4.1. We have improved the writing of our paper in terms of layout, presentation and clarity.

  3. Fairness Setting : LEAPGen utilizes the same resources as ConvPrompt and LEAPGen-lite(no descriptors with lightweight generators) utilizes less resources than ConvPrompt and comparable resources to existing SOTAs (sec. 5.1, sec. 5.2.g, Appendix E). All methods are evaluated in the same setting and run with respective best (recommended by official paper/code) hyperparameters. Thus we believe we have satisfied the fairness aspect in our study.

  4. The presence of GPT :

  • GPT is not part of our method (sec 4.2.a, and fig. 2), It is an optional part to generate descriptors following ConvPrompt. Even without GPT-generated descriptors. our method works excellently and significantly outperforms the existing SOTAs, as demonstrated by LEAPGen-lite (sec. 5.2.d) and LEAPGen-CN (Appendix D.4).

  • Descriptor impact to the model performance has been demonstrated by the performance difference between LEAPGen and LEAPGen-CN (Appendix D.4). We also extend our analysis of the performance of LEAPGen vs ConvPrompt in various types of descriptors (Appendix D.5).

  • LEAPGen-lite proves the significant impacts of our ideas i.e. (1) language as input for prompt generation, (2) task-wise generators, (3) soft task-id predictor, and (4)learning with auxiliary despite without GPT-generated descriptors, and utilize least parameters, running time and storage.

We believe, these highlights resolve your concerns and prove the significant impacts of our ideas, without being bound to the presence of GPT.

Best Regards,

Author of Submission 6684

审稿意见
8

The paper proposes a new method for prompt based continual learning. Unlike previous methods that focus on task-specific prompts or growing prompt components, this work generates prompts using language descriptions of top-k similar classes from all the previously trained tasks to an input image. The generated prompts are plugged in to the Key and Value pairs of the backbone model for the given input image. In addition, the combined encoded language descriptions of the top-k similar classes are appended to the input image embeddings. Result on CIFAR100, ImageNet-R and CUB datasets with pretrained ViT-B/16 backbone shows significant improvements across accuracy and forgetting metrics over previous works. Extensive ablation study on different components provides better insights on the proposed approach.

优点

• Introduction is clear in terms of discussing drawbacks of prior works and motivating the proposed approach. Contributions are listed clearly.

• I appreciate Section 3 and Figure 1 as it gives provides brief recap of the previous SOTA methods and laid some foundation for discussing the proposed method.

• Leveraging language descriptions and using all language descriptors to find top-k similar classes to generate the prompts is novel in the context of continual learning.

• The strategy to incorporate inter-class loss helps the method to achieve better results.

• Comparisons are made against various previous works and significant improvements are shown with their proposed method.

• Appreciate the detailed ablation study to understand benefits of different components.

缺点

I don’t find any major weakness but have set of questions (see below) to improve the understanding. However, I find that learning with auxiliary data and inter-task loss are critical components of the method (as shown in ablation study), which are not sufficiently emphasized in the contributions.

问题

  1. Between L155-160, under the context of previous task-specific approaches, it is mentioned that two different tasks could produce similar key vectors. It would make a stronger point to justify this statement with some empirical analysis.

  2. What is the value of k for top-k descriptors and how many Generators are chosen for each task? Can we chose a single generator for each task? How {E_i}^t of size |C^t|gets chosen for k set of generators?

  3. From the Figure 2 or its caption, it is not clear what is the component that is compared against input x to get the cosine similarity. The sentence “search space for top-k” in the figure gives some hint, but the arrows for entire blue block rising the confusion. Please clarify it in the figure or caption.

  4. Why it is interesting to solve equation 3, if the task is to search for E_i in all tasks [1, 2, ….t]? Similarly, why consider K^t and why not just use {L_i}^t ?

  5. Under the ablation study, what does absence of generator mean? How do you get prompts in this approach without generator?

  6. Is the MLP head (classifier) also trained for each task? How can it produce list of softmax values of all classes for equation 7?

  7. Please explain why inter-task loss help improve FAA and CAA? What does it tell us about the learned parameters?

  8. Has the importance of inter-task loss already explored in previous continual learning papers?

Minor suggestions:

a. In Figure 1, reposition the labels “fixed”, “trainable”, and “generated” since these labels affect entire Figure, not just the green block.

b. In section 3.b, I suggest to use emphasized text style for “pool-based approach”, “task-specific prompt approach”, and “growing component approach”.

c. In Line 242, correct the typo “tem” -> “them”

d. Line 267 mentions t_hat, but eq. 3 uses notation as t, but not t_hat.


Final review: I thank authors for their responses. I find that paper proposes a novel method to leverage language descriptors within the context of continual learning. Authors have conducted thorough experimental evaluation and demonstrated notable improvements across multiple datasets. I read the fellow reviewers concerns and I find that authors have sufficiently addressed the major concerns and revised their manuscript accordingly. In particular, regarding the utilization of GPT, authors have justified that their method does not necessarily rely on compute heavy GPTs or LLMs, and can also work without GPT-generated descriptors. Overall, I believe that this paper would be interesting to the research community and suggest towards acceptance.

评论

Q6. Is the MLP head (classifier) also trained for each task? How can it produce list of softmax values of all classes for equation 7?

Our Response: Thank you for your question. Yes, the MLP head is trained in every task. Please note that following standard practice in continual learning, the MLP head gradually adds output neurons when entering a new task. For example, in the case of class incremental learning with 10 classes per task. In the first task, the MLP head has 10 outputs, when entering task 2, the output neurons are added by 10 (now 20 outputs), and so on. Therefore It can produce logits for all and ongoing learned classes. In equation 7, we compare all the logits to the labels, while in equation 6 (intra-task loss) we only consider the logists from currently learned classes, excluding previously learned classes.

Q7. Please explain why inter-task loss help improve FAA and CAA? What does it tell us about the learned

Our Response: Thank you for your question. Inter-task loss improves the discriminative ability between currently learned classes and previously learned classes (from previous tasks). The newly added parameters/weights are driven to produce different representations between the currently learned classes and previously learned classes. Thus, the inter-task loss contributes to improving CAA and FAA. However, at the same time, inter-task loss increases the risk of forgetting since it changes (adjusts) model's parameters that optimal for previous learned classes.

Q8. Has the importance of inter-task loss already explored in previous continual learning papers?

Our Response: Thank you for your question. Our literature study shows that inter-task loss is the default cross-entropy loss that is widely utilized for continual learning. Following HidePrompt, we emphasize the importance of intra-task and inter-task loss to improve within-task prediction and inter-task prediction. However, our method and HiDe-Prompt have different mechanisms for intra-task loss. The new joint loss function including intra-task loss, inter-task loss, cosine similarity to KtK^t (Lt\mathcal{L}_t), and cosine similarity to LctL_c^t (Lc\mathcal{L}_c) as presented in eq.10 is unique and it is one of our contributions.

Sa. In Figure 1, reposition the labels “fixed”, “trainable”, and “generated” since these labels affect entire Figure, not just the green block.

Our Response: Thank you for the suggestion. We have revised Figure 1, we moved the labels outside the green block.

Sb. In section 3.b, I suggest to use emphasized text style for “pool-based approach”, “task-specific prompt approach”, and “growing component approach”.

Our Response: Thank you for the suggestion. We have bolded the mentioned texts in our revised paper.

Sc. In Line 242, correct the typo “tem” -> “them”

Our Response: Thank you for the correction. We have corrected the typo in our revised manuscript.

Sd. Line 267 mentions t_hat, but eq. 3 uses notation as t, but not t_hat.

Our Response: Thank you for the correction. We have revised the symbol into tt (without hat).

评论

Q1. Between L155-160, under the context of previous task-specific approaches, it is mentioned that two different tasks could produce similar key vectors. It would make a stronger point to justify this statement with some empirical

Our Response: Thank you for your question. We have added extra analysis on the similarity between task keys as presented in Appendix D1. Our analysis shows the high cosine similarity of task keys after completing all task (10 tasks) learning .

Q2. What is the value of k for top-k descriptors and how many Generators are chosen for each task? Can we chose a single generator for each task? How {E_i}^t of size |C^t|gets chosen for k set of generators?

Our Response: Thank you for your question. The value of k is the same as the number of generators, and it is assigned as a user parameter. We choose k=3 for our experiment. Our sensitivity analysis (sec 5.2 point e.3) shows the performance of our method in various k. Yes, we can use a single generator as well. In our method, using a single generator means we choose k=1 which also means we choose only a single descriptor for prompt generation. How to select top-k EiE_i? Given a predicted task tt, we choose top-k EiE_i from the predicted (current task) task i.e. EctE_c^t, and from the previous task Ec1E_c^1, Ec2E_c^2,... Ect1E_c^{t-1}. The top-k selected descriptor embedding i.e. E1E_1, E2E_2... EkE_k will be inputted into k generators of task tt i.e. G1tG_1^t, G2tG_2^t... GktG_k^t. We explain the complete components and procedures in sec. 4 and Figure 2.

Q3. From the Figure 2 or its caption, it is not clear what is the component that is compared against input x to get the cosine similarity. The sentence “search space for top-k” in the figure gives some hint, but the arrows for entire blue block rising the confusion. Please clarify it in the figure or caption.

Our Response: Thank you for your question. We compare x to LciL_c^i that is associated with Eci,i[1,2,...t]E_c^i, i \in [1,2,...t]. We have adjusted the font in Figure 2 so that the string "search space for top-k" and symbol LciL_c^i have the same size. We set both of them in italics for better cohesion between them.

Q4. Why it is interesting to solve equation 3, if the task is to search for E_i in all tasks [1, 2, ….t]? Similarly, why consider K^t and why not just use {L_i}^t ?

Our Response: Thank you for your question.

  • We search for top k EiE_i as the input for prompt generators. Along with the EiE_i from the expected class, we look for other k-1 descriptors that are closest to the expected class which may be located in previous tasks, not in the same task. For example, descriptors of class "pear" are closer to the descriptors of class "apple" than to the descriptors of class "palm tree". Class "apple" is located in the previous task, while class "palm tree" is in the same task as class "pear". Thus, it is better to select other k-1 descriptors from previous tasks. Why not search from all tasks? In that case, our method will always use the same (latest) generators that might be incompatible (un-optimal) for all tasks. Besides, It will increase the chance of wrong-selected descriptors.

  • The use of Lit{L_i}^t only has the risk of misprediction due to similar keys from different classes from different tasks. For example, images of class 'Scottish terrier' have high similarity with images of class 'West Highland white terrier' from a different task. The use of KtK^t only has the risk of similar key vectors from different tasks as we mentioned earlier. Thus, task prediction using Lit{L_i}^t and KtK^t together has less misprediction risk than using Lit{L_i}^t and KtK^t only.

Q5. Under the ablation study, what does absence of generator mean? How do you get prompts in this approach without generator?

Our Response: Thank you for your question. The absence of a generator means the prompt components are produced directly from selected (the top-k) descriptors i.e. PCi=EiPC_i = E_i, and the prompt is produced as P=ΣikPCi=ΣikPCi=ΣikEiP = \Sigma_i^k PC_i = \Sigma_i^k PC_i = \Sigma_i^k E_i.

审稿意见
6

This paper proposes a novel method LEAPGen (Language as Prompt Generator) for continual learning that leverages language knowledge for prompt generation. LEAPGen uses language descriptors for each class as input for prompt generation, rather than shared embeddings. For each task, LEAPGen employs a fixed number of generator networks, a learnable task-level key, and learnable class-level parameters associated with the descriptor embeddings. To generate a prompt, it first predicts the task ID using a soft prediction mechanism, then selects top-k matching descriptor embeddings based on cosine similarity to the input. Experimental results on datasets like CIFAR100, ImageNet-R, and CUB demonstrate that LEAPGen significantly outperforms existing state-of-the-art methods in terms of accuracy and forgetting measure, addressing key limitations of previous prompt-based continual learning approaches.

优点

  • LEAPGen uses language descriptors as input for prompt generation, leveraging richer semantic information compared to previous methods.

  • As a prompt-based method, it doesn't require storing and replaying old data.

  • By using fixed language descriptors as input, LEAPGen potentially reduces catastrophic forgetting compared to methods that continuously update shared embeddings.

  • Experimental results show significant improvements over state-of-the-art methods.

缺点

  • The paper does not explicitly specify some key details about the pre-trained models used. Which Sentence Transformer does the paper employ? Whether this is a pre-trained language model and if any fine-tuning of these models occurs during the continual learning process.

  • The paper does not provide a clear comparison of the increase in parameters due to the addition of generators for new tasks. Based on the paper, the generator is not a lightweight module. As the proposed method requires adding new generators for some new tasks, even if not all tasks, this could lead to a significant growth in the number of parameters as the number of tasks increases. And also an unfair comparison with related works.

问题

see above

评论

Q1/W1. The paper does not explicitly specify some key details about the pre-trained models used. Which Sentence Transformer does the paper employ? Whether this is a pre-trained language model and if any fine-tuning of these models occurs during the continual learning process.

Our Response: Thank you for your concern.

  • Following common practice in prompt-based methods, all the evaluated methods in tables 2 and 3 utilize the same model i.e. ViT-B/16 that is pre-trained on imagenet21K. Following ConvPrompt, we utilized SentenceTransformer pre-trained on Phase-BERT as our text encoder. Other language-guided methods i.e. LGCL and GMM utilize CLIP as a text encoder. There is no fine-tuning process for SentenceTransformer in our method. The model is solely employed to encode the descriptors into embedding before starting a training process in each task.

  • We explain the detailed setting for all method in Appendix E.

Q2/W2. The paper does not provide a clear comparison of the increase in parameters due to the addition of generators for new tasks. Based on the paper, the generator is not a lightweight module. As the proposed method requires adding new generators for some new tasks, even if not all tasks, this could lead to a significant growth in the number of parameters as the number of tasks increases. And also an unfair comparison with related works.

Our Response: Thank you for your concern.

  • We added a new experiment and analysis on the trade-offs between cost and performance that comprehensively evaluate the proposed methods and existing SOTAs in terms of performance (accuracy and forgetting), number of parameters, and running time. We also compare our proposed method with the large version of SOTAs that executes a higher number of parameters. Please see Appendix D3. Our analysis shows in comparison to the standard versions and large versions of SOTAs, our method achieves better performance (average accuracy and average forgetting), and requires a shorter running time.

  • Even in comparison with the SOTAs in a smaller number of parameters, our proposed method still achieves a smaller running time. The Smallest number of parameters i.e. HiDe-Prompt doesn't imply shorter running time as It requires mode additional operations i.e. drawing (augmenting) prototypes and constructing uninformed representations that run on many epochs.

  • To ensure fairness, we evaluated all the methods by applying the best setting i.e. as mentioned in the references and official code manual, including the preferred hyper-parameters. Please note that increasing the number of parameters i.e. by increasing the number of generators, or prompt length doesn't always improve the model performance. As shown in the table above, our sensitivity analysis in sec 5.2. point e and figure 4, after some points, the increase in prompt length or number of generators decreases the model performance. Thus, we believe, we have evaluated the consolidated methods in a fair way i.e. by applying their respective official setting and parameters.

评论

We would like to follow up on our response and revised manuscript. We would appreciate it if Reviewer j5sH could look at our revised manuscript, and offer additional comments.

评论

Dear Reviewer J5sH,

Could you kindly review the rebuttal thoroughly and let us know whether the authors have adequately addressed the issues raised or if you have any further questions.

Best,

AC of Submission6684

评论

Dear Reviewer J5sH,

We would like to follow up on our response to your comments Could you please kindly check our revised manuscript?

Here are the pointers to your key points:

W1. Details of ViT and Text Encoder: Appendix E. Detailed Experimental Setting

W2. Increase of Parameters:

  • Section 5.2.g. Analysis of Parameters, Running Time and Storage

  • Appendix D.3. TradeOffs Between Cost and Performance including detailed growing parameters on each task

  • Section 5.2.d., Tables 2,3,4, and Figure 3 i.e. Analysis of LEAPGen-lite (lightweight generators and without descriptors) performance that proves the significant performance of our ideas is not bounded by LLM-generated descriptors or the complexity (parameters and size) of the generators.

Thank you.

Best Regards,

Authors of Submission 6684

评论

Thank you for these detailed experiments. From the revisited paper, I now clearly understand that the increase in parameters is reasonable, and comparable with other research works. These results answer my questions, and I decide to increase the rate accordingly.

评论

Dear Reviewer J5sH,

Thank you for the time and effort to review our revised manuscript, and increase the score for our paper. We surely appreciate it.

Best Regards,

Author of Submission 6684

评论

W2. Increase of Parameters:

  • Section 5.2.d., Tables 2,3,4, and Figure 3 i.e. Analysis of LEAPGen-lite (lightweight generators and without descriptors) performance

5.2.d) LEAPGen-lite’s Performance: As shown in Table 2-4 and Figure 3 Despite utilizing far smaller (2.67% params) generators and without class descriptors generated by LLM, LEAPGen-lite still outperforms the existing methods significantly i.e. 4.7-25% FAA and 3-11% CAA in CIFAR100, and 3-30% FAA and 2-26% CAA in ImageNet-R dataset. LEAPGen-lite also achieves a low for- getting rate in these 2 datasets for all task-settings. In the CUB dataset, LEAPGen-lite archives a comparable performance to HiDe-Prompt and outperforms other SOTA with a significant margin i.e. 6-20% FAA and 3.3-14% CAA. This evidence proves our ideas i.e. language embedding as input for prompt generation, task-wise generators, soft task-id predictor, and learning with auxiliary data are not bounded by the generated descriptors and the size of generators.

Here is the snapshot of the tables, please see our revised paper for the full tables.

MethodImageNet-R
FAACAAFFMCFM
5 Tasks @40c
L2P64.62 ± 0.3268.01 ± 0.423.94 ± 0.163.55 ± 0.20
DualPrompt69.71 ± 0.1172.78 ± 0.143.32 ± 0.162.78 ± 0.25
CODA-P74.89 ± 0.3679.71 ± 1.278.89 ± 0.657.65 ± 0.98
LGCL69.93 ± 0.2172.91 ± 0.193.04 ± 0.362.50 ± 0.38
HiDe-Prompt75.40 ± 0.2778.88 ± 0.043.15 ± 0.462.64 ± 0.16
PGP69.71 ± 0.1572.77 ± 0.073.36 ± 0.232.85 ± 0.25
EvoPrompt77.27 ± 0.4081.67 ± 0.181.79 ± 0.311.41 ± 0.32
CPrompt78.65 ± 0.0082.44 ± 0.006.00 ± 0.005.49 ± 0.00
ConvPrompt79.36 ± 0.0882.93 ± 0.243.42 ± 0.052.36 ± 0.16
LEAPGen-lite82.44 ± 0.6384.37 ± 0.900.43 ± 0.080.17 ± 0.06
LEAPGen82.79 ± 0.3285.06 ± 0.290.51 ± 0.040.18 ± 0.07
ImageNet-R
L2P62.50 ± 0.5167.05 ± 0.475.01 ± 0.404.41 ± 0.43
DualPrompt68.59 ± 0.2472.18 ± 0.204.61 ± 0.073.70 ± 0.18
CODA-P73.77 ± 0.5079.38 ± 1.487.94 ± 0.086.72 ± 0.79
LGCL68.65 ± 0.2572.57 ± 0.194.75 ± 0.333.38 ± 0.58
HiDe-Prompt75.75 ± 0.4079.27 ± 0.172.29 ± 0.272.33 ± 0.17
PGP68.62 ± 0.1472.19 ± 0.204.53 ± 0.403.63 ± 0.35
EvoPrompt76.00 ± 0.2680.97 ± 0.304.22 ± 0.423.59 ± 0.52
CPrompt76.32 ± 0.5381.50 ± 0.306.10 ± 0.755.60 ± 1.35
ConvPrompt77.08 ± 0.2681.47 ± 0.104.17 ± 0.043.11 ± 0.17
LEAPGen-lite82.38 ± 1.0485.14 ± 0.523.01 ± 1.192.13 ± 0.60
LEAPGen84.09 ± 0.9385.54 ± 0.651.46 ± 1.252.11 ± 1.21
ImageNet-R
L2P57.40 ± 0.3163.33 ± 0.2110.76 ± 0.457.88 ± 0.17
DualPrompt65.19 ± 0.1770.31 ± 0.297.30 ± 0.185.16 ± 0.34
CODA-P70.55 ± 0.7177.08 ± 1.028.23 ± 0.866.95 ± 0.70
LGCL64.96 ± 0.6770.18 ± 0.377.35 ± 0.655.05 ± 0.32
HiDe-Prompt-81.60 ± 0.48-2.23 ± 0.38
PGP65.24 ± 0.2570.36 ± 0.267.17 ± 0.215.09 ± 0.25
EvoPrompt74.93 ± 0.6479.92 ± 0.136.72 ± 0.905.67 ± 0.26
CPrompt74.23 ± 0.1779.82 ± 0.515.98 ± 0.245.54 ± 0.48
ConvPrompt73.93 ± 0.3678.92 ± 0.374.87 ± 0.573.57 ± 0.25
LEAPGen-lite83.67 ± 0.3985.65 ± 0.331.06 ± 0.240.47± 0.14
LEAPGen87.03 ± 0.1287.81 ± 0.482.17 ± 0.172.54 ± 0.77

Please Note: The analysis of LEAGen-lite empirically proves that our method doesn't rely on GPT/LLM or complex generator networks (high number of parameters)

评论

W2. Increase of Parameters:

  • Appendix D.3. TradeOffs Between Cost and Performance including detailed growing parameters on each task

(1). Performance vs #Parameters and Running Time:

MethodMetricsValueAVGRunTime(h)
12345678910
HiDe-PromptAvg.Accuracy85.2282.9382.3579.7478.8577.8177.3776.3876.3175.7579.27-
HiDe-Prompt-LargeAvg.Accuracy85.7683.1282.0379.4978.5576.5576.0074.8274.9874.3078.56-
ConvPromptAvg.Accuracy89.5386.2484.4081.8880.7980.0478.8078.1577.7477.0881.47-
ConvPrompt-LargeAvg.Accuracy90.1286.2283.9681.4580.5279.7376.4074.1974.56-80.79-
LEAPGen-liteAvg.Accuracy89.7388.6686.9486.4386.0483.0382.7582.5282.9482.3885.14-
LEAPGenAvg.Accuracy89.8387.6386.4685.9784.7184.7984.0183.7784.1284.0985.54-
HiDe-PromptAvg.Forgetting3.291.502.412.192.532.202.252.302.292.332.33-
HiDe-Prompt-LargeAvg.Forgetting-3.051.892.142.153.252.912.932.662.702.63-
ConvPromptAvg.Forgetting-2.571.572.602.933.153.533.444.074.173.11-
ConvPrompt-LargeAvg.Forgetting-4.653.394.624.695.568.9110.5510.50-6.61-
LEAPGen-liteAvg.Forgetting1.110.750.580.564.133.372.922.703.012.132.13-
LEAPGenAvg.Forgetting-3.971.961.382.632.241.991.771.601.462.11-
HiDe-Prompt#Parameters (M)0.150.150.150.150.150.150.150.150.150.150.155.40
HiDe-Prompt-Large#Parameters (M)12.2912.2912.2912.2912.2912.2912.2912.2912.2912.2912.2915.15
ConvPrompt#Parameters (M)0.550.600.700.800.850.890.991.091.191.280.892.11
ConvPrompt-Large#Parameters (M)5.185.826.657.287.918.609.8110.9412.0113.238.7415.28
LEAPGen-lite#Parameters (M)0.020.030.050.070.080.100.110.130.150.160.090.53
LEAPGen#Parameters (M)6.216.236.246.266.276.296.316.326.346.356.281.79

(2). Detailed Running Time: T.Desc, Tr, and Inf denote time for generating descriptors, training, and inference respectively.

MethodTime (h)
CIFAR100 10TImageNet-R 10TCUB 10T
T.Desc.InfTr+InfTotalT.Desc.InfTr+InfTotalT.Desc.InfTr+InfTotal
HiDe-Prompt-0.0374.634.63-0.0195.405.40-0.0173.943.94
ConvPrompt0.530.0232.012.541.070.0331.042.111.090.0358.089.17
LEAPGen-lite-0.0161.21.2-0.0280.530.53-0.0280.380.38
LEAPGen0.530.0170.981.511.070.0250.721.791.090.0250.321.41

(3). Performance vs Storage:

MethodPerformanceStorage(MB)
FAACFA
HiDe-Prompt75.7579.27334
HiDe-Prompt-Large74.3078.56797
ConvPrompt77.0881.47346
ConvPrompt-Large74.56*80.79632
LEPGen-lite82.3885.14332
LEAPGen84.0985.54567
评论

Dear Reviewer J5sH,

We would like to follow up on our response. Could you please kindly review our latest manuscript regarding your concern? Thank you.

Here are the pointers to your key points:

We also copied the updates (or snapshots) here, thus you can review them directly.

W1. Details of ViT and Text Encoder: Appendix E. Detailed Experimental Setting

All the evaluated methods utilize ViT B/16 pre-trained on ImageNet 21K as the backbone model. LGCL utilizes pre-trained CLIP text encoder, while ConvPrompt utilizes SentenceTransformer pre-trained on BERT as its text encoder. LEAPGen: Our proposed method is implemented on top of the ViT backbone pre-trained on ImageNet-21K. The prompt structure is as defined in section 4. The prompt length is set to 30, and the prefix tuning layers is set to 7 i.e. [0,1,2,3,4,5,6] for all main experiments. We utilize Adam optimizer with a cosine learning rate scheduler. For CIFAR100 dataset, We set 0.01 initial learning rate and 3, 10, and 10 epochs for 5-task, 10-task, and 20-task settings respectively. For ImageNet- R dataset, We choose 5, 10, and 20 epochs for 5-task, 10-task, and 20-task settings respectively. The initial learning rate is chosen from the best of 0.04,0.05,0.06. For CUB dataset, We choose 20 epochs and 0.005 initial learning rate. Similar to ConvPrompt, we utilize SentenceTransformer as text encoder. All the pre-trained models i.e. ViT, SentenceTransformer and CLIP(LGCL) are kept frozen (not fine-tuned).

W2. Increase of Parameters:

  • Section 5.2.g: Analysis of parameters, running time, and storage
MethodDesc#Params(M)Running Time (h)Storage (MB)
T.DescInfTr+InfTotal
HiDe-Prompt-0.15-0.0195.405.40334
ConvPromptv1.281.070.0331.042.11346
LEAPGen-lite-0.16-0.0280.530.53332
LEAPGenv6.351.070.0250.721.79567

T.Desc, Tr, and Inf denote time for generating descriptors, training, and inference respectively, detailed in Appendix D3.

5.2.g) Parameters, Running Time, and Storage: The table above compares the number of parameters, running time (ImageNet-R), and storage of our methods and existing SOTAs. Despite having a higher number of parameters and storage, LEAPGen consumes less running time than existing SOTAs both training+inference and total running time. LEAPGen-lite consumes the least costs in total running time and storage and requires relatively low parameters and inference time. LEAPGen and ConvPrompt require additional time to generate descriptors that increase their total simulation time. Despite having the least parameters, Hide-Prompt requires the longest training and total times since it needs extra operations to generate uninstructed class representations.

评论

We would like to express our gratitude to all the reviewers for their constructive and insightful comments on our paper. We are honored that all the reviewers (ueBy, J5sH, UGi2, and bACr) found our work to be novel/new. We also appreciate reviewers (ueBy, and bACr) for pointing out the detailed writing of our paper as a positive point. We have revised our paper, following the reviewers' feedback. We have improved the quality and presentation of our paper with the summary of changes as follows:

(1). We added a detailed overview of our method as presented in section 4.1. to improve e better presentation and understanding of our proposed method. We emphasized the uniqueness of our method to existing SOTAs in the section.

(2). We replace our ablation study with a step-by-step scenario to improve the clarity of each component contribution, as presented in section 5.2.d.

(3). We added a new empirical experiment and analysis on the high similarity of task key (K^t) as presented in Appendix D.1. The high similarity of task keys motivates us to develop a more accurate task-id prediction mechanism rather than the conventional way.

(4). We added a new empirical experiment and analysis on the high similarity of language embedding generated by the encoded string "the photo of class name" as used in the existing method. The high similarity between classes motivates us to develop a more discriminative way of language-guided learning rather than utilizing the embedding directly as references or loss anchors. Please see Appendix D2.

(5). We added a new experiment and analysis on the trade-offs between cost and performance that comprehensively evaluate the proposed methods and existing SOTAs in terms of performance (accuracy and forgetting), number of parameters, and running time. We also compare our proposed method with the large version of SOTAs that executes a higher number of parameters. Please see Appendix D3.

(6). We added a new experiment and analysis on the "class name" as descriptors where we have no class descriptors generated by GPT or other LLMs. Our experiment results as presented in Appendix D4. shows that our method outperforms the existing SOTAs with a significant margin in 3 datasets despite substituting class descriptors with the class name.

(7). We elaborated the detailed setting for all consolidated methods as presented in Appendix E.

(8). We revised Figures 1 and 2 for a better understanding of our paper.

(9). We improve the readability and presentation of our paper by ensuring a clear space between a text to another text, caption, figure, or table.

(10), Last but not least, We revise the writing errors and typos.

The changes w.r.t. feedback from reviewers 1(ueBy), 2(J5sH), 3(UGij), and 4 (bACr) are highlighted in red, green, orange, and violet colors respectively.

评论

Dear Reviewers and Area Chairs,

We added additional updates in our latest manuscript that fully address the comments/concerns of reviewers J5sH, UGi2, and bACr as follows.

  1. We added experiments and analysis of the lite version of LEAPGen i.e. LEAPGen-lite that utilizes lightweight Conv1d generators and without class descriptors (it uses class names only). LEAPGen lite has far smaller parameters than ConvPrompt (12.7% of ConvPrompt #params) and is similar to Hide-Prompt parameters. Our experiment shows that LEAPGen-lite outperforms the existing SOTAs i.e. 4.7-25% FAA and 3-11% CAA in CIFAR100, 3-30% FAA and 2-26% CAA in ImageNet-R, and 6-20% FAA and 3.3-14% CAA in CUB. This evidence proves that the significant performance of our ideas (language as input for prompt generation, task-wise generators, soft task-id predictor, and learning with auxiliary) is not bounded by LLM-generated descriptors or the complexity (parameters and size) of the generators. Please kindly see the results and analysis in Tables 2,3,4, Figure 3, and Section 5.2.d.

  2. We added a summarized analysis on #Parameters, Running Time, and Storage as presented in section 5.2.g, detailed in Appendix D.3. The analysis shows that LEAPGen spends moderate inference time, but lower training+inference time and total running time, than ConvPrompt and Hide-Prompt, despite having more parameters and storage. LEAPGen-lite requires far smaller cost in all aspects than ConvPrompt and smaller running time and storage than HiDe-Prompt, with just a slightly higher number of parameters i.e. 0.16M vs 0.15M. We move the analysis of LEAPGen performance on various types of generators to the Appendix for the exchanged space in the main paper.

Both additional updates are presented in brown color in our latest manuscript.

Thank you.

Best Regards

Author of Submission 6684

评论

Dear Program Chairs (PC), Senior Area Chairs (SAC), Area Chairs (AC), and Reviewers,

As the discussion phase has ended,

First, we thank reviewers PC, SAC, AC, and Reviewers for your effort and time in organizing the review and discussion of our paper. Through these processes, we have improved the technical quality and presentation of our paper significantly. We thank the reviewers ueBy and J5sH for reviewing our paper and/or its revision thoroughly, confirming the novelty and significance of our works.

Second, we have revised our paper following the reviewers' advice in both writing i.e. layout, presentation, details, tables, figures, etc, and technical aspects i.e. method overview, ablation study, the trade-off between performance and cost, empirical/numerical evidence of task similarity, language embedding similarity, setting details, etc as we mentioned in our previous official comments.

Third, we would like to confirm that we have addressed all concerns and questions of the reviewers in our revised paper and responses. However, we would like to emphasize a few key points that reviewers UGi2 and/or bACr may not have reviewed thoroughly in the discussion phase, as follows:

  1. Fairness Setting : LEAPGen utilizes the same resources as ConvPrompt and LEAPGen-lite(no descriptors with lightweight generators) utilizes less resources than ConvPrompt and comparable resources to the other SOTAs (sec. 5.1, sec. 5.2.g, Appendix E). All methods are evaluated in the same setting and run with respective best (recommended by official paper/code) hyperparameters. Thus we believe we have satisfied the fairness aspect in our study.

  2. The Non-Mandatory Presence of GPT :

  • GPT is not part of our method (sec 4.2.a, and fig. 2), It is one of the options to generate class descriptors following ConvPrompt. Our method accommodates both class descriptors (as in ConvPrompt) and class names (as in LGCL) language text. Even without GPT-generated descriptors. our method works excellently and significantly outperforms the existing SOTAs, as demonstrated by LEAPGen-lite (sec. 5.2.d) and LEAPGen-CN (Appendix D.4).

  • GPT is utilized in an online way (through the Internet), by querying it using API and Python script. It is a common practice nowadays as in our simulation (please see https://platform.openai.com/docs/guides/). Thus we never download and save the GPT model in our storage. LEAPGen indeed saves the descriptors embedding (numerical vector), but not the GPT model. The embedding adds to the whole model's storage size but is still in a reasonable amount, and we have included it in our storage analysis.

  • Descriptor generation by GPT (before training) via only query indeed takes a fair amount of time. However, despite spending extra time on it, our method still has a lower total running time than existing SOTAs, i.e. ConvPrompt and HidePrompt. (sec. 5.2.g, Appendix D.3)

  • Descriptor impact to the model performance has been demonstrated by the performance difference between LEAPGen and LEAPGen-CN (Appendix D.4). We also extend our analysis of the performance of LEAPGen vs ConvPrompt in various types of descriptors (Appendix D.5).

  1. LEAPGen-lite proves the significant impacts of our ideas i.e. (1) language as input for prompt generation, (2) task-wise generators, (3) soft task-id predictor, and (4)learning with auxiliary despite without GPT-generated descriptors (only utilizes class names), and utilize least parameters, running time and storage. (sec. 5.2.d, sec. 5.2.g, Appendix D.3)

We believe these key points resolve all concerns regarding the fairness setting and GPT (non-mandatory) presence, and prove the significant impacts of our ideas i.e. (1) language as input for prompt generation, (2) task-wise generators, (3) soft task-id predictor, and (4) learning with auxiliary, without being bound to the presence of GPT.

I think that's all, once again thank you to all the committee, we pray that the hard work and dedication of the committee will be paid off by the success and impact of this year's ICLR, Thank you.

Best Regards,

Author of Submission 6684

AC 元评审

(a) The paper proposes LEAPGen, a novel method for continual learning that generates prompts using language descriptors for classes instead of shared embeddings, with a learnable task key and class-level parameters. LEAPGen predicts the task ID and selects top-k matching descriptor embeddings to generate prompts. Experimental results on CIFAR100, ImageNet-R, and CUB show that LEAPGen outperforms state-of-the-art methods in accuracy and mitigating forgetting.

(b) Strengths: The strengths of the paper lie in its clear and well-motivated introduction, which effectively highlights limitations of prior works and outlines the proposed approach. LEAPGen introduces a novel use of language descriptors for prompt generation, leveraging richer semantic information and avoiding the need to store or replay old data, thereby reducing catastrophic forgetting. The method achieves significant improvements over state-of-the-art methods and incorporates a unique inter-class loss strategy to enhance performance. The experimental results are robust, with thorough comparisons against existing methods and detailed ablation studies to validate the contributions of individual components. Section 3 and Figure 1 provide a concise recap of prior methods, laying a solid foundation for the proposed approach.

(c) Weaknesses: The paper has several weaknesses, including a lack of clarity about key details, such as the specific Sentence Transformer used, whether it is pre-trained, and whether any fine-tuning occurs during the continual learning process. The scalability of the proposed method is a concern, as adding new generators for new tasks could significantly increase the number of parameters, leading to potential unfair comparisons with related works. The writing and structure require improvement, as the compact format affects readability, and some components, like the theorem, seem irrelevant and better suited for supplementary materials. The ablation study lacks rigor and should more systematically validate the contributions of each module starting from a standard baseline. Additionally, the use of ChatGPT for generating descriptions raises concerns about the fairness of comparisons and the source of performance gains, prompting the need for transparency and public availability of these descriptions.

(d) The most important reasons for acceptance are the solid methods and compelling performance, along with the sufficient theoretical & empirical analyses. Since this paper has received 2 negative scores (i.e., 3), the AC has checked very carefully about the details and discovered that the major reasons for low scores, i.e., the inclusion of GPT for accurate description, are well addressed by the authors. The authors have designed a lite version (LEAPGen-lite) of the proposed approach which does not rely on GPT entirely. The LEAPGen-lite still shows significantly superior performance to previous sota methods, proving the significant impacts of the ideas. In the AC's view, all the issues have been addressed during the rebuttal period. By the way, Reviewer UGi2 who scores 3 fails to participate in the rebuttal period despite reminders from the AC.

One more suggestion for the authors to improve the presentation quality: In the current form of this manuscript, the authors show 4 principles. Four principles are too many; for a highly impactful paper, one principle is sufficient—for example, the residual idea in ResNet. The AC suggests the authors reconsider the core problem this paper aims to address, propose a single principle, and then present multiple methods to support this principle. This approach would enhance the clarity of the motivation and increase the paper's overall impact.

审稿人讨论附加意见

(a) Reviewer ueBy notes no major weaknesses but raises questions to improve understanding. They highlight that the auxiliary data and inter-task loss, critical components demonstrated in the ablation study, are underemphasized in the stated contributions. Emphasizing these aspects could strengthen the presentation of the method. The authors have successfully addressed the issues.

(b) Reviewer J5sH highlights that the paper lacks clarity on key details about the pre-trained models used, such as the specific Sentence Transformer and whether fine-tuning occurs during continual learning. They also point out the absence of a clear comparison of parameter increases due to adding generators for new tasks, noting that the generators are not lightweight. This could result in significant parameter growth and potentially unfair comparisons with related works. Given the rebuttal, the reviewer clearly understands that the increase in parameters is reasonable, and comparable with other research works. He decides to increase the score to 6.

(c) Reviewer UGi2 notes that the paper's writing and overly compact structure hinder readability, suggesting an initial overview and moving the weakly relevant theorem to the supplementary materials. They find the ablation study insufficiently convincing and recommend validating modules step-by-step from a standard baseline. Concerns are raised about the fairness of experimental comparisons involving ChatGPT and the extent of performance gains attributed to accurate descriptions, with a suggestion to make ChatGPT-generated descriptions publicly available to enhance the paper's impact. Unfortunately, the reviewer UGi2 fails to participate in the rebuttal period despite reminders from the AC. The AC has checked the rebuttal and discovered that the authors have addressed the concerns accordingly.

(d) Reviewer bACr finds the comparison of prompt similarities (lines 183-184) flawed, as evaluating similarity on a letter-by-letter basis is inappropriate and fails to reflect semantic differences. They highlight concerns about the reliance on ChatGPT and a Sentence Transformer, which introduce additional resources and create unfair comparisons with other methods, limiting real-world applicability. Although the method achieves state-of-the-art performance, gains are minor on fine-grained datasets like CUB200, and the paper's layout requires improvement for better readability. After the initial rebuttal, the reviewer still has the concerns regarding the resources of pre-computations from GPT. However, the AC notes that the LEAPGen-lite proposed by the authors has already removed the GPT-generation, demonstrating that the concern regarding GPT is not an issue.

最终决定

Accept (Poster)