PaperHub
5.5
/10
Poster4 位审稿人
最低2最高4标准差0.7
3
3
4
2
ICML 2025

Continual Generalized Category Discovery: Learning and Forgetting from a Bayesian Perspective

OpenReviewPDF
提交: 2025-01-21更新: 2025-07-24
TL;DR

VB-CGCD, a variational Bayes-based framework for Continual Generalized Category Discovery, achieved a +15.21% improvement outperforming state-of-the-art methods.

摘要

关键词
Continual LearningGeneralized Category DiscoveryVariational Bayes

评审与讨论

审稿意见
3

This paper introduces Variational Bayes Continual Generalized Category Discovery (VB-CGCD), a Bayesian framework designed to address key challenges in Continual Generalized Category Discovery (C-GCD), including label bias, pseudo-label errors, and the learning-forgetting tradeoff. VB-CGCD comprises four key components: (1) self-supervised offline fine-tuning, (2) self-corrective re-labeling, (3) variational Bayesian distribution estimation with a nearest-class-mean classifier, and (4) covariance-driven early stopping. Experimental results demonstrate that VB-CGCD outperforms existing methods across standard and newly proposed benchmarks with minimal labeled data.

给作者的问题

  1. When the number of new categories is unknown, how does VB-CGCD estimate it effectively?

  2. Can you further explain how the class merging is conducted? Specifically, how are “the latent variables of old classes and the pseudo-labeled distribution merged into a unified MVN-NCM classifier”?

  3. According to the ablation study results in Table 11, early stopping and re-labeling seem to hinder the learning of new classes. Does this suggest that the proposed method prioritizes stability (performance on old classes) over plasticity (performance on new classes) to achieve a tradeoff?

论据与证据

  1. The problem formulation in Eq. (1) requires that the total losses on previous t-1 tasks with parameters leanred after task t-1, does not be increased after learning task t. How is this gaurantee ensured? Does this optimization objective always have a valid solution? It seems that VB-CGCD could not fulfill this constraint, as shown in Figure 2 (a), the accuracy of previous tasks decreases after learning new tasks.

  2. The paper states: “To mitigate feature drift, we perform fine-tuning exclusively during the offline stages and keep the backbone network parameters frozen throughout all sessions, including offline and online classification learning.” However, freezing the backbone may hinder the model’s plasticity, particularly when significant distribution shifts occur between sessions. Further discussion on the trade-off between stability and adaptability would strengthen the paper.

  3. One of the core ideas, i.e., modeling each class as a multivariate normal distribution (MVN), has been explored in previous continual learning works [1-4].

[1] Fecam: Exploiting the heterogeneity of class distributions in exemplar-free continual learning. NIPS2023

[2] Learning semi-supervised gaussian mixture models for generalized category discovery. ICCV 2023

[3] Steering prototype with prompt-tuning for rehearsal-free continual learning. WACV 2024

[4] Happy: A Debiased Learning Framework for Continual Generalized Category Discovery, NIPS 2024

方法与评估标准

Yes.

理论论述

The paper presents the theoretical upper bound (Table 9) of model accuracy at the final session, but does not provide formal proofs. This upper bound might come from experiments.

实验设计与分析

Variational inference-based methods are generally computationally expensive. However, the paper does not compare the computational cost of VB-CGCD with other methods, focusing only on parameter sizes. A computational efficiency analysis would enhance the evaluation.

补充材料

I briefly reviewed the supplementary section.

与现有文献的关系

The key contributions of this paper aim to address the challenge of learning from both known and novel categories in an evolving data stream. By providing a practical solution, it aligns with broader efforts in the field to advance continual generalized category discovery.

遗漏的重要参考文献

Several continual learning works [5-9] that leverage variational inference should be discussed in the related work section for a more comprehensive comparison.

[5] Variational continual learning, ICLR 2018

[6] Variational auto-regressive gaussian processes for continual learning, ICML 2021

[7] Generalized variational continual learning, ICLR 2021

[8] Continual learning via sequential function-space variational inference, ICML 2022

[9] Continual variational autoencoder learning via online cooperative memorization, ECCV 2022

其他优缺点

Other stengths.

The improvement over previous methods is substantial.

其他意见或建议

  1. It is recommended to include punctuation in each numbered equation.

  2. Please ensure the correct quotation marks are used throughout the paper.

作者回复

Claims And Evidence:

The constrained optimization equation formally represents the objective of "learning new classes without forgetting old ones." We observed that maximizing overall accuracy inevitably leads to a decline in the classification performance of previous classes, as illustrated in P2 of Figure 2. Therefore, considering the interplay between old and new classes, we believe that this issue is a multi-objective optimization problem. Accordingly, we relax the original constraint by imposing a covariance constraint, which approximates a balance between preventing forgetting and effectively learning new classes.

With advancements in self-supervised pre-training, the effectiveness of self-supervised fine-tuning in small-scale scenarios has diminished. On the other hand, due to the presence of classification errors, employing supervised fine-tuning during online sessions can introduce a significant number of misclassifications, which in turn leads to a decline in overall performance. Meanwhile, we did not entirely freeze the backbone; instead, we performed fine-tuning with labeled data during the offline phase, which was performed using verified ground-truth labels.

Most methods use the MVN as a sampling distribution for data replay or to assist clustering, rather than directly as a classification prototype. In contrast, VB-CGCD approximates the MVN through variational inference, and we believe that this strategy is the key to our improved performance.

Experimental Designs Or Analyses:

We acknowledge that Variational Inference can be computationally intensive. However, our VI process does not need to reach full convergence; only a few steps (1000 epochs in our experiments) are sufficient to distinguish different categories. Additionally, by not requiring fine-tuning of the backbone network during training, we only need to extract features once for each sample, performing a single forward pass through the backbone, which significantly reduces training time. Consequently, compared to other methods that necessitate frequent fine-tuning of the backbone network, our training time is substantially reduced. The running time during training on one RTX-8000 is shown as follows:

DatasetsOfflineFine-Tuning (mins)OfflineSession (mins)OnlineSession (mins)Overall (mins)
C100673.031.5197
TINY1122.271.58145
IN1001403.361.22169
CUB12.092.011.1239

Compared to HAPPY (over 25 hours of training) and PromptCCD (over 40 hours of training), VB-CGCD demonstrates significantly superior training efficiency. After training, VB-CGCD achieves inference computational efficiency comparable to other prototype-based approaches, normally in milliseconds.

Q&A

As a dynamically extensible incremental learning classifier, VB-CGCD can be integrated with advanced clustering algorithms, including those capable of dynamically estimating the number of categories. Since VB-CGCD does not require a preset number of categories, it can be combined with many off-the-shelf methods for estimating unknown category counts. We employed the classical silhouette score to implement a method for estimating the number of unknown categories and integrated it into VB-CGCD. For experimental results, please refer to Response 2 of Reviewer 1.

VB-CGCD models each category as an independent probability distribution and maintains a collection of all category distributions (i.e., a matrix of latent variables) in memory. After estimating the pseudo-label distribution, we incorporate it into the distribution collection (prototype matrix). During re-labeling, the distances between all samples and this collection are computed to assign labels accordingly. Subsequently, the pseudo-distribution is removed from the collection, and a new distribution is re-estimated. The merge operation is purely an array manipulation, which accounts for VB-CGCD’s inherent scalability, and the reason for being able to handle unknown number of categories.

Since the primary goal of continual learning is to enhance overall learning performance, we agree that preventing overfitting to new classes is necessary to mitigate the forgetting of previously acquired knowledge. In fact, many continual learning methods, whether by adding regularization constraints or employing data replay, aim to some extent to slow down the learning of new classes to prevent forgetting old ones. Without imposing any constraints and allowing unrestricted learning of new classes, the model's classification accuracy on new classes may continue to improve, but the accuracy on old classes could drop drastically, as illustrated in phase P3 of Figure 2. Therefore, to maximize overall accuracy, continual learning must strike a balance between stability (preventing forgetting) and plasticity (learning new categories). This trade-off inherently implies that optimizing one aspect will inevitably come at the expense of the other.

审稿人评论

Thank’s for the rebuttal, which addresses some of my concerns. However, I still believe that the optimization objective in Eq. (1) may not be well-aligned with the proposed method. Introducing a relaxation to the constraint could be beneficial, yet this aspect appears to be missing in the original paper. In addition, while the authors provide an explanation regarding the effect of self-supervised pre-training, they do not directly address my concern about freezing the model after the offline stage, which may limit the model’s plasticity. Lastly, the absence of discussion on relevant works [1–9] weakens the paper’s comprehensiveness. Therefore, I would prefer to maintain my original recommendation at this stage.

作者评论

1. Eq(1)

Eq(1) is a general optimization objective for class incremental learning, which originates from the classical work on continual learning GEM [1]. In our discussion of the P2 phase in Figure 2, we argue that this objective can only achieve a suboptimal outcome (akin to the P1 phase). Eq (1) is not directly related to our work and was not the focus of our optimization; it was included merely as a general definition in the class incremental learning problem formulation.

We will avoid this ambiguity in the revised version.

[1] Gradient Episodic Memory for Continual Learning. NeurIPS 2017.

2. Model plasticity

We employ the pretrained model as a frozen backbone network for feature extraction, which encodes raw images into features. The model's plasticity is principally derived from the Bayesian Neural Network (BNN)-based prototype classifier constructed atop this backbone.

(1) This BNN classifier parametrically models each category's distribution as learnable multivariate Gaussian distributions. The strong approximation capacity of multivariate Gaussians endows the classifier with plasticity for incremental learning.

(2) The per-class independent distribution fitting mechanism enables dynamic expansion, allowing the model to adaptively accommodate arbitrary numbers of novel categories across domains.

Notably, baselines like HAPPY and PromptCCD also adopt backbone freezing with only final-block tuning. Their plasticity primarily stems from MLP classification heads, analogous to our BNN's functional role. Compared to an MLP classification head, BNN offers more flexible dynamic expansion and enhanced adaptability. Besides, many SOTA CIL approaches freeze the backbone while solely training classifiers atop it [1]. This design philosophy preserves the strong generalization capabilities of pretrained models. Representative methods, including FeCAM, L2P [2], and DualPrompt [3], strictly maintain the frozen pretrained backbone without any fine-tuning.

[1] Class-Incremental Learning: A Survey. TPMI, 2024.

[2] Learning to Prompt for Continual Learning. CVPR 2022.

[3] DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning, ICCV 2023.

To demonstrate the plasticity of our model, we continued learning on IN100 after having completed incremental learning on C100 with the pretrained backbone and BNN classifier. The model sequentially learns IN100 categories while retaining C100 knowledge, thereby validating its adaptability.

S0S1S2S3S4S5
AllAllOldNewAllOldNewAllOldNewAllOldNewAllOldNew
82.2481.4281.9476.381.0481.3178.080.080.8270.279.5579.8276.078.679.3168.6

The result shows that VB-CGCD achieved a final accuracy of 78.6, demonstrating its ability to effectively classify both C100 and IN100, which underscores its strong cross-domain adaptability.

3. Missing Related Works

GPC [2] employs a GMM for clustering and estimates unknown category numbers through split-merge mechanisms, focusing primarily on clustering rather than classification. FeCAM [1] utilizes Gaussian prototypes obtained via statistical methods, making them non-learnable and sensitive to data scale variations. In contrast, VB-CGCD employs a learnable BNN to establish Gaussian prototypes, offering greater generality and effectively handling covariance differences between classes.

CPP [3] uses class means as prototypes (prompts) and integrates a Transformer as the classifier, whereas VB-CGCD employs Gaussian distributions along with a non-parametric distance function, enhancing robustness and interpretability. HAPPY [4] utilizes Gaussian prototypes as a replay mechanism to mitigate forgetting, while VB-CGCD is replay-free and directly utilizes prototypes for classification.

VCL [5], GVCL [7], VAR-GP [6], and S-FSVI [8] are methods that approach continual learning as a sequence of tasks, utilizing variational inference to regularize parameter updates via the KL divergence. These techniques bridge Bayesian inference with continual learning, providing insights into the trade-off between plasticity and stability. These methods employ variational inference for regularization in learning likelihood probabilities, while VB-CGCD leverages variational inference to directly learn generative models of data distributions, enabling classification through distance functions. OCM [9] uses variational autoencoders to learn data distributions, serving as samplers for replay mechanisms to mitigate forgetting, whereas VB-CGCD operates without the need for replay.

The revised version will include these in the related work, clarifying technical distinctions and providing a comprehensive discussion.

We hope that our responses have addressed your concerns and would greatly appreciate it if you could consider raising our score. Also, let us know if there are any more concerns.

审稿意见
3

This paper addresses Continual Generalized Category Discovery (C-GCD), or say, iGCD, a task where a model must incrementally learn new classes from unlabeled data streams while preserving knowledge of previously learned classes, a challenge exacerbated by mixed-class data streams and catastrophic forgetting. C-GCD is a task that involves: IL, semi-supervised learning, and class discovery.

In this work, the authors Variational Bayes C-GCD (VB-CGCD), a Bayesian framework that uses variational inference to model class distributions, align covariances between old and new classes, and mitigate forgetting through a covariance-aware nearest-class-mean (NCM) classifier and an early stopping mechanism.

The proposed method, VV-CGCD is evaluated on well-established GCD benchmarks: CIFAR-100, TinyImageNet, ImageNet-100, and CUB-200. The results show improvements over the prev. arts.

给作者的问题

Q1: Sensitivity to Hyperparameters: The method involves several hyperparameters, such as the regularization term and the early stopping threshold. The paper does not provide a detailed sensitivity analysis, leaving it unclear how robust the method is to variations in these hyperparameters across different datasets or tasks.

Q2: How would the method behave in other modalities (non-image data)?

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Yes.

实验设计与分析

Yes.

补充材料

Yes, I reviewed the full supp.

与现有文献的关系

C-GCD methods can be used to discover novel classes/concepts, which can benefits a spectrum of related fields, such as tissue discoevry and protein discovery.

遗漏的重要参考文献

There is a significant lack of discussion on related work in this study.

First, the coverage of the Novel Class Discovery (NCD) and Generalized Category Discovery (GCD) literature is very limited, with only two works from each category being discussed. However, a broader spectrum of NCD techniques is highly relevant to this work, as they serve as the foundation for both GCD and Category-wise GCD (C-GCD). A more comprehensive discussion is needed to highlight the technical and methodological similarities between this study and existing NCD and GCD approaches.

Second, there is a notable omission of discussions and comparisons regarding Class-incremental NCD, which forms the basis of C-GCD. Foundational works in this domain [1, 2, 3] are neither reviewed nor compared, despite their relevance. Addressing these gaps would strengthen the contextualization of this work within the broader research landscape.

[1] Joseph K J, Paul S, Aggarwal G, et al. Novel class discovery without forgetting[C]//European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2022: 570-586.

[2] Roy S, Liu M, Zhong Z, et al. Class-incremental novel class discovery[C]//European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2022: 317-333.

[3] Liu M, Roy S, Zhong Z, et al. Large-scale pre-trained models are surprisingly strong in incremental novel class discovery[C]//International Conference on Pattern Recognition. Cham: Springer Nature Switzerland, 2024: 126-142.

其他优缺点

Strengths:

  1. the paper is well-written and well-presented. It is easy to read.
  2. the proposed Bayesian C-GCD framework is novel, interesting, and highly generalizable.
  3. the improvements on the evaluated benchmarks look promising and significant.
  4. the abaltion study is sufficient.

Weakness:

  1. the benchmarks used for evaluation are significantly limited. The proposed method is based on pre-trained DINO. Note that CIFAR, ImageNet, CUB, are all used in the representation learning of DINO. Although DINO does not use labels to supervise the model, nearly none of the "novel" classes the model meets at each incremental session are really novel because these concepts, or even visual content are senn by the encoder during its pre-training. Therefore, I cannot hold high confidence about the effectiveness of the proposed method with current evaluation. Evalution on truely novel benchmarks are needed.
  2. Insufficient discussion and comparison with related work. iNCD methods should be adapted and compared.
  3. dependence on Initial Clustering Quality: The performance of VB-CGCD relies heavily on the quality of the initial clustering used to generate pseudo-labels. Poor clustering results, especially in high-dimensional spaces, could lead to inaccurate pseudo-labels, which may propagate errors and degrade the model's performance over time.

其他意见或建议

NA

作者回复

Weakness:

We acknowledge that DINO-ViT-B16 was pre-trained on the ImageNet dataset, though its training was based on self-supervised learning using unlabeled data, and it was not trained on CIFAR or CUB. Since all our baselines including HAPPY (SOTA) utilize DINO for feature extraction, VB-CGCD also uses DINO to ensure a fair comparison with other methods. Additionally, to further assess the generalizability of our approach, we have conducted experiments on Stanford Cars, FGVC-Aircraft, CORE50, and Food-101 datasets.

DatasetsS0S1S2S3S4S5
AllAllOldNewAllOldNewAllOldNewAllOldNewAllOldNew
StanfordCars72.5569.4171.0645.3867.0867.7556.4665.4566.1853.5162.3563.7737.6860.9261.2255.39
FGVCAircraft82.3276.2379.2255.4774.9575.2872.3470.1571.9154.5165.1367.8443.6760.6662.6458.03
Core5099.9296.5099.9279.5791.1195.9561.4088.9191.1173.2984.9388.8054.1282.2884.8458.85
Food10184.7881.6781.8081.0479.5179.8777.3278.2577.8581.0475.4176.3068.273.3073.7669.12

Compared to other baselines, VB-CGCD outperforms them by a decent margin in terms of average overall accuracy, as shown below:

MethodsStanfordCarFGVCAircraft
MetaGCD54.6747.16
HAPPY62.7953.10
VB-CGCD65.0469.42

Thank you for your valuable feedback. We will include a discussion on FRoST (Class-iNCD) in the related work section. FRoST employs a Gaussian distribution as the sampling distribution for its replay mechanism, incorporating samples from previously learned classes during training to mitigate forgetting. This represents a typical replay-based approach. In contrast, VB-CGCD does not utilize a replay mechanism but instead relies directly on prototypes for classification. The experimental comparison between FRoST and VB-CGCD is as follows:

DatasetsS0S1S2S3S4S5
AllAllOldNewAllOldNewAllOldNewAllOldNewAllOldNew
C100FRoST90.3676.8779.5863.3065.3168.8843.9058.0161.0936.5049.2750.9036.2048.0348.17
TINYFRoST85.8675.1578.5658.1065.6467.8352.5051.3254.3130.4048.2252.1416.9040.1542.73
IN100FRoST96.2087.5092.9660.2079.6383.3757.2076.7877.0075.2066.1868.6546.4063.8266.40
CUBFRoST90.2677.0383.9543.5350.7753.4634.3346.4249.3126.0939.4041.4723.0834.5535.12

In terms of final overall accuracy, VB-CGCD outperforms FRoST by an average of 30.76%, demonstrating its superior performance.

In unsupervised scenarios, most methods are influenced by the performance of the clustering algorithm. Our strategy is to minimize errors introduced by the clustering algorithm from the classifier's perspective to enhance performance. To ensure fair comparisons with other methods, we have employed the fundamental k-means algorithm. Of course, our approach can also be integrated with more advanced clustering algorithms to further improve performance.

Q&A

Across all datasets and tasks, we employed a consistent set of hyperparameters, including learning rate and training epochs. Specifically, the fine-tuning coefficient λ was uniformly set to 0.001. Regarding the early stopping strategy, we posit that equalizing the average covariance between new and old classes effectively mitigates class bias. Therefore, we monitor the covariance of both new and old classes during training and halt training when they are equal, i.e., when R=0. We have conducted a sensitivity analysis on the R. Please refer to Response 4 of Reviewer 2.

Since our method performs density estimation solely on features, it can be broadly applied to various modalities. We conducted experiments on audio and text classification tasks.
For audio we evaluated our method on esc50, speechcommands, and audiomnist datasets. For text, we evaluated stackoverflow, and clinc.

DatasetsS0S1S2S3S4S5
AllAllOldNewAllOldNewAllOldNewAllOldNewAllOldNew
ESC5072.9875.1072.0396.6678.0574.68100.078.6878.0582.9280.3378.05100.078.078.9370.45
Speechcommands85.2580.4382.3772.1178.2779.1372.6771.7576.9245.3870.5970.5471.0169.1869.5266.44
AUDIO-MNIST96.2296.4096.2297.2696.9496.40100.097.0596.7498.8596.9496.7198.6896.496.7893.75
Clinc98.6694.2298.2274.2292.2594.081.7790.8391.8783.5590.7190.6691.1188.6690.3273.77
Stackoverflow89.086.8387.285.086.2886.6684.084.7585.5779.084.8884.2590.084.884.7785.0

The experimental results indicate that VB-CGCD demonstrates good performance in these tasks as well, further validating its generalizability and robustness to other modalities.

审稿意见
4

This paper studies the task of Continual Generalized Category Discovery (CGCD). It analyzes C-GCD’s forgetting dynamics through a Bayesian lens, revealing that covariance misalignment between old and new classes drives performance degradation. To solve these issues, this paper proposes Variational Bayes C-GCD (VB-CGCD) integrates variational inference with covariance-aware nearest-class-mean classification. VB-CGCD adaptively aligns class distributions while suppressing pseudo-label noise via stochastic variational updates.

给作者的问题

See Other Strengths And Weaknesses.

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

This paper derives the method from the perspective of variational bayes, which is clear.

实验设计与分析

Yes.

补充材料

Yes.

与现有文献的关系

This paper makes novel contributions of bayes variational inference for continual generalized discovery.

遗漏的重要参考文献

Please discuss the recent work of C-GCD Happy [R1] in the Related Work.

References: [R1]. Happy: A Debiased Learning Framework for Continual Generalized Category Discovery. NeurIPS 2024.

其他优缺点

Strengths:

  1. This paper is well-motivated and easy to follow.
  2. The writing and the diagram are clear. I appreciate Figure 1.
  3. The proposed variational Bayes-based method is novel to the community of category discovery.
  4. The performance gains are remarkable.

Weakness:

  1. Discussion of related work Happy [R1] is missing in the Related Work.
  2. Why are some results of Happy in Table1 different from the reported results in the original paper [R1]?
  3. Please include a brief introduction of the method (i.e., spirit, pipeline, etc) in the caption of figure 1.
  4. The experiments of hyperparameters should be added, including ϵ\epsilon in early stopping.
  5. Could the authors provide some experiments that support the claim that self-supervised loss during continual learning introduces feature drift?

References: [R1]. Happy: A Debiased Learning Framework for Continual Generalized Category Discovery. NeurIPS 2024.

其他意见或建议

No.

作者回复

We appreciate the reviewer’s valuable feedback. We have regarded HAPPY as an important SOTA but inadvertently missed its discussion in the related work section. We will include the discussion of HAPPY in the revised version, with a particular emphasis on its technique of using Gaussian distributions for prototype sampling, to provide a more comprehensive overview of the relevant work.

We reproduced the experimental results of HAPPY and found that our reproduced results outperform those reported in the original paper. We reported the reproduced results in our paper.

We will update the overall process in the caption of Table 1 as follows:
"(a) Offline Stage: Labeled data is used to fine-tune the backbone network. After fine-tuning, features are extracted, and variational inference is performed to obtain and store the class distribution.
(b) Online Stage: Unlabeled data is first processed by the feature network to extract features, which are then clustered to assign pseudo-labels. These pseudo-labels are combined with a first round of variational inference to obtain an approximate distribution, then combined with the stored class distribution from the previous session for re-labeling."

Across all datasets, we employed the same hyperparameter configuration, including conventional parameters such as epochs and learning rate, as detailed in Appendix C.3.
Regarding the fine-tuning hyperparameter λ, we found that the contribution of self-supervision to performance is almost negligible. This is primarily because, compared to self-supervised learning on extremely large-scale pretraining datasets, the fine-tuning dataset has a limited impact on performance, with the cross-entropy loss playing a more crucial role. Therefore, we set λ to a very small value of 0.001.

As for the early stopping hyperparameter, as discussed in Figure 2, we argue that ensuring equal average covariance between new and old classes can effectively mitigate class bias. Consequently, our early stopping strategy dictates training to be stopped once the covariances of new and old classes are equal (i.e., when R equals 0).
We conducted experiments on the CIFAR 100 dataset, focusing on the parameter R, adjusting its value to achieve different trade-offs between model plasticity and stability.

Comparisons in overall accuracy as R varies

S1S2S3S4S5
R=385.5582.279.3877.7475.61
R=286.8583.380.278.4176.13
R=188.485.983.4182.5781.11
R=088.486.083.6182.7281.23
R=-187.0383.7581.3180.1877.91
R=-287.0183.7181.2379.9477.71

Comparisons in the accuracy of old classes as R varies

S1S2S3S4S5
R=391.6285.3181.9879.1777.44
R=291.3286.6383.1580.0178.11
R=190.1887.284.6482.8181.92
R=089.4687.0684.7582.8581.84
R=-187.3484.281.9179.5777.91
R=-287.3084.1081.8279.3177.56

Comparisons in the accuracy of new classes as R varies

S1S2S3S4S5
R=355.263.561.266.359.1
R=264.563.359.565.658.3
R=179.578.174.880.773.8
R=083.179.675.681.775.7
R=-185.581.177.185.177.9
R=-285.681.477.185.079.0

The experimental results indicate that the parameter R serves to balance the accuracy between new and old classes, with the overall accuracy being highest when R equals 0. We will include these details in the updated version of the paper.

In continual learning, new data is introduced at each session. If the backbone network is continuously fine-tuned, the features extracted from the same input using different versions of the network (e.g., h₁ and h₂) will inevitably differ, i.e., h₁(x) ≠ h₂(x). We designed an experiment on CIFAR-100, where we continuously perform self-supervised fine-tuning using data from each session. Subsequently, we evaluate feature drift by conducting supervised learning.

SSL-finetunedoverallS0S1S2S3S4S5
S085.3785.383.084.985.290.184.0
S184.5584.9581.983.484.886.784.0
S283.8483.9482.082.2882.488.083.6
S385.0285.3481.884.184.689.184.0
S484.4584.7682.583.583.588.183.2
S585.0184.9981.984.685.688.684.4

The experimental results indicate that as fine-tuning progresses, the accuracy of each session, as well as the overall accuracy, fluctuates within a 2% range, demonstrating the presence of feature drift. Moreover, this drift does not have a consistently positive impact on accuracy and does not lead to significant performance improvements. Since our classification relies on storing the distribution of old classes, any drift may result in inaccurate classifications. Therefore, we chose to avoid continuously fine-tuning the pre-trained network.

审稿意见
2

This manuscript investigates the problem of Continual Generalized Category Discovery, addressing the challenges of mixed new and old categories and high uncertainty in unlabeled data under a continual learning setting. The authors propose a new variational Bayesian framework that utilizes offline fine-tuning and self-correcting re-labeling to mitigate label bias and label noise. By employing covariance-aware early stopping, the framework balances new-category plasticity and old-category stability.

给作者的问题

Please see weaknesses.

论据与证据

Yes

方法与评估标准

Yes

理论论述

Yes

实验设计与分析

Yes

补充材料

Yes

与现有文献的关系

No

遗漏的重要参考文献

No

其他优缺点

Strengths:

1.The paper is well written, with strong readability and logical clarity.

2.The newly designed framework demonstrates impressive performance, achieving state-of-the-art results.

Weaknesses:

1.The experimental design is not sufficiently comprehensive. The ablation study only covers the removal of entire modules but does not analyze specific design components within those modules. For instance, the paper mentions using Mahalanobis similarity to account for covariance structures and mitigate the curse of dimensionality. However, the benefit of introducing Mahalanobis similarity into the proposed framework has not been clearly demonstrated.

2.Although the paper states that existing methods can effectively estimate the number of new categories when it is unknown, there are no experiments or analyses for scenarios in which the number of new categories in each session is indeed unknown. Such a setting would be crucial to validate the effectiveness of the proposed framework in real-world conditions.

3.The framework reduces interference by distinguishing between new and old categories, offering an improvement over purely distance-based methods. However, the paper lacks a more in-depth comparison with open-set recognition and GCD approaches that also differentiate known from unknown classes. (e.g. Extreme Value Theory in open-set recognition and DPN[1] in GCD )

I would be happy to increase my score if the authors can resolve all concerns.

[1] Generalized Category Discovery with Decoupled Prototypical Network

其他意见或建议

None

作者回复

We conducted an ablation study on the benefit of introducing Mahalanobis distance, and the results obtained by replacing the Mahalanobis with the Euclidean distance are as follows:

DatasetsMethodsAllS0S0AllS5S5MfMd
C100w/oMahalanobis90.5276.0610.174.96
Tinyw/oMahalanobis87.6470.339.767.5
IN100w/oMahalanobis93.685.068.0486.56
CUBw/oMahalanobis83.7855.790.8228.15

Compared to VB-CGCD with Mahalanobis distance, the overall final accuracy decreased by approximately 5.56 on average. This is because Euclidean distance, being a special case of Mahalanobis, is susceptible to collapse in high-dimensional spaces. By incorporating covariance information, Mahalanobis distance effectively mitigates this issue, thereby achieving a more optimal classification boundary. We appreciate your valuable feedback and will incorporate this aspect into the revised version of our paper.

Since our proposed approach primarily serves as an incremental learning classifier and is decoupled from the clustering module, it allows integration with any clustering method (including those that estimate the number of categories).
As mentioned in the paper, there are many off-the-shelf methods [1][2] available for estimating the number of categories.
Nonetheless, we employed the classic silhouette score to implement a method for estimating the number of unknown classes, which we integrated into VB-CGCD.
The results are as follows:

DatasetsS0S1S2S3S4S5
AllAllOldNewAllOldNewAllOldNewAllOldNewAllOldNew
C10091.6882.4887.258.979.0882.3159.776.2578.9057.775.2476.0768.673.1374.9456.8
Tiny88.3285.5586.8279.282.4184.0172.879.8781.4169.177.2078.8264.274.7776.3460.6
IN10094.3291.9393.1286.089.4590.4383.688.688.2890.886.1187.6274.084.5685.0879.8
CUB85.7279.9883.6860.9165.9867.4257.4163.9165.9449.4757.5458.2451.7654.4057.5425.61

Due to errors in estimating the number of unknown classes, the clustering error is exacerbated, ultimately leading to an average overall accuracy reduction of approximately 5.66. Nonethless, it demonstrates the scalability of our method to handle scenarios with an unknown number of types, and our performance still outperforms the SOTA method, HAPPY, which also uses the silhouette score.

MethodsAllOldNew
HAPPY68.8072.4045.74
VB-CGCD77.2379.8860.34

[1] DeepDPM: Deep Clustering with an Unknown Number of Clusters. Meitar Ronen, Shahaf E. Finder, Oren Freifeld. CVPR 2022.
[2] PromptCCD: Learning Gaussian Mixture Prompt Pool for Continual Category Discovery. Fernando Julio Cendra, Bingchen Zhao, Kai Han. ECCV 2024.

Our approach distinguishes between known and unknown classes by approximately estimating the unknown distribution, thereby effectively reducing the interference of known classes on the learning of new ones—a strategy that is similar to certain open-set methods. However, there are two key differences between GCD and CGCD: 1) GCD does not need to address the forgetting problem inherent in continual learning; 2) In the unsupervised learning phase of CGCD, supervised data is inaccessible, making it unsuitable for semi-supervised approaches. For example, during training, DPN relies on using L_{known} as a part of the loss function, which renders it inapplicable in CGCD scenarios. Moreover, methodologically, DPN uses the mean as its prototype; however, in high-dimensional spaces, this approach is prone to feature collapse. In contrast, utilizing a Gaussian distribution is more robust. We will further emphasize this distinction in the related work section.

Additionally, we evaluated VB-CGCD on the same datasets which were evaluated in DPN ‘s paper, and the results are shown below.

bankingstackoverflowclinc
AllKnownNovelAllKnownNovelAllKnownNovel
DPN72.9680.9348.6084.2385.2981.0789.0692.9777.54
VB-CGCD75.5582.8853.1584.283.7385.690.9795.1678.19

The experimental results indicate that VB-CGCD achieves performance comparable to DPN. However, it is important to note that VB-CGCD does not need access to any labeled data during the unsupervised learning phase, which is a distinct difference from DPN.

最终决定

This paper introduces a variational Bayes-based framework for Continual Generalized Category Discovery (C-GCD), designed to address the challenges of incrementally learning new classes from unlabeled data streams while retaining knowledge of old, previously-seen classes. The authors analyze the forgetting dynamics in C-GCD from a Bayesian perspective and identify covariance misalignment between old and new classes as a key driver of performance degradation. Based on this insight, they introduce VB-CGCD, which integrates variational inference of Gaussian parameters and covariance-aware nearest-class-mean classification. Experimental results on standard benchmarks demonstrate that VB-CGCD significantly outperforms state-of-the-art methods. Of particular note is the performance of the proposed approach on a benchmark with only 10% labeled data in the first, offline supervised training session.

Reviewer opinion of the paper was initially mixed. Reviewers appreciated the readability and clarity of the technical narrative, the impressive performance improvement over baselines on standard benchmarks, and the novelty of the Bayesian C-GCD framework. On the other hand, they raised concerns over insufficiently comprehensive experiments, limited discussion of related work on incremental Novel Class Discovery, and the dependence on known number of classes on the quality of clustering.

The authors engaged with all reviewer concerns in their rebuttal, providing additional results and promising to incorporate more comprehensive discussion of related work and new results in the revised manuscript. Given that the standard practice in GCD and CGCD is to assume a known number of classes (and relegate estimation of this to a generic clustering algorithm), this criticism of the work is not well-founded -- although the authors are encouraged to include the comparison with HAPPY using silhouette score in the revised manuscript.

In terms of issues regarding coverage of the literature in the Related Work section, given page limits and need to narrowly zoom in on the real contributions for conference papers, it is very hard to range far afield and include everything even in the last year that has anything to do with variational inference, continual learning, or GCD. The authors articulated in rebuttal their motivations for excluding some of the mentioned works, and will update the Related Work to include some of the more relevant ones.

After addressing these critical issues in rebuttal, the consensus that remains is that this is a solid and novel contribution which yields excellent improvements over the state-of-the-art.