PaperHub
7.3
/10
Poster4 位审稿人
最低4最高5标准差0.5
4
4
5
5
2.0
置信度
创新性3.3
质量3.0
清晰度3.0
重要性2.8
NeurIPS 2025

Minimal Semantic Sufficiency Meets Unsupervised Domain Generalization

OpenReviewPDF
提交: 2025-05-01更新: 2025-10-29

摘要

关键词
minimal sufficient representationssemantic learningunsupervised domain generalization

评审与讨论

审稿意见
4

This manuscript introduces MS-UDG, an unsupervised domain generalization method that learns minimal semantic representations by disentangling semantics from variations without requiring labels. While it shows improved performance over prior SSL and UDG methods, its contributions are largely incremental, with limited novelty. Additionally, inaccuracies in table annotations and weak theoretical insights diminish its overall contribution.

优缺点分析

Strength

  1. The experimental evaluation is comprehensive, including both quantitative benchmarks and qualitative visualizations.
  2. The paper attempts to provide theoretical foundations for the proposed contrastive learning framework.

Weakness

  1. The manuscript offers limited theoretical novelty, as it relies on the well-established concept of Minimal Sufficient Semantic Representation (Line-145) without introducing new insights. Moreover, information-theoretic frameworks have already been extensively explored in domain generalization [#1,#2,#3,#4]. The proposed method exhibits little distinction from prior work, as it merely applies existing ideas without providing any theoretical advancements.
  2. Compared to existing work [#5], this manuscript primarily proposes incremental improvements, including introducing a mask, applying self-attention, and incorporating additional features for contrastive learning. Overall, the manuscript reads more like an extension or combination of [#5] and [#6] rather than presenting fundamentally new contributions.
  3. In Table 1, under the Label Fraction: 10% setting, MS-UDG achieves the second-best performance in the photo and art domains, yet the authors mark it as the best. Similar inconsistencies appear in other parts of the manuscript, where MS-UDG results are incorrectly highlighted as the top-performing method despite ranking second. These misannotations may mislead readers and should be carefully corrected.
  4. In the ablation study shown in Table 5 for the sketch domain, the performance when all losses are combined is worse than using each loss individually.
  5. In Figure 3, the t-SNE visualization shows that the semantic features from MS-UDG exhibit poor category separability.
  6. It is well known that self-supervised learning (SSL) benefits unsupervised tasks; this is not a novel contribution to the manuscript. The claim of operating without domain and category labels is also expected in the unsupervised setting and should not be regarded as a unique contribution.

[#1] INSURE: an Information theory iNspired diSentanglement and pURification modEl for domain generalization[J]. IEEE Transactions on Image Processing, 2024.

[#2] Rethinking domain generalization: Discriminability and generalizability[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2024.

[#3] How Does Distribution Matching Help Domain Generalization: An Information-theoretic Analysis[J]. IEEE Transactions on Information Theory, 2025.

[#4] Invariant information bottleneck for domain generalization[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2022

[#5] Towards unsupervised domain generalization[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022

[#6] Rethinking minimal sufficient representation in contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16041–16050, 2022

问题

Please refer to the weakness for more questions. I am open to changing the score if the concerns are well addressed.

Additional question

  1. The concepts of sufficiency and minimality are directly adopted from existing work. It remains unclear what novel theoretical or practical contributions this manuscript provides beyond these established definitions.

局限性

yes

最终评判理由

Thank the authors for their response. While I appreciate the clarifications provided, using an existing theoretical solution in SSL for the first time in DG does not, in my view, constitute a strong theoretical/practical contribution, especially since SSL has already been justified in tackling DG problems in previous works. The contribution does not yet meet the bar expected for NeurIPS on my side, and I do not believe that this work will have a noticeable impact to the community. Nonetheless, after carefully considering the completeness of this work and the ratings from other reviewers, I can increase my score to borderline acceptance.

格式问题

N/A

作者回复

We thank the reviewer for the detailed and insightful feedback on our manuscript. We reply on each point raised and address them as follows:

W1,W6,Q1: Theoretical novelty

W1,W6,Q1-Ans. First, we clarify that minimal sufficient representation builds on the established statistical concept of minimal sufficient statistics. Information bottleneck (IB) previously applied these principles to supervised learning, addressing the research problems of that time. Some methods extend this to the SSL scenario to discuss learning a minimal sufficient representation in unsupervised learning.

Our framework adapts minimal sufficient statistics for learning domain-invariant representations in unsupervised domain generalization (UDG). Crucially, UDG and supervised DG face fundamentally different theoretical challenges. PAC learning risk bounds for supervised domain generalization don't apply to self-supervised models where labels aren't available during training. This theoretical gap requires new frameworks tailored to the unsupervised setting. Our work provides the first information-theoretic approach developed specifically for UDG. Two learning objectives are induced from our theoretical framework, with empirical results validating this theoretical foundation across multiple benchmarks. In addition, our framework first relates the representation in SSL model to the downstream generalization bounds in UDG.

In the end, we appreciate the reviewer thinking about the theoretical improvement. Our theoretical discovery on UDG parallels how IB-based methods [7,8] use the same statistical foundation for domain generalization—these works appeared at top conferences and demonstrated how classical concepts can tackle modern problems. In addition, the analogy can extend to methodological relationships: ERM(empirical risk minimization,i.e., CrossEntropy for classification) provides the baseline for domain generalization methods, while SSL serves as the foundation for UDG approaches like ours. We will clarify and emphasize the unique contributions in the revised version.

W2: Technical improvements

W2-Ans. We thank the reviewer for the detailed technical discussion. Our paper's core contributions span two interconnected areas: 1) information-theoretically grounded optimization objectives for UDG; 2) practical algorithms that implement these theoretical insights.

Our approach differs fundamentally from existing UDG methods. The UDG baseline [5] employs adversarial training with domain labels to achieve domain invariance. This strategy operates independently of information-theoretic principles and targets different learning objectives than our framework. Meanwhile, H. Wang et al. [6] focus on capturing non-shared task-relevant information across domains to avoid overfitting, which contrasts with our goal of disentangling semantic content within the shared representation space for domain generalization.

These methodological differences reflect deeper theoretical distinctions. While previous work addresses domain invariance through adversarial objectives or captures domain-specific information, our framework tackles the problem by formalizing what constitutes an optimal representation for generalization: one that preserves semantic sufficiency while achieving minimality. This information-theoretic perspective leads to novel optimization strategies that haven't been explored in prior UDG, DG, and SSL literatures.

W3,W4: 1. Misannotation and 2. table explaination

W3,W4-Ans.

Misannotation: We thank the reviewer for identifying the annotation errors. These have been corrected in the revised manuscript.

Table explaination: The tabulated results demonstrate our method's effectiveness across domains. Our loss design aims to improve generalization consistently across diverse domains rather than optimizing for individual domain performance. While the sketch domain shows a marginal decrease when using the complete loss formulation versus Lmax\mathcal{L}_{max} alone, the remaining domains (photo, art, cartoon) all demonstrate substantial improvements. This trade-off yields the highest overall average accuracy (45.26), confirming that our unified framework successfully balances performance across the full domain spectrum rather than overfitting to specific domain characteristics.

W5: Explanation of the t-SNE visualization

W5-Ans. The t-SNE plots examine domain invariance properties by showing how our semantic representations S\mathcal{S} exhibit reduced domain clustering compared to baseline methods. The representation vv and vanilla MAE representations display clear domain separability, which validates our disentanglement approach, domain-specific information concentrates in the variation subspace, while semantic content becomes more domain-agnostic. This analysis targets the fundamental question of whether our method successfully separates domain-related variations from semantic content, rather than evaluating downstream classification performance. The visualization demonstrates that semantic representations avoid the domain-specific clustering patterns that typically hinder generalization, which aligns with our theoretical framework's emphasis on learning domain-invariant semantic features.

[1] INSURE: an Information theory iNspired diSentanglement and pURification modEl for domain generalization[J]. IEEE Transactions on Image Processing, 2024.

[2] Rethinking domain generalization: Discriminability and generalizability[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2024.

[3] How Does Distribution Matching Help Domain Generalization: An Information-theoretic Analysis[J]. IEEE Transactions on Information Theory, 2025.

[4] Invariant information bottleneck for domain generalization[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2022

[5] Towards unsupervised domain generalization[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022

[6] Rethinking minimal sufficient representation in contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16041–16050, 2022

[7] Ahuja, Kartik, et al. "Invariance principle meets information bottleneck for out-of-distribution generalization." Advances in Neural Information Processing Systems 34 (2021): 3438-3450.

[8] Du, Yingjun, et al. "Learning to learn with variational information bottleneck for domain generalization." European conference on computer vision. Cham: Springer International Publishing, 2020.

评论

Thank the authors for their response. While I appreciate the clarifications provided, using an existing theoretical solution in SSL for the first time in DG does not, in my view, constitute a strong theoretical/practical contribution, especially since SSL has already been justified in tackling DG problems in previous works. The contribution does not yet meet the bar expected for NeurIPS on my side, and I do not believe that this work will have a noticeable impact to the community.

Nonetheless, after carefully considering the completeness of this work and the ratings from other reviewers, I can increase my score to borderline acceptance. It is suggested to highlight the potential impact of this work.

评论

We appreciate the reviewer for the thoughtful suggestion. We will highlight the potential impact of our work in the revised manuscript.

审稿意见
4

Authors propose an Unsupervised Domain Generalization (UDG) algorithm with theoretical analyses from the perspective of information theory. They also conduct extensive experiments to show its effectiveness.

优缺点分析

Strengths:

  • There are theoretical justifications for the proposed algorithm.
  • Experimental results clearly demonstrate the effectiveness of the proposed algorithm.

Weaknesses:

  • Newer and more recent SSL methods could be added as baselines, e.g. MoCo-v3 [1], SMoG [2], ReLICv2 [3], BAM [4]

[1] Chen X, Xie S, He K. An empirical study of training self-supervised vision transformers[C]//Proceedings of the IEEE/CVF international conference on computer vision. 2021: 9640-9649.

[2] Pang B, Zhang Y, Li Y, et al. Unsupervised visual representation learning by synchronous momentum grouping[C]//European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2022: 265-282.

[3] Tomasev N, Bica I, McWilliams B, et al. Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?[J]. arXiv preprint arXiv:2201.05119, 2022.

[4] Shalam D, Korman S. Unsupervised Representation Learning by Balanced Self Attention Matching[C]//European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2024: 269-285.

问题

Please refer to weaknesses.

局限性

Yes

最终评判理由

My concerns are well addressed in rebuttal, so I decide to raise my score.

格式问题

N/A

作者回复

We thank the reviewer for the positive recognition of the theoretical justification and the effectiveness of our proposed algorithm based on the experimental results.

Regarding the suggestion to include newer SSL methods such as MoCo-v3 [1], SMoG [2], ReLICv2 [3], and BAM [4] as additional baselines, we appreciate this valuable recommendation. We have conducted additional experiments incorporating SSL methods MoCo-v3, SMoG, and BAM within UDG settings, as outlined below. Our method demonstrates competitive performance when compared to these SSL approaches.

Regarding ReLICv2, the publicly available implementation provides only linear evaluation components without the complete SSL pretraining pipeline. Despite our efforts to contact the original authors for implementation details, time constraints prevented us from developing a full reimplementation from scratch. We acknowledge this limitation in our current evaluation and commit to including ReLICv2 in future benchmark extensions once the implementation becomes available.

We plan to publicly release our UDG benchmark implementations for these SSL methods to facilitate future research and ensure reproducible comparisons within the community.

All reported experiments utilize the official codebases of the respective SSL methods, adapted to conform with standard UDG evaluation protocols while maintaining their core algorithmic components.

target domainphotoartcartoonsketchavg
SMoG70.8148.7245.7648.7153.50
BAM52.1654.3346.7258.8953.03
MoCo-v370.6164.1054.0352.2160.24
Ours74.1061.9063.7373.6668.35

Table 1. The results on PACS under 10% label fraction. We report the accuracy for every domain and the average accuracy for all domains.

target domainclipartinfographquickdrawpaintingrealsketchoverallavg.
SMoG41.7517.4117.1524.2235.5026.5226.3727.09
BAM74.9630.8825.6637.0746.2447.6743.9843.75
MoCo-v374.9030.5630.9852.5462.3152.8549.3350.69
Ours79.7030.0140.1153.7363.7765.8253.3755.52

Table 2. The results on DomainNet under 10% label fraction. We report the accuracy for every domain and the average accuracy for all domains.

评论

Thanks for your rebuttal. I decide to raise my score to 4.

评论

Thank you very much for the feedback. We truly appreciate your time and valuable input.

审稿意见
5

This paper explores Unsupervised Domain Generalization (UDG), a task aimed at improving the generalization of unsupervised learning models like Self-Supervised Learning (SSL) without using category labels. While prior UDG methods rely on domain labels to separate semantics from variations, this approach is limited by the unavailability of such labels in real-world scenarios. The authors propose a new framework called Minimal-Sufficient UDG (MS-UDG), which learns semantic representations by optimizing for both sufficiency (retaining shared semantic information) and minimality (removing irrelevant variations). Based on information theory, MS-UDG combines an InfoNCE-based objective with novel components to disentangle semantics and variations, achieving state-of-the-art performance on UDG benchmarks without needing category or domain labels.

优缺点分析

Strengths:

  1. This paper is well motivated.

  2. The experimental results are good.

  3. Overall, the proposed method is simple and effetive.

Weaknesses:

  1. Line 166 I(x_1;x_2;T) = I(x_1;x_2;S), what does S stands for? According to the following text, S may stand for semantic. But what does semantic information in X_{ssl} means? SSL should contain no label, so it hard to understand.

  2. Line 116~117 downstream supervised data P(X_{sup}, Y_{sup}) is used for fine-tuning? Is it the source domain ine Line 270?

问题

Please refer to the Weaknesses.

局限性

yes

最终评判理由

My issues are addressed. After reading all reviewers' comments and authors responses, I prefer to maintain my initial rating.

格式问题

No

作者回复

We thank the reviewer for the thoughtful and positive feedback, including sound motivation, effectiveness of the method ,and promising experimental results. Below, we address each comment and question in detail.

W1: Explanation of semantic factor S\mathcal{S}

W1-Ans. While self-supervised models cannot anticipate specific downstream tasks during pretraining, we can systematically identify categories of information that consistently impede cross-domain generalization, such as task-irrelevant information. In UDG settings, task-irrelevant information typically encompasses domain-specific characteristics such as imaging style, background patterns, lighting conditions, and acquisition artifacts, collectively represented as V\mathcal{V} in our framework. Given our decomposition Z=SV\mathcal{Z}=\mathcal{S}\oplus\mathcal{V}, the semantic component S\mathcal{S} captures the complementary information: fundamental visual structures, object relationships, and content-based features that remain consistent across domains and prove valuable for downstream recognition tasks.

This principled separation between semantic content and domain variations builds upon established theoretical foundations in UDG literature [1,2]. The practical significance becomes evident in real-world scenarios where training data inevitably spans multiple domains. For instance, medical datasets may aggregate scans from different institutions, each using varied imaging protocols, scanner manufacturers, and acquisition parameters. The resulting domain shifts introduce spurious correlations that can mislead models away from clinically relevant features. Our framework addresses these challenges by explicitly optimizing representations to retain semantic sufficiency while systematically reducing dependence on domain-specific information.

W2: Clarify the finetuning data

W2-Ans. Both the SSL pretraining data XsslX_{ssl} and supervised finetuning data XsupX_{sup} originate from source domains, following the standard UDG protocol where no target domain data is accessible during training. This setup mirrors classical domain generalization but operates without category labels during the representation learning phase. The key challenge lies in learning generalizable features from source domain variations that transfer effectively to unseen target distributions at test time. Our experimental setup strictly adheres to the benchmark protocols established in [1,2], ensuring fair comparison with existing methods. Specifically, we use the same train/validation/test splits and domain partitions as prior UDG works, where source domains provide both unlabeled data for SSL pretraining and labeled data for downstream task finetuning, while target domains remain completely held out until evaluation.

[1] Zhang, An, et al. "Disentangling Masked Autoencoders for Unsupervised Domain Generalization." European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2024.

[2] Zhang, Xingxuan, et al. "Towards unsupervised domain generalization." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.

审稿意见
5

The paper addresses Unsupervised Domain Generalization (UDG) for image classification by formalizing the task through an information-theoretic framework. The authors propose learning Minimal Sufficient Semantic Representations that capture shared semantic information across multiple augmented views of the same image while eliminating semantically irrelevant information. The approach includes theoretical proof demonstrating that this formulation reduces out-of-distribution risk, which directly addresses the UDG challenge.

优缺点分析

Strengths:

  • Writing Quality - The paper is exceptionally well-written with clear and well-motivated technical components.
  • Comprehensive Evaluation - The experimental comparison against existing methods is extensive and thorough, providing strong empirical validation.
  • Complete Ablation - The ablation studies effectively justify the use of both loss components, demonstrating that their combination yields superior performance.
  • Theoretical Foundation - The paper provides comprehensive theoretical analysis, introducing all necessary concepts regarding minimal sufficient semantic representations before presenting the method and its implementation.

Weaknesses:

  • Missing Failure Analysis - The paper lacks a failure analysis section examining cases where competing methods outperform MS-UDG. Such analysis would provide valuable insights into the method's limitations and help readers understand when and why the approach may not be optimal.

问题

Can you provide failure case analysis explaining why certain competing methods outperform MS-UDG in specific scenarios, and what insights does this offer about your method's limitations ?

局限性

yes.

最终评判理由

I appreciate the authors' response addressing my concerns regar sectionding the limitations of MS-UDG in some specific scenarios. The provided explanation has resolved my initial question, and I will therefore keep my original score.

格式问题

no concerns.

作者回复

We appreciate the reviewer for the positive recognition of our work, including the writing quality, comprehensive evaluation, complete ablation studies, and theoretical foundation. Below, we provide a detailed analysis of the failure cases and discuss the insights gained from competing methods that sometimes outperform MS-UDG in specific scenarios.

Missing Failure Analysis We conducted a comparative analysis with two advanced UDG methods, BSS and CycleMAE. The results highlight the strengths and limitations of each approach.

  • BSS (Fourier Augmentation Method): BSS benefits from mixing domain information through augmentation techniques, which sometimes enables it to outperform our method in domains with simpler and more uniform textures, such as sketches, quickdraw, and clipart. This improvement may not be suitable for rich-style domains, like art and real, since these domains contain richer low-frequency information compared with other domains. The Fourier augmentation in BSS may not significantly alter the style, which limits the effectiveness of data augmentation. Without disentanglement or other generalization techniques, the performance of BSS on rich-style domains is often inferior to that on other domains.

  • CycleMAE (Generation-Based Method): Generation-based methods, such as CycleMAE, learn informative representations through reconstruction. However, domains like sketches often result in trivially low values of reconstruction losses because of simple backgrounds and domain styles. As a result, CycleMAE tends to focus on learning representations from richer-style domains, such as real images, art, and infographics. This bias toward more complex domains can cause CycleMAE to outperform MS-UDG in those specific areas, while in simpler-style domains, it underperforms compared to MS-UDG..

Even though these two methods occasionally perform better than our method through their own advantages, our MS-UDG method strikes a balance between these two approaches, which can enhance the generalization ability. LmaxL_{max} aims to learn the domain-related representation through reconstruction and LminL_{min} helps disentangle the domain-related information from sufficient semantic representations. As a result, our method achieves state-of-the-art performance on average, while still maintaining robustness across a diverse set of domains.

We will add these discussions to the revised manuscript. Thank you for the advice for better insights into those methods.

评论

I appreciate the authors' response addressing my concerns regar sectionding the limitations of MS-UDG in some specific scenarios. The provided explanation has resolved my initial question.

评论

Thank you for your thoughtful feedback. We sincerely appreciate your time and insightful comments on our work.

最终决定

The paper introduces a new method for unsupervised domain generalization (UDG), building upon contrastive learning principles but proposing explicit loss components to achieve minimal sufficient semantic disentanglement. The authors presents their method from an information-theoretic perspective, distinguishing between semantic and variation components of the learned representation, and to optimize explicitly for minimal sufficient semantics. All reviewers agree that this is a solid paper with thorough experimental results and sufficiently well-motivated theory. The empirical evaluation is comprehensive, and the theoretical grounding connects well with the proposed training objectives. The ablation studies also demonstrate the usefulness of the loss components. The reviewer comments were also well addressed during the rebuttal. Overall, this is a good contribution for the important problem of semantic disentanglement and unsupervised domain generalization, a nice balance between reasonably-well motivated theory and training loss components that seem to demonstrate good experimental results.