PaperHub
6.4
/10
Poster4 位审稿人
最低3最高5标准差0.7
4
5
3
4
4.3
置信度
创新性2.8
质量3.0
清晰度2.8
重要性3.0
NeurIPS 2025

ComRank: Ranking Loss for Multi-Label Complementary Label Learning

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29

摘要

关键词
multi-label learningcomplementary label leaningweakly supervised learning

评审与讨论

审稿意见
4

This work introduces a ranking loss framework for multi-label complementary label learning (MLCLL). Different from existing methods that are based on the assumption that complementary labels follow a uniform distribution, the proposed framework considers the instance-dependent assumption and introduces a complementary ranking loss for the MLCLL task.

优缺点分析

This work introduces a ranking loss framework for multi-label complementary label learning (MLCLL). Different from existing methods that are based on the assumption that complementary labels follow a uniform distribution, the proposed framework considers the instance-dependent assumption and introduces a complementary ranking loss for the MLCLL task.

Strengths:

  1. The writing of this work is clear and easy to follow.
  2. The presented mathematical proofs are rigorous.
  3. The motivation of this work is convictive and experimental results verify the effectiveness of the proposed complementary ranking loss.

Weaknesses:

  1. My main concern focuses on the section of "Complementary Label Distribution". Regarding the uniform distribution assumption, it’s limitations illustrated in this work are easy to follow and persuasive. However, the introduced biased distribution assumption in Eq. (5) is quite challenging to follow and interpret. From my understanding from Eq. (5), p(k=yx)p(k={\overline y} | x) is inversely proportional to p(kYx)p(k \in Y | x). However, it is quite different from the biased distribution assumption. I wonder if there is a misunderstanding on my part. It would be helpful if the authors could offer a more detailed and intuitive explanation for Eq. (5).
  2. Another concern of mine is about "Assumption 4.4". I think that Assumption 4.4 is not suitable for a multi-label learning setting, it is more suitable from the single-label learning setting. As the example provided on Page 4, Line 162, if there are more than on animals in an image, a complementary label provided by annotators cannot be solely based on the image’s features. However, Assumption 4.4 plays an important role in this work, I wonder if there is a misunderstanding on my part. It would be helpful if the authors could offer a more detailed and intuitive explanation for Assumption 4.4.

I would be willing to give a higher score if the authors can address the above-mentioned concerns in the response.

问题

Please refer to "Strengths And Weaknesses"

局限性

Please refer to "Strengths And Weaknesses"

最终评判理由

My concerns have been addressed, and the clarifications provided helped me better understand the contribution. Therefore, I have decided to raise my overall score.

格式问题

Please refer to "Strengths And Weaknesses"

作者回复

Thank you for reviewing our paper and providing valuable insights. We appreciate your feedback and would like to respond to the weaknesses raised.

R1: My main concern focuses on the section of "Complementary Label Distribution". ... It would be helpful if the authors could offer a more detailed and intuitive explanation for Eq. (5).

The core idea of Eq. (5) is to model a biased complementary label distribution. Specifically, it assumes that labels outside the relevant label set have non-uniform probabilities of being selected as complementary labels, in contrast to the uniform assumption where all irrelevant labels are equally likely to be chosen.

Eq. (5) expresses the intuition that the probability of an irrelevant label being chosen as a complementary label depends on its correlation with the relevant labels: the more similar a label is to the relevant labels, the less likely it is to be selected as a complementary label [1]. Thus, the probability of selecting ans irrelevant label kk as a complementary label is modeled to be inversely proportional to its likelihood of being relevant, i.e., p(kYx)p(k \in Y \mid x). All the above motivates our assumption in Eq. (5):

p(k=yˉx)1p(kYx),ifkYp(k = \bar{y} \mid x) \propto \frac{1}{p(k \in Y \mid x)}, if k \notin Y

For example, given an image of a body of water, "desert" (a semantically distant label) is more likely to be excluded than "lake" (which is similar to a possible true label like "river"). This example illustrates how label similarity influences complementary label selection, leading to a biased (non-uniform) distribution. We will revise the manuscript to clearly explain the assumption expressed in Eq. (5), and improve the intuitive explanation.

R2: Another concern of mine is about "Assumption 4.4". ... It would be helpful if the authors could offer a more detailed and intuitive explanation for Assumption 4.4.

Assumption 4.4 aims to reflect a more realistic and relaxed generation mechanism in MLCLL, where annotators typically do not have access to the full relevant label set when assigning a complementary label. According to previous work [2,3,4], this aligns with typical weakly supervised scenarios, where annotations rely on salient or dominant features rather than full label information.

In some scenarios—such as an image containing multiple animals—annotators may struggle to exclude all relevant categories based solely on features. However, our assumption is motivated by the observation that annotators often eliminate obviously irrelevant labels by inspecting the input instance (e.g., excluding "building" from an image showing animals), without needing to infer the full set of relevant labels. This behavior supports modeling yˉ\bar y as conditionally independent of YY given x x.

We adopt this assumption as a tractable approximation that reflects limited annotator knowledge and allows us to derive a theoretically sound risk estimator under biased complementary label distributions. We will revise the manuscript to more clarify this motivation.

References

[1] Gao et al., Complementary to multiple labels: A correlation-aware correction approach, TPAMI, 2024.

[2] Chen et al., Beyond class-conditional assumption: A primary attempt to combat instance-dependent label noise, AAAI, 2021.

[3] Xia et al., Part-dependent label noise: Towards instance-dependent label noise, NeurIPS, 2020.

[4] Zhu et al., A second-order approach to learning with instance-dependent label noise, CVPR, 2021.

评论

Thank you for the authors' detailed response. My concerns have been adequately addressed, and the clarifications provided helped me better understand the contribution. Therefore, I have decided to raise my overall score.

评论

Thank you for your encouraging support, and thank you again for your time and many helpful comments! We will revise our paper accordingly!

审稿意见
5

This paper proposes a novel algorithm for multi-label complementary label learning (MLCLL), supported by solid theoretical analysis and strong empirical results.

However, it suffers from certain conceptual inconsistencies (see Issue 1–2), incomplete consideration of related work (see Issue 3), and potential flaws in the experimental design (see Issues 4–5).

Overall, although I raised a lot of issues, the merits outweigh the flaws. I suggest a weak accept.

优缺点分析

Strengths

  1. Clear motivation. Due to practical concerns such as privacy protection, MLCLL is an important problem to study. This paper is among the first to point out that in real-world scenarios, complementary labels are not sampled uniformly at random, but tend to be semantically opposed to true labels.

  2. Strong theoretical foundation. The proposed method is proven to be Bayes-consistent, which is a desirable theoretical property.

  3. Extensive experiments. The paper provides a comprehensive empirical evaluation across a wide range of baselines, datasets, metrics, and ablation analyses.

Weaknesses

  1. Potential contradiction in modeling assumptions. Although Assumption 4.4 (instance-dependence of complementary labels) facilitates tractable loss construction and avoids unobservable dependencies on the true label set Y, it potentially conflicts with the authors' earlier intuition that complementary labels are selected with semantic exclusion from the true labels—e.g., “desert” being chosen over “lake” when the true label is "water". This highlights a fundamental tension between model identifiability and the realism of annotator behavior modeling.

  2. Theoretical comparison with URE is insufficient. While the proposed ComRank loss is theoretically proven to be Bayes-consistent in preserving the ranking order of label posterior probabilities, it does not offer an unbiased estimation of the original classification risk E[L(g(x),Y)] as URE does under the uniform assumption. Therefore, the theoretical advantage of ComRank is conditional and context-specific, and it does not universally dominate URE in terms of consistency with classification-oriented loss functions. To summarize, the paper criticizes URE for its bias but praises ComRank for its consistency. However, consistency alone does not imply superiority. Could the authors provide analysis or evidence regarding the (lack of) unbiasedness of ComRank?

  3. Missing discussion on related ranking-based multi-label learning (MLL) methods. The paper lacks a discussion of prior work on ranking-based approaches in MLL, which may lead to an overstatement of novelty. For example:[1] shows that in probabilistic MLL (PML), ranking between positive and negative labels can help disambiguate false positives. [2] shows that in MLML settings, missing labels can be assumed to rank between positive and negative labels. I suggest the authors add a paragraph in the related work section discussing ranking-based MLL methods.

Additionally, Section 3 (Preliminaries) is somewhat redundant. It uses two equations to explain what is essentially standard background. I recommend simplifying this section by presenting only one general MLL risk minimization formula and using a comparative table to illustrate the differences between MLL and MLCLL in terms of labels, learners, and loss definitions—along with symbol definitions.

  1. Unrealistic experimental assumption about complementary labels. The experiments assume that each instance has exactly one complementary label. This may not reflect realistic scenarios and could bias the evaluation in favor of the proposed method. Specifically, the proposed method relies on pairwise ranking between positive and complementary labels. With only one complementary label, the computational cost is O(K−1), but if the number of complementary labels grows to K/2, the complexity increases to O(K^2/4). Please clarify the implications of this.

  2. Evaluation metrics are incomplete. The experiments rely solely on ranking-based metrics, while classification-based metrics are not reported. Some baselines are designed for MLL or PML, which may not be directly comparable with the proposed method. If including these baselines is necessary, more recent and ranking-aware baselines should be included to ensure fairness and meaningful comparison.

  3. Minor issues in Appendix A. The #Domain refers to data source rather than the data type, all datasets used are tabular data. The #Domain of BibTeX should be Text rather than Images.

[1] Partial multi-label learning, AAAI, Vol. 32, 2018.

[2] A ranking-based problem transformation method for weakly supervised multi-label learning, PR, 2024

问题

See above weaknesses (especially 1–5) for necessary revisions or clarifications. The overall direction is promising, but the paper would benefit from addressing the above concerns to improve its clarity, fairness, and theoretical rigor.

局限性

yes

最终评判理由

My concerns have been resolved. The paper is technically acceptable.

格式问题

no

作者回复

Thank you for reviewing our paper and providing valuable insights. We will reorganize the preliminaries section and address the issues in Appendix A. Below, we respond to the remaining concerns raised in the review.

W1: Potential contradiction in modeling assumptions.

We are glad to have the opportunity to clarify that Assumption 4.4 does not contradict our earlier intuition regarding semantic exclusion. In fact, both true and complementary labels are selected by annotators based on the same source of information, the observed instance x x. Annotators assign true labels by identifying what is clearly present in the instance, and assign complementary labels by excluding what seems clearly absent. Since both decisions depend on the same features of x x, an implicit semantic exclusion naturally arises between complementary and true labels, even though the annotator may not know the full true label set YY.

For example, if an image contains water, an annotator may not know whether the true label is 'river' or 'lake', but they can still confidently exclude 'desert' as a complementary label based on visual cues. This illustrates that semantic exclusion behavior can emerge indirectly due to shared dependence on instance features—precisely what Assumption 4.4 formalizes. This assumption provides a practical foundation for modeling complementary label selection without requiring access to YY.

W2: Theoretical comparison with URE is insufficient. ... Could the authors provide analysis or evidence regarding the (lack of) unbiasedness of ComRank?

We would like to clarify that ComRank is not intended to derive URE, nor to serve as a risk estimator. Therefore, it is not applicable to analyze ComRank in the same manner as URE-based methods, i.e., deriving unbiasedness of ComRank. In fact, ComRank and URE-based methods reflect different learning paradigms. While URE-based methods follow a risk-recovery framework that reconstructs classification risk via surrogate losses and distributional assumptions, ComRank directly optimizes a pairwise ranking loss induced by complementary labels, without relying on such assumptions. Additionally, our goal is not to claim that ComRank universally dominates URE, but to highlight a limitation of URE: it may suffer from inconsistency when complementary labels are biased. In contrast, ComRank remains Bayes consistent without requiring strong assumptions on the label distribution. We will include this discussion in the revised version.

W3: Missing discussion on related ranking-based multi-label learning (MLL) methods.

Thank you for the insightful comment. We will add the related work of ranking-based MLL methods in the revised version. Specifically, both [1,2] you suggested show how ranking can serve as an effective inductive bias under weak supervision. Beyond these, ranking losses have also been explored in standard supervised MLL settings. For example, [3] and [4] studied ranking-based surrogate losses with theoretical consistency guarantees. [5] applied calibrated label ranking for multi-label prediction, and [6] investigated pairwise ranking formulations for label dependencies. We will revise the related work section to include the above discussions.

W4: The experiments assume that each instance has exactly one complementary label. This may not reflect realistic scenarios and could bias the evaluation in favor of the proposed method. If the number of complementary labels grows to K/2, the complexity increases to O(K^2/4).

In response to this concern, we conducted additional experiments where each training instance is uniformly assigned K/2K/2 complementary labels. Indeed, the computational complexity increases with the number of complementary labels. However, our implementation can avoid explicit pairwise loops to reduce the impact of the O(K2/4)O(K^2/4) complexity. The loss can be computed using matrix operations and masking, which makes it efficient even for large label spaces. To empirically validate its efficiency, we provide the running time(in 10² seconds) across datasets for large label spaces. As shown in the table below, ComRank achieves comparable or superior speed to most baselines, indicating strong scalability.

Dataset#Label ClassesCCMNL-UWGDFMAERu(g)R_u(g)ComRank
enron532.943.223.793.723.143.15
rcv1-s11014.133.674.123.563.763.64
bibtex1595.113.764.073.693.893.91
bookmark20833.2517.1117.1818.6817.2317.19
avg-11.366.947.297.417.016.97

Additionally, the following table reports the average precision of methods under K/2K/2 complementary labels, where ComRank achieves better performance than most baselines. While the complexity increases with more complementary labels, the performance gains we observe make this trade-off both acceptable and worthwhile. We will update the manuscript to include these analyses.

MethodL-UWCCMNGDFMAERu(g)R_u(g)ComRank
scene0.4810.5500.8450.5730.8490.849
yeast0.7150.6910.7510.7280.7500.749
enron0.4810.4910.6660.6250.6520.673
rcv1-s10.4500.4430.6550.6060.6870.690
bibtex0.4550.5040.7950.6930.7830.799
bookmark0.6220.5240.7090.7100.7140.731

W5.1: Classification-based metrics are not reported.

We add hamming loss as a classification-based evaluation metric for experiments, where each instance is associated with a single complementary label that is uniformly sampled. To ensure fairness, we selected the optimal 0-1 threshold from prediction scores to prediction labels. As shown in the table below, our method achieves the best performance across all datasets, demonstrating that it also performs well under classification-oriented metrics.

DatasetCCMNL-UWMAEGDFRu(g)R_u(g)ComRank
scene0.2420.8200.8200.1520.1630.136
yeast0.3150.6850.6850.5800.3040.227
enron0.5460.2100.5560.3400.1980.165
rcv1-s10.4560.8120.8850.4200.1150.111
bibtex0.5960.1910.8700.7740.0800.059
bookmark0.1500.7700.7250.9160.0840.063
nuswideBoW0.1350.8460.8640.8650.1350.124

W5.2: Some baselines are designed for MLL or PML, which may not be directly comparable with the proposed method. If including these baselines is necessary, more recent and ranking-aware baselines should be included to ensure fairness and meaningful comparison.

We selected the following recent MLCLL and ranking-aware methods for comparison. Their details are summarized below:

  • MLCL [7]: A recent method for multi-label complementary label learning.
  • LogRank [3]: A log-based ranking loss inspired by surrogate risk formulations.
  • SoftmaxRank: A baseline we design by applying softmax function over predicted label scores and comparing their pairwise differences.

We conduct experiments on datasets with uniform complementary labels, where each instance is associated with a single complementary label, and the experimental settings follow those described in the paper. As shown by the average precision results below, our method, ComRank, achieves the best performance on most datasets, further validating its proven Bayes consistency. We will include this discussion in the revised paper.

New baselinesMLCLLogRankSoftmaxRankComRank
scene0.8140.4190.6770.785
yeast0.7310.7120.7200.729
enron0.4350.5470.5670.627
rcv1-s10.5470.2630.2800.605
bibtex0.2480.4530.4500.677
bookmark0.5570.1970.5860.618
nuswideBoW0.5300.4850.4880.583

References

[1] Xie et al., Partial multi-label learning, AAAI, 2018.

[2] Li et al., A ranking-based problem transformation method for weakly supervised multi-label learning, Pattern Recognition, 2024.

[3] Dembczyński et al., Consistent multilabel ranking through univariate losses, ICML, 2012.

[4] Li et al., Improving pairwise ranking for multi-label image classification, CVPR, 2017.

[5] Zhang et al., Binary relevance for multi-label learning: An overview, Frontiers of Computer Science, 2018.

[6] Fürnkranz et al., Multilabel classification via calibrated label ranking, Machine Learning, 2008.

[7] Gao et al., Complementary to multiple labels: A correlation-aware correction approach, TPAMI, 2024.

评论

I appreciate your revision plan and look forward to seeing the improved camera-ready paper. Additionally, I suggest removing the lengthy and somewhat redundant Preliminaries section to keep the presentation concise.

审稿意见
3

This paper proposed a ranking loss framework for Multi-label complementary label learning under both uniform and biased cases. The theoretical analysis is satisfying.

优缺点分析

Strengths:

  1. The paper presents a well-organized structure with clear logical flow and professional presentation, meeting the high standards of academic writing in this field.
  2. The theoretical analysis demonstrates particular rigor, offering valuable insights into the limitations of Unbiased Risk Estimator (URE) within the MLCLL context. This contribution significantly enhances our theoretical understanding of the problem.
  3. The work introduces a novel weakly supervised multi-label learning paradigm that addresses an important gap in current research, with demonstrated practical relevance to real-world applications.

Weaknesses:

  1. The assertion that the proposed framework can directly capture label correlation information from rankings (Lines 45-46) requires stronger empirical validation. Including results or analyses that explicitly demonstrate this capability would strengthen the paper’s validity.
  2. The relationship between Theorem 4.5 (URE under biased distribution) and the proposed ComRank requires deeper explanation. The author should explicitly state the theoretical connections.
  3. The formulation of Assumption 4.4 (instance-dependent assumption) requires further clarification to ensure rigorous theoretical grounding.

问题

Please see the weaknesses which need for more evidences and explanations.

Additionally, please answer the following questions:

  1. The author should provide more empiricial analysis to prove "In MLCLL, based on the fact that complementary labels are known to be irrelevant, while non-complementary labels may include relevant ones, it is generally desirable for the predicted probabilities of complementary labels to rank lower than those of non-complementary labels."

  2. I hope to see more experiment results on more challenging MLL benchmark datasets, including VOC and COCO, to strengthen the empirical evaluation.

  3. Since the author used the exponential function as the surrogate loss, I expect to see more ablation studies about the impact of selecting different forms of surrogate loss, including log loss, sigmoid loss, and softmax loss.

局限性

The relationship between Theorem 4.5 (URE under biased distribution) and the proposed ComRank requires deeper explanation. The author should explicitly state the theoretical connections.

最终评判理由

This paper have provided a novel weakly supervised multi-label learning paradigm that addresses an important gap in current research, with demonstrated practical relevance to real-world applications. However, the discussion on the introduce of noise in the proposed instance-dependent setting is far from satisfactory to me. To my knowledge, the instance-dependent setting in complementary label learning is quite hard to feature, most of which will turn out to be a simple biased setting. As a result, I will keep my initial score.

格式问题

NA

作者回复

Thank you for your meticulous and insightful review of our paper. Below are our responses to the weaknesses and other questions.

W1: The assertion that the proposed framework can directly capture label correlation information from rankings (Lines 45-46) requires stronger empirical validation.

To claims this assertion, we measure the ability of each method to preserve label correlations. We choose to compute the Pearson correlation matrix[1], which can reflect the overall level of inter-label correlation. We compute it for (1) the test label sets and (2) their prediction scores of each method, and measure their difference using Frobenius norm distance (lower is better), shown in the following table. All experimental settings remain the same as in the paper, training on data with uniform complementary labels. ComRank shows better correlation preservation than most baselines, indicating that the ranking-based design in ComRank still helps retain meaningful label dependencies.

Notably, ComRank achieves the lowest Frobenius distance on the scene dataset, which aligns with its highest average precision score in Table 1 of the paper. This suggests that preserving label correlation is not only theoretically desirable but also empirically linked to better predictive performance. We will add this analysis in the revised version.

DatasetCCMNL-UWGDFMAERu(g)R_u(g)ComRank
scene6.446.381.836.421.361.82
yeast4.454.685.795.923.706.34
enron14.9212.8814.3314.843.333.65
rcv1-s112.4215.0611.3215.043.204.95
bibtex15.1814.9813.6915.221.802.71
bookmark13.3815.0911.3815.091.2711.26
nuswideBoW14.5214.459.2214.584.938.96

W2&L: The relationship between Theorem 4.5 (URE under biased distribution) and the proposed ComRank requires deeper explanation.

Theorem 4.5 serves as the motivation for designing ComRank. It reveals that the URE under a biased complementary label distribution is inherently different from the URE derived under the uniform distribution assumption. Since all existing URE-based methods rely on the uniform complementary label assumption, their theoretical guarantee of unbiasedness breaks down when the label distribution is biased. This significantly limits the generality and robustness of URE-based approaches in real-world applications, where biased label distributions are common.

To avoid strong dependency on the complementary label distribution, we propose ComRank inspired by ranking-based learning. ComRank directly encourages the model to assign lower scores to complementary labels. As shown in Theorem 5.2, this approach achieves Bayes consistency under both uniform and biased distributions. We will include this analysis in the revised version.

W3: Assumption 4.4 (instance-dependent assumption) requires further clarification to ensure rigorous theoretical grounding.

Assumption 4.4 aims to reflect the practical setting where annotators lack access to the relevant label set YY when selecting complementary labels, and base their decisions solely on the observed instance xx. This aligns with typical weakly supervised scenarios, where annotations rely on salient or dominant features rather than full label information[2,3,4]. While instances like images containing multiple objects may challenge annotators' ability to exclude all relevant labels, empirical evidence suggests that they can reliably eliminate clearly irrelevant labels (e.g., excluding “building” from an image of animals) by inspecting the instance alone. This supports modeling yˉ\bar{y} as conditionally independent of YY given xx.

Q1: Need more empiricial analysis to prove "In MLCLL, based on the fact that complementary labels are known to be irrelevant, while non-complementary labels may include relevant ones, it is generally desirable for the predicted probabilities of complementary labels to rank lower than those of non-complementary labels."

By encouraging complementary labels to rank lower than non-complementary labels, the model learns useful information, which eventually can translate into encouraging irrelevant labels to rank lower than the relevant labels. Additionally, to support our claim, we reported the metric of ranking loss [5] for all methods in test sets which measures the proportion of relevant labels that are ranked lower than irrelevant labels (smaller values indicate better performance). Notably, ComRank consistently achieves the lowest scores, which provides strong support for the hypothesis. Considering ComRank shows best performance in our paper, the above results confirm that such a ranking structure is not only intuitively desirable but also empirically realizable by well-trained models. We will include this analysis in the revised version to strengthen the empirical analysis.

DatasetL-UWCCMNPMLMDPARDMAEGDFRu(g)R_u(g)ComRank
scene0.5230.4660.4670.1720.4880.1530.1750.136
yeast0.2450.2940.2210.3550.2310.2190.2500.215
enron0.4380.4810.2620.3700.3870.2310.3950.221
rcv1-s10.2720.3440.4470.2990.3070.2750.3690.267
bibtex0.4910.4790.4600.2800.4260.2060.3080.180
bookmark0.2800.3680.6610.2690.2870.1960.2800.187
nuswideBoW0.2870.3400.2740.3170.2740.2060.1960.192

Q2: More experiment results on more challenging MLL benchmark datasets, including VOC and COCO, to strengthen the empirical evaluation.

Due to time limitations, we conducted experiments on the VOC dataset with uniform complementary labels, where the results of average precision are shown below. ResNet18 is used to classify and experimental settings are the same as in the paper. The average precision of our method surpasses all baselines, demonstrating the effectiveness of ComRank. We will update it and COCO in the revised version.

MethodL-UWCCMNMAEGDFRu(g)R_u(g)ComRank
VOC0.3270.3900.3140.4320.4310.436

Q3: Ablation studies about the impact of selecting different forms of surrogate loss, including log loss, sigmoid loss, and softmax loss.

The following table reports the average precision from an ablation study on different surrogate losses under uniform complementary labels. The surrogate losses include:

  • Log loss: Lˉ(g(x),yˉ)=k=1,kyˉKlog(1+(gyˉ(x)gk(x)))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} log\left(1 + \left(g_{\bar{y}}( x) - g_k(x)\right)\right).
  • Sigmoid loss: Lˉ(g(x),yˉ)=k=1,kyˉK(gyˉ(x)gk(x))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} \left(g_{\bar{y}}( x) - g_k( x)\right). Since predicted probability gg in MLL is already produced through an output sigmoid function to keep gg in [0,1], we directly use the difference between gg values.
  • Softmax loss: Lˉ(g(x),yˉ)=k=1,kyˉK(gyˉsoftmax(x)gksoftmax(x))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} \left(g^{softmax}_{\bar{y}}( x) - g^{softmax}_k( x)\right). Here, gsoftmaxg^{softmax} refers to the version of gg where the output function is changed from sigmoid to softmax.

Our method achieves competitive performance in most cases, validating its effectiveness and proven Bayes consistency.

Surrogate lossLogSigmoidSoftmaxComRank
scene0.4190.6870.6770.785
yeast0.7120.7320.7200.729
enron0.5470.5910.5670.627
rcv1-s10.2630.4750.2800.605
bibtex0.4530.6490.4500.677
bookmark0.1970.6290.5860.618
nuswideBoW0.4850.4900.4880.583

References

[1] Pearson et al., Mathematical contributions to the theory of evolution. III. Regression, heredity, and panmixia, Philosophical Transactions of the Royal Society of London. Series A, 1896.

[2] Xia et al., Beyond class-conditional assumption: A primary attempt to combat instance-dependent label noise, AAAI, 2021.

[3] Xia et al., Part-dependent label noise: Towards instance-dependent label noise, NeurIPS, 2020.

[4] Yao et al., A second-order approach to learning with instance-dependent label noise, CVPR, 2021.

[5] Zhang et al., A review on multi-label learning algorithms, IEEE TKDE, 2014

评论

Assumption 4.4 suggests the anntation of complementary label is only dependent on the instance not the ground-truth, which may introduce noise to the complementary label. Can you add a discussion on this issue and how can you deal with it accordingly?

评论

Thank you for your insightful comment. Assumption 4.4 specifies that the distribution of the complementary label yˉ\bar y is conditioned only on the instance xx, i.e., p(yˉx)=p(yˉx,Y) p(\bar{y} \mid x) = p(\bar{y} \mid x, Y). We acknowledge that, in isolation, this assumption may raise concerns: without explicitly referencing the ground-truth label set YY, yˉ\bar{y} might inadvertently coincide with a true label, introducing label noise.

To address this, we combine Assumption 4.4 with a structural constraint—namely in the paper, that the probability of selecting a true label as a complementary label is zero, via the commonly used uniform assumption (Assumption 4.1) [1,2] and our proposed biased distribution (Assumption 4.3). That is, p(k=yˉx)=0p(k = \bar{y} | x) = 0 if kYk \in Y. These assumptions jointly ensure that yˉ\bar{y} remains disjoint from YY, thus preventing noise under the defined setting.

This noise-free formulation is standard in the foundational literature of complementary label learning [1,2,3,4] and enables tractable theoretical analysis. Our goal here is to understand the idealized theoretical behavior of MLCLL methods before tackling annotation noise.

That said, we fully agree that in practical annotation scenarios, such assumptions may not hold—e.g., annotators might mistakenly choose a relevant label as a complementary one due to ambiguity or fatigue. This is an important limitation and a promising direction for future research, especially toward building robust MLCLL methods under noisy or imperfect labeling conditions. Possible approaches include explicitly modeling the noise transition process, designing loss functions tolerant to mislabeled complementary annotations, or using a small set of clean labels for calibration. Such extensions may be beyond the scope of this paper. In this work, we intentionally focus on the limitations of existing methods under biased complementary label distributions and propose a new ranking-based loss that is Bayes consistent, which we view as a necessary foundation before tackling more complex, noise-prone scenarios.

[1] Learning from complementary labels. NeurIPS, 2017.

[2] Unbiased risk estimator to multi-labeled complementary label learning. IJCAI, 2023.

[3] Discriminative complementary-label learning with weighted loss. ICML, 2021

[4] Learning with multiple complementary labels. ICML, 2020.

评论

Thanks for the response. However, all the papers, i.e., [1]-[4], are based on the instance-independent setting. In such case, it is easy to perform a noise-free assumption. Conversely, in your instance-dependent assumption, I cannot find the rationale for the noise-free assumption, i.e., Assumption 4.1 and 4.3, under the scenario that the ground-truth label is unknown to the annotator. In my opinion, the complementary setting partially stems from the privacy protection application, where the ground-truth is known but is not suitable to conceal in the training dataset. I highly encourage the authors to investigate the noisy case and consider the anti-noise techniques in the future research.

评论

Thank you very much for your thoughtful suggestion. In future work, we will investigate how noise-robust techniques can be incorporated into the complementary label learning framework to improve its resilience and generalizability on noisy data. We sincerely thank you for bringing up this valuable perspective, which will help guide our further research.

审稿意见
4

The paper addresses multi-label complementary label learning. It identifies limitations in existing methods relying on unbiased risk estimators, which assume uniform distribution of complementary labels—an assumption often violated in real-world scenarios with annotation biases, and these methods also underutilize label correlations.To tackle this, the paper proposes ComRank, a ranking loss framework that encourages complementary labels to be ranked lower than non-complementary ones, capturing pairwise label relationships. Experiments show ComRank outperforms existing methods on various datasets with both uniform and biased complementary labels. Its contributions include: analyzing why URE-based methods fail under biased distributions; introducing ranking loss to MLCLL via ComRank, which has theoretical guarantees under both uniform and biased cases and shows strong experimental performance.

优缺点分析

Strengths (1) Strong theoretical foundation: Comprehensively analyzes URE limitations and proves Bayes consistency of ComRank under both uniform and biased complementary label distributions. (2) Advances MLCLL: Introduces ranking loss to model label correlations, filling gaps in URE-based methods. (3) Rigorous experiments: Uses diverse datasets with uniform/biased labels, outperforming baselines across metrics like Average Precision and Ranking Loss. Weaknesses (1) Lacks computational complexity analysis, critical for scaling to large datasets. (2) Neglect label sparsity effects: Multi-label data often has sparse labels, but the paper doesn't test ComRank under varying sparsity. With extreme sparsity, few non-complementary labels may undermine its ranking logic, leaving its stability in such common real-world scenarios unproven.

问题

(1)Is there an error in defining a sample of Cartesian product used in MLL in the Preliminaries of Part 3? According to the definition in the text, it is a single label sample. Is it more accurate to define Y={1,2,..., K} as Y ∈ {0,1} ^ K . (2)The paper claims ComRank “directly captures label correlation information from rankings”, but this mechanism is not explicitly validated. Maybe you clarify it by Comparing ComRank’s performance on datasets with strong vs. weak label correlations. If correlation strength correlates with ComRank’s improvement over baselines, it would validate the claim.

局限性

The paper only mentions that ComRank is restricted to the single complementary label setting, with extending to multiple complementary labels noted as future work. It fails to discuss other critical limitations, such as the sensitivity to the choice of surrogate loss (the paper uses an exponential function without exploring alternatives like logistic loss) and scalability issues with large label spaces, where the pairwise comparisons in the loss function could lead to high computational costs.

最终评判理由

Given the original quality of the paper and the response experiments, this paper could be accepted.

格式问题

No issues

作者回复

Thank you for the time and effort you devoted to reviewing our paper. The error in Q1 will be modified in the revision. Below we provide our responses to other weaknesses and questions.

W1 & L2: Lacks computational complexity analysis, critical for scaling to large datasets. Fail to discuss scalability issues with large label spaces, where the pairwise comparisons in the loss function could lead to high computational costs.

In this paper, each instance is assigned a single complementary label, resulting in a computational complexity of O(n(K1))O(n(K-1)) for the proposed ComRank method, where nn and KK are the number of instances and classes, respectively. Indeed, the computational complexity increases with the number of complementary labels. However, our implementation can avoid high computational costs caused by pairwise comparisons in the loss function. The loss can be computed using matrix operations and masking, which makes it efficient for large label spaces. To empirically validate its efficiency, we provide the running time (in 10² seconds) across datasets with large label spaces, where each instance is associated with K/2K/2 complementary labels. As shown in the table below, ComRank achieves comparable or superior speed to most baselines, indicating scalability.

Dataset#Label ClassesCCMNL-UWGDFMAERu(g)R_u(g)ComRank
enron532.943.223.793.723.143.15
rcv1-s11014.133.674.123.563.763.64
bibtex1595.113.764.073.693.893.91
bookmark20833.2517.1117.1818.6817.2317.19
avg-11.366.947.297.417.016.97

W2: Neglect label sparsity effects: Multi-label data often has sparse labels, but the paper doesn't test ComRank under varying sparsity. With extreme sparsity, few non-complementary labels may undermine its ranking logic, leaving its stability in such common real-world scenarios unproven.

To verify the effect of ComRank on the data with sparse labels, we conduct experiments on datasets with varying label densities. In the table below, the '#Density' column represents the label density of each dataset, where smaller values indicate higher sparsity. We report the average precision of different methods, which is also shown in Table 1 of the paper. ComRank achieves the best performance in all cases, with densities ranging from 0.082 to 0.303, which indicates its stability under sparse labeling. The reason lies in the pairwise ranking design, which relies on relative comparisons between complementary and non-complementary labels, rather than absolute label counts. We will include this analysis in the revised version.

Dataset#DensityL-UWCCMNMAEGDFRu(g)R_u(g)ComRank
scene0.1790.3950.4580.4320.7590.7340.780
yeast0.3030.6850.6460.6980.7120.6790.715
enron0.1890.3750.3370.4270.6200.4110.634
rcv1-s10.1110.4450.4090.4270.4710.3630.491
bibtex0.0820.2370.2590.2870.6140.4130.658
bookmark0.0830.5060.3970.5060.6190.5120.628

Q2: "directly captures label correlation information from rankings" is not explicitly validated. Maybe you clarify it by Comparing ComRank's performance on datasets with strong vs. weak label correlations. If correlation strength correlates with ComRank's improvement over baselines, it would validate the claim.

We agree that comparing performance across datasets with different label correlation strengths is intuitive. However, this approach might be confounded by uncontrolled factors such as feature difficulty, sample size, and label distribution. Specifically, label correlations are inherent properties of each dataset and cannot be adjusted or isolated within the same dataset for controlled comparison. Comparing across datasets with varying correlation strengths may thus lead to unfair or misleading conclusions.

To address this issue in a fair and interpretable way, we instead measure the ability of each method to preserve label correlations. We compute the Pearson correlation matrix[1], which can reflect the overall level of inter-label correlation. We compute it for (1) the test label sets and (2) their prediction scores of each method, and measure their difference using Frobenius norm distance (lower is better), shown in the following table. All experimental settings remain the same as the paper, i.e., training on data with uniform complementary labels. ComRank shows better correlation preservation than most baselines, indicating that the ranking-based design in ComRank still helps retain meaningful label dependencies.

Notably, ComRank achieves the lowest Frobenius distance on the scene dataset, which aligns with its highest average precision score in Table 1 of the paper. This suggests that preserving label correlation is not only theoretically desirable but also empirically linked to better predictive performance. We will add this analysis in the revised version.

DatasetCCMNL-UWGDFMAERu(g)R_u(g)ComRank
scene6.446.381.836.421.361.82
yeast4.454.685.795.923.706.34
enron14.9212.8814.3314.843.333.65
rcv1-s112.4215.0611.3215.043.204.95
bibtex15.1814.9813.6915.221.802.71
bookmark13.3815.0911.3815.091.2711.26
nuswideBoW14.5214.459.2214.584.938.96

L1: Fail to discuss the sensitivity to the choice of surrogate loss

In response, we conducted experiments on different surrogate losses, including

  • Log loss: Lˉ(g(x),yˉ)=k=1,kyˉKlog(1+(gyˉ(x)gk(x)))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} log\left(1 + \left(g_{\bar{y}}(x) - g_k(x)\right)\right).
  • Sigmoid loss: Lˉ(g(x),yˉ)=k=1,kyˉK(gyˉ(x)gk(x))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} \left(g_{\bar{y}}( x) - g_k( x)\right). Since predicted probability gg in MLL is already produced through an output sigmoid function to keep gg in [0,1], we directly use the difference between gg values.
  • Softmax loss: Lˉ(g(x),yˉ)=k=1,kyˉK(gyˉsoftmax(x)gksoftmax(x))\bar{\mathcal{L}}(g( x), \bar{y}) = \sum_{k=1, k \ne \bar{y}}^{K} \left(g^{softmax}_{\bar{y}}( x) - g^{softmax}_k( x)\right). Here, gsoftmaxg^{softmax} refers to the version of gg where the output function is changed from sigmoid to softmax.

The following table shows the average precision of different surrogate losses on the dataset with uniform complementary labels. Our proposed method, ComRank, achieves comparable perfomance in most cases, which also validates its effectiveness.

Surrogate lossLogSigmoidSoftmaxComRank
scene0.4190.6870.6770.785
yeast0.7120.7320.7200.729
enron0.5470.5910.5670.627
rcv1-s10.2630.4750.2800.605
bibtex0.4530.6490.4500.677
bookmark0.1970.6290.5860.618
nuswideBoW0.4850.4900.4880.583

Reference

[1] Pearson et al., Mathematical contributions to the theory of evolution. III. Regression, heredity, and panmixia, Philosophical Transactions of the Royal Society of London. Series A, 1896.

评论

I appreciate the additional analyses and experiments provided in the rebuttal, which offer some clarifications. I look forward to seeing them further refined and integrated into the camera-ready paper.

最终决定

This paper makes a solid contribution to the multi-label learning literature, particularly by highlighting theoretical deficiencies in prior work on using complementary labels and providing a new framework to address them. The theoretical analysis is well-motivated, and the paper raises important points that advance understanding in this area.

During review, some concerns were raised about the theoretical framework and the strength of certain assumptions (especially Assumption 4.4), as well as the connection between the theory and the eventual algorithm. While these concerns are valid discussion points, the authors provided strong and thoughtful responses that clarified their reasoning and addressed most of the issues. Incorporating these clarifications and discussion into the final version will further strengthen the contribution.