PaperHub
4.3
/10
Rejected4 位审稿人
最低3最高6标准差1.3
3
5
3
6
3.3
置信度
正确性2.0
贡献度2.3
表达2.3
ICLR 2025

Stochastic Order Learning: An Approach to Rank Estimation Using Noisy Data

OpenReviewPDF
提交: 2024-09-23更新: 2025-02-05

摘要

关键词
rank estimationlabel noiseorder learning

评审与讨论

审稿意见
3

This paper solves the rank estimation problem under noisy data. The main novelty of this work is handling the input data with label noise, which seems limited, as other works can also handle noise data. Moreover, the motivation of the adopted techniques is also not clear. The experimental evaluation of this paper is good.

优点

This paper proposes a model to handle the noisy rank learning problem. The experimental evaluation is good.

缺点

Many descriptions are not clear. For example: What is the difference between rank estimation and order learning? In Eq. (1), the ground truth label is represented by consecutive integers, while the observed noisy rank is assumed to be the sum of the true rank with some Gaussian noise. When the Gaussian noise is added, will the observed rank still be integers? Besides, why does the noise follow a Gaussian distribution? In line 137, what is the relationship between e and e_x?

问题

Please see the weakness. Also, please describe the motivation for the proposed method.

评论

Thank you for your constructive comments. Please find our responses below.


Rank estimation vs. order learning

Rank estimation is a "task" that aims to predict the ordered class of an object. On the other hand, as stated in L106, order learning is an approach to rank estimation problem. Specifically, order learning algorithms exploit ordering relationship between instances for rank estimation.

Gaussian noise

  • As stated in L137, the noise is assumed to have a "discrete" Gaussian distribution. Therefore, the observed rank rxr_x in Eq.(1) will still be integers after the Gaussian noise is added.

  • In practice, when annotators make errors in ranking, they are more likely to choose neighboring ranks than far-away ranks. As stated in L140-142, the symmetric(ps=psp_s=p_{-s}), and unimodal(psptp_s \geq p_t for 0st0\leq s \leq t) property of a Gaussian noise distribution well-models this trait. Thus, we chose the noise to follow a Gaussian distribution.

  • As stated in L136, exe_x is the label error of an instance xx and e\mathbf{e} is the random variable underlying each error exe_x. We are not sure of what exactly the point of the question here is. We will do our best to answer your question if you could specify it.

Motivation

Our motivation was to make a noise-robust algorithm suitable for rank estimation. Although many noise-robust classification methods have been developed, they yield poor results when applied to rank estimation. Most rank estimation methods, on the other hand, do not consider label noise. Specifically, the state-of-the-art method[1] designs an embedding space that arranges instances according to deterministic ranks. Using deterministic ranks fails to consider the possibility of rank label errors. In order to handle noisy data, we were motivated to use stochastic variables to model the label errors. We derived an embedding space in which instances are arranged based on probabilistically associated ranks instead of deterministic ranks.

[1] Geometric order learning for rank estimation. In NIPS, 2022.


If you have any additional concerns, please let us know. We will address them faithfully. Thank you again for your constructive comments.

审稿意见
5

This paper introduces a novel algorithm called Stochastic Order Learning (SOL) for reliable rank estimation in the presence of label noise. The proposed method models label errors as random variables and formulates a desideratum to minimize dissimilarity from stochastically related centroids. By employing discriminative loss and stochastic order loss, the algorithm constructs an embedding space that effectively arranges instances according to their true ranks. Additionally, it incorporates outlier detection and relabeling schemes to refine noisy data, demonstrating robust performance across various benchmark tasks.

优点

  1. The proposed algorithm in this paper demonstrates reasonable formulation for rank estimation in the presence of label noise, with clear mathematical derivations.

  2. The experimental section is adequately designed, covering multiple benchmark datasets to validate the algorithm's performance.

缺点

  1. The novelty of the paper is relatively moderate. While it builds upon existing research, it lacks significant innovative contributions.

  2. The Dataset Section can be placed in the Appendix to allow more space for detailed analysis of experimental results.

  3. The paper lacks comparisons with algorithms from the past two years. Including recent benchmarks would provide a more comprehensive evaluation for the proposed method. Besides, the number of comparison algorithms is too small. Adding more comparison algorithms will further highlight its advantages.

问题

See Weaknesses

评论

Thank you for your constructive review and valuable suggestions. We have revised the paper to address your comments faithfully, and highlighted the revised parts in blue. Please find our responses below.


Novelty

SOL is the first order-learning algorithm to handle label noise by designing a stochastic approach. GOL[1] designs an embedding space that arranges instances according to their ranks. However, the ranks assigned are deterministic, failing to take label noise into account. To model label errors, we use stochastic variables and we develop the discriminative loss and the stochastic order loss to derive an embedding space in which instances are arranged based on probabilistically associated ranks. Furthermore, we propose outlier detection and relabeling schemes to reduce the overall noise level of a given dataset. Experiments on various datasets prove that our algorithm shows the best performance for rank estimation tasks under label noise.

[1] Geometric order learning for rank estimation. In NIPS, 2022.

More comparisons

As you suggested, we compare our algorithms with more comparative methods[1,2,3] including recent benchmarks. Below lists the comparisons on the MORPH II dataset. The proposed SOL still achieves the best scores. Recent noise robust-classification methods[1,2] cannot distinguish between subtle differences across ordered ranks, thus they yield poor performances. The results for other three datasets(CLAP2015, AADB, RSNA) are similar as well. We included and discussed these comparisons in the revised paper. Please see Table 1 on page 6 and Tables 12, 13, 14 on page 16 and 17.

κ=0.2\kappa=0.2κ=0.3\kappa=0.3κ=0.4\kappa=0.4κ=0.5\kappa=0.5
AlgorithmMAECSMAECSMAECSMAECS
SPR [1]8.44641.718.88134.799.23936.899.99328.14
ACL [2]9.01736.759.49235.619.31435.749.74334.16
C-Mixup [3]3.06382.263.39377.213.39576.843.41576.65
POE2.98982.883.09380.333.25379.233.58074.41
MWR2.57090.072.69389.252.85187.163.08984.34
GOL2.51690.892.67189.072.86185.973.07884.70
SOL w/o refinement2.50791.262.65788.712.84286.892.99584.79
SOL2.48991.352.66389.622.82687.702.98685.88

[1] Scalable penalized regression for noise detection in learning with noisy labels. In CVPR, 2022.

[2] Active negative loss functions for learning with noisy labels. In NIPS, 2023.

[3] C-Mixup: Improving generalization in regression. In NIPS, 2022.

Dataset section

We agree with you that moving the dataset section to the Appendix adds more space for experimental results. Following your suggestion, we moved the detailed dataset descriptions to Appendix E on page 16 of the revised paper. Thank you for your constructive comment.


We have revised our paper to address your comments faithfully. We hope that this revision resolves your concerns. If you have any additional comments, please let us know.

审稿意见
3

This paper presents a novel algorithm to address noise in rank ordering within learning-to-rank tasks. Extensive experiments are conducted to demonstrate the effectiveness of the proposed method.

优点

This paper addresses a practical, real-world setting, and the experiments effectively demonstrate the method's effectiveness.

缺点

  1. The problem of noisy classification is well-studied, particularly in image classification (e.g., [1]). However, the authors do not clearly explain how their approach to learning with rank labels differs from or improves upon standard classification methods.
  2. The writing quality requires improvement. Several notations are undefined, and some equations are presented without adequate derivation or verification.

[1] Xiao, Tong, et al. "Learning from massive noisy labeled data for image classification." Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.

问题

  1. Where is the definition of d(,)d(\cdot, \cdot) in Eq.(3)
  2. How do the authors ensure that the noise exe_x remains within the range of true labels in Equations Eq.(1) and Eq.(2)
  3. In line 155, could the authors clarify why μrx+s\mu_{r_{x + s}} the probablity psp_s?
  4. According to Equation (5), does the algorithm require scanning all data?
  5. Is the probability in Equation (2) known or assumed?
评论

Thank you for your constructive review. We have revised the paper to address your comments faithfully, and highlighted the revised parts in blue. Please find our responses below.


Comparison to classification methods

Unlike classification tasks, data in rank estimation have "ordered" classes. In other words, labels of ordered data are continuous, thus the difference between classes can be very small. As stated in L41-42, classification methods suffer from discriminating subtle differences across ordered classes. The table below compares the proposed algorithm to recent classification methods on the MORPH II dataset. The proposed SOL significantly outperforms the classification methods in noisy rank estimation tasks. We have included these comparisons in the revised paper. Please see Table 1 on page 6 and Tables 12, 13, 14 on page 16 and 17.

κ=0.2\kappa=0.2κ=0.3\kappa=0.3κ=0.4\kappa=0.4κ=0.5\kappa=0.5
AlgorithmMAECSMAECSMAECSMAECS
SPR [1]8.44641.718.88134.799.23936.899.99328.14
ACL [2]9.01736.759.49235.619.31435.749.74334.16
SOL w/o refinement2.50791.262.65788.712.84286.892.99584.79
SOL2.48991.352.66389.622.82687.702.98685.88

[1] Scalable penalized regression for noise detection in learning with noisy labels. In CVPR, 2022.

[2] Active negative loss functions for learning with noisy labels. In NIPS, 2023.

Writing

With all due respect, we do not agree with the reviewer. Please provide specific examples. We will do our best to revise the paper.

Definition of d(,)d(\cdot,\cdot)

In L156, d(,)d(\cdot, \cdot) is defined. It denotes the Euclidean distance between two points in the embedding space.

Noise range

In real-world problems, generally, the range of true labels is unknown. Hence, we do not impose constraints to ensure that rxr_x, the rank given after noise exe_x is added, is within the range of true labels in Eq.(1). However, we clip the values so that rx0r_x \geq 0, to avoid impractical (i.e. negative) ranks.

Relationship of μrx+s\mu_{r_x+s} and psp_s

If the label noise of instance xx is ss, the true rank of xx would be equal to rx+sr_x+s. Since psp_s denotes the probability that the label error of xx is ss, xx is associated with μrx+s\mu_{r_x+s} with probability psp_s. If we have misunderstood your comment on clarifying the relationship between μrx+s\mu_{r_x+s} and psp_s, please let us know and we will address them more faithfully.

Equation (5)

Yes, all data is scanned when computing the centroids in Eq.(5).

Probability in Equation (2)

As stated in L137, the random variable e\mathbf{e} is assumed to have a discrete Gaussian distribution. Thus, the probability in Eq.(2) is assumed.


If you have any additional concerns, please let us know. We will address them faithfully. Thank you again for your feedback.

评论

Thank you for the authors' responses. My concerns are as follows:

  1. While the authors demonstrate subtle differences between ordered learning and classification problems, I still believe that their training processes are fundamentally similar. Could the authors clarify why classification algorithms cannot address ordered learning, or why ordered learning is necessary? Additionally, this aspect seems to be missing in the current submission.

  2. The proposed method appears to have limited effectiveness. Since scanning the entire dataset is required in the proposed algorithm, it is inefficient. Especially, when dealing with large volumes of data.

  3. In Equation 2, could the authors clarify how the variance σ\sigma of the Gaussian distribution is obtained?

评论

Thank you for your feedback. We appreciate it greatly.


Difference from classification

Rank estimation (or ordinal regression), which order learning algorithms aim to solve, is fundamentally different from ordinary classification. Unlike classification, different errors have different severities in rank estimation; misclassifying a 20-year-old as a 10-year-old is severer than mistaking it for a 22-year-old. Thus, for reliable rank estimation, it is important to avoid such severe errors. However, classification-based approaches do not consider this ordinal property in rank estimation datasets, and thus yield poor performances when applied to noisy rank estimation tasks. We clarified this by adding more explanation in the introduction of the revised paper.

Also, the training method of SOL is different from that of classification methods such as SPR and ACL. While they employ cross-entropy based losses which equally treat all noise (or label deviation), SOL exploits the discriminative loss and the stochastic order loss to leverage the ordinal relationship of instances/classes. Experimental results including Table 1, 12, 13, and 14 show that SOL outperforms classification-based approaches by meaningful margins.

Efficiency

Below, we measure the time required to compute the centroids in Eq.(5). We use an RTX 4090 GPU. Even for the RSNA dataset which consists of 12,611 training samples, it takes only a few minutes. It is fast enough for most use cases, since centroid update is performed only once at each epoch. Thus, as shown in the table below, the centroid update accounts for only a small portion of the overall training time. Moreover, even with very large datasets, we can easily reduce the computation time by subsampling the training instances for the centroid update.

MORPH IICLAP2015AADBRSNA
Centroid computation6.10s5.08s39.18s286.06s
Training 1 epoch60.17s52.07s145.43s1160.66s

For large scale training, memory efficiency is also important. We compare the number of parameters of SOL with that of conventional methods. SOL requires the smallest number of parameters, showing its potential for large scale applications. We have included these results in the revised paper. Please see Tables 17 and 18 in the Appendix.

ACLMWRGOLSOL
# of parameters134.68M139.41M14.75M14.72M

σ\sigma in Eq.(2)

In practice, prior knowledge of noise is not given. To simulate such practical settings, we also assume that noise information is unknown. Hence, σ\sigma in Eq. 2 is a hyper-parameter in SOL and determined empirically. We have already provided the analysis of σ\sigma in Table 9 (also in the table below). In general, SOL provides decent results at σ=1\sigma = 1 regardless of ground-truth noise variance. Therefore, we use σ=1\sigma=1 as our default option for all experiments.

σ=0.5\sigma=0.5σ=1\sigma=1σ=1.5\sigma=1.5
MAECSMAECSMAECS
κ=0.2\kappa=0.22.55790.622.50791.262.52990.80
κ=0.3\kappa=0.32.68288.622.65788.712.68989.44
κ=0.4\kappa=0.42.85086.982.84286.892.86986.70

If you have any additional concerns, please let us know. We will do our best to resolve them.

审稿意见
6

To solve the rank estimation problem with label noise, this manuscript proposes the stochastic order learning (SOL) method. Specifically, label errors are represented as random variables that follow a Gaussian distribution. Based on this, the stochastic dissimilarity of instance x from rank r is defined and the corresponding discriminative loss and stochastic order loss are designed to train a network to build an embedding space where instances are arranged according to their ranks. Furthermore, after network training, the SOL method identifies outliers that are likely to have extreme label errors and relabel them to refine the dataset.

优点

  1. This manuscript models label errors as stochastic variables and derives embedding space constraints to arrange instances based on their probabilistically associated ranks.
  2. This manuscript proposes outlier detection and relabeling schemes to pinpoint and mitigate instances with extreme label errors, thereby diminishing the dataset's overall noise level.

缺点

  1. The reason for setting formulas 12-14 lacks description. Specifically, why is the formula defined in this way? It is necessary to explain the meaning of setting rxrys+tr_x-r_y-s+t. In other words, whether I can consider that if the label noise of samples x and y is s and t, respectively, the distance between the true labels of the two samples is rxrys+tr_x-r_y-s+t.
  2. In the experiments, there are few comparative methods and no comparison with the state-of-the-art methods. If there are few recent papers about rank estimation with label noise, it can be compared with related methods mentioned in this paper. For example, if rank estimation is regarded as a classification task, is the SOL method better than the approach in the paper [1]? Besides, on datasets with a rank less than 5, whether the SOL method is still performing significantly and better than the method in the paper [2].

[1] Xichen Ye, Xiaoqiang Li, Songmin Dai, Tong Liu, Yan Sun, and Weiqin Tong. Active negative loss functions for learning with noisy labels. In NIPS, 2023.

[2] Huan Liu, Jiankai Tu, Anqi Gao, and Chunguang Li. Distributed robust support vector ordinal regression under label noise. Neurocomputing, pp. 128057, 2024.

问题

  1. The proposed method settings tend to have a risk of labeling errors for each sample. More likely, only a small amount of samples in the dataset is mislabeled. In this situation, is the performance of the method still effective? In addition, the proposed method is based on the key assumption that the label noise of each sample follows the same discrete Gaussian distribution. It is more natural that the standard deviation of noise variables varies among different samples. In the future, a fine-tuning mechanism for setting standard deviations for each sample can be considered while taking into account computational complexity.
  2. There are many hyperparameters in the manuscript. Can an adaptive mechanism be designed to select the optimal values of these parameters for different datasets?
  3. In the relabeling mechanism, the rank of each detected outlier is adjusted by the same magnitude, that is, half of the average absolute difference between noisy and estimated rankings overall training instances, which seems counterintuitive. It is reasonable to combine outlier detection with the relabeling mechanism to adjust the rank of the outlier.
评论

Thank you for your positive review and insightful suggestions. We have revised the paper to address your comments faithfully, and highlighted the revised parts in blue. Please find our responses below.


Formulas (12)~(14)

As you pointed out, the difference between the true labels of xx and yy is rxrys+tr_x-r_y-s+t when the label noise of xx and yy is ss and tt, respectively. Following Eq.(11), the ordering relationship is xyx \prec y when rxrys+t<τr_x-r_y-s+t < -\tau. Since we model label noise as stochastic variables, we can compute the probability of xyx \prec y as Formula (12) by using Eq.(2). We have clarified this point in L235-239 of the revised paper.

More comparisons

As suggested, we compare our algorithms with more noise-robust methods [1,3,4] on the MORPH II, CLAP2015, AADB, and RSNA datasets. The table below shows the comparison results on the MORPH II dataset. The proposed SOL still achieves the best scores. The results for other three datasets are similar as well. We have included all these results in the revised paper. Please see Table 1 on page 6 and Tables 12, 13, 14 on page 16 and 17.

For method RSVOR[2], we were not able to re-implement it successfully — RSVOR yields very poor results only — because its source code is not available and the authors haven’t responded to our question about implementation details. Moreover, RSVOR is designed for tabular datasets which are very different from vision datasets used in SOL. Hence, the application of RSVOR is not straightforward. We will try to fine-tune the implementation of RSVOR as much as possible and update the results.

κ=0.2\kappa=0.2κ=0.3\kappa=0.3κ=0.4\kappa=0.4κ=0.5\kappa=0.5
AlgorithmMAECSMAECSMAECSMAECS
ACL [1]9.01736.759.49235.619.31435.749.74334.16
SPR [3]8.44641.718.88134.799.23936.899.99328.14
C-Mixup [4]3.06382.263.39377.213.39576.843.41576.65
SOL w/o refinement2.50791.262.65788.712.84286.892.99584.79
SOL2.48991.352.66389.622.82687.702.98685.88

[1] Active negative loss functions for learning with noisy labels. In NIPS, 2023.

[2] Distributed robust support vector ordinal regression under label noise. Neurocomputing, pp. 128057, 2024.

[3] Scalable penalized regression for noise detection in learning with noisy labels. In CVPR, 2022.

[4] C-Mixup: Improving generalization in regression. In NIPS, 2022.

Partially corrupted data

We measured the performances of SOL and GOL when only a small amount of samples are mislabeled. In this test, we use the CLAP2015 dataset and randomly select 10% of samples for label corruption. As shown in the table below, SOL is still effective even when a small amount of samples have label errors. We have added these results in the revised paper. Please see Table 15 and its description on page 17.

Algorithmκ=0.2\kappa=0.2κ=0.3\kappa=0.3κ=0.4\kappa=0.4κ=0.5\kappa=0.5
GOL3.4423.5403.5903.690
SOL3.4203.5053.5493.639

Noise distribution

We agree with you. Please note that SOL is the first attempt to extend the concept of order learning to noise-robust rank estimation, which is a relatively under-researched area. Hence, we focus on developing a simple but reasonable algorithm instead of covering all details in real-world scenarios. It is a future research issue to set the noise distribution adaptively according to each instance. We have clarified this point in the revised paper. Please see the last paragraph on page 17.

Hyper-parameters

We observed that as long as the hyper-parameters do not deviate much from the proposed default settings, decent performances are yielded. In Table 10 of page 14, we listed all the hyper-parameter values used for all datasets. Please note that the values do not vary much for the different datasets. We do agree, however, that designing an adaptive mechanism may be helpful for improving the scalability of SOL to various datasets. We will investigate it in future work. Thank you for your constructive comment.

Relabeling mechanism

During the development of SOL, we have tried a relabeling scheme where outliers are relabeled by different magnitudes. The table below compares the performances according to the relabeling scheme on CLAP2015 at κ=0.4\kappa=0.4. Although relabeling with different magnitudes also show performance improvement, the proposed technique yields better results. It is because the proposed relabeling scheme prevents drastic rank changes, as stated in L314. Please see Table 16 on page 17 for more details.

Relabeling mechanismMAECS
No relabeling4.05873.68
Different magnitudes4.01272.75
Proposed4.00273.68

We have made every attempt to address your comments in the revised paper. We hope that you find this revision satisfactory. If you have additional concerns, please let us know. Thank you again for your positive comments.

评论

Thank you for the response. Most of my questions have been clarified, and I will keep my score.

评论

Thank you for your time and effort for reviewing our paper. We appreciate your positive comments and review.

评论

We would like to thank all reviewers for their time and efforts for providing constructive reviews. We will upload our response to each question or comment as soon as possible.

AC 元评审

Thanks for your submission to ICLR.

This paper received four reviews, which were borderline tending to reject. On the positive side, the reviewers appreciated that this paper addresses a practical problem. But they had various concerns, including i) issues with the presentation / missing descriptions, ii) issues with the experiments, including missing comparisons to other methods, iii) concerns about the effectiveness / novelty of the approach.

The author response clarified some of these issues, but the negative reviewers were unwilling to raise their scores. I think the rebuttal did address several of these issues, but I also think (particularly given the overall lack of enthusiasm with the current paper) that the paper could use some additional refinement to address these issues. With more work it seems that the paper could definitely be publishable.

审稿人讨论附加意见

Two of the reviewers participated in the discussion, but two did not. The more negative reviewer who participated maintained their score after the discussion period. I took a look at the concerns of the other reviewers, as well as the rebuttal. Though the rebuttal did resolve some issues, I think the paper still needs work before it is ready to be published.

最终决定

Reject