Revive Re-weighting in Imbalanced Learning by Density Ratio Estimation
摘要
评审与讨论
This paper presents a dynamic re-weighting method for imbalanced learning. The author defines the ratio of the balanced data set distribution to the training set distribution, and tries to estimate it with an iterative update method. The effectiveness of this method is proved by experiments.
优点
- This paper points out a problem with distribution differences, which leads to the potential missing feature patterns in general re-weighting methods.
- This paper proposes a new method, which approximates the ratio of the balanced data set distribution to the training set distribution using methods of density ratio estimation. As far as I know, a dynamic re-weighting strategy is novel in this field.
- The experimental introduction of this paper is clear, and extensive experiments have been carried out, which validates the effectiveness of the proposed method.
缺点
- The formula derivation in Sec. 3.3 can be more detailed. It is suggested to explain how formula (7) is obtained in the appendix.
- The introduction may have overlooked some key articles. For example, the article mentions Wang et al. 's article at the end of Sec.3.3, but does not discuss this paper in the introduction section.
- Does the new method enjoy the same theoretical boundaries as the general reweighting method? It is recommended to provide more analysis.
- Besides, there are some typos in the details:
- In the experimental section, 'class[390,385] 'may be a typo.
- In table 3, the interpretation of Tr_{Few} is supposed to be there.
问题
Please refer to Weaknesses.
Besides, I am also curious about some experimental details. Do the authors use other techniques such as RandAug or mixup?
局限性
No. Although the authors say they clarify the limitations in Sec.3.4, I find they mainly highlight the efficiency of the proposed method. More discussion about limitations and potential negative societal impact is recommended.
The formula derivation in Sec. 3.3 can be more detailed. It is suggested to explain how formula (7) is obtained in the appendix.
Thank you for the suggestion. We will include detailed derivations of Eq. (7) in the appendix of revision, which we simply summarize the deduction as follows for clarity:
Back to Eq. (5), replace the expectations over and by and , respectively. For each class , We can obtain
, where
Then, taking the derivative of with respect to and setting it to zero, we can obtain the estimation of density ratio in imbalanced learning as follows
Solving equation above with respect to , we can obtain the solution as
The introduction may have overlooked some key articles. For example, the article mentions Wang et al. 's article at the end of Sec.3.3, but does not discuss this paper in the introduction section.
We will take the reviewers' advice to carefully revise the second paragraph of the Introduction (lines 32-34), specially including a discussion of the work by Wang et al. and other related studies. We present the corresponding revision as follows for your reference.
However, such subsequent improvements can alleviate but still cannot effectively push that forward. Wang et al. [2023] obtains a fine-grained generalization bound for re-weighting in imbalanced learning through the data-dependent contraction technique. Limited research has focused on the intrinsic limitations...
Does the new method enjoy the same theoretical boundaries as the general reweighting method? It is recommended to provide more analysis.
Thank you for your suggestion. To address the reviewer's concern, we present a generalization bound sketch for the re-weighting-based methods and provide the possible insights with the empirical verifiction in the following. Note that, about its formal version with the complete hypothesis and the deduction procedure, we will update to the manuscript in the final.
Given the model and a reweighting loss function , for any , with probability at least over the training set , the following generalization bound holds for the risk on the balanced distribution
where is positively correlated with the empirical reweighting risk of training set. denotes the empirical complexity of the function set . denotes the minimal prediction on the ground-truth class in the training set. refers to the weight of class of the reweighting loss . The formal theory and the proof will presented in the revision.
We can get some insights from the above generalization bound:
- Why reweighting is necessary: helps to rebalancing the imbalanced term to get a sharper bound.
- Why dynamic reweighting is necessary: The term changes dynamically with model training. Therefore, we need a that can adapt dynamically to the changes of .
- Why RDR works: From Figure 1 in the attached PDF, we observe that the dynamic trend of the RDR weight aligns well with , denoted as . This demonstrates that our RDR can adapt to the dynamic changes in , maintaining a sharp bound during dynamic training.
Besides, there are some typos in the details: In the experimental section, 'class[390,385] 'may be a typo. In table 3, the interpretation of Tr_{Few} is supposed to be there. Do the authors use other techniques such as RandAug or mixup?
Thank you for pointing it out. We have carefully proofread the whole manuscript, corrected typos, and added a table about necessary explanations of some notations or terms (please refer to the notation table in the attached PDF file). denotes average trace of Hessian matrix over Few classes. These changes will be included in the revision. We used RandAug in our experiments but did not include mixup, for fair competition.
More discussion about limitations and potential negative societal impact is recommended.
Thanks for your suggestion. We will enrich the discussion about limitation and social impact in the revision. Specially, we will highlight the potential computational complexity issue when scaling up to a super large number of classes (e.g., in face recognition, retail product recommendation and landmark detections), which should pay additional attention to some corresponding lightweight techniques. Regarding the potential negative societal impact, we should clearly recognize that overly rebalancing to minority groups during training may bring the desctructive effect on the learning of majority groups, which is also not the desirable goal of imbalanced learning. All rebalancing techniques should build on top of a proper range for fairness and we should avoid some improper abuse by the malicious minority groups.
Dear Reviewer 24eD,
We genuinely appreciate your detailed feedback and the insightful comments you've provided on our manuscript.
We have provided more clarifications and explanations as suggested. Additionally, we have discussed limitations and potential negative societal impact.
Please let us know if anything is unclear. We truly appreciate this opportunity to improve our work and shall be grateful for any feedback you could give to us.
Best Regards,
The authors of Submission 4464
Thanks for your rebuttal! My concerns have been clarified. Hence, I will increase my rating accordingly.
Dear Reviewer 24eD,
We greatly appreciate your time and effort in reviewing our responses and contributing to the enhancement of this paper. We will carefully follow your suggestions to incorporate all the points of our rebuttal in the revised version.
Best,
The authors of Submission 4464
The paper introduces a novel approach called Re-weighting with Density Ratio (RDR) to address the challenges posed by imbalanced data distributions in machine learning. The RDR approach aims to mitigate overfitting on majority classes and enhance adaptability across diverse datasets by continuously updating the weights in response to observed shifts in class density. Extensive experiments on various large-scale, long-tailed datasets demonstrate that the RDR method significantly improves the model's generalization capabilities, particularly under severely imbalanced conditions. The analysis of the weight changes during training reveals that the method increasingly focuses on minority classes as training progresses, initially learning common features across all categories and then targeting learning towards minority samples to enhance generalizability. The paper also provides an ablation study to further validate the effectiveness of the proposed approach.
优点
-
The paper introduces a novel approach called Re-weighting with Density Ratio (RDR) to address the challenges posed by imbalanced data distributions in machine learning.
-
Extensive experiments on various large-scale, long-tailed datasets demonstrate that the RDR method significantly improves the model's generalization capabilities, particularly under severely imbalanced conditions.
-
The analysis of the weight changes during training reveals that the method increasingly focuses on minority classes as training progresses, initially learning common features across all categories and then targeting learning towards minority samples to enhance generalizability.
-
The paper provides an ablation study to further validate the effectiveness of the proposed approach.
-
The results show that RDR generally outperforms other methods, including Inverse Frequency (1/n) and SAM variants, in both the Many and Few classes, indicating that RDR can efficiently address the overfitting issues for Few classes.
缺点
-
The paper does not provide a detailed theoretical analysis or justification for the proposed Re-weighting with Density Ratio (RDR) method, beyond the intuition that it can mitigate overfitting on majority classes and enhance adaptability across diverse datasets.
-
I am interested in how RDR might perform in the presence of extreme imbalance, noisy data, or other challenging scenarios. The current experiment is well-established but dataset itself is relatively simple.
-
The paper discusses reweighting/non-reweighting for classification problems. I suggest the authors also briefly discuss reweighting methods in imbalanced regression problems, e.g., VIR [1] for reweighting problems and ConR [2] for non-reweighting problems.
[1] Variational Imbalanced Regression: Fair Uncertainty Quantification via Probabilistic Smoothing, NeurIPS 2023
[2] ConR: Contrastive Regularizer for Deep Imbalanced Regression, ICLR 2024
Summary I think the theoretical analysis or at least insights is needed for acceptance, so my suggest score is 5, as the experiment part is excellent in this paper.
问题
see above
局限性
authors discussed in sec 3.4
The paper does not provide a detailed theoretical analysis or justification for the proposed Re-weighting with Density Ratio (RDR) method, beyond the intuition that it can mitigate overfitting on majority classes and enhance adaptability across diverse datasets.
Thank you for your suggestion. To address the reviewer's concern, we present a generalization bound sketch for the re-weighting-based methods and provide the possible insights with the empirical verifiction in the following. Note that, about its formal version with the complete hypothesis and the deduction procedure, we will update to the manuscript in the final.
Given the model and a reweighting loss function , for any , with probability at least over the training set , the following generalization bound holds for the risk on the balanced distribution
where is positively correlated with the empirical reweighting risk of training set. denotes the empirical complexity of the function set . denotes the minimal prediction on the ground-truth class in the training set. refers to the weight of class of the reweighting loss . The formal theory and the proof will presented in the revision.
We can get some insights from the above generalization bound:
- Why reweighting is necessary: helps to rebalancing the imbalanced term to get a sharper bound.
- Why dynamic reweighting is necessary: The term changes dynamically with model training. Therefore, we need a that can adapt dynamically to the changes of .
- Why RDR works: From Figure 1 in the attached PDF, we observe that the dynamic trend of the RDR weight aligns well with , denoted as . This demonstrates that our RDR can adapt to the dynamic changes in , maintaining a sharp bound during dynamic training.
I am interested in how RDR might perform in the presence of extreme imbalance, noisy data, or other challenging scenarios. The current experiment is well-established but dataset itself is relatively simple.
Thank you for the suggestion. In the following, we conduct more experiments. The results are shown in Table 2 and Table 3 in the attached PDF.
- We include the results of two new datasets: CIFAR-10-LT-NL and CIFAR-100-LT-NL, with both class imbalance and label noise. We can see that on more complex datasets, our method achieves consistent and significant improvements.
- Additionally, note that the imbalance factors constructed on ImageNet-LT and Places-LT in the submission are 256 and 996, respectively. These datasets are indeed extremely imbalanced. We will explicitly indicate the imbalance factors of these two datasets in the experiment tables in the revision. Furthermore, we conduct experiments on the more imbalanced CIFAR-10-LT and CIFAR-100-LT datasets, specifically with imbalance factors of 200 and 500. Our method consistently achieves significant improvements.
The paper discusses reweighting/non-reweighting for classification problems. I suggest the authors also briefly discuss reweighting methods in imbalanced regression problems, e.g., VIR [1] for reweighting problems and ConR [2] for non-reweighting problems. [1] Variational Imbalanced Regression: Fair Uncertainty Quantification via Probabilistic Smoothing, NeurIPS 2023 [2] ConR: Contrastive Regularizer for Deep Imbalanced Regression, ICLR 2024
Thank you for the valuable suggestion and the recommendation. We will include a brief discussion on imbalanced regression problems in the revision as follows:
We shall note that the primary focus of this study is on imbalanced classification. Beyond imbalanced classification that has discretized label space, an noteworthy area, imbalanced regression that has a continuous label space is also very common in real applications [a,b]. In this direction, the empirical label distribution often does not accurately reflect the true label density in regression tasks, which limits the effectiveness of traditional reweighting techniques [a,c]. Label Distribution Smoothing (LDS) [a] and Variational Imbalanced Regression (VIR) [c] propose using kernel smoothing and other techniques to estimate an accurate label density distribution. Ranking Similarity (Ranksim) [b] leverages local and global dependencies by encouraging the correspondence between the similarity order of labels and features. Balanced Mean Squared Error (Balanced MSE)[d] extends the concept of Balanced Softmax[e] to regression tasks to achieve a balanced predictive distribution. Contrastive Regularizer (ConR)[f] improves contrastive learning techniques to translate label similarities into the feature space. Considering the different rebalancing paradigms compared with that in imbalanced classification, and the limited space, we will leave the potential extension of our RDR to this area in the future explorations.
[a] Delving into Deep Imbalanced Regression. ICML 2021.
[b] RankSim: Ranking Similarity Regularization for Deep Imbalanced Regression. ICML 2022.
[c] Variational Imbalanced Regression: Fair Uncertainty Quantification via Probabilistic Smoothing. NeurIPS 2023.
[d] Balanced MSE for Imbalanced Visual Regression. CVPR 2022.
[e] Balanced Meta-Softmax for Long-Tailed Visual Recognition. NeurIPS 2020.
[f] ConR: Contrastive Regularizer for Deep Imbalanced Regression. ICLR 2024.
I would like to thank the authors for their detailed response. I will take all the reviews and responses into consideration.
It seems that the other reviewers have not responded yet. After reviewing the authors' responses to the other reviewers, I have decided to raise my score. However, I cannot promise that the authors' responses address the concerns of the other reviewers, so the authors should not use my improved score as a reference or evidence.
In summary, I appreciate the authors' responses to all of us. Specifically, I feel that the authors have addressed my questions, so I have decided to increase my score.
Dear Reviewer MG5X,
We sincerely appreciate you taking the time to review our responses and contributing to improve this paper. We will carefully follow reviewer's advice to incorporate all the addressed points with additional experiments in the updated version.
We promise not to use your improved score as evidence to persuade other reviewers, but focus on truly addressing their concerns.
Thank you once again for your dedicated and valuable contribution in reviewing our paper!
Best,
The authors of Submission 4464
The paper presents a weighting strategy in order to handle class imbalance. Contrary to existing method, they propose to adapt the weight throughout the training procedure.
Their method estimates the discrepancy between the sample distribution and the balanced sample distribution for parameterization w and updates the estimate through the training.
The authors use two resnet architectures to evaluate their contribution on multiple datasets. They also compare to other baselines and show significant gain.
优点
- The paper develops a novel approach for handling class imbalance.
- The methodology is derived theoretically from the problem formulation
- The authors propose an analysis of the complexity of the method and empirically evaluate the training time.
- The methodology is evaluated on multiple datasets and compared to multiple baselines.
缺点
- The paper is sometimes difficult to read:
- Row 125, the authors refer to the distribution of training set, which get parameterized by w. Thus, my understanding is that the authors refer to the distribution of the training set "captured by the model".
- row 134, P_bal = pi P.. P_bal is the distribution of y in the balanced case ? But should therefore be 1/number classes... and P, should just be the class proportion and we should have P = pi P_bal ?
- LDAM and LA terms are not defined at first. First definition of LA is at row 208
- row 212 "trategies" => strategies
问题
- Could you clarify the term, P, etc. You often refer to it as "real world data distribution", but the distribution of x does not depends on any other classes (imbalanced or not) ?
局限性
The paper would benefit from a clarification in notification (what is the true data distribution, what is the feature distribution, what is an estimate of what quantity, etc.). I believe the contribution is novel and worth it.
Row 125, the authors refer to the distribution of training set, which get parameterized by w. Thus, my understanding is that the authors refer to the distribution of the training set "captured by the model".
Yes, you are right. We use the distribution parameterized by to represent "the distribution captured by the model". We will enrich its description for clarity in the revision.
row 134, P_bal = pi P.. P_bal is the distribution of y in the balanced case ? But should therefore be 1/number classes... and P, should just be the class proportion and we should have P = pi P_bal ?
Thank you for pointing out this oversight: the term in the equation should indeed be . We would like to clarify here that this mistake is limited to this equation and does not affect the correctness of our subsequent derivations. We greatly appreciate the reviewer’s careful review and will make the correction.
LDAM and LA terms are not defined at first. First definition of LA is at row 208
row 212 "trategies" => strategies
LDAM and LA respectively refer to Label-Distribution-Aware margin loss and Logit Adjustment, two prevalent long-tailed learning methods. We will include their full names at the first mention. For the typo, we have carefully proofread the whole manuscript again and corrected all possible mistakes. These changes will be included in the revision.
Could you clarify the term, P, etc. You often refer to it as "real world data distribution", but the distribution of x does not depends on any other classes (imbalanced or not) ?
Sorry for the possible confusion. Here is a clarification of :
represents the real world data distribution, or more directly, the training set distribution. Specifically, is the joint distribution of the training data (x, y), and the training set is obtained by IID sampling from . and represent the marginal distributions of the training data and , respectively. and represent the corresponding conditional distributions. For other distribution notations parameterized by , they mean the corresponding distributions captured by the model, as the reivewer's understanding.
We also place a notation table for clarity in the attached PDF file, which will be placed into the manuscript if possible (or in the appendix if not). If you have any further questions or need any clarifications, please let us know. We will carefully improve the manuscript.
Dear Reviewer xrua,
We sincerely appreciate your constructive feedback and positive evaluation of our submission.
We have provided more clarifications and explanations, made necessary corrections, and improved definitions as suggested.
Please let us know if anything is unclear. We truly appreciate this opportunity to improve our work and shall be most grateful for any feedback you could give to us.
Best Regards,
The authors of Submission 4464
Dear authors,
I thank you for your answers. After considering your propositions and mostly the enhanced readability (e.g. symbol table), I have decided to raise my note.
Regards
Dear Reviewer xrua,
We sincerely appreciate you taking the time to review our responses and contributing to improve this paper. We will carefully follow your advice to incorporate all the points of our rebuttal in the updated version.
Best,
The authors of Submission 4464
We thank reviewers for your valuable feedback, and appreciate the great efforts made by all reviewers, ACs, SACs and PCs.
Please refer to our detailed responses to each reviewer, where we addressed each question and concern point by point. In the attached PDF, we have included a notation summary table for improved readability and clarification, additional tables for experimental results, along with the empirical validation of the theoretical insights presented in the rebuttal.
We appreciate all reviewers’ time again. We are looking forward to your reply!
The authors address the critical and challenging problem of class imbalance in machine learning. They propose a method that adaptively weighs the importance of classes throughout training. The main idea is to use an estimate of the density ratios (existing vs. balanced) to re-weight the samples in the loss. The advantage of the method is demonstrated using several real datasets with imbalance ratios ranging from 10 to 256. Overall, the reviewers are positive about the paper, with the main criticism being the presentation. This work provides an interesting solution to the problem that seems to work well. For these reasons, I recommend acceptance of this paper, and I highly encourage the authors to implement the reviewers' suggestions and work towards improving the writing and presentation.