PaperHub
7.3
/10
Poster4 位审稿人
最低3最高6标准差1.1
3
5
4
6
4.3
置信度
创新性3.0
质量2.8
清晰度2.8
重要性2.8
NeurIPS 2025

TS-MOF: Two-Stage Multi-Objective Fine-tuning for Long-Tailed Recognition

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29

摘要

关键词
Long-tailed learningMulti-objective optimization

评审与讨论

审稿意见
3

This paper proposes TS-MOF, a two-stage framework for long-tailed recognition (LTR). Stage 1 pre-trains a feature backbone using standard cross-entropy loss. Stage 2 freezes the backbone and fine-tunes multiple classifier heads via multi-objective optimization (MOO). In Stage 2, the authors proposed two interesting ideas: R-PLA and RD-PCGrad for long-tailed recognition.

优缺点分析

Strengths

  1. The method underwent rigorous evaluation across three benchmark datasets under varying imbalance ratios.
  2. R-PLA and RD-PCGrad are novel and interesting.

Weaknesses

  1. While the two strategies (R-PLA and RD-PCGrad) in the fine-tuning stage are interesting, the two-stage framework itself appears non-novel. What specific problems do these two strategies solve? The writing logic suffers from a weak connection between the paper’s motivation and these strategies. The authors should significantly reorganize the Introduction. Since the core effectiveness stems from the two MOO strategies, the Abstract and Introduction should highlight them as primary innovations rather than overemphasizing the two-stage framework.

2.The paper lacks schematic diagrams, reducing accessibility. Adding visual illustrations (e.g., workflow of R-PLA/RD-PCGrad) would significantly improve intuitive understanding for readers.

  1. The multi-head structure resembles multi-expert models (e.g., SHIKE). However, the reported performance of SHIKE (46.9% on CIFAR100-LT IR=100 in Table 1) contradicts its original paper (56.3%). The authors must: (i) Clarify this discrepancy, (ii) Analyze how TS-MOF fundamentally differs from multi-expert ensembles in methodology and objectives.

  2. Critical implementation details are missing — specifically, which re-balancing strategies (e.g., LDAM? BCL?) were integrated into the multi-head MOO framework.

  3. Ablations are critically underdeveloped. The authors should: (1) Isolate the impact of individual components (R-PLA, RD-PCGrad); (2)Compare variants (e.g., alternative task-weighting schemes, gradient conflict resolution methods); (3) Benchmark performance across different strategy combinations.

问题

See Strengths And Weaknesses

局限性

See Strengths And Weaknesses

最终评判理由

After rebuttal and discussion, I still tend to reject this paper, so I maintain my initial decision.

格式问题

See Strengths And Weaknesses

作者回复

Response to Reviewer Fe6u

Thank you very much for your careful review and insightful comments, which are truly beneficial for improving the rigor of our paper.

W1. On the Novelty of the Two-Stage Framework:

Thank you for your valuable suggestions regarding the writing logic. This has led us to deeply reconsider how to best present our core contributions.

We will clarify that the core advantage of our work lies in the synergy between our proposed new components and the two-stage framework:

  • R-PLA and RD-PCGrad, as the technical foundation of our method, can stably resolve gradient conflicts among multiple strategies, ensuring the optimization process rapidly moves towards a Pareto-improving direction. This allows the insights from different long-tail strategies to be efficiently integrated.
  • The two-stage decoupled framework, in turn, drastically reduces the cost of this strategy integration. Compared to other multi-expert or multi-objective methods that require end-to-end training, our method only needs to fine-tune the lightweight classifier heads, significantly reducing computational overhead and training time.

In short, the new components ensure the "effectiveness of fusion," while the two-stage framework ensures the "efficiency of fusion." We will follow this logic to clarify our innovations and motivation in the revised manuscript. Thank you again for your profound insights.

W2. Adding a Framework Schematic:

Apologies, but due to rebuttal guidelines, we're unable to directly include images.​ We will add a full-process schematic diagram in the manuscript, including Stage 1 pre-training and Stage 2 multi-expert fine-tuning, as well as detailed illustrations of R-PLA weight calculation and RD-PCGrad gradient coordination. This will significantly improve the method's comprehensibility and clearly highlight our contributions.

W3. Regarding the SHIKE Performance and Comparison to Multi-Expert Models:

We apologize for the misunderstanding and would like to provide a clarification. Our paper uses a unified framework to ensure compatibility, allowing different expert strategies to be incorporated into the fine-tuning process. In our experiments, we used the backbone from LOS and retained the original hyperparameters for other strategies. The specific hyperparameters for the backbone and our MOF components were kept consistent across different combinations to diligently validate the method's generality and its performance improvement over baselines. Under this unified framework, the baseline performance of different methods may vary. We will emphasize this point in the subsequent revision. Additionally, we have done our best to supplement experiments to clearly demonstrate the performance gains over the baseline.

Comparison to Multi-Expert Models: Our motivation differs from that of multi-expert methods. Multi-expert models (such as SHIKE and SADE) can work well independently and are often meticulously designed by researchers. In contrast, our work is a study on a gradient-based multi-objective optimization method that is strongly relevant to long-tail learning. Its goal is to accelerate the utilization of these expert strategies and speed up the fine-tuning process. Therefore, it can potentially absorb benefits from different representation learning methods and expert strategies that favor balance. When combined with other strategies, it can be viewed as a new type of long-tailed multi-expert method. We will add comparisons and discussions with classic multi-expert models (such as SHIKE and SADE) in the main text.

W4 & W5. Supplementing Implementation Details and Ablation Studies:

Thank you for your suggestions. We will clarify these points in the main text to enhance comprehensiveness. The strategies fused by MOF in our main experiments are KPS, BCL, and LOS (as hinted in Figure 2), and we will explicitly state this in the revised text. Furthermore, we have diligently supplemented necessary experiments:

R-PLARD-PCGradEOSSIR=10IR=50IR=100
HeadMediumTailAllHeadMediumTailAllHeadMediumTailAll
72.347.3--64.674.243.522.952.476.546.918.148.6
74.760.8--70.475.350.444.559.678.249.138.656.1
74.661.5--70.575.250.545.859.778.849.038.956.4
74.360.3--69.975.551.043.559.778.449.836.956.0
74.660.1--70.175.350.743.959.678.349.237.455.9
74.762.0--70.875.250.747.560.279.049.239.956.8

Table: Ablation experiments of TS-MOF three modules with IR=10/50/100 on CIFAR100-LT Benchmarks.

Component Variant Analysis: We have conducted experiments comparing our components with alternative methods to demonstrate their superiority.

ModulesVariant Methodmacs/MTime/minγ1\gamma_1γ2\gamma_2γ1/γ2\gamma_1/\gamma_2Accuracy(%)
R-PLAentropy3.5408.451.0001.0010.99956.3
confidence3.5408.781.0000.9641.03853.5
uniform3.5408.441.0001.0020.99856.3
pla(ours)3.5408.461.0001.0001.00056.8
RD-PCGradmgda3.5408.361.0000.9061.10456.4
gradnorm3.5408.761.0000.8641.15755.2
uncertainty3.5407.461.0001.0150.98556.3
pcg(ours)3.5407.571.0001.0001.00056.8

Table: Comparison of computational cost and model performance between TS-MOF and conventional multi-task optimization in the second stage.

Strategy Combination Benchmark: To showcase the general applicability of our framework, we have benchmarked the performance of various strategy combinations.

MethodIR=10IR=50IR=100
HeadMediumTailAllHeadMediumTailAllHeadMediumTailAll
CE+BS72.449.9--40.774.241.26.940.877.143.54.041.5
CE+KPS72.547.2--39.973.839.912.142.072.438.118.142.9
KPS+BCL72.342.2--38.274.137.215.642.377.737.410.842.0
CE+BS+CE-DRW72.248.4--40.274.440.28.240.977.541.93.841.1
CE+BS+LDAM-DRW72.250.4--40.874.041.511.242.277.343.45.742.1
CE+BCL+CE-DRW72.139.1--37.174.635.72.137.577.938.92.939.9
CE+BS+KPS+CE-DRW72.349.2--40.574.341.09.541.677.343.15.542.0
CE+BS+KPS+LDAM-DRW72.050.8--40.971.641.125.446.174.743.314.344.1
TS-MOF(CE+BS)\text{TS-MOF}_\text{(CE+BS)}73.054.1--42.41.742.4_{1.7}74.846.624.148.57.748.5_{7.7}77.949.813.747.25.747.2_{5.7}
TS-MOF(CE+KPS)\text{TS-MOF}_\text{(CE+KPS)}73.457.6--43.73.843.7_{3.8}74.544.735.851.79.751.7_{9.7}77.443.528.149.76.849.7_{6.8}
TS-MOF(KPS+BCL)\text{TS-MOF}_\text{(KPS+BCL)}74.057.4--43.85.243.8_{5.2}75.043.438.852.410.152.4_{10.1}77.742.526.849.05.049.0_{5.0}
TS-MOF(CE+BS+CE-DRW)\text{TS-MOF}_\text{(CE+BS+CE-DRW)}73.554.3--42.62.442.6_{2.4}75.247.023.848.67.748.6_{7.7}78.849.413.347.16.047.1_{6.0}
TS-MOF(CE+BS+LDAM-DRW)\text{TS-MOF}_\text{(CE+BS+LDAM-DRW)}73.755.1--42.92.142.9_{2.1}75.547.527.950.38.150.3_{8.1}78.849.522.450.28.150.2_{8.1}
TS-MOF(CE+BCL+CE-DRW)\text{TS-MOF}_\text{(CE+BCL+CE-DRW)}73.547.9--40.53.440.5_{3.4}75.741.612.043.15.643.1_{5.6}78.443.27.042.82.942.8_{2.9}
TS-MOF(CE+BS+KPS+CE-DRW)\text{TS-MOF}_\text{(CE+BS+KPS+CE-DRW)}74.057.7--43.93.443.9_{3.4}76.248.937.654.212.654.2_{12.6}78.450.229.552.710.752.7_{10.7}
TS-MOF(CE+BS+KPS+LDAM-DRW)\text{TS-MOF}_\text{(CE+BS+KPS+LDAM-DRW)}74.558.0--44.13.244.1_{3.2}76.349.238.554.78.654.7_{8.6}78.550.930.953.45.753.4_{5.7}

Table: Benchmark performance across different strategy combinations with IR=10/50/100 on CIFAR100-LT Benchmarks.The subscript is the improvement of the relative linear weighted combination.

评论

Dear Reviewer Fe6u,

Thank you again for your insightful and highly constructive review of our submission. We have submitted a detailed rebuttal and, following your guidance, have systematically addressed each of your key concerns with substantial new experiments and firm commitments for the revision.

To facilitate your review, we've summarized our main efforts below:

  1. On Writing Logic and Novelty (W1): We have embraced your suggestion and committed to restructuring the introduction. We propose a clearer narrative: our new components ensure the "effectiveness of fusion," while the two-stage framework ensures the "efficiency of fusion," to better highlight our core contributions.

  2. On the Lack of Schematics (W2): We have committed to adding the schematic diagrams you recommended in the final version. These will illustrate the overall workflow and our core modules (R-PLA/RD-PCGrad) to enhance the paper's readability.

  3. On SHIKE Performance and Multi-Expert Comparison (W3): We clarified the performance discrepancy, explaining it stems from our unified experimental framework. We also articulated the fundamental difference in motivation between our method (an efficient multi-objective optimization tool) and traditional multi-expert ensembles.

  4. On Critical Details and Ablation Studies (W4, W5): This is where our response is most extensive. We have provided extremely thorough supplementary experiments that precisely address all your requests:

    • We specified the exact strategies fused in our experiments (KPS, BCL, LOS).
    • We provided an independent ablation study for our core components (New Table 1).
    • We added a performance comparison against alternative methods to demonstrate our components' superiority (New Table 2).
    • We included a benchmark of various strategy combinations to showcase the framework's general applicability (New Table 3).

We are eager to resolve any outstanding concerns. Should you have any further questions or require additional clarifications, we would be very happy to address them to the best of our ability within the remaining discussion period.

Thank you once again for your professional guidance and valuable time.

审稿意见
5

This paper proposes a practical solution to address the long-tailed distribution problem in machine learning. The TS-MOF framework, through its ingenious two-stage design, enables multiple long-tailed learning strategies to work synergistically, achieving excellent performance across multiple benchmark datasets.

优缺点分析

S1: The design of decoupling feature learning from classifier adaptation greatly simplifies the optimization of complex long-tailed problems, allowing the MOO process to act more stably and efficiently on classifier heads.

S2: R-PLA and RD-PCGrad are the core highlights of this work. R-PLA's adaptive task weighting mechanism can dynamically adjust task weights based on real-time performance patterns, while RD-PCGrad ensures stable resolution of gradient conflicts and constructive fusion, effectively promoting synergy among different LTR strategies.

S3: The paper achieves significant results on multiple benchmark datasets, demonstrating its efficiency.

W1: Although the MOO components enhance robustness, there are still some hyperparameters in R-PLA and RD-PCGrad. I want see some sensitivity analysis of how these parameters affect model performance and convergence.

问题

Q1: Does εsimε_{sim} have a role other than ensuring numerical stability when computing sj,es_{j,e}? For example, in cases of extremely low similarity, does it affect weight allocation preferences?

局限性

Yes

最终评判理由

After reading the authors' response, I would like to maintain my score.

格式问题

N/A

作者回复

Of course. Here is a response to Reviewer uDJM, addressing questions W1 and Q1 using the provided information.


Response to Reviewer uDJM

Thank you for your positive feedback and for highlighting the core strengths of our work. We are grateful for your insightful questions and provide our responses below.

W1: On Hyperparameter Sensitivity Analysis

Thank you for this valuable suggestion. We agree that analyzing the sensitivity of the hyperparameters in R-PLA and RD-PCGrad is important for understanding their robustness. We have conducted a sensitivity analysis for these key parameters and present the results below.

The data indicates that while the modules' performance shows some dependence on hyperparameters, our method remains robust across a reasonable range of values, consistently outperforming baselines. This suggests that meticulous tuning is not required to achieve strong results, although optimal performance can be found with experimental combinations.

ModuleParameter NameParameter ValueManyMediumFewAccuracy (%)
R-PLApla_beta_clip_min0.178.2049.7437.8056.12
0.377.9149.6038.1056.06
0.578.2649.6038.4356.28
pla_beta_clip_max1.078.1748.8637.3755.67
2.078.2646.0013.7047.60
pla_similarity_threshold0.178.4048.4037.6055.66
0.378.4349.3437.1055.85
0.578.2048.7438.5756.00
RD-PCGradpcg_epsilon1e-0678.4348.9138.6356.16
1e-0478.1450.4038.7356.61
pcg_conflict_threshold-0.178.8351.6035.9356.43
0.178.3149.1736.2755.50
0.378.1449.5139.0356.39

Table: Hyperparameter analysis for R-PLA and RD-PCGrad modules on the CIFAR100-LT benchmark (IR=100).

Q1: On the Role of ϵsim\epsilon_{\text{sim}}

Your intuition is correct; ϵsim\epsilon_{\text{sim}} has a subtle secondary role beyond ensuring numerical stability. While its primary function is to prevent division by zero when the norms of the performance vectors are near zero, it also acts as a stabilizer in cases of extremely low similarity.

When the performance vectors are nearly orthogonal, their dot product is close to zero. Without ϵsim\epsilon_{\text{sim}}, the denominator would also be close to zero, making the similarity value highly sensitive to small numerical fluctuations. By ensuring the denominator has a small positive lower bound, ϵsim\epsilon_{\text{sim}} dampens this sensitivity and prevents drastic swings in the similarity value (and thus the weight allocation) due to minor noise. In essence, it helps to regularize the similarity computation, ensuring that the weight allocation remains stable and meaningful even when strategies are performing very differently.

审稿意见
4

This paper presents TS-MOF (Two-Stage Multi-Objective Fine-tuning), a novel decoupled-training expert-based framework for long-tailed recognition. TS-MOF decouples training into two stages: (1) Feature Learning: standard supervised training to learn good representations ; (2) Multi-head Classification Learning: freeze the learned representation and spawn multiple specialized “expert” heads, each fine-tuned with a different long-tailed method (e.g., BCL [26], LOS [20]). Key contributions in the second stage include: (a) Refined Performance Level Agreement (R-PLA): Dynamically weights each expert’s task loss based on per-class performance on the training set ; (b) RD-PCGrad: a robust deterministic variant of PCGrad [23] that resolves gradient conflicts between task objectives via gradient modulation. At inference time, a dynamic expert aggregation module weights and combines expert predictions per class, using performance on a validation set. Experiments on standard long-tailed benchmarks show that TS-MOF achieves higher balanced accuracy than single-objective baselines, with only a brief fine-tuning stage.

优缺点分析

Strengths

  • S1 - Novel decoupled-training framework: introduces a two-stage learning approach that leverages multiple existing long-tailed approaches in parallel experts, with training-time adaptive task weighting and inference-time aggregation.

  • S2 - Theoretical insights: Provides analysis on gradient stability, Pareto improvements, convergence behaviour, and robustness of R-PLA and RD-PCGrad.

Weaknesses

  • W1 - Reliance on validation set: Requires a (probably) balanced holdout set to select Stage 1 weights (L160–161) and to compute per-class aggregation weights (Eq. 8–9). In extreme imbalance scenarios, such a validation set may be impractical/unfeasible, and it isn't a common assumption in the LTR literature.

  • W2 - Missing component ablations: Lacks experiments isolating the contributions of R-PLA, RD-PCGrad, and the expert aggregation module.

  • W3 - Comparison to expert-based SOTAs: Does not discuss or compare with recent skill-diverse multi-expert models (e.g., BalPoE [A], SADE [B], MDCS [C]) that specialize experts with multiple objectives end-to-end.

  • W4 - Task Selection Strategy: While Appendix shows a large task combination search, the paper picks the “best” combination based on test-set performance. A more scalable, principled selection mechanism is missing from the paper.

  • W5 - Deterministic task ordering: The method for defining the fixed task order in RD-PCGrad is not specified.

  • W6 - Computational Overhead: No empirical comparison of Stage 2’s compute or memory cost versus vanilla multi-task optimization.

Additional references:

[A] Sanchez Aimar, Emanuel, et al. Balanced Product of Calibrated Experts for Long-Tailed Recognition. CVPR 2023.

[B] Zhang, Yifan, et al. Self-Supervised Aggregation of Diverse Experts for Test-Agnostic Long-Tailed Recognition. NeurIPS 2022.

[C] Zhao, Qihao, et al. MDCS: More Diverse Experts with Consistency Self-Distillation for Long-Tailed Recognition. ICCV 2023.

问题

  • Q1 - RD-PCGrad vs. PCGrad [23]: Why choose a fixed task ordering rather than PCGrad’s random ordering? What evidence shows that RD-PCGrad changes improve robustness?

  • Q2 - Validation-Set requirement: In the absense of a balanced validation set, can TS-MOF work with an imbalanced holdout set matching training distribution? If so, have you tested sensitivity to validation set imbalance?

  • Q3 - R-PLA computation: In Eq. 4, how do you estimate early performance vectors (index j) before computing cosine similarity with the final vector (index N) within the same epoch?

  • Q4 - Expert-head architectures: What is the architecture of each classification head (linear layer, MLP, etc.)? How does this compare to expert-based models such as RIDE [22]?

  • Q5 - Hyperparameter sensitivity: Which TS-MOF hyperparameters require careful tuning (as indicated in Section 10, L600–601)?

While the TS-MOF framework presents a novel take into decoupled-learning and expert-based frameworks, the absence of key ablations, the impractical balanced-validation-set assumption, unclear task-selection and ordering procedures, and missing comparisons to relevant expert-based methods leave important questions unanswered. A focused rebuttal that supplies the above ablations/comparisons and clearly quantifies practical costs will significantly strengthen confidence in TS-MOF’s utility—and could increase my score. Conversely, if these gaps remain unaddressed, my score will stay at 3.

局限性

yes

最终评判理由

Despite my initial concerns, the authors have clarified several key aspects: their method does not strictly depend on a balanced holdout dataset; each contribution has a relevant role per the ablation study; it outperforms recent expert-based baselines in expanded comparisons; it shows low computational overhead; it exhibits low hyperparameter sensitivity; and they’ve explained their task selection and ordering decisions. Overall, these improvements convince me to raise my score, leaning toward acceptance.

格式问题

No formatting issues detected.

作者回复

Response to Reviewer BSGz

Thank you for your detailed and critical feedback. We have done our best to supplement new experiments to substantially improve the paper.

W1 & Q2. On the Dependency on a Balanced Validation Set:

Thank you for your recognition of our experimental section. We would like to clarify that our method does not strictly depend on a balanced validation set. For the extreme case you are concerned about, such as when a validation set is unavailable, we can also sample from the training set. To explicitly verify this, we have added a new experiment on CIFAR100-LT where we construct an imbalanced validation set by sampling from the training set. The results (shown below) indicate that although there is a slight drop in performance compared to using the original (more balanced) validation set, TS-MOF still significantly outperforms the SOTA baseline (LOS), demonstrating its applicability and robustness.

MethodIR=10IR=50IR=100
HeadMediumTailAllHeadMediumTailAllHeadMediumTailAll
LOS71.962.3--69.072.451.440.258.070.352.336.653.9
TS-MOF(from Train Set)74.161.8--69.674.850.247.159.478.149.438.756.2
TS-MOF(from Val Set)74.762.0--70.875.250.747.560.279.049.239.956.8

Table 1: Comparison of model performance on CIFAR100-LT benchmarks using an imbalanced validation set (from Train Set) and the original validation set.

W2. On the Supplementation of Ablation Studies:

Thank you for your suggestion to supplement the experimental section. We have conducted separate ablation experiments for R-PLA, RD-PCGrad, and the Expert Aggregation Module (EOSS). The following are the results.

R-PLARD-PCGradEOSSIR=10IR=50IR=100
HeadMediumTailAllHeadMediumTailAllHeadMediumTailAll
72.347.3--64.674.243.522.952.476.546.918.148.6
74.760.8--70.475.350.444.559.678.249.138.656.1
74.661.5--70.575.250.545.859.778.849.038.956.4
74.360.3--69.975.551.043.559.778.449.836.956.0
74.660.1--70.175.350.743.959.678.349.237.455.9
74.762.0--70.875.250.747.560.279.049.239.956.8

Table 2: Ablation experiments of TS-MOF's three modules with IR=10/50/100 on CIFAR100-LT benchmarks.

W3. Comparison with Multi-Expert SOTAs:

Thank you for your valuable suggestion. We have done our best to supplement several multi-expert baseline methods within the limited time, such as BalPoE, SADE, and MDCS. All comparisons will be cited and added to the main text. As seen from the table below, our TS-MOF method still demonstrates high performance under the same settings.

MethodIR=10IR=50IR=100
HeadMediumTailAllHeadMediumTailAllHeadMediumTailAll
BalPoE71.255.6--66.474.445.730.454.777.849.022.851.2
SADE72.349.5--64.675.042.210.549.978.749.013.248.7
MDCS71.951.5--65.675.944.534.355.578.747.928.853.0
TS-MOF74.762.0--70.875.250.747.560.279.049.239.956.8

Table 3: Comparison for CIFAR100-LT Benchmarks, showcasing the model performance of three multi-expert methods: BalPoE, SADE, and MDCS.

W4. On the Task Selection Mechanism:

Thank you for your question. The appendix aims to demonstrate that our framework can achieve performance improvements through rapid multi-objective fine-tuning across different strategy combinations. Our goal is not to perform a search for a single SOTA performance on the test set, but rather to showcase the general advantage of performance enhancement brought by the two-stage learning + multi-objective fine-tuning based on R-PLA and RD-PCGrad.

To demonstrate this general advantage, we avoided hyperparameter tuning as much as possible. All hyperparameters for the individual strategies themselves were kept identical to their settings in the original papers and public code. The hyperparameters specific to our method (such as those for the Stage 1 representation learning, R-PLA, and RD-PCGrad) were kept consistent across different strategy groups. We have also supplemented additional experiments on these hyperparameters, which you can refer to in Table 5.

W5 & Q1. Deterministic Task Ordering in RD-PCGrad:

In our implementation, the task order is determined by the sequence in which the strategies are passed. This deterministic approach ensures better reproducibility of our experimental results. We have also supplemented a performance comparison between RD-PCGrad and other potential multi-objective methods to illustrate its advantages; please see W6 for details due to space constraints.

W6. On Computational Overhead:

Compared to end-to-end multi-expert learning or multi-objective optimization methods, we only fine-tune for the last 20 epochs and only update the classifier head parameters. This reduces computational costs in real-world applications. Additionally, as per your suggestion, we have compared our method with other potential multi-objective approaches, with the results shown below. Macs represents the computational cost (in millions of operations), Time is the training time (in minutes), γ1\gamma_1 is the ratio of total trained parameters, and γ2\gamma_2 is the ratio of training time. The results indicate that the computational complexity and time of TS-MOF are roughly equivalent to ordinary methods, while delivering more efficient and accurate performance.

ModulesVariant Methodmacs/MTime/minγ1\gamma_1γ2\gamma_2γ1/γ2\gamma_1/\gamma_2Accuracy(%)
R-PLAentropy3.5408.451.0001.0010.99956.3
confidence3.5408.781.0000.9641.03853.5
uniform3.5408.441.0001.0020.99856.3
pla(ours)3.5408.461.0001.0001.00056.8
RD-PCGradmgda3.5408.361.0000.9061.10456.4
gradnorm3.5408.761.0000.8641.15755.2
uncertainty3.5407.461.0001.0150.98556.3
pcg(ours)3.5407.571.0001.0001.00056.8

Table 4: Comparison of computational cost and model performance between TS-MOF and conventional multi-task optimization in the second stage.

Q3. Explanation of Eq. 4:

Thank you for this sharp question. We clarify the R-PLA computation process here: for the same training batch, we first compute the performance vectors {a1,...,aN}\{\mathbf{a}_1, ..., \mathbf{a}_N\} for all N tasks through a single forward pass. After all vectors are computed, we then designate the last vector aN\mathbf{a}_N as the reference and compute the cosine similarity of the other vectors aj\mathbf{a}_j with it to determine the weights. Therefore, all performance information is available when calculating the weights, and there is no temporal conflict.

Q4. On Classifier Head Architecture:

In the main text, our model and classifier head architectures are consistent with prior works (e.g., LOS). For broader scenarios, such as when users wish to extend to more strategy combinations for rapid fine-tuning, the choice of model architecture and hyperparameters can be guided by our response in W4. We will add more detail on this in the revised manuscript.

Q5. On Introduced Hyperparameters:

The main discussion on hyperparameters is referenced in W4. Additionally, we have done our best to supplement discussions and experiments on other hyperparameters. We hope this answers your questions.

ModulesParameter NameParameter ValueManyMediumFewAccuracy(%)
R-PLApla_beta_clip_min0.178.2049.7437.8056.12
0.377.9149.6038.1056.06
0.578.2649.6038.4356.28
pla_beta_clip_max1.078.1748.8637.3755.67
2.078.2646.0013.7047.60
pla_similarity_threshold0.178.4048.4037.6055.66
0.378.4349.3437.1055.85
0.578.2048.7438.5756.00
pla_normalizationnone78.2049.2037.8355.94
softmax78.2350.0338.3756.40
RD-PCGradpcg_reductionsum78.1749.2938.5356.17
pcg_random_orderTrue78.2649.5735.5355.40
pcg_epsilon1e-0678.4348.9138.6356.16
1e-0478.1450.4038.7356.61
pcg_conflict_threshold-0.178.8351.6035.9356.43
0.178.3149.1736.2755.50
0.378.1449.5139.0356.39

Table 5: Hyperparameter analysis for the R-PLA and RD-PCGrad modules in the TS-MOF framework with IR=100 on CIFAR100-LT benchmarks.

评论

Dear Reviewer BSGz,

Thank you again for your time and for the insightful feedback you provided on our submission. We have now submitted a detailed rebuttal and, following your guidance, have conducted substantial new experiments to systematically address each of the key concerns you raised.

To facilitate your review, we've summarized our main efforts below:

  1. On the validation set dependency (W1, Q2): Addressing your concern about requiring a balanced validation set, we have added new experiments showing that our method remains robust and effective even when using an imbalanced validation set.
  2. On the key ablation studies (W2): We have now provided the complete ablation studies you requested, clearly quantifying the individual contributions of our core components like R-PLA and RD-PCGrad.
  3. On the comparison with SOTA models (W3): We now include a direct comparison with the relevant state-of-the-art models you mentioned (e.g., BalPoE, SADE), with results demonstrating the competitiveness of our method.
  4. On the computational overhead (W6): In response to your question about computational costs, we have provided a quantitative analysis showing that our method achieves superior performance with overhead comparable to conventional approaches.
  5. On clarifying methodological details (W4, W5, Q1, Q3, Q5): We have offered detailed explanations for the rationale behind our task selection, the motivation for deterministic ordering in RD-PCGrad, the R-PLA computation process, and hyperparameter sensitivity.

Should you have any further questions or require additional clarifications, we would be very happy to address them to the best of our ability within the remaining discussion period.

Thank you once again for your professional guidance and valuable time.

评论

I’d like to thank the authors for addressing my concerns and clarifying several key points: their method does not strictly depend on a balanced holdout dataset and remains robust with imbalanced validation data with a reasonable trade-off in performance; comprehensive ablation studies confirm the need of each component; the extended comparisons show it outperforms recent expert-based baselines; it maintains reasonable computational overhead and exhibits relatively low hyperparameter sensitivity; and the authors have clearly explained their task selection and ordering decisions. These collective improvements have convinced me to raise my score.

评论

Dear Reviewer BSGz, Thank you very much for your thoughtful follow-up and for taking the time to carefully review our rebuttal. We are very encouraged to hear that our responses and new experiments have addressed your concerns. We are especially grateful for your decision to raise the score. Your constructive feedback has been invaluable in strengthening our paper. Thank you again for your time and support.

审稿意见
6

This paper proposes TS-MOF, a two-stage multi-objective fine-tuning framework for long-tailed recognition. The approach integrates multiple loss functions via a multi-head architecture and resolves gradient conflicts using a modified version of PCGrad (RD-PCGrad), while dynamically reweighting tasks based on a refined performance-level agreement mechanism (R-PLA). Experiments on several long-tailed benchmarks (CIFAR10, ImageNet- iNaturalist) demonstrate consistent improvements over competitive baselines.

优缺点分析

Strengths

  1. The proposed two-stage framework is well-motivated and aligns with prior findings that feature decoupling improve long-tailed generalization;
  2. RD-PCGrad introduces a deterministic and task-aware projection strategy, offering a practical solution to inter-task gradient interference;
  3. R-PLA dynamically adjusts task importance in a label- and batch-aware manner, enabling adaptive optimization focus across head and tail classes;
  4. The method shows consistent gains across CIFAR100-LT, ImageNet-LT, and iNaturalist, with a relatively lightweight fine-tuning phase.

Weaknesses

  1. While Prop. 1 states that the final gradient gfinalg_{final} is Pareto-improving or non-worsening, the paper does not analyze how far this direction can deviate from the true Pareto front in the proof. In complex objective landscapes, the projection step may inadvertently suppress beneficial gradient components, and a discussion of such trade-offs is missing;
  2. The use of the cube root in Eq. (5) is not well justified. Based on the proof of Prop. 2, any concave, increasing function could preserve the robustness properties. The choice of the cube root lacks explanation on whether it confers specific benefits (e.g., smoother gradients or empirical stability);
  3. Although the proposed method is evaluated in a two-stage training pipeline, the core techniques (RD-PCGrad and R-PLA) are not inherently limited to such a setup. Gradient conflict resolution might be even more critical in single-stage end-to-end training. The paper would benefit from a more apparent justification for the choice of a two-stage framework or a discussion of broader applicability.

问题

Please see the weaknesses.

局限性

The authors have adequately addressed the limitations, and no potential negative impact is found.

最终评判理由

The concerns are well-addressed. The motivation is well-considered, balancing theoretical justification with empirical support and practical utility. Therefore, I suggest accepting this paper.

格式问题

N/A

作者回复

Response to Reviewer LH8y

Thank you, Reviewer LH8y, for your recognition of our work and for your constructive feedback. We address your questions point by point below.

1. On the Deviation of RD-PCGrad from the True Pareto Front:

Thank you for this insightful question. Our method ensures a Pareto-improving direction for stability, which involves a deliberate trade-off. We formally analyze this below.

Formal Analysis of the Trade-off

Given two conflicting gradients gAg_A and gBg_B (i.e., gATgB<0g_A^T g_B < 0), RD-PCGrad modifies gAg_A by removing its conflicting component pAB=projgB(gA)p_{A \to B} = \text{proj}_{g_B}(g_A).

a. Suppressed Benefit & Cost: The suppressed component pABp_{A \to B} is indeed beneficial for minimizing its own loss LAL_A. The magnitude of this benefit is proportional to: (gATgB)2/gB20(g_A^T g_B)^2 / \|g_B\|^2 \ge 0 This is the cost of our projection. However, this component is equally harmful to task LBL_B. By removing it, we prioritize the "Do No Harm" principle, which is essential for preventing the "seesaw effect" in LTR and ensuring stable, balanced progress.

b. Deviation Analysis: The deviation from task A's optimal direction is proportional to the relative magnitude of the suppressed component, which can be expressed as: ρ=pAB/gA=cos(θAB)\rho = \|p_{A \to B}\| / \|g_A\| = |\cos(\theta_{AB})| This shows the deviation is directly tied to the conflict angle θAB\theta_{AB} between the gradients. In LTR, where conflicts between head- and tail-focused strategies are inherent, this deviation is a necessary compromise. Our goal is a balanced solution, not the extreme optimization of any single strategy. Thus, this shift towards balance is a feature, not a bug, for solving the long-tail problem.

2. On the Rationale for Using a Cube Root in Equation (5):

Thank you for this insightful question. We acknowledge that, in theory, any monotonically increasing concave function can ensure robustness. We chose the cube root based on the specific advantages it offers in terms of weight dynamics and gradient smoothness, which collectively enhance the model's empirical stability.

First, compared to other concave functions like the square root, the cube root provides a better trade-off in sensitivity. It is more sensitive in the low-similarity region (when strategy performance differs greatly), providing a stronger "incentive" signal for underperforming strategies to catch up. Concurrently, it is less sensitive in the high-similarity region (when strategies perform consistently), providing greater stability and preventing unnecessary adjustments to already well-performing strategies.

Second, the cube root leads to smoother gradient flow. Since similarity is a function of model parameters, the weighting function introduces an additional meta-optimization gradient during backpropagation. The derivative properties of the cube root make this gradient flow smoother than that of other concave functions (like the square root), especially in the extreme case where similarity approaches zero. This effectively reduces the risk of exploding gradients.

We will also add an ablation study in the appendix comparing different weighting functions to empirically support our choice.


3. On the Motivation for the Two-Stage Framework:

Thank you for this insightful question. We chose the two-stage framework deliberately because it offers significant advantages for our MOO components when tackling the long-tail problem:

  1. Ultimate Efficiency: Compared to end-to-end training, our method only applies multi-objective optimization to the lightweight classifier heads during a short final fine-tuning stage, drastically reducing computational cost and time.
  2. Feature Protection: By freezing the backbone, we protect the valuable general-purpose features learned in the first stage from being corrupted by strong re-balancing signals, thus preserving performance on head and medium classes.
  3. Training Stability: Confining the complex gradient conflict resolution to the low-parameter classifier heads greatly reduces the risk of training instability, making the optimization process more robust and controllable.
  4. Plug-and-Play Modularity: Our framework allows researchers to easily "plug and play" any existing long-tail strategy without complex end-to-end adaptation, significantly improving the method's ease of use and extensibility.

Therefore, the combination of the two-stage framework and our proposed strategies enables us to achieve an efficient, stable, and low-cost fusion of multiple long-tail strategies.

评论

Thank you for the thoughtful and comprehensive rebuttal. Your responses have fully addressed all my concerns. I appreciate that the motivation for your work is well-considered, balancing theoretical justification with empirical support and practical utility. My assessment remains positive.

最终决定

This paper presents TS-MOF, a two-stage long-tailed recognition framework. Stage 1 learns features; Stage 2 freezes the backbone and fine-tunes multiple classifier heads with multi-objective optimization. Two new components, R-PLA (adaptive task weighting) and RD-PCGrad (deterministic gradient conflict resolution), coordinate several LTR strategies. The method shows strong results on CIFAR100-LT, ImageNet-LT, and iNaturalist.

Reviewers highlight consistent gains across standard benchmarks, a clear motivation for decoupling feature learning from classifier adaptation, and practical components (R-PLA and RD-PCGrad) that make multiple LTR heads work together. One review rates the paper accept with excellent quality and significance. Another notes useful theoretical analysis around gradient stability and robustness. The fine-tuning phase is lightweight relative to training from scratch, which improves usability.

The negative review centers on presentation and missing detail rather than correctness: the two-stage scaffold is not new, the Introduction underplays that the real advances are R-PLA and RD-PCGrad, and the paper would benefit from schematic diagrams for both modules. Important ablations are thin: isolate R-PLA, RD-PCGrad, and the aggregation module; compare against alternative task-weighting and gradient-projection baselines; report sensitivity of key hyperparameters. After rebuttal, the reviewer maintained a borderline-reject due to the scope of revisions (writing logic, diagrams, ablations, method details, expert comparisons). Other reviews remain positive, emphasizing strong results and practical value of R-PLA and RD-PCGrad.

On balance, the empirical evidence across three benchmarks and multiple imbalance ratios, together with two concrete, generally useful MOO techniques, makes the strengths outweigh the weaknesses. The main concerns are fixable with added analysis, clearer writing, and targeted ablations, rather than fundamental flaws. I therefore recommend acceptance, contingent on addressing the items discussed during rebuttal in the camera-ready.