PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
5
4
4
4
3.3
置信度
创新性3.0
质量3.5
清晰度3.3
重要性3.0
NeurIPS 2025

Rethinking Fine-Tuning when Scaling Test-Time Compute: Limiting Confidence Improves Mathematical Reasoning

OpenReviewPDF
提交: 2025-05-07更新: 2025-10-29
TL;DR

We show that limiting a model's confidence during training can improve test-time scaling in mathematical reasoning.

摘要

关键词
Test-time computeinference-time computecoveragepass@Nreasoninglarge language modelformal mathfinetuningoverfittingoverconfidence

评审与讨论

审稿意见
5

This paper investigates an misalignment between the standard fine-tuning objective for large language models (cross-entropy loss) and the widely used test-time strategy of pass@N sampling. The authors demonstrate, both theoretically and empirically that training with cross-entropy loss can lead to model overconfidence. This overconfidence, while improving pass@1 accuracy somewhat paradoxically harms performance when sampling a larger number of solutions (pass@N for large N), as the model becomes less exploratory and repeatedly generates the same incorrect high-confidence answers. To address this, the paper introduces Direct Coverage Optimization, a loss function that directly optimizes the pass@N metric during training. By acting as a confidence regularizer, DCO prevents the model from becoming overconfident in its solutions. The authors show that DCO not only rescues the degrading pass@N performance but also establishes a Pareto-optimal frontier, improving results on mathematical reasoning benchmarks like MATH and theorem-proving benchmarks like MiniF2F, for both direct-answer and chain-of-thought settings.

优缺点分析

Strengths

  1. The paper addresses a highly relevant and overlooked problem in LLM training for test-time scaling. The core finding that more training via standard CE loss can actively harm performance for certain inference strategies is somewhat counter-intuitive at first but makes sense on further consideration. The paper compellingly argues for a tighter co-design of training and inference protocols.

  2. The paper is exceptionally well-written and clear. The authors do an excellent job of building intuition for the problem, starting with a simple two-problem example and generalizing with theoretical lemmas (4.2, 4.3). The proposed solution, DCO is elegant and directly motivated by the identified issue of overconfidence. The connection between the DCO gradient and the "overconfidence regularizer" (Figure 2b) provides a clear and satisfying explanation of why the method works.

  3. The empirical validation is good. The authors test their hypothesis across multiple scenarios including direct answers, chain-of-thought reasoning on the MATH dataset, and automated theorem proving on LeanDojo. The inclusion of comparisons to both standard CE loss and a reinforcement learning baseline (GRPO) is good. The visualizations of confidence distributions (Figure 1) and performance frontiers (Fig 2) are effective and strongly support the paper's claims.

Weaknesses

  1. The central weakness of the paper is the question of whether the core findings and the effectiveness of DCO hold at the scale of state-of-the-art models. While the authors include results for a Llama-3-70B model in the appendix (Section C.3, Figures 7 and 8) even this is not near the scale of state-of-the-art. The bulk of the results, figures, and discussion are based on the 8B model. I wonder as models become significantly larger and more capable, does this overconfidence issue persist with the same severity? It is plausible that larger models might inherently have better-calibrated uncertainty, potentially diminishing the performance gap between CE and DCO.

  2. The entire premise of the pass@N strategy and the DCO objective relies on the existence of a binary verifier to determine the correctness of a generated sample. The authors acknowledge this limitation in their discussion. However, this is a major constraint that limits the applicability of the proposed method to specific domains like math, coding, and formal theorem proving. The approach is not directly usable for more open-ended tasks.

问题

  1. The results for Llama-3-70B in the appendix are encouraging as they show a similar trend to the 8B model. Could you provide more insight into the magnitude of this effect at scale? Does the performance gap between DCO and CE loss widen, shrink, or remain constant when moving from 8B to 70B? To strengthen the paper's claim of generalizability, I strongly encourage you to integrate the 70B results more centrally into the main paper, perhaps directly compare the confidence distributions and performance frontiers across scales.

  2. The results show that the optimal DCO hyperparameter N' is closely tied to the test-time pass@ N. This suggests that one might need to train multiple models to be prepared for different inference-time compute budgets. What is your recommendation if the exact test-time N is not known in advance? Is there a "sweet spot" N' that provides robust performance across a wide range of test N values? Or is ensembling models trained with different N' values (as you did for theorem proving) the most practical approach?

  3. While the work focuses on domains with clear verifiers, have you considered how DCO might be adapted for tasks where verification is "soft," such as using a reward model's score? How might the DCO loss function behave if the correctness signal p(y|x) is not binary (0 or 1) but a continuous score from a reward model?

局限性

yes

最终评判理由

Overall this is a good paper addressing the misalignment between cross-entropy based training and pass@N metrics during inference. The rebuttal addressed my concerns regarding the choice of N' and how the method can be used with soft verification.

格式问题

None

作者回复

We sincerely thank Reviewer j14W for the thorough and insightful review. We appreciate that the reviewer found the paper to be exceptionally well-written, the DCO formulation elegant, and the empirical validation strong and thorough. Below, we directly address the reviewer’s important questions and provide additional clarification and evidence to strengthen our contributions.

Overconfidence Severity at Scale (Larger Models)

We thank the reviewer for raising this important problem. Indeed, one of our motivations for investigating the Llama-3-70B model was precisely to explore whether the severity of overconfidence diminishes or intensifies as models grow larger and more capable. Our empirical findings indicate that overconfidence may become more severe at larger scales: For Llama-3-8B model, the maximum improvement of DCO frontier over CE loss happens at N=541N=541 with an improvement of 26.326.3%, while for Llama-3-70B model, the maximum improvement over CE loss occurs at N=258N=258 with a larger improvement of 29.329.3%. Interestingly, the performance gap between DCO and CE not only persists but grows slightly larger at the 70B scale. We suspect that the larger model, because it has much more parameters, may be more prone to overfitting and thus exhibit greater overconfidence under CE training. We acknowledge that this observation is based specifically on the Llama-3 series, and broader validation across more model families would further strengthen generalizability. To clarify, we are not suggesting that larger models inherently exhibit overconfidence without fine-tuning; rather, we aim to highlight a critical failure mode resulting from the misalignment between CE training and pass@N evaluation. We believe this phenomenon of training-test misalignment and the resulting overconfidence is likely a general issue.

We also recognize that the pass@1 performance of the model on the benchmark is not near state-of-the-art, because we use the base model instead of an instruction-tuned model. We intentionally selected base models for two reasons: (1) to avoid potential data contamination or interference from post-training and (2) because lower pass@1 accuracy in the base models provides more meaningful room for improvement to demonstrate the effectiveness of scaling test-time compute.

We will integrate these key results and comparisons in the revised paper to better address scalability concerns.

Robust choices of N’ under unknown test-time compute vs. ensembling

We appreciate the reviewer’s observation that the test-time NN might not be known at training time. To address the possibility of having “sweet-spot” choices of NN’ for training, we evaluate the robustness of DCO to mismatches between NN′ (the DCO objective parameter) and NN (the NN used to evaluate pass@N at test time). We define a choice of NN’ to be optimal for a given test-time NN, if its pass@N performance is within 1% of the DCO frontier at that NN (see Figure 2a for the frontier and individual DCO curves). Based on experiments fine-tuning Llama-3-8B on direct-answer generation for the MATH dataset, we observe:

  • N=1N’=1 is optimal for 1N31\leq N\leq 3.
  • N=16N’=16 is optimal between 2N202\leq N\leq 20.
  • N=256N’=256 is optimal between 180N385180\leq N\leq 385.
  • N=4096N’=4096 is optimal for N1151N\geq 1151.

These results indicate that while matching N=NN’=N is a standard and intuitive approach, a fixed choice of NN’ can result in near-optimal performance for a range of test-time NN. Selecting a few strategic NN’ values can thus yield near-optimal performance across broad ranges.

We also believe that ensemble may be beneficial for certain use cases such as theorem proving. As highlighted by the reviewer, an ensemble of models trained with different NeffN_{eff} values can actually surpass the performance of a single model when using larger NN in theorem proving. Thus, training multiple specialized models can be an advantageous strategy rather than purely a limitation. Of course, both the robustness of DCO to the test-time NN and the benefits due to ensembling models at various choices of NN’ are likely task-dependent.

Regarding the question of whether a single model could effectively handle multiple inference budgets at test time, we would like to suggest a potential strategy: explicitly conditioning the model on the target sampling budget NN during training and testing (i. e. learning the distribution of p(yx,N)p(y|x,N)), thus enabling a single model to flexibly adapt to various inference budgets. While exploring this direction in depth remains beyond the scope of our current work, we acknowledge it as a valuable direction for future research.

Settings with “soft” verification

We agree with the reviewer regarding the importance of exploring extensions of DCO to domains where verification is not strictly binary but rather “soft”. We acknowledge that restricting the scope to verifiable problems is one of the limitations as we stated at the end of our paper. Extending DCO to handle continuous verification signals represents a promising direction for future research. Given the substantial density and content of our current manuscript, we leave such extensions to future work and hope the reviewer could understand.

We want to thank the reviewer again for their thoughtful comments and constructive suggestions. We remain eager to address any additional questions or concerns the reviewer might have.

评论

Dear Reviewer j14W,

We hope this message finds you well! We sincerely thank you for your thorough and positive evaluation of our submission. We are especially grateful for your recognition of the clarity of our presentation, the elegance of the proposed DCO method, and the strength of our empirical validation.

We wanted to kindly follow up regarding our responses to your important questions about the scalability of the identified overconfidence issue and the robustness of our approach across different inference-time budgets. In response to your insightful feedback, we have further highlighted results for the Llama-3-70B model, demonstrating that the issue of overconfidence persists or even becomes more pronounced at larger scales, reinforcing our conclusions. Additionally, we have discussed practical considerations for selecting optimal NN' values and potential strategies, such as model ensembling or conditioning on inference budgets, to manage uncertainty about the exact inference-time compute available.

If there remain any unresolved points or if further clarifications are needed, please let us know! We would be delighted to provide additional details.

Thank you again for your valuable insights, thoughtful questions, and constructive suggestions, which have greatly contributed to improving our paper.

评论

Dear authors,

Thank you for the detailed response - you have clarified the questions I had. The results regarding N' are interesting and it would be useful to include in the paper.

Regards, j14W

评论

Dear Reviewer j14W,

Thank you very much for your kind response and helpful feedback. We are delighted to hear that our clarifications addressed your questions effectively. We appreciate your suggestion regarding the results on NN', and we will certainly incorporate these findings into the revised manuscript to further strengthen our work.

Once again, thank you for your thoughtful insights and valuable contributions to improving our paper.

审稿意见
4

The paper investigates the misalignment between cross-entropy (CE) fine-tuning and test-time compute scaling (e.g., pass@N). It shows that prolonged CE training induces overconfidence, degrading pass@N performance despite improving pass@1. The authors attribute this to a tradeoff between exploitation (high confidence) and exploration (diverse sampling). To realign training with pass@N, they propose Direct Coverage Optimization (DCO), a loss that limits model confidence via a regularization term attenuating gradients for high-confidence predictions. Experiments on MATH (direct answers/CoT) and LeanDojo (theorem proving) demonstrate that DCO rescues pass@N scaling and achieves Pareto-optimal frontiers. Extensions to step-wise DCO for proof trees and approximate DCO for CoT further validate the framework.

优缺点分析

Strength

  • Novel Insight: Identifies a counterintuitive phenomenon—longer CE training harms pass@N performance—and traces it to overconfidence, supported by theory (Lemmas 4.2–4.3) and empirical validation (Table 1, Figure 1).

  • Practical Impact: The ensemble of step-wise DCO models (Table 2) boosts theorem-proving success by +5.2% (MathLib) and +4.1% (MiniF2F) over CE with matched compute, highlighting real-world utility.

  • Theoretical Depth: Lemmas 4.2–4.3 formally characterize the confidence-exploration tradeoff, and Appendix A provides complete proofs.

Weakness

  • Approximate DCO for CoT requires online Monte Carlo estimation, increasing compute by 10× vs. standard fine-tuning.

  • The paper links overconfidence to reduced pass@N performance but does not quantify how DCO affects output diversity.

问题

See the Weakness.

局限性

yes

最终评判理由

The detailed response solves my concern.

格式问题

N/A.

作者回复

We thank the reviewer for their thorough review and insightful feedback and suggestions. We are pleased that the reviewer recognizes the novelty and practical relevance of our findings and appreciates the depth of our theoretical analysis. Below, we directly address the reviewer's concerns and provide additional evidence and clarifications.

The effect of DCO on output diversity

We thank the reviewer for suggesting this point. We agree with the reviewer that the effect of the DCO objective on output diversity is important to understand. To quantify this, we estimated the mean Shannon entropy of the model outputs using Monte Carlo sampling with 4096 samples per problem. Specifically, we fine-tuned a Llama-3-8B-base model on the MATH dataset to generate direct answers using DCO losses with different values of N’ and compute the mean Shannon entropy of the model distribution on the test set of MATH and AIME24. We will explicitly incorporate these diversity results into our revised paper. We summarize the results in the table below:

Table 1. Mean and standard deviation of Shannon entropy for models trained with DCO of varying N’ values.

DatasetN=1N’=1 (CE loss)N=8N’=8N=64N’=64N=512N’=512N=4096N’=4096
MATH2.27±1.432.27\pm1.433.23±1.163.23\pm1.164.22±0.914.22\pm0.915.52±0.715.52\pm0.716.71±0.586.71\pm0.58
AIME242.74±1.272.74\pm1.274.11±1.174.11\pm1.174.39±0.914.39\pm0.915.89±0.415.89\pm0.417.10±0.227.10\pm0.22

The table clearly demonstrates that the entropy consistently increases with larger NN’. This indicates that models trained with higher NN’ indeed produce more diverse outputs, thus directly confirming that DCO naturally encourages diversity. Intuitively, superior pass@N performance at larger NN requires distributing probability mass across multiple candidate solutions, reducing overconfidence and enhancing coverage.

Computational overhead in the CoT setting

We agree with the reviewer regarding the significant computational overhead involved in Monte Carlo estimation in CoT settings. While acknowledging this limitation, we briefly note here that, although this overhead is significant relative to standard SFT, costly online inference is also required in RL frameworks for CoT training, as discussed in Appendix B. Importantly, practitioners can significantly reduce training costs by using fewer Monte Carlo samples, trading precision in estimating success probability for increased computational efficiency.

We thank the review again for their thoughtful questions and constructive feedback. We hope our detailed responses address all concerns, and we remain available to provide further clarifications if needed.

评论

Dear Reviewer 52g7,

We hope this message finds you well! Thank you very much for your detailed and constructive feedback on our submission. We are pleased to hear that you found our theoretical insights novel and our practical results impactful.

We wanted to kindly follow up regarding our responses to your comments, specifically the quantification of output diversity under the DCO objective as per your suggestion. We greatly appreciate this suggestion as it helps better demonstrate that DCO diversifies the output distribution. As described in our rebuttal, we conducted additional analyses measuring the mean Shannon entropy, confirming that DCO indeed enhances output diversity. We also acknowledged and discussed strategies for managing the computational cost associated with Monte Carlo estimations in CoT scenarios. We will make sure to incorporate these points into our revised manuscript.

Please let us know if there remain any unresolved points or if you require further clarification. We would be more than happy to provide additional information.

We would greatly appreciate it if you could revisit your evaluation in light of our responses and possibly update your score to reflect these improvements.

Thank you once again for your thoughtful and valuable review, which significantly contributes to strengthening our paper.

评论

Thank the authors for the detailed response which solves my concern. I keep my positive score.

评论

Thank you very much for your time and thoughtful engagement. We're glad to hear that our response has addressed your concern. We truly appreciate your positive evaluation and hope you will continue to support the acceptance of our paper, as the concerns raised have been resolved.

审稿意见
4

This paper focus on the problem that after few epochs of training, while the pass@1 result of model increases, pass@N decreases. It claims a mismatch between how we train LLMs (with cross-entropy loss that optimizes single best output) and how we test them (by pass@N), and the standard fine-tuning makes the model too confident at one guess and hurts performance when we sample many times. They propose a new loss objective, Direct Coverage Optimization (DCO), which directly maximizes the chance of getting at least one correct answer in N samples, showing gains in pass@N on math problems and theorem proving benchmarks.

优缺点分析

Strengths:

  • The paper provides a insightful observation and points out a real gap between training and inference and explains why it matters, with clear proofs show why cross-entropy leads to overconfidence and how DCO fixes it.
  • The performance of DCO and its variants are good on MATH, AIME, MiniF2F, and Mathlib, cross multiple domains.
  • The method applies to both step-wise proof search and chain-of-thought, showing robustness.

Weaknesses:

  • Training with DCO may need more sampling and GPU time, which is acceptable.

问题

In the paper, the authors train with N’ = N to match the test-time sampling budget. But retraining a separate model for every target N can be very costly. How robust is a model trained for one N’ when tested at different budgets (e.g., training at N’=16 but testing at N=64 or N=4)? Have you explored stragety to support multiple budgets without full retraining for each?

局限性

yes

格式问题

No

作者回复

We sincerely thank Reviewer asva for their thorough and insightful feedback. We are pleased that the reviewer found our identification of the gap between training and inference insightful, and recognized the robustness and strong empirical performance of our proposed DCO. Below, we address the reviewer's main concerns and questions directly, providing additional clarifications and evidence to strengthen our contribution.

Robustness of DCO

We appreciate the reviewer’s observation that training a separate model for each choice of NN' is computationally expensive. Indeed, the central theme for this paper is that we cannot simply train a single model to be optimal for every NN without conditioning on NN at test time.

The reviewer raises a question about the robustness of DCO to mismatches between NN' (the DCO objective parameter) and NN (the NN used to evaluate pass@N at test time). We would like to remind the reviewer that there is an example plot in Figure 6, where we highlight and show how robust a model trained with a specific N=256N=256 is with respect to the Pareto frontier. To further investigate the robustness, we define an NN' as optimal for a given NN if its pass@N performance is within 1% of the DCO frontier at that NN (see Figure 2a). Based on experiments fine-tuning Llama-3-8B on direct-answer generation for the MATH dataset, we observe:

  • N=1N’=1 is optimal for 1N31\leq N\leq 3.
  • N=16N’=16 is optimal between 2N202\leq N\leq 20.
  • N=256N’=256 is optimal between 180N385180\leq N\leq 385.
  • N=4096N’=4096 is optimal for N1151N\geq 1151.

These results strongly indicate that while matching N=NN'=N is a standard and intuitive approach, a fixed choice of NN' can result in near-frontier performance for a range of test-time NN. Selecting a few strategic NN' values can yield near-optimal performance across broad ranges, reducing computational overhead.

Moreover, as demonstrated in our theorem proving results, an ensemble of models trained with different NN' values can actually surpass the performance of a single model using larger NN. Thus, training multiple models and forming an ensemble can be an advantageous strategy rather than purely a limitation.

Strategies for Supporting Multiple Test-time Budgets

The reviewer raises an important question regarding potential strategies to support multiple test-time budgets without complete retraining. Indeed, a promising approach is to explicitly condition the model on NN during training and testing (i. e. p(yx,N)p(y|x,N)), enabling a single model to flexibly adapt to various inference budgets. While exploring this direction in depth remains beyond the scope of our current work, we acknowledge it as a valuable direction for future research.

Computational overhead in the CoT setting

We agree with the reviewer regarding the significant computational overhead involved in Monte Carlo estimation in CoT settings. While acknowledging this limitation, we briefly note here that, although this overhead is significant relative to standard SFT, costly online inference is also required in RL frameworks for CoT training, as discussed in Appendix B. Importantly, practitioners can significantly reduce training costs by using fewer Monte Carlo samples (NMCN_{MC}), trading precision in estimating success probability for increased computational efficiency.

We thank the reviewer again for their valuable suggestions and insightful questions, which help improve our paper substantially. We hope our responses clearly address their concerns and further strengthen their confidence in our findings. Please don’t hesitate to reach out if anything remains unclear or if additional questions arise.

评论

Dear Reviewer asva,

We hope you are doing well! We greatly appreciate your insightful feedback and positive evaluation of our submission, especially your recognition of the identified gap between training and inference and the robustness of our proposed DCO method.

We wanted to kindly follow up to ensure that our rebuttal has fully addressed your questions, particularly regarding the robustness of DCO to mismatched training and testing sampling budgets and strategies to support multiple test-time budgets without extensive retraining. As outlined in our response, we've demonstrated that a carefully selected fixed NN' can achieve near-frontier performance across a range of test-time NN. We also acknowledge and propose explicitly conditioning the model on NN as an exciting avenue for future work.

If there remain any unresolved points or further clarifications you might require, please let us know. We would be very happy to provide additional details.

We would greatly appreciate it if you could revisit your evaluation in light of our responses and potentially update your score to reflect the improvements we've incorporated.

Thank you once again for your valuable insights and consideration, which substantially contribute to the quality and clarity of our paper.

评论

Thanks for your clarification and it makes sense. I will keep my rating.

审稿意见
4

The paper investigates a subtle but important mismatch between how LLMs are fine-tuned and how they are later used with heavy test-time sampling (e.g. pass@N). The authors first show that training with standard cross-entropy (CE) pushes the model toward over-confident, low-diversity distributions that hurt coverage when many samples are drawn. On MATH, for instance, pass@256 accuracy falls from 65% after one epoch to 45% after four epochs despite pass@1 steadily improving. They trace the effect to a theory-backed exploit/explore trade-off and introduce Direct Coverage Optimization (DCO), a simple loss that directly maximizes the probability a correct answer appears in N samples while automatically regularizing over-confidence. Results on math are quite promising.

优缺点分析

Strengths:

  • Solid theoretical analysis explaining why CE induces over-confidence;
  • Writing is clear; figures/tables make trade-offs visually clear
  • Method is simple and easy to implement
  • Topic very relevant today

Weakness:

  • Related work on inference-aware fine-tuning (e.g. BonBon, BOND) is acknowledged but could be contrasted more deeply, especially on diversity vs. confidence control
  • There are many previous loss functions that try to regularize cross entropy to avoid over confidence. Focal loss, which is commonly used in computer vision, is one such example. Could you discuss (or maybe even compare to) those losses?

问题

  • Regarding N: Could you provide a rule-of-thumb (perhaps tied to available inference budget) to pick these hyper-parameters without sweeping? A calibration plot would make adoption easier.
  • The epsilon part is a bit sketchy. Is it really necessary? What's the rule of thumb for that?
  • As sequence length gets longer, the prob of the sequence naturally gets lower. How does that affect things?

局限性

yes

格式问题

no

作者回复

We sincerely thank the reviewer for their thorough and insightful comments. We are encouraged that the reviewer found the theoretical analysis solid, our writing clear, and our topic timely and relevant. Below, we directly address the reviewer's concerns and questions, with additional empirical results and clarifications to strengthen our contributions.

Comparison to Focal Loss

The reviewer correctly notes that various works have also proposed loss functions aiming to reduce overconfidence in the cross entropy (CE) loss. However, we would like to emphasize that the overconfidence we identified arises specifically from a train-test mismatch in sampling strategies. While there are other methods/loss functions to alleviate overconfidence, our DCO loss is directly designed to address this misalignment. To concretely demonstrate this difference, we followed the reviewer’s suggestion and conducted a thorough comparison with Focal Loss (FL) [Lin et al., 2017], FLγ(x,y)=(1p^(yx))γlogp^(yx)\ell^\gamma_\text{FL}(x, y) = -(1-\hat{p}(y \mid x))^\gamma \log \hat{p}(y \mid x), optimizing over its hyperparameter γ\gamma for each value of NN. Our results are as follows:

Table 1. We fine-tune a Llama-3-8B-base model on the MATH dataset to generate direct answers with CE loss, FL with different γ\gamma and the DCO losses with different NN, and test the model on the test set of MATH. We have reported pass@N performance for the CE baseline and DCO frontier (as plotted in Figure 2a). Here, we include the FL frontier, which selects the best performing γ\gamma for each NN. The FL frontier consistently outperforms the CE baseline but underperforms the DCO frontier across all values of NN. This further highlights the advantage of DCO in improving coverage under heavy sampling.

N=1N=8N=64N=512N=1024N=2048N=4096
CE0.070.180.330.490.540.590.63
Focal Loss frontier0.070.220.440.630.680.720.75
DCO frontier0.070.240.540.750.800.840.87

Table 2. Same as Table 1, but test on the out-of-distribution AIME24.

N=1N=8N=64N=512N=1024N=2048N=4096
CE0.020.040.090.190.230.290.36
Focal Loss frontier0.020.040.120.240.330.450.55
DCO frontier0.020.040.110.480.640.790.89

These results clearly show that while heuristically designed objectives like FL can indeed outperform vanilla CE loss, DCO loss consistently achieves significantly superior performance, particularly at larger NN values critical for practical applications. We will include these comparisons explicitly in the revised paper. Specifically, we will add a corresponding plot illustrating that the FL frontier consistently lies above the CE baseline but below the DCO frontier across nearly all values of NN on both in-distribution MATH test set and out-of-distribution AIME24.

We fully agree that providing clear guidelines for choosing NN’ greatly enhances the practicality of our approach, given that training a separate model for each choice of NN' is computationally expensive. The main argument of this paper is that one should align the training objective with the test-time strategy. That said, we study the calibration of the DCO objective and find that DCO can be robust to some extent to mismatches between NN' (the DCO objective parameter) and NN (the NN used to evaluate pass@N at test time). We define an NN' as optimal for a given NN if its pass@N performance is within 1% of the DCO frontier at that NN (see Figure 2a). Based on experiments fine-tuning Llama-3-8B on direct-answer generation for the MATH dataset, we observe:

  • N=1N’=1 is optimal for 1N31\leq N\leq 3.
  • N=16N’=16 is optimal between 2N202\leq N\leq 20.
  • N=256N’=256 is optimal between 180N385180\leq N\leq 385.
  • N=4096N’=4096 is optimal for N1151N\geq 1151.

We also remind the reviewer that we proposed a rule of thumb for selecting NeffN_{\text{eff}} in the case of theorem proving (see the paragraph starting at line 269). If one has an estimate of task difficulty (e.g., proof length kk), we suggest using Neff=N1/kN_{\text{eff}} = N^{-1/k}, where NN is the inference-time sampling budget. Additionally, if an inference-time budget (in FLOPs or wall-clock time) is available, a practical strategy is to estimate the distribution of cost per sample and then choose the largest NN' such that the expected total cost remains within budget.

Clarification of the role of ϵ\epsilon in the lemmas

We thank the reviewer for raising this important point. We clarify that ϵ\epsilon captures uncertainty regarding exact probability values for lower-ranked terms. The aim of Lemma 4.2 is to establish an upper bound for the model confidence p1^\hat{p_1}* associated with the highest-ranked answer under the optimal policy that maximizes coverage. The precise value of p1^\hat{p_1}* inherently depends on all probabilities pip_{i}. The introduction of the ϵ\epsilon term explicitly acknowledges uncertainty regarding the exact values of probabilities pip_{i} for indices i>ki>k, capturing our incomplete knowledge of these lower-ranked terms. Thus, ϵ\epsilon is a theoretically motivated component ensuring robustness rather than a practically burdensome hyperparameter.

Comparison to other inference-aware tuning methods

We appreciate the suggestion to expand our discussion regarding BonBon and BOND. Both methods elegantly address the problem of distilling the Best-of-NN (BoN) policy, thereby avoiding the cost of NN inference passes at test-time, closely related to RLHF objectives. However, their work focuses on improving the pass@1 performance from the BoN policy, while we focus on directly improving the pass@N performance. We will incorporate these distinctions clearly in our revised related-work section, highlighting the complementary contexts in which each method applies.

Effect of sequence length on the probability

We recognize the reviewer’s concern regarding longer sequences naturally receiving lower probabilities. Importantly, our method is designed such that sequence length does not inherently penalize coverage: In direct-answer tasks, DCO optimizes directly for the correct final answer, independent of sequence length. For formal math, we apply DCO at the step-level granularity, avoiding any length-related bias. In Chain-of-Thought scenarios, our method marginalizes over reasoning traces, ensuring no direct penalization of longer sequences. Thus, sequence length inherently does not disadvantage our proposed approach.

We greatly appreciate Reviewer 4Nd6's valuable feedback, which significantly strengthens our paper. We hope our detailed clarifications and additional experiments address all concerns and increase the reviewer’s confidence in our work. We would be happy to clarify any points that are still unclear or any additional questions.

Reference:

[Lin et al., 2017] Lin, Tsung-Yi, et al. "Focal loss for dense object detection." Proceedings of the IEEE international conference on computer vision. 2017.

评论

We hope this message finds you well! We sincerely appreciate your thorough review and insightful feedback on our submission, as well as the opportunity to address your concerns during the rebuttal period.

We wanted to kindly follow up to confirm whether our responses and additional experiments have adequately addressed your questions and concerns. Specifically, we have run additional experiments comparing against the focal loss as per your suggestion and found that our proposed DCO loss is superior. We have also clarified the role of ϵ\epsilon in our lemmas, as well as the impact of sequence length. Additionally, we will incorporate more thorough discussions around other inference-aware methods in the manuscript. If there remain any unresolved points or if further clarifications are needed, we would be more than happy to provide additional information.

We would greatly appreciate it if you could revisit your evaluation in light of our responses and possibly update your score to reflect the improvements we've incorporated.

Thank you again for your valuable input and consideration. Your constructive review greatly helps us further improve the quality of our paper.

最终决定

In this paper, the authors note that large language models are often evaluated on pass@k evaluations (i.e. whether one sample is correct from k samples conditioned on a prompt) but they are trained on pass@1. They observe that models trained on standard objectives actually degrade in performance on pass@k. As such, they propose a new training loss they call Direct Coverage Optimization (DCO). They show empirically that DCO improves on pass@k evaluations compared to cross-entropy and the recent GRPO method. They also show theoretically why pass@k performance degrades with existing strategies.

Overall the reviews are borderline leaning towards accept (4, 4, 4, 5). The reviewers seemed to find the paper "exceptionally" well written and clear. They found the work well-motivated, topically "very relevant" and theoretically "solid". They noted that the method presented was simple and seemed empirically convincing. The reviewers who gave scores of 4, noted very little in terms of weaknesses and overal seemed positive. Unfortunately, that didn't give the authors a significant amount to address and, in-addition, those reviewers didn't seem to engage meaningfully in discussion despite what seems like an in-depth response. Two of these reviewers noted that the proposed algorithm required additional sampling and thus was computationally less efficient (however, GRPO does as well). The reviewer that recommended accept gave a thorough review and engaged with the authors. Given the lack of any significant concerns raised by some of the reviewers, I would deduce that their scores are perhaps lower than they should be. Therefore, the recommendation is to accept (i.e. up-weighing the 5 score).

The Accept review noted that there remains a question whether the results generalize to the current largest SOTA models. Naturally, that is difficult to empirically verify, but seems like an important potential limitation.